Skip to main content
Run any Skill in Manus
with one click

evaluate

// Evaluates RAG retrieval and LLM-as-judge metrics (faithfulness, relevancy, context precision). Triggers: measure RAG quality, knowledge gap, RAG eval, golden dataset.

$ git log --oneline --stat
stars:143
forks:18
updated:May 6, 2026 at 11:14
SKILL.md
readonly