Skip to main content
Run any Skill in Manus
with one click

eval-run

// Execute skill evaluation against test cases, score with judges, and report results. Requires eval.yaml (generated by /eval-analyze). Use when the user wants to test a skill, run eval, benchmark, compare models, detect regressions, check skill quality, or verify changes didn't break anything. Triggers on "run eval", "test the skill", "evaluate", "benchmark", "check for regressions", "how does my skill perform", "score the skill", "run the tests", "run my evals", "compare against baseline", "did I break anything", "test my changes". Also called by /eval-optimize for automated iterations.

$ git log --oneline --stat
stars:6
forks:8
updated:May 6, 2026 at 13:03
File Explorer
13 files
SKILL.md
readonly