Skip to main content
Run any Skill in Manus
with one click
$pwd:

agentic-eval

// Patterns and techniques for evaluating and improving AI agent outputs through iterative critique and refinement. Use this skill when: implementing self-critique loops, evaluator-optimizer pipelines, rubric-based scoring, LLM-as-judge evaluation, adversarial evaluation, judge-and-refine cycles, structured output quality improvement, cross-agent evaluation, plan quality review, spec validation before planning, document consistency checking, stage transition quality gate, architect review of plan or spec, verifying a downstream agent can consume an upstream artifact, or inter-stage artifact consistency checks. Do NOT use for standard code review (use code-security-review skill), general refactoring (use refactor skill), or security audits.

$ git log --oneline --stat
stars:0
forks:1
updated:May 6, 2026 at 09:16
File Explorer
4 files
SKILL.md
readonly