Skip to main content
تشغيل أي مهارة في Manus
بنقرة واحدة
$pwd:

llm-evaluation

// Implement comprehensive evaluation strategies for LLM applications using automated metrics, human feedback, and benchmarking. Use when testing LLM performance, measuring AI application quality, or establishing evaluation frameworks.

$ git log --oneline --stat
stars:34,024
forks:3,690
updated:٧ مارس ٢٠٢٦ في ١٥:٥٣
SKILL.md
readonly