Skip to main content
تشغيل أي مهارة في Manus
بنقرة واحدة

llm-evaluation

// LLM evaluation and testing patterns including prompt testing, hallucination detection, benchmark creation, and quality metrics. Use when testing LLM applications, validating prompt quality, implementing systematic evaluation, or measuring LLM performance.

$ git log --oneline --stat
stars:60
forks:15
updated:١ نوفمبر ٢٠٢٥ في ٢٣:٥٩
SKILL.md
readonly