Ask me what skills you need
What are you building?
Tell me what you're working on and I'll find the best agent skills for you.
This skill should be used when the user asks to "implement LLM-as-judge", "compare model outputs", "create evaluation rubrics", "mitigate evaluation bias", or mentions direct scoring, pairwise comparison, position bias, evaluation pipelines, or automated quality assessment.
This skill covers production-grade techniques for evaluating LLM outputs using LLMs as judges. It synthesizes research from academic papers, industry practices, and practical implementation experience into actionable patterns for building reliable evaluation systems.
Key insight: LLM-as-a-Judge is not a single technique but a family of approaches, each suited to different evaluation contexts. Choosing the right approach and mitigating known biases is the core competency this skill develops.
Activate this skill when:
Evaluation approaches fall into two primary categories with distinct reliability profiles:
Direct Scoring: A single LLM rates one response on a defined scale.
npx skills add sickn33/antigravity-awesome-skills --skill advanced-evaluationHow clear and easy to understand the SKILL.md instructions are, rated from 1 to 5.
The SKILL.md content is hard to understand and quite ambiguous.
How directly an agent can act on the SKILL.md instructions, rated from 1 to 5.
The SKILL.md is hard to act on; an agent would not know what to do.