Ask me what skills you need
What are you building?
Tell me what you're working on and I'll find the best agent skills for you.
Run metric-driven iterative optimization loops. Define a measurable goal, build measurement scaffolding, then run parallel experiments that try many approaches, measure each against hard gates and/or LLM-as-judge quality scores, keep improvements, and converge toward the best solution. Use when optimizing clustering quality, search relevance, build performance, prompt quality, or any measurable outcome that benefits from systematic experimentation. Inspired by Karpathy's autoresearch, generalized for multi-file code changes and non-ML domains.
Run metric-driven iterative optimization. Define a goal, build measurement scaffolding, then run parallel experiments that converge toward the best solution.
Use the platform's blocking question tool when available (AskUserQuestion in Claude Code, request_user_input in Codex, ask_user in Gemini). Otherwise, present numbered options in chat and wait for the user's reply before proceeding.
<optimization_input> #$ARGUMENTS </optimization_input>
If the input above is empty, ask: "What would you like to optimize? Describe the goal, or provide a path to an optimization spec YAML file."
Reference the spec schema for validation:
references/optimize-spec-schema.yaml
Reference the experiment log schema for state management:
references/experiment-log-schema.yaml
For a first run, optimize for signal and safety, not maximum throughput:
references/example-hard-spec.yaml when the metric is objective and cheap to measurereferences/example-judge-spec.yaml only when actual quality requires semantic judgmentnpx skills add EveryInc/compound-engineering-plugin --skill ce-optimizeHow clear and easy to understand the SKILL.md instructions are, rated from 1 to 5.
The SKILL.md content is hard to understand and quite ambiguous.
How directly an agent can act on the SKILL.md instructions, rated from 1 to 5.
The SKILL.md is hard to act on; an agent would not know what to do.