Ask me what skills you need
What are you building?
Tell me what you're working on and I'll find the best agent skills for you.
Integrate a new text-to-speech model into vLLM-Omni from HuggingFace reference implementation through production-ready serving with streaming and CUDA graph acceleration. Use when adding a new TTS model, wiring stage separation for speech synthesis, enabling online voice generation serving, debugging TTS integration behavior, or building audio output pipelines.
HF Reference -> Stage Separation -> Online Serving -> Async Chunk -> CUDA Graph
(Phase 1) (Phase 2) (Phase 3) (Phase 4) (Phase 5)
Goal: Understand the reference implementation and verify it produces correct audio.
config.json fields, model_type, sub-model configs<|voice|>, <|audio_start|>, <|im_end|>, etc.)npx skills add vllm-project/vllm-omni --skill add-tts-modelHow clear and easy to understand the SKILL.md instructions are, rated from 1 to 5.
Mostly clear, but there are still a few confusing or poorly structured parts.
How directly an agent can act on the SKILL.md instructions, rated from 1 to 5.
Partially actionable with several concrete steps, but still missing important details.