Every day, thousands of decisions happen inside AI context windows. Your skill vs. competitors. Find out who wins — and why.
"Advanced web search with real-time results and structured output..."
"Basic web scraping tool for fetching page contents..."
"The winning skill clearly specified real-time capabilities and structured output, which directly matched the task of finding current news. The competitor only mentioned basic page fetching without real-time or formatting guarantees."
Developers don't search for tools anymore — they ask their AI assistant. Your skill is either chosen or invisible.
Three steps to know if your skill wins
Paste your skill description, upload a .md file, or search the skills.sh repository.
Your skill enters the arena against real competitors. AI agents simulate choosing between them.
Get your selection rate, detailed reasoning for each decision, and insights to improve.
Benchmark, compare, and optimize your agent skills
Compare your skill directly against competitors in realistic agent scenarios.
See exactly how often AI agents choose your skill — with clear percentages.
Rank your skill against others. See where you stand in the arena.
AI-powered reasoning explains why your skill was chosen — or wasn't.
Run evaluations locally or in CI/CD. pip install skills-arena.
Test across Claude, GPT, and more to see how different agents evaluate your skill.
Run evaluations locally, in CI/CD, or anywhere Python runs.
Find out if AI agents will choose your skill.
Two free evaluations — no signup needed.