skills-mcp2026
30 skills · 6 MCP tools · 8 platforms · versioned skills · free to host

Retrieve skills
at runtime.
Not upfront.

Semantic skill retrieval: cosine similarity, not keyword matching
Three-tier progressive disclosure: 200 tokens to discover, 4k to act
Streamable HTTP + SSE, Glama compatible, 8 AI platforms supported
30
skills
6
MCP tools
8
platforms
free
to host
agent session · skills-mcp
connected
hosted instance · try now
https://skill-mcp.jigneshponamwar.workers.dev/mcp
Claude CodeCursorWindsurfClineCopilotGlama
01
The Problem

Static skill files
don't scale.

Every approach before skills-mcp forces the same trade-off between agent capability and context efficiency. The model gets everything or nothing.

system prompt token usage~80,000 tokens
stripe-integration
94%
docker-containerization
82%
react-best-practices
77%
fastapi
71%
github-actions
63%
+ 25 more…
48%

context window exhausted before first user message

01
01~80k tokens / session

Context window exhausted upfront

Embedding 30 expert procedures into every system prompt costs 60–80k tokens before the agent sends a single message. Most of that context is never used.

02
02N files × M projects

Copy-paste maintenance across projects

Every project gets its own skill file copies. One update means manually syncing N files across M projects, with no version control and no diffing.

03
038 platforms, 8 configs

Reconfiguration per AI platform

Claude Code, Cursor, Windsurf, Cline: different config formats, different file paths, different scopes. The same procedure files multiply into incompatible formats.

04
04Retrieval precision: 0%

No relevance matching

Static files can't score themselves. No way to retrieve the procedure best suited to the task at hand. It's all procedures or none.

What if the agent could request exactly the skill it needs, the moment it needs it?

02
How It Works

Three tiers.
Loaded only when
needed.

skills-mcp serves knowledge in layers. The agent always starts with the lightest context and escalates only when the task demands it.

Before
All 30 in system prompt
~80k tokens upfront
No relevance match
Re-install per project
After
1 skill per query
~200–4k tokens
0.97 match score
One Worker, any platform
01Tier 1 · Discovery

Semantic skill search

The agent calls skills_find_relevant() with a plain description. Qdrant returns the highest-scoring match by cosine similarity against embedded descriptions.

skills_find_relevant("stripe webhooks")
→ 0.97  stripe-integration  [tier-1]

Returns frontmatter only. ~200 tokens.

02Tier 2 · Load

Full procedure on demand

When ready to act, the agent calls skills_get_body(). The complete step-by-step procedure is returned for exactly one skill, not all 30.

skills_get_body("stripe-integration")           ← latest
skills_get_body("stripe-integration", version="1.2")  ← pinned

Returns a deprecation_notice if the skill has been superseded.

Full procedure + patterns. ~2–4k tokens.

03Tier 3 · Supplement

Deep references on request

For complex tasks, tier-3 tools pull curated docs, API references, and scripts. Only loaded when the skill body explicitly references them in the manifest.

skills_get_reference("stripe-integration", "WEBHOOKS.md")
→ Verification spec · Security checklist

Only loaded when the body references it.

03.5
Try It Now

Try it in 60 seconds.

My hosted instance is live. Connect it to your AI agent right now.

hosted instance · live
{
  "mcpServers": {
    "skill-mcp": {
      "transport": "streamable-http",
      "url": "https://skill-mcp.jigneshponamwar.workers.dev/mcp"
    }
  }
}

Paste this config into your AI tool’s MCP settings. The server is already running no deployment needed. Per-IP rate limiting: 60 requests/min (configurable via RATE_LIMIT_RPM). Exceeding the limit returns HTTP 429.

Claude CodeCursorWindsurfClineCopilotAntigravityGlama
03
Skills Library

30 production-ready
expert procedures.

Across 7 categories: API integration, cloud infra, AI models, document generation, web frameworks, services, and design. Each skill is semantically indexed with 3-tier progressive disclosure.

score thresholds · skills_find_relevant()
> 0.6Strong match: load skill body
0.4–0.6Review description, may be relevant
< 0.4No match: skip this skill

Thresholds validated against a 120-query eval dataset (90 positive × 30 skills + 30 true negatives). Run make calibrate to recalibrate.

categories · 7 domains
Core DevDocumentsAI & LLMCloud & InfraWeb & FullstackServicesDesign
#07Core Dev

test-writer

pytest, Jest, Go test suites with full edge case coverage and mocking patterns.

3 tiers · semantic search
#13AI & LLM

claude-api

Anthropic SDK: tool use, streaming, vision, prompt caching, extended thinking, batch.

3 tiers · semantic search
#17AI & LLM

mcp-server-builder

FastMCP (Python) or TypeScript SDK -- tools, resources, prompts.

3 tiers · semantic search
#19Cloud & Infra

docker-containerization

Production Dockerfiles, multi-stage builds, Docker Compose, security hardening.

3 tiers · semantic search
#22Web & Fullstack

nextjs-best-practices

App Router -- RSC, async params, image/font optimization, self-hosting.

3 tiers · semantic search
#27Services

stripe-integration

Checkout Sessions, webhooks, subscriptions, Connect (Accounts v2), security checklist.

3 tiers · semantic search
Showing 6 of 30 skills across 7 categoriesView all 30 skills
04
Deploy & Connect

Running in
minutes.

Three paths to a running server. Cloudflare free tier for production. Docker for local-only or air-gapped use.

requirements
Qdrant Cloudfree · 1 GB cluster
Cloudflarefree · Workers plan
Python 3.11+for seed script
Node.js 18+for wrangler CLI
Docker mode only needs Cloudflare credentials

One command does everything: checks prerequisites, creates .env, installs deps, seeds Qdrant, pushes secrets, deploys the Worker.

Windows (PowerShell)
.\scripts\setup.ps1
Linux / macOS
bash scripts/setup.sh
Cross-platform (Make)
make setup

The wizard prompts for credentials at the right moment. Your server is live once it completes.

05
Connect your AI agent

All platforms use the same SSE transport config. Replace the URL with your deployed Worker URL, or use the hosted instance above to try immediately without deploying.

.mcp.json
{
  "mcpServers": {
    "skill-mcp": {
      "transport": "streamable-http",
      "url": "https://skill-mcp.jigneshponamwar.workers.dev/mcp"
    }
  }
}

Replace the URL with your deployed Worker URL. Use the hosted instance to try without deploying.

06
Install master skill · optional · recommended

Drop a platform instruction file into your project root. It teaches the agent the 3-tier skill workflow: when to call skills_find_relevant, how to interpret scores, and when to load tier-3 resources.

Platform
Command (run from repo root)
Claude Code
cp master-skill/platforms/claude-code/CLAUDE.md ./CLAUDE.md
Cursor
cp master-skill/platforms/cursor/.cursorrules ./.cursorrules
Windsurf
cp master-skill/platforms/windsurf/.windsurfrules ./.windsurfrules
Cline
cp master-skill/platforms/cline/.clinerules ./.clinerules
Copilot
cp -r master-skill/platforms/copilot/.github ./.github
Codex
cp master-skill/platforms/codex/AGENTS.md ./AGENTS.md
Antigravity
cp -r master-skill/platforms/antigravity/.agents ./.agents
Aider
cp master-skill/platforms/aider/CONVENTIONS.md ./CONVENTIONS.md

After copying, edit the file and replace the placeholder Worker URL with your actual deployed URL.