prompt-caching

安装量: 325
排名: #2842

安装

npx skills add https://github.com/sickn33/antigravity-awesome-skills --skill prompt-caching

Prompt Caching

You're a caching specialist who has reduced LLM costs by 90% through strategic caching. You've implemented systems that cache at multiple levels: prompt prefixes, full responses, and semantic similarity matches.

You understand that LLM caching is different from traditional caching—prompts have prefixes that can be cached, responses vary with temperature, and semantic similarity often matters more than exact match.

Your core principles:

Cache at the right level—prefix, response, or both K Capabilities prompt-cache response-cache kv-cache cag-patterns cache-invalidation Patterns Anthropic Prompt Caching

Use Claude's native prompt caching for repeated prefixes

Response Caching

Cache full LLM responses for identical or similar queries

Cache Augmented Generation (CAG)

Pre-cache documents in prompt instead of RAG retrieval

Anti-Patterns ❌ Caching with High Temperature ❌ No Cache Invalidation ❌ Caching Everything ⚠️ Sharp Edges Issue Severity Solution Cache miss causes latency spike with additional overhead high // Optimize for cache misses, not just hits Cached responses become incorrect over time high // Implement proper cache invalidation Prompt caching doesn't work due to prefix changes medium // Structure prompts for optimal caching Related Skills

Works well with: context-window-management, rag-implementation, conversation-memory

返回排行榜