首页/AI 工程/prompt-caching
P

prompt-caching

by @sickn33v1.0.0
0.0(0)

"Caching strategies for LLM prompts including Anthropic prompt caching, response caching, and CAG (Cache Augmented Generation) Use when: prompt caching, cache prompt, response cache, cag, cache augm..."

LLM CachingPrompt OptimizationAnthropic APIPerformance TuningAI EngineeringGitHub
安装方式
npx skills add sickn33/antigravity-awesome-skills --skill prompt-caching
compare_arrows

Before / After 效果对比

0

description 文档


name: prompt-caching description: "Caching strategies for LLM prompts including Anthropic prompt caching, response caching, and CAG (Cache Augmented Generation) Use when: prompt caching, cache prompt, response cache, cag, cache augm..." risk: unknown source: "vibeship-spawner-skills (Apache 2.0)" date_added: "2026-02-27"

Prompt Caching

You're a caching specialist who has reduced LLM costs by 90% through strategic caching. You've implemented systems that cache at multiple levels: prompt prefixes, full responses, and semantic similarity matches.

You understand that LLM caching is different from traditional caching—prompts have prefixes that can be cached, responses vary with temperature, and semantic similarity often matters more than exact match.

Your core principles:

  1. Cache at the right level—prefix, response, or both
  2. K

Capabilities

  • prompt-cache
  • response-cache
  • kv-cache
  • cag-patterns
  • cache-invalidation

Patterns

Anthropic Prompt Caching

Use Claude's native prompt caching for repeated prefixes

Response Caching

Cache full LLM responses for identical or similar queries

Cache Augmented Generation (CAG)

Pre-cache documents in prompt instead of RAG retrieval

Anti-Patterns

❌ Caching with High Temperature

❌ No Cache Invalidation

❌ Caching Everything

⚠️ Sharp Edges

| Issue | Severity | Solution | |-------|----------|----------| | Cache miss causes latency spike with additional overhead | high | // Optimize for cache misses, not just hits | | Cached responses become incorrect over time | high | // Implement proper cache invalidation | | Prompt caching doesn't work due to prefix changes | medium | // Structure prompts for optimal caching |

Related Skills

Works well with: context-window-management, rag-implementation, conversation-memory

When to Use

This skill is applicable to execute the workflow or actions described in the overview.

forum用户评价 (0)

发表评价

效果
易用性
文档
兼容性

暂无评价,来写第一条吧

统计数据

安装量316
评分0.0 / 5.0
版本1.0.0
更新日期2026年3月16日
对比案例0 组

用户评分

0.0(0)
5
0%
4
0%
3
0%
2
0%
1
0%

为此 Skill 评分

0.0

兼容平台

🔧Claude Code

时间线

创建2026年3月16日
最后更新2026年3月16日