inject
'Inject relevant knowledge into session context from .agents/ artifacts. Triggers: "inject knowledge", "recall context", SessionStart hook.'
npx skills add boshu2/agentops --skill injectBefore / After 效果对比
0 组description 文档
name: inject description: 'Inject relevant knowledge into session context from .agents/ artifacts. Triggers: "inject knowledge", "recall context", SessionStart hook.' skill_api_version: 1 user-invocable: false metadata: tier: background dependencies: [] internal: true
DEPRECATED (removal target: v3.0.0) — Use
ao lookup --query "topic"for on-demand learnings retrieval, or see.agents/AGENTS.mdfor knowledge navigation. This skill and theao injectCLI command still work but are no longer called from hooks or other skills.
Inject Skill
On-demand knowledge retrieval. Not run automatically at startup (since ag-8km).
Inject relevant prior knowledge into the current session.
How It Works
In the default manual startup mode, MEMORY.md is auto-loaded by Claude Code and no startup injection occurs. Use /inject or ao inject for on-demand retrieval when you need deeper context.
In lean or legacy startup modes (set via AGENTOPS_STARTUP_CONTEXT_MODE), the SessionStart hook runs:
# lean mode (MEMORY.md fresh): 400 tokens
ao inject --apply-decay --format markdown --max-tokens 400 \
[--bead <bead-id>] [--predecessor <handoff-path>]
# legacy mode: 800 tokens
ao inject --apply-decay --format markdown --max-tokens 800 \
[--bead <bead-id>] [--predecessor <handoff-path>]
This searches for relevant knowledge and injects it into context.
Work-Scoped Injection
When --bead is provided (via HOOK_BEAD env var from Gas Town):
- Learnings tagged with the same bead ID get a 1.5x score boost
- Learnings matching bead labels get a 1.2x boost
- Untagged learnings still appear but ranked lower
Predecessor Context
When --predecessor is provided (path to a handoff file):
- Extracts structured context: progress, blockers, next steps
- Injected as "Predecessor Context" section before learnings
- Supports explicit handoffs, auto-handoffs, and pre-compact snapshots
Manual Execution
Given /inject [topic]:
Step 1: Search for Relevant Knowledge
With ao CLI:
ao inject --context "<topic>" --format markdown --max-tokens 1000
Without ao CLI, search manually:
# Recent learnings
ls -lt .agents/learnings/ | head -5
# Recent patterns
ls -lt .agents/patterns/ | head -5
# Recent research
ls -lt .agents/research/ | head -5
# Global learnings (cross-repo knowledge)
ls -lt ~/.agents/learnings/ 2>/dev/null | head -5
# Global patterns (cross-repo patterns)
ls -lt ~/.agents/patterns/ 2>/dev/null | head -5
# Legacy patterns (read-only fallback, no new writes)
ls -lt ~/.claude/patterns/ 2>/dev/null | head -5
Step 2: Read Relevant Files
Use the Read tool to load the most relevant artifacts based on topic.
Step 3: Summarize for Context
Present the injected knowledge:
- Key learnings relevant to current work
- Patterns that may apply
- Recent research on related topics
Step 4: Record Citations (Feedback Loop)
After presenting injected knowledge, record which files were injected for the feedback loop:
mkdir -p .agents/ao
# Record each injected learning file as a citation
for injected_file in <list of files that were read and presented>; do
echo "{\"learning_file\": \"$injected_file\", \"timestamp\": \"$(date -Iseconds)\", \"session\": \"$(date +%Y-%m-%d)\"}" >> .agents/ao/citations.jsonl
done
Citation tracking enables the feedback loop: learnings that are frequently cited get confidence boosts during /post-mortem, while uncited learnings decay faster.
Knowledge Sources
| Source | Location | Priority | Weight |
|--------|----------|----------|--------|
| Learnings | .agents/learnings/ | High | 1.0 |
| Patterns | .agents/patterns/ | High | 1.0 |
| Global Learnings | ~/.agents/learnings/ | High | 0.8 (configurable) |
| Global Patterns | ~/.agents/patterns/ | High | 0.8 (configurable) |
| Research | .agents/research/ | Medium | — |
| Retros | .agents/learnings/ | Medium | — |
| Legacy Patterns | ~/.claude/patterns/ | Low | 0.6 (read-only, no new writes) |
Decay Model
Knowledge relevance decays over time (~17%/week). More recent learnings are weighted higher.
Key Rules
- Runs automatically - usually via hook
- Context-aware - filters by current directory/topic
- Token-budgeted - respects max-tokens limit
- Recency-weighted - newer knowledge prioritized
Examples
SessionStart Hook Invocation (lean/legacy modes only)
Hook triggers: session-start.sh runs at session start with AGENTOPS_STARTUP_CONTEXT_MODE=lean or legacy
What happens:
- Hook calls
ao inject --apply-decay --format markdown --max-tokens 400(lean) or--max-tokens 800(legacy) - CLI searches
.agents/learnings/,.agents/patterns/,.agents/research/for relevant artifacts - CLI applies recency-weighted decay (~17%/week) to rank results
- CLI outputs top-ranked knowledge as markdown within token budget
- Agent presents injected knowledge in session context
Result: Prior learnings, patterns, research automatically available at session start without manual lookup.
Note: In the default manual mode, MEMORY.md is auto-loaded by Claude Code and this hook emits only a pointer to on-demand retrieval commands (ao search, ao lookup).
Manual Context Injection
User says: /inject authentication or "recall knowledge about auth"
What happens:
- Agent calls
ao inject --context "authentication" --format markdown --max-tokens 1000 - CLI filters artifacts by topic relevance
- Agent reads top-ranked learnings and patterns
- Agent summarizes injected knowledge for current work
- Agent references artifact paths for deeper exploration
Result: Topic-specific knowledge retrieved and summarized, enabling faster context loading than full artifact reads.
Troubleshooting
| Problem | Cause | Solution |
|---------|-------|----------|
| No knowledge injected | Empty knowledge pools or ao CLI unavailable | Run /post-mortem to seed pools; verify ao CLI installed |
| Irrelevant knowledge | Topic mismatch or stale artifacts dominate | Use --context "<topic>" to filter; prune stale artifacts |
| Token budget exceeded | Too many high-relevance artifacts | Reduce --max-tokens or increase topic specificity |
| Decay too aggressive | Recent learnings not prioritized | Check artifact modification times; verify --apply-decay flag |
forum用户评价 (0)
发表评价
暂无评价,来写第一条吧
统计数据
用户评分
为此 Skill 评分