sf-ai-agentforce-testing
执行Agentforce多轮对话测试、验证topic和action覆盖率、分析测试结果并提供修复建议,支持CLI自动化
npx skills add jaganpro/sf-skills --skill sf-ai-agentforce-testingBefore / After 效果对比
1 组手动设计测试用例、逐个执行多轮对话测试、人工记录响应结果、手动分析topic和action覆盖情况、编写测试报告和修复Bug,一个中等复杂度Agentforce代理的全面测试需要2-3天,且难以保证覆盖所有边缘场景
自动生成全面测试用例覆盖所有topic和action、执行结构化多轮对话测试、自动记录和分析响应质量、生成详细的覆盖率报告和失败用例列表、一键验证修复效果,一个中等复杂度代理的全面测试只需2-3小时,且覆盖率和准确性显著提升
description SKILL.md
sf-ai-agentforce-testing
sf-ai-agentforce-testing: Agentforce Test Execution & Coverage Analysis
Use this skill when the user needs formal Agentforce testing: multi-turn conversation validation, CLI Testing Center specs, topic/action coverage analysis, preview checks, or a structured test-fix loop after publish.
When This Skill Owns the Task
Use sf-ai-agentforce-testing when the work involves:
-
sf agent testworkflows -
multi-turn Agent Runtime API testing
-
topic routing, action invocation, context preservation, guardrail, or escalation validation
-
test-spec generation and coverage analysis
-
post-publish / post-activate test-fix loops
Delegate elsewhere when the user is:
-
building or editing the agent itself → sf-ai-agentforce or sf-ai-agentscript
-
running Apex unit tests → sf-testing
-
creating seed data for actions → sf-data
-
analyzing session telemetry / STDM traces → sf-ai-agentforce-observability
Core Operating Rules
-
Testing comes after deploy / publish / activate.
-
Use multi-turn API testing as the primary path when conversation continuity matters.
-
Use CLI Testing Center as the secondary path for single-utterance and org-supported test-center workflows.
-
Fixes to the agent should be delegated to sf-ai-agentscript when Agent Script changes are needed.
-
Do not use raw
curlfor OAuth token validation in the ECA flow; use the provided credential tooling.
Script path rule
Use the existing scripts under:
~/.claude/skills/sf-ai-agentforce-testing/hooks/scripts/
These scripts are pre-approved. Do not recreate them.
Required Context to Gather First
Ask for or infer:
-
agent API name / developer name
-
target org alias
-
testing goal: smoke test, regression, coverage expansion, or bug reproduction
-
whether the agent is already published and activated
-
whether the org has Agent Testing Center available
-
whether ECA credentials are available for Agent Runtime API testing
Preflight checks:
-
discover the agent
-
confirm publish / activation state
-
verify dependencies (Flows, Apex, data)
-
choose testing track
Dual-Track Workflow
Track A — Multi-turn API testing (primary)
Use when you need:
-
multi-turn conversation testing
-
topic re-matching validation
-
context preservation checks
-
escalation or action-chain analysis across turns
Requires:
-
ECA / auth setup
-
agent runtime access
Track B — CLI Testing Center (secondary)
Use when you need:
-
org-native
sf agent testworkflows -
test spec YAML execution
-
quick single-utterance validation
-
CLI-centered CI/CD usage where Testing Center is available
Quick manual path
For manual validation without full formal testing, use preview workflows first, then escalate to Track A or B as needed.
Recommended Workflow
1. Discover and verify
-
locate the agent in the target org
-
confirm it is published and activated
-
confirm required actions / Flows / Apex exist
-
decide whether Track A or Track B fits the request
2. Plan tests
Cover at least:
-
main topics
-
expected actions
-
guardrails / off-topic handling
-
escalation behavior
-
phrasing variation
3. Execute the right track
Track A
-
validate ECA credentials with the provided tooling
-
retrieve metadata needed for scenario generation
-
run multi-turn scenarios with the provided Python scripts
-
analyze per-turn failures and coverage
Track B
-
generate or refine a flat YAML test spec
-
run
sf agent testcommands -
inspect structured results and verbose action output
4. Classify failures
Typical failure buckets:
-
topic not matched
-
wrong topic matched
-
action not invoked
-
wrong action selected
-
action invocation failed
-
context preservation failure
-
guardrail failure
-
escalation failure
5. Run fix loop
When failures imply agent-authoring issues:
-
delegate fixes to sf-ai-agentscript
-
re-publish / re-activate if needed
-
re-run focused tests before full regression
Testing Guardrails
Never skip these:
-
test only after publish/activate
-
include harmful / off-topic / refusal scenarios
-
use multiple phrasings per important topic
-
clean up sessions after API tests
-
keep swarm execution small and controlled
Avoid these anti-patterns:
-
testing unpublished agents
-
treating one happy-path utterance as coverage
-
storing ECA secrets in repo files
-
debugging auth with brittle shell-expanded
curlcommands -
changing both tests and agent simultaneously without isolating the cause
Output Format
When finishing a run, report in this order:
-
Test track used
-
What was executed
-
Pass/fail summary
-
Coverage gaps
-
Root-cause themes
-
Recommended fix loop / next test step
Suggested shape:
Agent: <name>
Track: Multi-turn API | CLI Testing Center | Preview
Executed: <specs / scenarios / turns>
Result: <passed / partial / failed>
Coverage: <topics, actions, guardrails, context>
Issues: <highest-signal failures>
Next step: <fix, republish, rerun, or expand coverage>
Cross-Skill Integration
Need Delegate to Reason
fix Agent Script logic sf-ai-agentscript authoring and deterministic fix loops
create test data sf-data action-ready data setup
fix Flow-backed actions sf-flow Flow repair
fix Apex-backed actions sf-apex Apex repair
set up ECA / OAuth sf-connected-apps auth and app configuration
analyze session telemetry sf-ai-agentforce-observability STDM / trace analysis
Reference Map
Start here
Execution / auth
Coverage / fix loops
Advanced / specialized
Templates / assets
Score Guide
Score Meaning
90+ production-ready test confidence
80–89 strong coverage with minor gaps
70–79 acceptable but coverage expansion recommended
60–69 partial validation only
< 60 insufficient confidence; block release
Weekly Installs271Repositoryjaganpro/sf-skillsGitHub Stars234First SeenJan 22, 2026Security AuditsGen Agent Trust HubPassSocketPassSnykPassInstalled oncodex263cursor263gemini-cli261opencode261github-copilot258amp255
forum用户评价 (0)
发表评价
暂无评价,来写第一条吧
统计数据
用户评分
为此 Skill 评分