首页/测试 & QA/sf-ai-agentforce-testing
S

sf-ai-agentforce-testing

by @jaganprov1.0.0
4.5(4)

执行Agentforce多轮对话测试、验证topic和action覆盖率、分析测试结果并提供修复建议,支持CLI自动化

testingai-agentstest-automationquality-assuranceai-engineeringGitHub
安装方式
npx skills add jaganpro/sf-skills --skill sf-ai-agentforce-testing
compare_arrows

Before / After 效果对比

1
使用前

手动设计测试用例、逐个执行多轮对话测试、人工记录响应结果、手动分析topic和action覆盖情况、编写测试报告和修复Bug,一个中等复杂度Agentforce代理的全面测试需要2-3天,且难以保证覆盖所有边缘场景

使用后

自动生成全面测试用例覆盖所有topic和action、执行结构化多轮对话测试、自动记录和分析响应质量、生成详细的覆盖率报告和失败用例列表、一键验证修复效果,一个中等复杂度代理的全面测试只需2-3小时,且覆盖率和准确性显著提升

description SKILL.md

sf-ai-agentforce-testing

sf-ai-agentforce-testing: Agentforce Test Execution & Coverage Analysis

Use this skill when the user needs formal Agentforce testing: multi-turn conversation validation, CLI Testing Center specs, topic/action coverage analysis, preview checks, or a structured test-fix loop after publish.

When This Skill Owns the Task

Use sf-ai-agentforce-testing when the work involves:

  • sf agent test workflows

  • multi-turn Agent Runtime API testing

  • topic routing, action invocation, context preservation, guardrail, or escalation validation

  • test-spec generation and coverage analysis

  • post-publish / post-activate test-fix loops

Delegate elsewhere when the user is:

Core Operating Rules

  • Testing comes after deploy / publish / activate.

  • Use multi-turn API testing as the primary path when conversation continuity matters.

  • Use CLI Testing Center as the secondary path for single-utterance and org-supported test-center workflows.

  • Fixes to the agent should be delegated to sf-ai-agentscript when Agent Script changes are needed.

  • Do not use raw curl for OAuth token validation in the ECA flow; use the provided credential tooling.

Script path rule

Use the existing scripts under:

  • ~/.claude/skills/sf-ai-agentforce-testing/hooks/scripts/

These scripts are pre-approved. Do not recreate them.

Required Context to Gather First

Ask for or infer:

  • agent API name / developer name

  • target org alias

  • testing goal: smoke test, regression, coverage expansion, or bug reproduction

  • whether the agent is already published and activated

  • whether the org has Agent Testing Center available

  • whether ECA credentials are available for Agent Runtime API testing

Preflight checks:

  • discover the agent

  • confirm publish / activation state

  • verify dependencies (Flows, Apex, data)

  • choose testing track

Dual-Track Workflow

Track A — Multi-turn API testing (primary)

Use when you need:

  • multi-turn conversation testing

  • topic re-matching validation

  • context preservation checks

  • escalation or action-chain analysis across turns

Requires:

  • ECA / auth setup

  • agent runtime access

Track B — CLI Testing Center (secondary)

Use when you need:

  • org-native sf agent test workflows

  • test spec YAML execution

  • quick single-utterance validation

  • CLI-centered CI/CD usage where Testing Center is available

Quick manual path

For manual validation without full formal testing, use preview workflows first, then escalate to Track A or B as needed.

Recommended Workflow

1. Discover and verify

  • locate the agent in the target org

  • confirm it is published and activated

  • confirm required actions / Flows / Apex exist

  • decide whether Track A or Track B fits the request

2. Plan tests

Cover at least:

  • main topics

  • expected actions

  • guardrails / off-topic handling

  • escalation behavior

  • phrasing variation

3. Execute the right track

Track A

  • validate ECA credentials with the provided tooling

  • retrieve metadata needed for scenario generation

  • run multi-turn scenarios with the provided Python scripts

  • analyze per-turn failures and coverage

Track B

  • generate or refine a flat YAML test spec

  • run sf agent test commands

  • inspect structured results and verbose action output

4. Classify failures

Typical failure buckets:

  • topic not matched

  • wrong topic matched

  • action not invoked

  • wrong action selected

  • action invocation failed

  • context preservation failure

  • guardrail failure

  • escalation failure

5. Run fix loop

When failures imply agent-authoring issues:

  • delegate fixes to sf-ai-agentscript

  • re-publish / re-activate if needed

  • re-run focused tests before full regression

Testing Guardrails

Never skip these:

  • test only after publish/activate

  • include harmful / off-topic / refusal scenarios

  • use multiple phrasings per important topic

  • clean up sessions after API tests

  • keep swarm execution small and controlled

Avoid these anti-patterns:

  • testing unpublished agents

  • treating one happy-path utterance as coverage

  • storing ECA secrets in repo files

  • debugging auth with brittle shell-expanded curl commands

  • changing both tests and agent simultaneously without isolating the cause

Output Format

When finishing a run, report in this order:

  • Test track used

  • What was executed

  • Pass/fail summary

  • Coverage gaps

  • Root-cause themes

  • Recommended fix loop / next test step

Suggested shape:

Agent: <name>
Track: Multi-turn API | CLI Testing Center | Preview
Executed: <specs / scenarios / turns>
Result: <passed / partial / failed>
Coverage: <topics, actions, guardrails, context>
Issues: <highest-signal failures>
Next step: <fix, republish, rerun, or expand coverage>

Cross-Skill Integration

Need Delegate to Reason

fix Agent Script logic sf-ai-agentscript authoring and deterministic fix loops

create test data sf-data action-ready data setup

fix Flow-backed actions sf-flow Flow repair

fix Apex-backed actions sf-apex Apex repair

set up ECA / OAuth sf-connected-apps auth and app configuration

analyze session telemetry sf-ai-agentforce-observability STDM / trace analysis

Reference Map

Start here

Execution / auth

Coverage / fix loops

Advanced / specialized

Templates / assets

Score Guide

Score Meaning

90+ production-ready test confidence

80–89 strong coverage with minor gaps

70–79 acceptable but coverage expansion recommended

60–69 partial validation only

< 60 insufficient confidence; block release

Weekly Installs271Repositoryjaganpro/sf-skillsGitHub Stars234First SeenJan 22, 2026Security AuditsGen Agent Trust HubPassSocketPassSnykPassInstalled oncodex263cursor263gemini-cli261opencode261github-copilot258amp255

forum用户评价 (0)

发表评价

效果
易用性
文档
兼容性

暂无评价,来写第一条吧

统计数据

安装量200
评分4.5 / 5.0
版本1.0.0
更新日期2026年3月27日
对比案例1 组

用户评分

4.5(4)
5
0%
4
0%
3
0%
2
0%
1
0%

为此 Skill 评分

0.0

兼容平台

🔧Claude Code

时间线

创建2026年3月27日
最后更新2026年3月27日