This Learning Path includes a set of structured, verifiable agent skills that enable automation of the performance analysis workflow by AI coding assistants (such as Codex, Claude, Cursor, or Copilot) and CI systems.
Traditional documentation describes what a human should do. Agent skills describe how an automated system should do it, with explicit definitions of inputs, actions, expected outputs, and validation criteria. Each skill defines required inputs and preconditions, the exact actions to perform, the files and artifacts that should be produced, and validation steps that confirm successful execution. This structure enables workflows that are reproducible, verifiable, and suitable for AI-assisted development, automated regression testing, and repeatable onboarding.
All agent skills are included in the same code repository as the performance analysis kit under:
Start by reading the skill catalog
Each skill is documented as a standalone, executable workflow. Available skills include:
01_setup_workspace.md - Initialize performance analysis environment02_build_runners.md - Build SME2-on/off runner binaries03_export_model.md - Export PyTorch model to ExecuTorch .pte04_run_profiling.md - Execute performance measurement pipeline (timing + trace)05_analyze_results.md - Generate operator-category breakdown06_validate_workflow.md - End-to-end smoke test07_report_generation.md - Generate comprehensive markdown report08_onboard_edgetam.md - Onboard EdgeTAM image encoder modelEach agent skill follows a consistent structure designed for both human readability and machine execution:
name and description metadataThe key feature: Each skill is self-contained and verifiable. Validation steps ensure that the skill completed successfully before downstream skills are executed. This makes the workflow robust when automated.
| Skill | Time | Prerequisites | Outputs |
|---|---|---|---|
setup_workspace | ~30 min | Python 3.9+, CMake 3.29+ | .venv/, executorch/ |
build_runners | ~20 min | Setup complete | executorch/cmake-out/android-arm64-v9a*/executor_runner (for mobile device testing) or executorch/cmake-out/mac-arm64*/executor_runner (developer accessibility) |
export_model | ~5 min | Setup complete | out_<model>/artifacts/*.pte |
run_profiling | ~10 min | Runners + model | out_<model>/runs/<platform>/*.etdump |
analyze_results | ~2 min | Performance measurement complete | out_<model>/runs/<platform>/analysis_summary.json |
validate_workflow | ~15 min | Setup complete | Full smoke test validation |
report_generation | ~1 min | Analysis complete | out_<model>/runs/<platform>/report.md |
Times are approximate and depend on host performance and network availability.
For AI assistants, use skills in this order:
First-time setup (required for fresh downloads):
01_setup_workspace.md - Initialize environment (run this first if you downloaded the repo)After setup is complete:
06_validate_workflow.md - Quick end-to-end test (recommended after setup)Ongoing usage: For regular model analysis:
02_build_runners.md (only if runners are missing or ExecuTorch changed)03_export_model.md04_run_profiling.md05_analyze_results.mdSkills are composable. You can chain them together for end-to-end automation, or use them individually for specific tasks. Each skill includes verification steps to ensure success before proceeding.
Agent skills are designed to be pulled directly into an AI agent’s context as executable instructions. Each skill is a standalone Markdown file (*.md) that the agent reads and follows step by step. The typical interaction model is:
Try this example User prompt in your AI agent: “Set up the performance analysis environment and run a smoke test”
Agent behavior:
01_setup_workspace.md and creates .venv/, executorch/02_build_runners.md and builds the required runner binaries.06_validate_workflow.md and runs the end-to-end smoke testThe agent doesn’t need to guess which commands to run or how to verify success. The skill files provide explicit instructions, making the performance analysis workflow reliable across human-driven, AI-driven, and fully automated environments.
In this section you learned how to use agent skills to automate profiling workflows, run skills in the correct order for setup and ongoing usage, and integrate agent skills with AI coding assistants for repeatable optimization.
This Learning Path as a whole has shown you how to analyze ExecuTorch performance on Arm using SME2, operator-level profiling, and agent-driven automation.
You are now ready to:
For additional guidance on extending this toolkit or integrating with production workflows, see the project repository and the agent skill catalog.