Loading...
Loading...
Set up and improve harness engineering (AGENTS.md, docs/, lint rules, eval systems, project-level prompt engineering) for AI-agent-friendly codebases. Triggers on: new/empty project setup for AI agents, AGENTS.md or CLAUDE.md creation, harness engineering questions, making agents work better on a codebase. ALSO triggers when users are frustrated or complaining about agent quality — e.g. 'the agent keeps ignoring conventions', 'it never follows instructions', 'why does it keep doing X', 'the agent is broken' — because poor agent output almost always signals harness gaps, not model problems. Covers: context engineering, architectural constraints, multi-agent coordination, evaluation, long-running agent harness, and diagnosis of agent quality issues.
npx skill4agent add 10xchengtu/harness-engineering harness-engineering| Signal | Action |
|---|---|
| Empty/new project | → Full project setup (Section 1) |
| User frustrated with agent | → Diagnose & fix harness gaps (Section 7) |
| Existing project needs improvement | → Assess & incrementally improve |
| Explicit harness question | → Reference relevant sections |
references/01-project-setup.mdreferences/02-context-engineering.mdreferences/03-constraints.mdreferences/05-eval-feedback.mdreferences/04-multi-agent.mdreferences/06-long-running.mdreferences/07-diagnosis.md| Layer | What | Reference |
|---|---|---|
| Project Setup | AGENTS.md, docs/, directory conventions | |
| Context Engineering | What info agents see, progressive disclosure, working state | |
| Constraints & Guardrails | Linters, type systems, architecture enforcement, safe autonomy | |
| Multi-Agent Architecture | Agent separation, coordination protocols, delegation patterns | |
| Eval & Feedback | Testing, grading, GC agents, observability | |
| Long-Running Tasks | Progress tracking, context resets, handoff artifacts | |
| Diagnosis | When agents fail — identify root cause in harness, not model | |