---
id: "framework-factory-agent-readiness"
type: "framework"
source_timestamps: ["00:11:43", "00:12:08"]
tags: ["dev-ops", "code-quality", "evaluation"]
related: ["concept-agent-environment-readiness", "entity-factory-ai"]
steps: ["Evaluate Style and Validation (Linting)", "Evaluate Build Systems", "Evaluate Testing infrastructure", "Evaluate Documentation clarity", "Evaluate Dev Environment reproducibility", "Evaluate overall Code Quality", "Evaluate Observability of the system", "Evaluate Security and Governance controls"]
sources: ["s41-nvidia-open-sourced"]
sourceVaultSlug: "s41-nvidia-open-sourced"
originDay: 41
---
# Factory.ai Agent Readiness Framework

## Origin

[[entity-factory-ai-d41]] developed this framework to evaluate how ready a codebase is to host autonomous AI agents. The underlying philosophy: **agents fail when the environment is poor, not because the LLM lacks reasoning.** This codifies [[concept-agent-environment-readiness]] and the behavioral claim [[claim-agents-are-lazy-developers]].

## The 8 Pillars

| # | Pillar | What It Measures |
|---|---|---|
| 1 | **Style and Validation** | Strict linting, formatter enforcement, style configs |
| 2 | **Build Systems** | Reproducible builds, deterministic toolchains |
| 3 | **Testing** | Coverage, fast feedback loops, isolation |
| 4 | **Documentation** | Clarity, currency, machine-readability |
| 5 | **Dev Environment** | Reproducibility (containers, devcontainers, nix) |
| 6 | **Code Quality** | Cohesion, modularity, dead-code hygiene |
| 7 | **Observability** | Logs, traces, metrics — every agent action visible |
| 8 | **Security and Governance** | Policy enforcement, secrets handling, auditability |

## How to Use

1. Score each pillar 1–5 against the target codebase.
2. Weak pillars are the **first** place to invest before adding more agent capability.
3. Track scores over time; treat them as engineering OKRs.

The most immediate concrete first step is [[action-implement-strict-linting]].

## Why It Works

Every pillar maps to a way agents "cheat." Loose linting → messy commits. No tests → no feedback. Poor docs → hallucinated API calls. No observability → invisible failures. The framework systematically closes shortcuts.

## Adjacent Benchmarks

- **SWE-Bench** (https://www.swebench.com/) — agents on real GitHub issues; environment quality > reasoning is a recurring finding.
- **AgentBench**, **WebArena** — environment-readiness emphasis.

## See Also

- [[concept-agent-environment-readiness]]
- [[entity-factory-ai-d41]]
- [[framework-rob-pike-agent-rules]] — the code-quality counterpart
- [[action-implement-strict-linting]] — concrete first step
