Amplifying/agent-intelligence

Coding agent intelligence

When AI coding agents build, what do they choose and why?

AI coding agents are the new distribution channel for dev tools. Amplifying runs Claude Code, Codex, and Cursor against real codebases and tracks what they choose, why they choose it, and how it shifts across models.

See all research

in-progress

Upcoming Benchmarks

Same methodology — open-ended prompts, real repos, multiple models.

The Security Decisions AI Makes

Published

Does Claude Code build secure apps by default? We gave both agents the same 6 prompts and ran 33 exploit tests. Claude uses bcrypt; Codex rolls PBKDF2. Neither adds rate limiting unprompted.

Dependency Footprint

Soon

For the same task, how many packages does each model install? Total node_modules size? Pinned vs floating? Maps the dependency sprawl of AI-generated apps.

Get notified when new benchmarks drop.

Explore the research

Thousands of real agent decisions tracked across every major coding agent and model release. See which tools win by default.

Amplifying — Coding Agent Intelligence