Amplifying/ai-benchmarks

amplifying/intelligence · for-investors

AI agents are reshaping market share. Who's winning?

Tens of thousands of real agent decisions, tracked across every major model release.

For VCs & Growth Equity

Which categories are still open? Which portfolio companies are invisible to agents? Multiple categories have no clear winner yet, making them fundable white spaces before agents lock in a default.

Get the Investor Report

For Public Market Investors

Which public company tech are agents adopting? Which are they skipping? Every model release shifts defaults. Those shifts show up here quarters before they hit reported revenue.

Get the Investor Report
Request Custom Analysis

By ticker, category, or portfolio.

Coverage areas

Beyond the 20 categories in our public benchmark, these are areas we're actively researching or already covering for institutional clients.

GPU / ML SDK

When AI agents write ML training and inference code, do they default to CUDA or ROCm?

NVIDIA (CUDA)AMD (ROCm)Intel (oneAPI)

PyTorch defaults to CUDA. That's not a hardware decision. It's an import statement. But import statements are where lock-in actually happens in agent-generated code. The question for AMD and Intel isn't benchmarks or pricing. It's how often, across thousands of AI projects, an agent changes one default. That number doesn't exist in any earnings call.

Cloud Provider

When agents provision infrastructure, which cloud do they default to?

AWSGoogle CloudAzure

Enterprise cloud decisions take months of vendor evaluation. An agent makes the same choice in a single scaffolding step, and it cascades: the provider determines the storage layer, the auth model, the deployment target. The three hyperscalers report revenue, not agent adoption. That gap matters because agent platform choices today predict where the next generation of workloads land.

Vector Database

Which vector DB do agents reach for when building RAG and AI-native apps?

PineconeWeaviateChromapgvector

Zero VC funding five years ago. Over $1B now. But the category's survival depends on a behavior nobody has measured: do agents treat vector search as a standalone infrastructure problem, or as a feature of the database they're already using? The answer probably varies by project complexity. That distribution is the whole game.

Observability at Scale

Sentry leads in our current data, but for production infrastructure, who wins?

DatadogGrafana LabsNew RelicSentry

Observability vendors sell to engineering leadership. Agents have no manager to approve a vendor. 22% of projects build custom logging. The relevant question isn't the level. It's the trajectory: a flat 22% is one story. A number that rises with each model generation is a very different one for Datadog.

LLM Orchestration

Do agents use LangChain, LlamaIndex, or just call APIs directly?

LangChainLlamaIndexDirect SDK calls

Orchestration frameworks assume that calling a model API is complex enough to need abstraction. But agents are themselves LLMs calling APIs. This is the one category where the tool and the user are the same technology. The $50M+ in VC funding is a bet that the abstraction layer adds value even when the consumer is an LLM. Whether that's true shows up in what agents actually import.

Message / Event Streaming

Do agents use managed Kafka, cloud-native queues, or lightweight alternatives for event-driven architectures?

Confluent (Kafka)Apache KafkaAWS SQS / SNSGoogle Pub/Sub

Kafka is powerful and complex. SQS is simple and effectively free at low scale. In most categories, agents gravitate toward the simpler tool. But Kafka isn't just complex. It's the industry standard. Whether agents respect that standard or route around it tells you something about Confluent's ~$10B valuation that quarterly earnings can't.

Infrastructure as Code

Terraform, Pulumi, or SST? What do agents default to for cloud provisioning?

TerraformPulumiSSTAWS CDK

Terraform uses HCL. Pulumi and CDK use TypeScript. Agents write TypeScript. That's a structural advantage, but for which TypeScript-native tool? IBM paid $5B for Terraform on the assumption that HCL's ecosystem is the moat. If agents don't write HCL, that assumption needs revisiting. The specific alternative they pick is a separate, equally important question.

Data Pipeline

When agents build data workflows, do they reach for Snowflake or Databricks?

SnowflakeDatabricksDuckDB

Snowflake and Databricks both require accounts, credentials, and network configuration. DuckDB runs in-process with zero setup. Agents prefer zero-config tools in smaller categories. Whether that preference holds at data-warehouse scale matters, because the pattern would suggest the modern data stack is infrastructure or overhead.

CDN / Edge Computing

When agents deploy, do they push to the edge by default, and if so, whose edge?

CloudflareFastlyVercel EdgeAWS CloudFront

Most edge computing analysis focuses on latency. The more interesting variable is distribution. Every AI-generated project that deploys to Cloudflare Workers is a customer acquired through code, not sales. Whether that's happening, and at what rate, is a distribution signal that doesn't appear in any existing dataset.

Auth / Identity Platform

Our data shows 48% Custom/DIY for auth. But which vendor captures the other 52%?

Auth0 (Okta)ClerkFirebase AuthSupabase Auth

Most categories settle into a clear build-or-buy ratio. Auth is split 48/52. That near-even balance means small shifts in agent behavior swing real vendor share. The question inside the 52% is whether Auth0 still holds default status or Clerk has quietly taken it. Okta paid $6.5B for Auth0. Where agents actually default is one signal worth having.

Get early results

Institutional partners get preliminary findings before public release.

Request Access

Do you know where your portfolio companies rank?

Standard reports updated with every model release. Custom analysis by ticker, category, or portfolio.

For Investors | AI Agent Ecosystem Intelligence — Amplifying