AI Security

AI-BOM Adoption: State of the Art in 2026

The AI Bill of Materials went from concept paper to procurement requirement in under two years. Here is what the current state of the art actually looks like.

Shadab Khan
AI Security Researcher
6 min read

The AI Bill of Materials moved from concept paper to procurement requirement faster than most of its proponents expected. Two years ago, AI-BOM was a niche conversation among ML security researchers and a few forward-looking standards bodies. Today, federal procurement language in the United States, sectoral regulators in the EU, and mature enterprise buyers all routinely ask for an AI-BOM as part of vendor onboarding. The speed of that shift has exposed both the genuine value of the concept and the places where the format-and-tooling ecosystem has not yet caught up.

Why AI-BOM Adoption Moved Fast

Software Bill of Materials adoption took the better part of a decade to reach its current level, and it is still uneven. AI-BOM adoption, by contrast, has moved in under two years. The acceleration is not mysterious — the SBOM conversation did the political work already. Regulators, buyers, and security teams already believed that supply chain transparency was necessary. Extending that belief to AI components required an analogy and a vocabulary, not a fresh argument.

The other accelerator has been the visible friction of not having an AI-BOM. Several high-profile supply chain incidents through 2025 — a compromised model on a major registry, a backdoored embedding library, a fine-tuning dataset found to contain poisoned samples — were incidents in which the affected organizations could not quickly answer the question "where is this component used?" That inability became an executive-level pain, and AI-BOM became a plausible answer.

What the Format Ecosystem Looks Like

Two formats dominate current practice, with one clear third. CycloneDX added AI/ML component types in a series of extensions that have now converged into the core specification, and it is the most widely adopted format in enterprise tooling. SPDX's AI profile has also stabilized and has strong traction in open-source communities and regulated government contexts. A handful of purpose-built formats exist, but none have achieved broad interoperability, and most vendors who started there are now emitting CycloneDX or SPDX alongside their native format.

The field set that actually matters in production AI-BOMs has stabilized around a shorter list than early specifications suggested. In practice, the fields that consumers read are:

  • Model identity and version, including hash of weights
  • Base model lineage, for fine-tuned or adapted models
  • Training data characterization at a category level, with references to provenance documentation
  • Evaluation results against named benchmarks and any disclosed safety evaluations
  • Dependencies at the framework, runtime, and tokenizer level
  • License and usage terms
  • Known limitations and intended-use statements

Additional fields exist in the specifications, but consumers who read AI-BOMs programmatically tend to focus on the list above, and vendors who populate only those fields well are, in practice, passing procurement.

Where AI-BOM Generation Actually Happens

AI-BOM generation has split into three patterns, roughly by model type.

Frontier and foundation models. The major labs publish AI-BOM-equivalent documentation as part of their model release or as an addendum to it. The level of detail varies — some labs publish extensive system cards that map cleanly to AI-BOM fields; others publish summaries that require follow-up for any but the simplest consumer. The trend is toward more detail under regulatory pressure.

Fine-tuned and enterprise-adapted models. This is where most of the AI-BOM generation tooling is concentrated. Platform vendors have added AI-BOM emission to their training pipelines, and the output is routinely stored alongside model artifacts. The quality of these outputs depends heavily on how disciplined the underlying pipeline is; a well-instrumented pipeline emits a high-quality AI-BOM as a byproduct, and a messy pipeline produces a messy AI-BOM.

Embedded and bundled models. Models that ship as part of a larger software product — the ML component in a security tool, the embedding model in a database, the small model bundled in a mobile SDK — are the hardest case. AI-BOM emission here has lagged, and it is where most of the interesting tooling innovation will happen over the next year.

The Consumer Side Has Caught Up Faster Than Expected

A year ago, a common objection to AI-BOM was that there were no consumers — that even if vendors emitted them, no buyer would actually use them. That objection has weakened considerably. Enterprise buyers now ingest AI-BOMs into the same inventory systems they use for SBOMs, and a small but growing number of organizations run automated policy evaluation against AI-BOM contents at procurement time.

The policies that get enforced are not yet sophisticated. Common examples include rejecting models trained on datasets that are on an internal disallow list, flagging models whose license terms conflict with a given use case, and alerting on base-model lineage that includes a known compromised checkpoint. These are coarse-grained controls, but they are real, and they are the precursors to more nuanced automated evaluation.

The Gaps That Remain

Three gaps are worth naming.

First, runtime AI-BOM — the question of what model and configuration is actually in production at any given moment — is still underdeveloped. Static AI-BOMs describe what was released; runtime discovery tools that confirm what is actually running are immature, and the gap creates an audit risk similar to the one SBOMs had before runtime SCA tools matured.

Second, dependency depth for AI components is still shallow in most AI-BOMs. A model's tokenizer, for example, has its own lineage and its own risks, and tokenizer-level issues are almost invisible in current AI-BOM outputs. The next specification revision is expected to address this, but tooling will lag.

Third, AI-BOM for agent configurations is essentially non-existent. An agent is not just a model; it is a model plus tools plus prompts plus policies plus memory. Describing the relevant components of an agent in an AI-BOM-like document is a conversation that is just starting, and it is where the frontier will be over the next eighteen months.

What Enterprise Security Teams Should Do Now

Enterprise security teams reviewing their own AI-BOM posture should be able to answer three questions. Does every model we run in production have an AI-BOM on file, in a standard format, ingested into our inventory system? Do we have policy checks that run automatically against those AI-BOMs and alert on violations? When an upstream incident occurs — a compromised model, a tainted dataset — can we produce an impact report from our inventory in under an hour?

Teams that answer yes to all three are ahead of the median. Teams that cannot answer any of the three should treat AI-BOM as a near-term program rather than an aspiration. The regulatory direction, the procurement direction, and the incident response direction all point the same way, and the cost of catching up later is measurably higher than the cost of catching up now.

AI-BOM is not a silver bullet, and nobody responsible has claimed it is. But it is a baseline, and in 2026, operating without one is no longer a defensible position for any organization that takes AI security seriously.

Never miss an update

Weekly insights on software supply chain security, delivered to your inbox.