AI Security

Griffin AI vs GPT-5: Enterprise Controls

Frontier models offer impressive enterprise features. Security programs need deeper controls than chat can provide—controls that live in the engine around the model.

Shadab Khan
Head of AI Research
7 min read

A fair question we hear from CISOs evaluating AI for security programs is some variant of: "GPT-5 has enterprise SSO, admin consoles, data residency options, and audit logs. Why isn't that enough?" The short answer is that enterprise controls over a chat product are not the same as enterprise controls over a security workflow, and the gap between them is where most of the interesting problems live.

As with every post in this series, the framing matters. Griffin AI uses frontier reasoning from Anthropic's Claude family under the hood. We are not building a replacement for GPT-5. We are building the engine that wraps frontier reasoning in the enterprise controls security programs actually require when the output drives real action inside the organization.

What enterprise controls look like on a frontier chat product

Credit first. The major frontier vendors have done real work on enterprise features. You get single sign-on, centrally managed seats, admin policies that restrict what users can do with the chat, data residency commitments, retention controls, and audit logs of conversations. For an organization using AI as a productivity tool—drafting emails, summarizing documents, exploring ideas—these controls are substantial and often sufficient.

The controls are designed around the mental model that the AI is a conversation surface. Users chat, admins manage the chat, and the controls govern the chat. That mental model is correct for the shape of the product. It is not the mental model of a security program.

The controls a security program actually needs

A security program has a different control surface. The AI is not a chat surface; it is a system that reads from and writes to the organization's security infrastructure. The controls required are correspondingly deeper:

  • Per-action authorization. Who is allowed to run a particular scan, open a particular ticket, or close a particular finding? This is not a user-level toggle; it is an action-level policy evaluated at the moment of the action.
  • Resource-level scoping. A user may be allowed to view findings for their team's services and not for others. The AI's tool calls must respect this, consistently and verifiably.
  • Change approvals. High-impact actions—closing a critical finding, approving a deployment gate, modifying a policy—often require human approval. The workflow has to integrate with the organization's existing approval processes.
  • Separation of duties. The user who requests a change should not be the user who approves it. The system should enforce this, not hope for it.
  • Immutable audit. Every query, every tool call, every data access must be recorded in a form that an auditor will accept. Chat logs are not audit logs.
  • Break-glass and revocation. If an account is compromised or an integration is abused, the organization needs to disable it immediately, across all workflows, without waiting for a vendor support ticket.

These are not features a chat product was built to provide. They are features a security platform has to be built around from day one.

How Griffin structures enterprise controls

Griffin's control surface is designed for the shape of a security workflow. The primitives are:

Identity and access. Griffin integrates with the organization's identity provider and honors group memberships, role assignments, and attribute-based policies. Every session carries an authenticated identity, and every action is evaluated against that identity's permissions. There is no "service account" shortcut that lets the AI act outside a user's scope.

Policy-driven action gating. Griffin's action layer is declarative. An administrator can write a policy stating, for example, "Findings with CVSS above 9.0 may not be auto-closed by AI without human approval." The engine enforces this regardless of what the model suggests. The model can recommend; the policy decides whether the recommendation becomes an action.

Integration-level scopes. When Griffin is connected to Jira, Slack, GitHub, or ServiceNow, the scopes granted to each integration are narrow and explicit. The AI cannot post to arbitrary channels, modify arbitrary repositories, or create arbitrary tickets. The tools it can call are bound to configured targets.

Approval workflows. High-impact actions flow through the organization's existing approval process. Griffin can draft a remediation plan; a designated reviewer approves it; the action executes. This is not an AI-native process invented for Griffin. It is the organization's existing process, extended to include AI-generated proposals.

Audit that assessors accept. Every user action, every AI action, every tool call, and every policy evaluation is recorded in an append-only audit log with full traceability. The log answers "what happened, who initiated it, what did the AI see, and what did the policy decide?"—in a format that holds up under SOC 2, ISO 27001, and equivalent frameworks.

Tenant isolation by default. Griffin's architecture isolates tenants at every layer: storage, compute, retrieval, and integration. A cross-tenant data leak is not prevented by careful prompt engineering; it is prevented by architecture.

A concrete example

Suppose an enterprise has a policy that only the security team can close findings marked as critical. A developer asks an AI assistant to close a critical finding that they believe is a false positive.

In a pure chat workflow with enterprise controls, the AI might refuse, might comply, or might provide instructions on how the developer could close the finding themselves. The outcome depends on prompt engineering, not policy. The audit trail, if any, is a chat log.

In a Griffin workflow, the developer's request reaches the action layer. The action layer evaluates the policy: "close_finding on severity=critical requires security-team role." The developer does not have that role. The action is denied, with a clear explanation routed back through the AI's response. An audit record is written capturing the attempt and the denial. The frontier model reasons about the denial and suggests the appropriate next step—filing a request for security review, not closing the finding.

The model is the same. The enterprise control surface is what changed the outcome.

Where frontier vendor controls still matter

It would be wrong to suggest that frontier model vendors do not care about enterprise concerns. They do, and the features they offer are real. Griffin depends on some of them—in particular, the data handling commitments that make it safe for us to route reasoning through a hosted frontier model. A Griffin deployment would not be defensible without the vendor's data handling posture underneath it.

The point is that the vendor's controls apply to the model's surface. They do not and cannot apply to the workflows your security team runs around the model. Those workflows need their own control plane, and that control plane is what Griffin provides.

Comparing the surfaces side by side

A useful way to think about it is that a frontier chat product offers controls over "what the user can do with AI." Griffin offers controls over "what AI can do with the organization's security systems." The first is necessary. The second is the one your auditor will ask about.

A second useful framing is scope. A chat product's admin console scopes usage. Griffin's policy engine scopes actions. The difference is that usage is reversible—if a user misuses a chat session, the impact is limited to the conversation. Actions are not reversible—a misplaced ticket, a closed finding, a deployed change have real-world consequences.

What to evaluate

When evaluating any AI tool for security work against enterprise standards, the useful questions are:

  • Does the tool integrate with our identity provider and respect group-based access controls?
  • Can we express policies that gate specific actions, not just specific users?
  • Is the audit log comprehensive enough for our compliance program to rely on?
  • How does the tool handle separation of duties for high-impact actions?
  • What is the isolation posture for tenant data, both at rest and during model inference?

These questions are not a checklist of features. They are an examination of whether the tool's architecture takes enterprise controls seriously. A frontier chat product with good answers to user-level questions may have no answers at all to action-level questions. That gap is where Griffin lives.

The positioning

Griffin AI uses frontier reasoning. The frontier model vendors are excellent partners and their enterprise features are real. Griffin adds the control plane that turns frontier reasoning into a security program your compliance team will stand behind. The model is not replaceable, and we do not try. The controls are also not optional, and we do not skip them.

Never miss an update

Weekly insights on software supply chain security, delivered to your inbox.