Skip to main content
Cybersecurity professional verifying data on a secure network, ensuring trusted access and accountability.

Editorial illustration for Trusted access enables defenders through verification and accountability

AI Security: Defenders' Trust in Verification Wins

Trusted access enables defenders through verification and accountability

2 min read

The rise of generative AI has reshaped how security teams operate, but the tools that protect us are only as strong as the people who wield them. While new language models can churn out code, phishing scripts, or deep‑fakes in seconds, defenders must keep pace without exposing critical systems to unnecessary risk. That tension has sparked a debate: should access be opened broadly to accelerate response, or locked down to prevent misuse?

Recent discussions suggest a middle ground—granting permissions that are both verifiable and auditable, so that every action can be traced back to a trusted source. As the underlying models become more capable, the protective measures must evolve in lockstep, mirroring the same trajectory of improvement. The challenge, then, is not just building smarter defenses, but ensuring those defenses are scalable, accountable, and grounded in clear trust signals.

This perspective frames the next point about how we intend to broaden legitimate defender access while maintaining rigorous verification.

Instead, we aim to enable as many legitimate defenders as possible, with access grounded in verification, trust signals, and accountability. - Defenses should be continually scaled with capability. As model capabilities increase, defenses need to scale alongside them.

We've seen steady improvements in agentic coding, which have direct implications for cybersecurity and we've adapted our approach in step. - We began cyber-specific safety training with GPT‑5.2, then expanded it with additional safeguards through GPT‑5.3‑Codex and GPT‑5.4, where we also classified the model as "high" cyber capability under our Preparedness Framework. In parallel, we increased support for defenders: launching a $10M Cybersecurity Grant Program, reached over 1,000 open source projects with Codex for Open Source(opens in a new window) which provides free security scanning, and continued to improve Codex Security.

Thousands of verified defenders will soon have TAC access. Hundreds of teams are also slated to join. The program rests on democratized access, iterative deployment, and resilience, yet how those principles translate at scale remains uncertain.

Fine‑tuning models for defensive use cases shows steady improvement, but the upcoming surge in OpenAI’s capabilities could outpace current safeguards. The quote stresses verification, trust signals, and accountability as foundations; whether those signals can keep pace with rapid model evolution is still unclear. Defenses are expected to grow alongside model power, a premise that the article acknowledges without detailing concrete metrics.

If verification processes hold, legitimate defenders may gain broader tools; if not, gaps could appear. Can verification keep up with the pace? The initiative’s ambition is evident, but measurable outcomes have yet to be reported.

Ultimately, the success of scaling TAC will depend on how effectively verification and accountability mechanisms can be maintained as AI models become more capable. Monitoring will be essential to assess real‑world impact.

Further Reading

Common Questions Answered

How does OpenAI approach trusted access for cybersecurity defenders?

OpenAI aims to enable legitimate defenders through a carefully managed access model based on verification, trust signals, and accountability. The approach involves scaling defenses alongside increasing model capabilities, with a focus on cyber-specific safety training and controlled deployment.

What are the key principles behind OpenAI's defender access program?

The program is built on three core principles: democratized access, iterative deployment, and resilience. Thousands of verified defenders are expected to receive TAC (Trusted Access Credentials) access, with hundreds of teams preparing to join the initiative.

What challenges does OpenAI face in scaling cybersecurity defenses with AI?

The primary challenge is balancing broad access with system security, ensuring that defenders can effectively respond to threats without exposing critical systems to unnecessary risks. There are ongoing uncertainties about how the principles of verification and accountability will translate at a larger scale.