OpenAI has significantly scaled its Trusted Access for Cyber program, making its latest GPT-5.5 model and a specialized variant, GPT-5.5-Cyber, available to thousands of vetted cybersecurity professionals and hundreds of teams as of . This expansion grants approved users access to models with intentionally reduced safeguards, specifically designed to accelerate critical defensive cyber activities like vulnerability research, defensive programming, and automated red-teaming, acknowledging the dual-use nature of advanced AI in the security domain.
- OpenAI is expanding its Trusted Access for Cyber program to thousands of individual defenders and hundreds of teams.
- Approved users gain access to GPT-5.5 and a specialized GPT-5.5-Cyber model with fewer safeguards for cyber-related tasks.
- GPT-5.5-Cyber is designed for advanced defensive workflows, including binary reverse engineering and vulnerability validation.
- The AI Security Institute (AISI) has evaluated GPT-5.5, noting its strong capabilities for cyber exploits, including comparison to Anthropic’s Mythos.
- This initiative aims to accelerate vulnerability research, defensive programming, and automated red-teaming of critical systems.
What changed
Previously, OpenAI’s models, including earlier GPT versions, operated under stringent safeguards designed to prevent misuse. While beneficial for general public use, these safeguards often triggered on legitimate cybersecurity activities that involve analyzing potentially malicious code or simulating attacks for defensive purposes. The core change is the expansion of the Trusted Access for Cyber program, which now provides vetted cybersecurity professionals with versions of GPT-5.5 and a new, more permissive model, GPT-5.5-Cyber, explicitly designed for these dual-use scenarios [1, 4, 7].
GPT-5.5-Cyber is a scaled-up version of OpenAI’s latest model, tailored for advanced defensive workflows such as binary reverse engineering and validating high-severity vulnerabilities [1, 4, 5]. This specialized access allows a smaller set of partners to explore complex scenarios where typical safeguards might impede legitimate security research [8]. The program’s scaling means access is no longer limited to a small alpha group but is being rolled out to thousands of individual defenders and hundreds of teams [5].
How it works
Access to GPT-5.5 and GPT-5.5-Cyber is granted through an application and vetting process for OpenAI’s Trusted Access for Cyber program. Once approved, users receive access to API endpoints or interfaces for these models. The key differentiator is the adjusted safeguard layer: these models have “reduced friction around safeguards which might trigger on dual-use cyber activity” [1]. This enables security professionals to perform tasks like analyzing malware, generating proof-of-concept exploits for vulnerabilities, or red-teaming critical systems without encountering the same content moderation barriers that a public-facing model would enforce [1].
The specialized GPT-5.5-Cyber model is particularly tuned for “advanced defensive workflows” [5]. This suggests fine-tuning on cybersecurity-specific datasets and potentially architectural optimizations to handle tasks like static and dynamic analysis of binaries, exploit development for defensive testing, and automated vulnerability discovery. The AI Security Institute (AISI) has evaluated GPT-5.5, noting its “strongest model we have tested” capabilities for cyber exploits, indicating its proficiency in understanding and generating code relevant to security vulnerabilities [2, 3].
Why it matters for operators
For cybersecurity operators, this is not just another AI announcement; it’s a strategic shift in how OpenAI views and supports the defensive cyber landscape. The explicit reduction of safeguards for vetted users acknowledges a critical reality: security tools often operate in a grey area that general-purpose AI models are designed to avoid. This move empowers defenders by providing them with powerful AI capabilities that can genuinely accelerate tasks like vulnerability research and automated red-teaming, which are currently bottlenecked by human expertise and manual effort [1].
However, operators must approach this with a clear-eyed understanding of the implications. While the promise of scaling defensive capabilities is immense, the dual-use nature of these models means that the same capabilities can, in theory, be leveraged by adversaries if access controls or vetting processes are compromised. Therefore, organizations integrating GPT-5.5-Cyber into their workflows must establish robust internal governance and ethical guidelines. This isn’t a plug-and-play solution; it’s a powerful tool that demands sophisticated operational control and continuous monitoring. The real value for an operator lies not just in the model’s raw power, but in how effectively it can be integrated into existing security pipelines to augment human analysts, not replace them, especially in the nuanced and high-stakes world of critical infrastructure protection [1]. Expect a new class of “AI-assisted red teamers” to emerge, capable of probing systems with unprecedented speed and depth, shifting the burden of defense to an even more proactive stance.
Benchmarks and evidence
During alpha testing, GPT-5.5-Cyber was already used to “scale automated red-teaming of critical systems and validate high-severity vulnerabilities” [1]. The AI Security Institute (AISI) evaluated GPT-5.5, noting it “may be the strongest model we have tested” for cyber exploits [2]. AISI’s evaluation further indicates that performance on “TLO” (likely a cyber-specific benchmark, though not explicitly defined in the sources) continues to scale with increased inference compute, suggesting that the model’s capabilities are not yet plateauing [3]. This continuous improvement across model generations at fixed token budgets underscores the ongoing advancement in AI’s ability to handle complex cyber tasks [3].
Risks and open questions
- Adversarial Access: While OpenAI emphasizes vetting, the inherent dual-use nature of these models means that any compromise of the Trusted Access program could provide sophisticated tools to malicious actors. How robust are the vetting and ongoing monitoring processes to prevent this?
- Ethical Guidelines and Misuse: Even with vetted users, the potential for unintended misuse or ethical dilemmas remains. What specific ethical guidelines are provided to users of GPT-5.5-Cyber, especially when dealing with sensitive vulnerability data or critical infrastructure?
- Transparency and Explainability: For security professionals, understanding why an AI identifies a vulnerability or suggests a particular exploit is crucial for validation and remediation. How transparent are the internal workings of GPT-5.5-Cyber, and what tools are provided for explainability?
- Dependency on a Single Vendor: Relying heavily on a proprietary model from a single vendor for critical defensive capabilities could introduce vendor lock-in risks and potential points of failure. What are the long-term implications for the cybersecurity industry’s ecosystem?
- Impact on Skill Development: While AI can augment human capabilities, there’s a risk that over-reliance could diminish foundational skills in vulnerability research and manual red-teaming among new professionals. How will this impact the development of human expertise?
Sources
- Scaling Trusted Access for Cyber with GPT-5.5 and GPT-5.5-Cyber | OpenAI
- r/OpenAI on Reddit: Scaling Trusted Access for Cyber with GPT-5.5 and GPT-5.5-Cyber
- Our evaluation of OpenAI’s GPT-5.5 cyber capabilities | AISI Work
- OpenAI rolls out advanced AI cyber model to challenge Anthropic’s Mythos – POLITICO
- Oracle CISO Perspective: Mythos, GPT 5.5-Cyber, and the CISO’s New Threat Model | ateam
- OpenAI says is scaling trusted access for cyber with GPT-5.5 and GPT-5.5-cyber | MarketScreener
- OpenAI makes GPT-5.5 more widely available to cyber defenders
- OpenAI rolls out new GPT-5.5-Cyber to vetted cybersecurity teams