OpenAI launches GPT-5.4-Cyber for approved security teams, expanding the Trusted Access program.
In summary, OpenAI is launching GPT-5.4-Cyber, a model specialized in defensive cybersecurity with reduced refusal limits and capabilities for binary reverse engineering, while also expanding its Trusted Access for Cyber program to thousands of accredited defenders. This decision follows Anthropic's move to limit its more advanced Mythos model to merely 11 organizations, creating a philosophical divide: OpenAI favors widespread verified access, while Anthropic chooses a more restricted deployment strategy.
OpenAI will provide its most advanced cybersecurity model to thousands of vetted defenders, introducing GPT-5.4-Cyber and broadening its Trusted Access for Cyber program in direct response to Anthropic’s announcement of Project Glasswing last week.
GPT-5.4-Cyber is a specialized version of GPT-5.4 tailored for defensive security applications. Its key feature is a lower refusal threshold: while standard models typically block sensitive inquiries related to vulnerability research, exploit assessments, or malware behavior, this variant is designed to respond to such queries if the user is validated as a legitimate security expert. Additionally, the model includes binary reverse engineering capabilities, allowing analysts to investigate compiled software for vulnerabilities without needing access to the source code.
Scaling up Trusted Access for Cyber
The model operates within OpenAI’s Trusted Access for Cyber (TAC) initiative, first introduced in February alongside a $10 million cybersecurity grant fund. TAC serves as a framework that regulates access to more advanced models based on verification stages. Users can authenticate at chatgpt.com/cyber, while enterprises can request team-wide access via an OpenAI representative. Security researchers requiring the most extensive capabilities can apply for an exclusive tier.
The April update transitions the program from a limited pilot to what OpenAI characterizes as “thousands of verified individual defenders and hundreds of teams responsible for defending critical software.” The company is introducing new tiers, allowing higher verification levels to unlock more powerful features. Users approved for the highest tier will gain access to GPT-5.4-Cyber, but there is a caveat: top-tier users may need to forgo Zero-Data Retention, meaning OpenAI will maintain oversight of how the model is utilized.
This approach marks a philosophical shift: instead of relying predominantly on model-level restrictions to curb misuse, OpenAI is adopting an access-control model that verifies the identity of the user before determining the model's responses. The company aligns this shift with three principles: democratized access through objective verification criteria, iterative deployment that updates safety mechanisms as risks develop, and ecosystem resilience supported by grants and open-source efforts.
Understanding the Anthropic context
OpenAI's timing is significant when viewed in relation to Anthropic's Project Glasswing, announced on April 7. Anthropic disclosed that its Claude Mythos Preview model autonomously identified thousands of zero-day vulnerabilities across major operating systems and web browsers, including a 27-year-old bug in OpenBSD and a 17-year-old remote code execution flaw in FreeBSD, which Mythos uncovered, exploited, and documented without human input.
In response, Anthropic has severely restricted access to Mythos Preview, making it available exclusively to 11 organizations, such as Apple, Google, Microsoft, AWS, Cisco, CrowdStrike, and JPMorgan Chase, through a $100 million defensive initiative. The model will not be made publicly available, and Anthropic has indicated it may remain that way due to the risk of misuse stemming from its exploit-generation abilities.
In contrast, OpenAI is opting for broader access. While GPT-5.4-Cyber is less proficient than Mythos in sheer vulnerability discovery, it is being made accessible to a much wider audience. The underlying rationale is that limiting powerful security tools to a few technology giants leaves the majority of organizations—including those tasked with protecting critical infrastructure, hospitals, municipal governments, and small security firms—without access to equally effective defensive technology.
Capabilities of GPT-5.4-Cyber
In addition to its reduced refusal boundaries, the model is designed for workflows that standard ChatGPT handles inadequately or outright denies. The standout feature is binary reverse engineering: security analysts can input compiled executables into the model to receive analyses regarding potential malware behavior, embedded vulnerabilities, and structural flaws. Such analyses typically necessitate specialized tools like IDA Pro or Ghidra, along with substantial manual expertise.
The model can also engage with dual-use inquiries, such as questions about attack techniques, exploit chains, and classes of vulnerabilities—topics that standard models often flag as risky. OpenAI states that earlier versions sometimes declined to address legitimate defensive inquiries, creating barriers for security professionals needing the model to analyze adversarial techniques for their defense strategies.
Codex Security, OpenAI’s automated code-scanning tool, enhances the model's capabilities. Since its launch, Codex Security has facilitated over 3,000 critical and high-severity vulnerability fixes across the open-source landscape, covering more than 1,000 open-source projects through a free scanning initiative.
The dual-use dilemma
Other articles
OpenAI launches GPT-5.4-Cyber for approved security teams, expanding the Trusted Access program.
OpenAI introduces GPT-5.4-Cyber, featuring binary reverse engineering for validated defenders, expanding access to thousands while competing with Anthropic's limited Mythos model.
