OpenAI Launches Safety Fellowship Amid Damning Investigation and Dissolved Teams
TECH

OpenAI Launches Safety Fellowship Amid Damning Investigation and Dissolved Teams

30+
Signals

Strategic Overview

  • 01.
    OpenAI announced the OpenAI Safety Fellowship on April 6, 2026, a pilot program running from September 14, 2026 to February 5, 2027 that provides mentorship, compute resources, and a monthly stipend to independent researchers working on AI safety and alignment.
  • 02.
    The fellowship targets researchers with proposals addressing safety evaluation, ethics, robustness, scalable mitigation strategies, privacy-preserving methods, agentic oversight, and high-severity misuse domains, with fellows expected to produce research papers or datasets.
  • 03.
    The announcement arrived on the same day The New Yorker published an 18-month investigation by Ronan Farrow and Andrew Marantz into OpenAI's safety record and Sam Altman's leadership, prompting widespread skepticism about the timing.
  • 04.
    OpenAI has dissolved three dedicated internal safety teams in under two years — the Superalignment team (May 2024), the AGI Readiness team (October 2024), and the Mission Alignment team (February 2026) — raising questions about the credibility of an externally-focused safety program.

The Timing Collision: A Fellowship Launched in the Shadow of an Investigation

The most striking feature of the OpenAI Safety Fellowship announcement is not its content but its calendar. On April 6, 2026, OpenAI published details of its new fellowship program supporting independent AI safety research. Within hours, Ronan Farrow and Andrew Marantz's 18-month investigation into OpenAI landed in The New Yorker, drawing on over 200 pages of internal documents and interviews with more than 100 people. The investigation raised serious questions about Sam Altman's trustworthiness and OpenAI's actual commitment to safety.

Farrow himself drew the connection publicly on X, writing that the fellowship announcement "arrives hours after our investigation described how OpenAI dissolved its superalignment and AGI-readiness teams and dropped safety from the list of its most significant activities on its IRS filings." His post received approximately 284,000 views and 3,600 likes, establishing the dominant public narrative around the fellowship as one of strategic timing rather than genuine institutional commitment. The crypto news outlet BSCN amplified the framing with its headline: "OPENAI ANNOUNCES SAFETY FELLOWSHIP HOURS AFTER ARTICLE INVESTIGATING CEO'S INTEGRITY." Whether the timing was coincidental or deliberate, the fellowship's launch is now inseparable from the investigation's revelations.

Three Teams in Two Years: OpenAI's Serial Dissolution of Internal Safety Groups

The fellowship announcement must be understood against a documented institutional pattern. Since mid-2024, OpenAI has dissolved three dedicated internal safety teams in rapid succession. The Superalignment team, created with a public pledge of 20% of OpenAI's compute resources, was disbanded in May 2024 after reportedly receiving only 1-2% of compute on the company's oldest hardware. Its co-leaders — Ilya Sutskever and Jan Leike — both departed, with Leike publicly stating that safety had taken a backseat to product development.

Five months later, in October 2024, the AGI Readiness team was dissolved when its head advisor Miles Brundage resigned. Then in February 2026, just two months before the fellowship announcement, OpenAI shut down its Mission Alignment team after only 16 months of operation. Each dissolution was accompanied by the departure of senior safety-focused personnel. This sequence raises a fundamental question: if OpenAI could not sustain internal safety teams with direct access to its models and infrastructure, what does an external fellowship with a five-month duration and limited compute access realistically accomplish? Critics argue the pattern suggests that safety work is valued primarily as a public relations asset rather than as a core organizational function.

External Fellowship vs. Internal Investment: A Structural Mismatch

The fellowship's structure reveals an interesting strategic choice. Rather than rebuilding internal safety capacity — which OpenAI has repeatedly built and dismantled — the company is now outsourcing safety research to independent scholars through a time-limited program. Fellows receive mentorship, compute resources, and a monthly stipend over approximately five months, and are expected to produce research papers or datasets. The program explicitly targets "the next generation of talent" and prioritizes proposals on safety evaluation, ethics, robustness, scalable mitigation, privacy-preserving methods, agentic oversight, and high-severity misuse.

This externalization model has both legitimate benefits and structural limitations. On the positive side, independent researchers may produce work that is less constrained by corporate incentives and more credible to regulators and the public. The research areas listed align with genuine frontier safety challenges. On the other hand, external fellows lack the deep model access that internal teams had, their five-month engagement is far shorter than the multi-year timelines safety research typically requires, and the program's outputs — papers and datasets — may have limited impact on OpenAI's actual deployment decisions. The EU AI Act's requirements for safety assessments of high-risk AI systems may also be a motivating factor, as fellowship-generated research could help demonstrate regulatory compliance without requiring permanent internal investment.

The AI Safety Talent War: OpenAI vs. Anthropic Fellowships

The fellowship does not exist in a competitive vacuum. Anthropic, founded by former OpenAI safety lead Dario Amodei, is simultaneously running its own Anthropic Fellows Program with applications open for May and July 2026 cohorts. This parallel creation of external safety fellowship programs by the two leading frontier AI companies signals that AI safety talent has become a strategic resource worth competing for.

The competition carries an ironic undertone. Amodei left OpenAI in part because of concerns about its safety culture — concerns he documented in over 200 pages of internal notes — and has characterized the fundamental problem at OpenAI as being "Sam himself." Now both companies are vying for the same pool of independent safety researchers. For prospective fellows, this competition is arguably beneficial: it creates more funded positions, more mentorship opportunities, and more institutional support for safety work. But it also raises the question of whether these programs will produce research that meaningfully influences how either company develops and deploys its models, or whether they primarily serve as talent scouting and reputation-building exercises in an increasingly scrutinized industry.

What the Internal Documents Reveal About Safety Culture

The New Yorker investigation provides crucial context that the fellowship announcement alone cannot. According to reporting, Altman internally expressed that "vibes don't match a lot of the traditional AI-safety stuff," suggesting a philosophical skepticism toward the very field the fellowship aims to support. Dario Amodei's 200-plus pages of internal notes reportedly paint a picture of an organization where safety concerns were systematically deprioritized.

Gary Marcus framed the stakes bluntly, asking whether anyone would want Altman "deciding, unilaterally, whether or not it is ok to release" a model capable of enabling bioweapons or cyberattacks. The investigation also revealed that OpenAI dropped safety from its list of most significant activities on its IRS filings — a bureaucratic detail that may carry more weight than any public announcement. These revelations suggest a gap between OpenAI's public safety narrative and its internal priorities. The fellowship, regardless of its operational merits, now exists within this credibility gap. Its ultimate significance will depend less on the quality of research it produces and more on whether OpenAI demonstrates sustained institutional commitment to safety beyond time-limited external programs.

Historical Context

2023-07-01
OpenAI pledged one-fifth of its computing power to a superalignment team focused on preventing AI-caused human extinction, but the team reportedly received only 1-2% of resources on the oldest hardware.
2024-05-17
OpenAI disbanded its Superalignment team after co-leaders Ilya Sutskever and Jan Leike departed. Leike publicly criticized that 'safety culture and processes have taken a backseat to shiny products.'
2024-09-25
CTO Mira Murati, Chief Research Officer Bob McGrew, and VP of Research Barret Zoph all departed OpenAI on the same day. The Mission Alignment team was created as a replacement for prior safety efforts.
2024-10-24
OpenAI disbanded its AGI Readiness team as head advisor Miles Brundage resigned.
2026-02-12
OpenAI dissolved its Mission Alignment team after just 16 months of operation, marking the third internal safety team to be disbanded.
2026-04-06
OpenAI announced the Safety Fellowship on the same day The New Yorker published Farrow and Marantz's 18-month investigation into the company's safety record and Altman's leadership.

Power Map

Key Players
Subject

OpenAI Launches Safety Fellowship Amid Damning Investigation and Dissolved Teams

OP

OpenAI

Organizer and funder of the Safety Fellowship; faces scrutiny over dissolving three internal safety teams while launching an external program.

SA

Sam Altman

CEO of OpenAI; subject of a concurrent New Yorker investigation questioning his trustworthiness and commitment to safety.

RO

Ronan Farrow and Andrew Marantz

New Yorker investigative journalists whose 18-month investigation into OpenAI published on the same day as the fellowship announcement.

AN

Anthropic

Competitor running a parallel Anthropic Fellows Program for AI safety research, with applications open for May and July 2026 cohorts.

IL

Ilya Sutskever

Former OpenAI co-founder and chief scientist who departed in 2024 and founded Safe Superintelligence Inc; authored internal memos critical of Altman's leadership.

THE SIGNAL.

Analysts

"Farrow directly challenged the fellowship's credibility by noting it was announced hours after his investigation revealed that OpenAI had dissolved its superalignment and AGI-readiness teams and removed safety from its IRS filings. His framing — linking the timing to reputation management — set the dominant narrative on social media, with his quote-tweet garnering approximately 284,000 views and 3,600 likes."

Ronan Farrow
Investigative Journalist, The New Yorker

"Marcus questioned whether Altman should hold unilateral safety decision-making power, asking: "If some future OpenAI model could enable a massive bioweapon or cyberattack, would you really want Altman deciding, unilaterally, whether or not it is ok to release the model?""

Gary Marcus
AI Researcher and Critic

"Amodei, who compiled over 200 pages of internal notes critical of OpenAI's safety culture, characterized the core issue succinctly: "The problem with OpenAI is Sam himself.""

Dario Amodei
CEO of Anthropic, former OpenAI safety lead

"Altman reportedly expressed skepticism toward traditional AI safety approaches internally, stating that "vibes don't match a lot of the traditional AI-safety stuff," raising questions about the sincerity of the fellowship's safety mission."

Sam Altman
CEO, OpenAI
The Crowd

"This announcement arrives hours after our investigation (newyorker.com/magazine/2026/) described how OpenAI dissolved its superalignment and AGI-readiness teams and dropped safety from the list of its most significant activities on its IRS filings"

@@RonanFarrow3600

"(8/11) Some former OpenAI researchers argue that the company has forfeited its original safety mission and accelerated an industry-wide race to the bottom. The piece details a set of public and internal safety commitments that former researchers say were abandoned."

@@RonanFarrow0

"TECH: OPENAI ANNOUNCES SAFETY FELLOWSHIP HOURS AFTER ARTICLE INVESTIGATING CEO INTEGRITY DEBUTS BY ME TOO JOURNALIST."

@@BSCNews0
Broadcast