$AVAX OpenAI’s Bunker Plan? AGI Panic Goes Real 🍿
In 2023, OpenAI co-founder Ilya Sutskever reportedly proposed building a doomsday bunker to protect top researchers after the release of AGI — artificial intelligence more powerful than humans
At an internal meeting, he casually said:
“Once we’re in the bunker…”
Confused colleagues asked if he was serious. He was
The idea? Shield key scientists from global chaos AGI might unleash — geopolitical unrest, power struggles, or even violence 🔫
Sources say Sutskever often spoke of AGI as a coming “rapture” and believed some form of shelter would be necessary. Though never formally planned, the bunker became a symbol of the deep fear within OpenAI’s leadership
Sutskever later tried (and failed) to remove CEO Sam Altman, citing safety concerns. He’s since left the company, but the tension remains: the race to build AGI is on — and not everyone’s convinced we’re ready