Ilya Sutskever, OpenAI’s former chief scientist, launches new AI company

Ilya Sutskever launches Safe Superintelligence Inc. after leaving OpenAI to develop safe superintelligent AI with a dedicated team and a for-profit model.

: Ilya Sutskever has started a new venture, Safe Superintelligence Inc. (SSI), following his departure from OpenAI. The company aims to develop safe superintelligent AI, working on balancing safety and capabilities. Unlike OpenAI, SSI is structured as a for-profit entity from the ground up.

Ilya Sutskever, who was the chief scientist and co-founder of OpenAI, has launched a new company called Safe Superintelligence Inc. (SSI) just a month after leaving. He is joined by former Y Combinator partner Daniel Gross and ex-OpenAI engineer Daniel Levy in this new venture, focusing on building safe superintelligent AI.

At OpenAI, Sutskever worked on AI safety, particularly concerning superintelligent AI systems, a task he handled with Jan Leike, who now works at Anthropic. Sutskever has long highlighted the potential dangers of superintelligent AI, emphasizing the importance of controlling and restricting such technology to ensure safety.

SSI is being designed as a for-profit entity, with an intense focus on developing safe superintelligent AI by solving technical problems through innovative engineering and scientific breakthroughs. The company is currently recruiting talent in Palo Alto and Tel Aviv and is expected to attract significant capital due to the high interest in AI and the team's expertise.