Ilya Sutskever, co-founder of OpenAI and one of the most esteemed AI researchers globally, has launched a new start-up named Safe Superintelligence (SSI) Inc.
This development comes merely a month after Sutskever’s departure from OpenAI, following an unsuccessful attempt to oust its CEO, Sam Altman.
SSI Inc. is positioned as “the world’s first straight-shot SSI lab, with one goal and one product: a safe superintelligence,” according to a statement released on X.
Tekedia Mini-MBA edition 16 (Feb 10 – May 3, 2025) opens registrations; register today for early bird discounts.
Tekedia AI in Business Masterclass opens registrations here.
Join Tekedia Capital Syndicate and invest in Africa’s finest startups here.
“We approach safety and capabilities in tandem, as technical problems to be solved through revolutionary engineering and scientific breakthroughs. We plan to advance capabilities as fast as possible while making sure our safety always remains ahead,” the statement said.
Sutskever has co-founded this pioneering initiative with Daniel Levy, a former OpenAI employee, and Daniel Gross, an AI investor and entrepreneur with stakes in prominent tech companies like GitHub and Instacart, as well as AI ventures including Perplexity.ai, Character.ai, and CoreWeave.
The founders emphasize that SSI’s mission is singular and undistracted by the need for revenue generation, allowing them to attract top talent dedicated solely to the development of safe superintelligence—an advanced form of AI that could surpass human cognitive abilities. This focus is intended to free the company from the short-term commercial pressures that can compromise safety and ethical considerations.
“Our singular focus means no distraction by management overhead or product cycles, and our business model means safety, security, and progress are all insulated from short-term commercial pressures,” Sutskever said on X.
SSI will operate with headquarters in both Palo Alto and Tel Aviv.
Sutskever’s departure from OpenAI followed a period of internal turbulence at the leading AI company. In November, OpenAI’s board, which included Sutskever at the time, made a controversial decision to oust Altman. The move, which shocked investors and staff alike, was quickly reversed, with Altman reinstated under a new board configuration, leading to Sutskever’s resignation in May. At the time of his departure, Sutskever hinted at an exciting new project that held personal significance for him.
The founders of SSI assert that their exclusive focus on developing safe superintelligence will insulate their work from the distractions of management overhead and product development cycles. This approach harks back to OpenAI’s original mission when it was founded in 2015 as a non-profit research lab aimed at creating beneficial superintelligent AI.
Under Altman’s leadership, OpenAI has transformed from a non-profit research institution into a rapidly expanding business. Despite this growth, the company has faced internal strife regarding its leadership direction and the prioritization of AI safety. Jan Leike, another recent OpenAI departure who worked closely with Sutskever, joined the rival start-up Anthropic, citing growing tensions over the diminishing emphasis on safety protocols at OpenAI.
The Rise of Spin-Offs
SSI is not the first significant spin-off from OpenAI. In 2021, Dario Amodei, formerly head of AI safety at OpenAI, founded Anthropic, a start-up committed to developing safe AI systems. Anthropic has since secured $4 billion in funding from Amazon and hundreds of millions more from venture capitalists, achieving a valuation exceeding $18 billion.
Widening The Safety Concern
The rapid proliferation of AI companies like SSI and Anthropic underlines the urgent need for robust AI regulation. As these companies push the boundaries of AI capabilities, the potential risks associated with superintelligent AI—ranging from ethical concerns to safety and security issues—become increasingly pronounced.
The influx of new AI enterprises necessitates comprehensive regulatory frameworks to ensure that advancements in AI technology are managed responsibly and ethically.
Regulators worldwide are beginning to recognize this need. The European Union’s AI Act, for instance, aims to create a legal framework to manage the risks associated with AI. Similarly, in the United States, there is growing bipartisan support for more stringent AI regulations to safeguard against potential misuse and ensure that AI development aligns with broader societal values.
The announcement of SSI has garnered significant attention within the tech community. Sutskever’s reputation as a leading AI researcher and his instrumental role in OpenAI’s early successes lend substantial credibility to the new venture. This move also highlights ongoing concerns about AI safety and governance within the rapidly evolving industry.