Will Ilya Sutskever Create Safe Superintelligence?

Picture of Julia McCoy

Julia McCoy

Creator and Co-founder

safe superintelligence

If you’ve been following the world of artificial intelligence, you know about Ilya Sutskever’s groundbreaking work as OpenAI’s chief scientist. But the latest chapter in his career involves a fascinating move: launching his own company, Safe Superintelligence Incorporated (SSI Inc).

What makes this move significant?

Why is the OpenAI co-founder now focused on safe superintelligence?

What’s on the horizon for Ilya Sutskever’s SSI?

We’re going to unpack the why, what, and how behind Sutskever’s bold new venture and explore the potential implications for the future of AI.

From OpenAI to SSI: A Quest for Safe Superintelligence

Ilya Sutskever is not a new name in AI. He was a co-founder at OpenAI, the research lab behind the viral sensation ChatGPT, along with other tech giants like Elon Musk and OpenAI CEO Sam Altman.

Sutskever’s departure from OpenAI raised a lot of eyebrows because of how fast the company grew from 2023 into 2024.

His departure led some people online (including Elon Musk) to question if there were short-term commercial pressures he wanted to avoid.

Perhaps Sutskever saw a problem on the horizon with Artificial General Intelligence (AGI), given how rapidly the industry is moving forward with minimal regulation.

What’s different about Ilya Sutskever’s new company is its singular focus: tackling the monumental task of building safe superintelligence.

Why “Safe Superintelligence” Matters

The idea of superintelligence — AI that surpasses humans in general intelligence — can feel like something straight out of a science fiction film. But as technology keeps accelerating, superintelligence isn’t remaining in the realm of fiction for long.

And that’s where the “safe” part becomes incredibly crucial.

Here’s the heart of the issue: imagine an AI so intelligent that it’s capable of outthinking and outmaneuvering any human.

If this superintelligence isn’t aligned with human values and ethics from the ground up, the potential risks could be significant.

The concept of AI going rogue isn’t new. But it’s not about killer robots like we’re used to seeing on the big screen.

It’s more about unintended consequences and unforeseen ripple effects that could occur when you have a system potentially more powerful than the ones that created it.

That’s the challenge and mission that Sutskever’s new company is aiming to solve, making it one of the most ambitious and important startups of our time.

Unpacking SSI Inc: A Closer Look

Founded in June 2024, Safe Superintelligence Incorporated is on a mission to “pursue safe superintelligence”.

SSI Inc. describes itself as “the world’s first straight-shot SSI lab with one goal and one core product: building safe superintelligence.”

There’s no ambiguity about their intentions; they’re aiming directly for the frontier of artificial superintelligence (ASI).

And, they plan to do so with an unwavering focus on the paramount importance of its alignment with human values and safety protocols.

Sutskever understands that building such groundbreaking technology isn’t a one-man show. He’s assembled a “cracked” team of some of the most talented engineers and researchers in the field of AI.

His co-founders, Daniel Gross and Daniel Levy, are forces to be reckoned with.

Gross comes with experience from building the Q&A search engine that was acquired by Apple back in 2013. Meanwhile, Levy was a critical contributor to large AI models during his time at OpenAI.

What Does This Mean for the Future of AI?

While only time will reveal what this new AI company ultimately achieves, its mere existence is already making waves throughout the field of artificial intelligence.

First, Sutskever’s venture increases awareness of a crucial issue in the AI race — the essential need for superintelligence development that puts ethical considerations and safety at the forefront, not just afterthoughts.

There’s always been an underlying tension between pushing the boundaries of innovation and making sure we aren’t building something that could ultimately harm us.

It’s very possible that in the coming years, other researchers, businesses, and policymakers will feel the pressure and choose to prioritize safety as more than just a box to tick. It’ll hopefully become increasingly clear that a superintelligence future demands rigorous ethical frameworks, responsible development practices, and collaboration on a global scale.


Ilya Sutskever’s new company, SSI Inc., represents an exciting, somewhat daunting, but crucial step in artificial intelligence development.

As AI technologies evolve at a speed that continuously surprises even the most ardent supporters, a focus on “safe superintelligence” is more critical than ever.

Become a Content Hacker

Collaborate with other growth-minded, savvy Content Hackers – plus a host of experts. The price of two coffees/month gets you in.