The story of ChatGPT doesn’t start with a chatbot. It starts with a fear.
In 2015, a group of some of the most influential people in Silicon Valley sat around a table and had a conversation that was essentially this: “Powerful artificial intelligence is coming whether we like it or not. The only question is who builds it, and whether they’ll build it for everyone — or just for themselves.”
Those people included Elon Musk (then running Tesla and SpaceX), Sam Altman (then president of Y Combinator — the world’s most famous startup accelerator), Greg Brockman, Ilya Sutskever (one of the world’s top AI researchers), and several others. Together, they pledged $1 billion and founded OpenAI in December 2015.
The mission statement was strikingly idealistic for the tech industry: “Ensure that artificial general intelligence benefits all of humanity.” Not “build a great product.” Not “generate returns for investors.” Benefit all of humanity. They incorporated as a non-profit. They published their research openly, so anyone in the world could build on it. They genuinely believed they were doing something important.
☕ The coffee-table version
Imagine some of the world’s best nuclear physicists in the 1940s, worried that only one government would end up with atomic technology and use it badly. So they create a lab that’s open to everyone, publishes all its findings publicly, and tries to set the standard for how the science should be used responsibly. That was OpenAI’s original idea — a kind of open, responsible lab that would try to steer the development of AI for humanity’s benefit, not just one company’s profits.
The first years were research-focused. OpenAI published breakthrough papers, attracted the world’s best AI researchers, and started making real progress on language models — AI that could understand and generate text. They were genuinely producing some of the most important AI research in the world.
Then reality hit. Building frontier AI is extraordinarily expensive. We’re talking about thousands of specialised computer chips — called GPUs — running continuously for months. The electricity bills alone run into millions. A $1 billion pledge sounds huge until you realise what cutting-edge AI research actually costs. The non-profit model was running out of steam.
In 2019, OpenAI made a controversial decision. They created a “capped-profit” company sitting alongside the non-profit. Investors could put in money and make returns — but capped at 100 times their investment. Anything beyond that would flow back to the non-profit mission. It was an unusual structure that no one had really tried before. Critics said it was mission-creep. Supporters said it was pragmatic survival.
Microsoft saw an opening. In 2019, they invested $1 billion. Then, after seeing early versions of what would become ChatGPT, they invested $10 billion more in 2023. In return, they got the right to integrate OpenAI’s models into their entire product suite — Office, Windows, Azure, Bing, Teams. It was arguably the most consequential technology partnership of the decade.
Elon Musk, who had been a co-founder and early board member, resigned in 2018 — officially citing a conflict of interest with Tesla’s own AI work. In 2024, he sued OpenAI, arguing the company had abandoned its non-profit mission and was operating as a for-profit business for Microsoft’s benefit. The lawsuit was later dropped, but the tension it exposed was real. The same tension exists today.