Sam Altman Says the GPT-5 Haters Got It All Wrong

sam-altman-says-the-gpt-5-haters-got-it-all-wrong

OpenAI’s August launch of its GPT-5 large language model was somewhat of a disaster. There were glitches during the livestream, with the model generating charts with obviously inaccurate numbers. In a Reddit AMA with OpenAI employees, users complained that the new model wasn’t friendly, and called for the company to restore the previous version. Most of all, critics griped that GPT-5 fell short of the stratospheric expectations that OpenAI has been juicing for years. Promised as a game changer, GPT-5 might have indeed played the game better. But it was still the same game.

Skeptics seized on the moment to proclaim the end of the AI boom. Some even predicted the beginning of another AI Winter. “GPT-5 was the most hyped AI system of all time,” full-time bubble-popper Gary Marcus told me during his packed schedule of victory laps. “It was supposed to deliver two things, AGI and PhD-level cognition, and it didn’t deliver either of those.” What’s more, he says, the seemingly lackluster new model is proof that OpenAI’s ticket to AGI—massively scaling up data and chip sets to make its systems exponentially smarter—can no longer be punched. For once, Marcus’ views were echoed by a sizable portion of the AI community. In the days following launch, GPT-5 was looking like AI’s version of New Coke.

Sam Altman isn’t having it. A month after the launch he strolls into a conference room at the company’s newish headquarters in San Francisco’s Mission Bay neighborhood, eager to explain to me and my colleague Kylie Robison that GPT-5 is everything that he’d been touting, and that all is well in his epic quest for AGI. “The vibes were kind of bad at launch,” he admits. “But now they’re great.” Yes, great. It’s true the criticism has died down. Indeed, the company’s recent release of a mind-bending tool to generate impressive AI video slop has diverted the narrative from the disappointing GPT-5 debut. The message from Altman, though, is that naysayers are on the wrong side of history. The journey to AGI, he insists, is still on track.

Numbers Game

Critics might see GPT-5 as the waning end of an AI summer, but Altman and team argue that it cements AI technology as an indispensable tutor, a search-engine-killing information source, and, especially, a sophisticated collaborator for scientists and coders. Altman claims that users are beginning to see it his way. “GPT-5 is the first time where people are, ‘Holy fuck. It’s doing this important piece of physics.’ Or a biologist is saying, ‘Wow, it just really helped me figure this thing out,’” he says. “There’s something important happening that did not happen with any pre-GPT-5 model, which is the beginning of AI helping accelerate the rate of discovering new science.” (OpenAI hasn’t cited who those physicists or biologists are.)

So why the tepid initial reception? Altman and his team have sussed out several reasons. One, they say, is that since GPT-4 hit the streets, the company delivered versions that were themselves transformational, particularly the sophisticated reasoning modes they added. “The jump from 4 to 5 was bigger than the jump from 3 to 4,” Altman says. “We just had a lot of stuff along the way.” OpenAI president Greg Brockman agrees: “I’m not shocked that many people had that [underwhelmed] reaction, because we’ve been showing our hand.”

OpenAI also says that since GPT-5 is optimized for specialized uses like doing science or coding, everyday users are taking a while to appreciate its virtues. “Most people are not physics researchers,” Altman observes. As Mark Chen, OpenAI’s head of research, explains it, unless you’re a math whiz yourself, you won’t care much that GPT-5 ranks in the top five of Math Olympians, whereas last year the system ranked in the top 200.

As for the charge about how GPT-5 shows that scaling doesn’t work, OpenAI says that comes from a misunderstanding. Unlike previous models, GPT-5 didn’t get its major advances from a massively bigger dataset and tons more computation. The new model got its gains from reinforcement learning, a technique that relies on expert humans giving it feedback. Brockman says that OpenAI had developed its models to the point where they could produce their own data to power the reinforcement learning cycle. “When the model is dumb, all you want to do is train a bigger version of it,” he says. “When the model is smart, you want to sample from it. You want to train on its own data.”

Altman and company seem stung by the criticism of GPT-5 and respond with disbelief that people think that the scaling hypothesis is moot. OpenAI has obviously not given up on massive scale. That’s the reason it’s spending hundreds of billions of dollars that it doesn’t have to build giant datacenters in Abilene, Texas and other locations. Brockman indicates that until those giant computation factories come online, there simply isn’t enough firepower to make the next big leaps. “The scaling challenge is hard,” he says. It’s like, really, really hard to execute, like building a bigger rocket—building a 2x bigger rocker is probably 10x harder.”

When I mention Marcus to Altman, he’s visibly offended. “Is that a real question?” he asks. Well, it’s not only Gary, I say. Altman straightens up in his chair. “What I can tell you with confidence is GPT-6 will be significantly better than GPT-5, and GPT-7 will be significantly better than GPT-6. And we have a pretty good track record on these.”

AGI or Bust

Altman spent many months this year talking about how AGI was imminent. Lately, however, OpenAI has been steering people away from the idea that AGI is a destination. Now it’s a process. Sounds logical, but the rhetorical tweak liberates the company from a deadline. “We had almost a category error of thinking of OpenAI as a project with a defined end date,” says Brockman. “We were thinking, ‘OK, if we just build AGI, and make it good for humanity, that’s what we’re here to do. That is not how we think about it anymore.” Now, he says it’s more of a never-ending rollout. “The mission is really about this continuous impact, and transforming the economy into this AI-powered world. And even if AGI is a mile marker—reasonably well defined, maybe a little bit of fuzziness—there’s this continuous exponential.”

When I bring up AGI with Altman, he says that the discussion might be less than useful, since people have such different ideas of what it means. Employing one of his favorite bits of interview jiu-jitsu, he asks us to share our definitions of AGI, as if they make any difference to OpenAI. In its charter, OpenAI defines AGI as “highly autonomous systems that outperform humans at most economically valuable work.” Altman now says his view has evolved beyond the charter. His thinking on AGI seems to center on scientific acumen. “We can wrap our heads around what it means for most economic work to happen,” he says. “But the scientific progress definition is really a big deal for the world. It’s hard to wrap our heads around that, so we talk about it less.”

GPT-5, then, is a modest step toward that milestone. “I would not claim that GPT-5 is like doing meaningful science, obviously not,” Altman says. “But there is a glimmer, and I think by 6 or 7, we’ll see more of it.”

Despite AGI’s “fuzziness,” as Brockman puts it, OpenAI is having a branding moment with the acronym. Literally in the room with Altman as he mused over definitions was his PR minder, whose laptop had a sticker that read, “FEEL THE AGI.” A merch kiosk on the first floor of the company’s headquarters sells T-shirts with the same legend. Hanging in the hallways are marketing posters highlighting issues on the road to AGI and explaining phenomena such as superintelligence and “AI puberty,” defined as “the transitional stage where a system shifts from narrow AI toward more general human-like intelligence, with unpredictable or awkward behavior.” Whether AGI is a process or a destination, OpenAI is linked with it forever. And it’s willing to spend hundreds of billions of dollars to scale its way there.

Additional reporting by Kylie Robison.


This is an edition of Steven Levy’s Backchannel newsletter. Read previous newsletters here.

Related Posts

Leave a Reply