08/14 2025
361
On August 7, GPT-5 debuted with four models (Regular, Mini, Nano, and Pro). However, just five days later, on August 12, Sam Altman announced on X that GPT-4o would resume its role as the default model for all paid users.
This rapid turnaround, from introduction to retraction, mirrors OpenAI's hurried rollback during the ChatGPT "downtime" incident in November 2023. Unlike the previous technical failure, this move was a strategic self-correction.
VentureBeat obtained backend logs revealing three significant issues in GPT-5's first week:
To stem the negative feedback, OpenAI swiftly reverted to the default model. Altman's assurance echoed as a promise: "If GPT-4o is ever removed again, we will give you ample notice."
In industry terms, this signifies that GPT-5 is not yet ready for full-scale production.
User "Model Attachment Disorder": The First "Fandomization" of AI Products
It may seem surprising that large AI models can evoke such strong emotional attachments.
Independent developer Alex tweeted about his VSCode plugin, noting that GPT-4o's coding style "feels like a silent, trusted partner." A Japanese illustrator even compiled GPT-4o's responses into a book titled "4o Poetry Collection." There was even a Change.org petition demanding the permanent retention of GPT-4o's "personality parameters."
This is no joke but a genuine "model personality stickiness" recently recognized by the OpenAI product team. When LLMs become daily tools for millions of creators, their "tone" becomes a critical factor in productivity.
Altman noted on internal Slack: "We underestimated users' sensitivity to 'personality consistency'."
Thus, the next GPT-5 version will introduce a "temperature dial":
This marks the first time an "AI skin system" has been introduced, not to change colors but to alter the essence.
Hidden Costs: The "Electricity Bill" of Inference Mode
How costly is GPT-5's "Thinking" mode?
These are just consumer-end costs. The enterprise API pricing is even more eye-opening:
Mode | Input / 1M tokens | Output / 1M tokens | Relative Increase from GPT-4o |
---|---|---|---|
GPT-5 Thinking | $15 | $60 | +400% |
GPT-5 Fast | $5 | $15 | +50% |
GPT-4o | $3 | $10 | Baseline |
Electricity bills, graphics cards, and carbon emissions have made "unlimited context" a luxury. According to exclusive data from the Bit.ly/4mwGngO salon:
OpenAI's VP of Infrastructure conceded in a private meeting: "The rise in inference costs is outpacing the decline predicted by Moore's Law."
Efficiency vs. Expansion: The "Triple Point" of Scaling Law
For the past five years, the AI industry adhered to the belief that "the larger the parameters, the better the performance." Now, we are reaching a critical juncture of "expansion-efficiency-sustainability":
As a result, three new paths have emerged:
In summary, "bigness" is no longer the sole selling point, and "efficiency" is the cornerstone of the next funding round.
OpenAI's "Multi-threaded" Future: One Launch Event, Three Business Models
Viewing GPT-5's rocky launch and rollback within OpenAI's broader commercial landscape reveals a "synchronized test" of three revenue streams:
This incident has reordered the priorities of these three models:
As AI Enters the Era of "Intensive Cultivation"
GPT-5's rocky start echoes Apple's removal of the headphone jack from the iPhone 7 in 2016:
In the next 12 months, we can expect:
AI is no longer a "brute force miracle" black box but a meticulously engineered business. Even Altman concedes:
"Our enemy is not competitors, but the laws of physics."