The "GPT-5 will show AGI" hype was always a ridiculously high bar for OpenAI, and I would argue that the quest for that elusive AGI threshold has been an unnecessary curse on machine learning and AI development in general. Who cares? Do we really want to replace humans? We should want better and more reliable tools (like Claude Code) to assist people, and maybe cover some of the stuff nobody wants to do. This desire for "AGI" is delivering less value and causing us to put focus on creative tasks that humans actually want to do, putting added stress on the job market.
The one really bad sign in the launch, at least to me, was that the developers were openly admitting that they now trust GPT-5 to develop their software MORE than themselves ("more often than not, we defer to what GPT-5 says"). Why would you be proud of this?
The idea that models “feel” smarter may be 100% human psychology. If you invest in a new product, admitting that it isn’t better than what you had is hard for humans. So, if users say a model “feels” smarter, we won’t know that it really is smarter.
Also, if users manage to improve quality of responses after using it for a while, who says they couldn’t have reached similar results if they stayed using the old tool, tweaking their prompts to make that model perform better?
AGI doesn't really replace humans, it merely provides a unified model that can be hooked up to carry out any number of tasks. Fundamentally no different than how we already write bespoke firmware for every appliance, except instead of needing specialized code for each case, you can simply use the same program for everything. To that extent, software developers have always been trying to replace humans — so the answer from the HN crowd is a resounding yes!
> We should want better and more reliable tools
Which is what AGI enables. AGI isn't a sentience that rises up to destroy us. There may be some future where technology does that, but that's not what we call AGI. As before, it is no different than us writing bespoke software for every situation, except instead of needing a different program for every situation, you have one program that can be installed into a number of situations. Need a controller for your washing machine? Install the AGI software. Need a controller for your car's engine? Install the same AGI software!
It will replace the need to write a lot of new software, but I suppose that is ultimately okay. Technology replaced the loom operator, and while it may have been devastating to those who lost their loom operator jobs, is anyone today upset about not having to operate a loom? We found even more interesting work to do.
I appreciate the well-crafted response, but respectfully disagree with this sentiment, and I think it's a subtle point. Remember the no free lunch theorems: no general program will be the best at all tasks. Competent LLMs provide an excellent prior from which a compelling program for a particular task can be obtained by finetuning. But this is not what OpenAI, Google, and Anthropic (to a lesser extent) are interested in, as they don't really facilitate it. It's never been a priority.
They want to create a digital entity for the purpose of supremacy. Aside from DeepMind, these groups really don't care about how this tech can assist in problems that need solving, like drug discovery or climate prediction or discovery of new materials (e.g. batteries) or automation of hell jobs. They only care about code assistance to accelerate their own progress. I talk to their researchers at conferences and it frustrates me to no end. They want to show off how "human-like" their model is, how it resembles humans in creative writing and painting, how it beats humans on fun math and coding competitions that were designed for humans with a limited capacity to memorize, how it provides "better" medical opinions than a trained physician. That last use case is pushing governments to outlaw LLMs for medicine entirely.
A lab that claims to push toward AGI is not interested in assisting mankind toward a brighter future. They want to be the first for bragging rights, hype, VC funding, and control.
Isn't it obvious? They have a huge vested interest in getting people to believe that it's very useful, capable, etc.
Unfortunately for a substantial number of people the answer to this question seems to be a resounding "yes"
The other 99% would like automation to make their lives easier. Who wouldn't want the promised tech utopia? Unfortunately, that's not happening so it's understandable that people are more concerned than joyous about AI.
Whether it's indicative of patterns beyond OpenAI remains to be seen, but I don't expect much originality from tech execs.
Only about 5 minutes of the whole presentation are dedicated to enterprise usage (COO in an interview sort of indirectly confirms that haven't figured it out yet). And they are cutting the costs already (opaque routing between models for non-API users is a clear sign of that). The term "AGI" is dropped, no more exponential scaling bullshit - just incremental changes over the time and only over select few domains. Actually it is a more welcoming sign and not concerning at all that this technology matures and crystallizes around this point. We will charitably forget and forgive all the insane claims made by Sam Altman in the previous years. He can also forget about cutting ties with Microsoft for that same reason.
Also note that they're losing money on their paid subscribers.
"Reflections" by Sam Altman, January 2025 - https://blog.samaltman.com/reflections
I find this pattern in tech hype really frustrating. Someone in a leadership role in a major tech company/VC promises something outrageous. Time passes and the promise never materializes. People then retcon the idea that "everybody knew that wasn't going to happen". Well either "everybody" doesn't include Elon Musk[1], Sam Altman, or Marc Andreessen[2] or these people are liars. No one seems to be held to their track record of being right or wrong, instead people just latch on to the next outrageous promise as if the previous one was fulfilled.
[1] https://electrek.co/2025/03/18/elon-musk-biggest-lie-tesla-v...
[2] https://dailyhodl.com/2022/06/01/billionaire-and-tech-pionee...
Did we just get scammed right in front of our eyes with an overhyped release and what is now an underwhelming model if the point was that GPT-5 was supposed to be trustworthy enough for serious use-cases and it can't even count or reason about letters?
So much for the "AGI has been achieved internally" nonsense with the VCs and paid shills on X/Twitter bullshitting about the model before the release.
It doesn't understand anything. It can't possibly "understand my codebase". It can only predict tokens, and it can only be useful if the pattern has been seen before. Even then, it will product buggy replicas, which I've pointed out during demos. I disabled the ai helpers in my IDEs because the slop the produce is not high quality code, often wrong, often misses what I wanted to achieve, often subtly buggy. I don't have the patience to deal with that, and I don't want to waste the time on it.
Time is another aspect of this conversation, with people claiming time wins, but the data not backing it up, possibly due to a number of factors intrinsic to our squishy evolved brains. If you're interested, go find gurwinder's article on social media and time - I think the same forces are at work in the ai-faithful.
I think most of us are in the camp that even though we don't need AI right now we believe we will not be valuable in the near future without being highly proficient with the tooling.
This reads to me like you don't think you're valuable right now either
Well done model routing is a tremendous leap forward to minimize the latency & improve the user experience.
E.g. I love Gemini 2.5 Pro. But it's darn slow (sorry GDM!). I love the latency I'm getting from 4o. The solution? Just combine them under one prompt, with well done model routing.
Is GPT5 router "good enough"? We'll see.
I think OpenAI is a smart company. And Sama is a tremendous leader. They're moving in the right direction.
Which reminds me that one of the most obvious failings of LLMs is they never say "I've been thinking about that and have a new idea." The thinking leaning thing needs work.
Maine was #89 (That is not a typo.) and Oregon was #1.
OpenAI as a company simply cannot exist without a constant influx of investor money. Burning money on every request is not a viable business model. Companies built on OpenAI/Anthropic are similarly deeply unprofitable businesses.
OpenAI needs to convert to a for-profit to get any more of the funding that Softbank promised (that its also unclear how Softbank itself would raise) or to get significant cash from anyone else. Microsoft can block this and probably will.
It all reminds me of that Paddy's Dollars bit from it's always sunny.
"We have no money and no inventory... there's still something we can do... that's still a business somehow..."
PhDs need to catch up!
That money they burned was on customer acquisition, building infrastructure, etc. The unit economics of paying to be driven to the airport or Benihanas was always net positive.
They weren't losing money on every customer, even paying ones. There just isn't a business model here.
I wouldn't say they had no edge. They had a huge advantage over traditional taxi companies. You can argue that a local Uber-like app could be easily implemented, that's where the investors came in to flood the markets and ensure other couldn't compete easily.
The situation is in no way similar to OpenAI's. OpenAI truly has no edge over Anthropic and others. AGI is not magically emerging from LLM's and they don't seem to have an alternative (nobody does but they promised it and got the big bucks so now it's their problem).
OpenAI competes with google, who can drop 50B/y into AI hype for very long time.
TBD. Some people did well while Uber gave money away, but Uber is not net profitable over its lifetime.
So let's imagine is 2040 and OpenAI is finally profitable. Now, Uber did this by increasing prices, firing some staff and paying smaller wages to drivers. And all this while having near-monopoly in certain areas. What realistic measures would they need to take in order to compete with, say, Google? Because I just wish them good luck with that.
I have no idea if OpenAI succeeds or not but I find arguments like yours difficult to understand. Most businesses are not using these systems to draw a map. Maybe the release of 5 is lackluster but it does not change that there is some value in these tools today and ignoring R&D (which is definitely a huge cost) they run at a profit.
how can you say such a hand wavy comment with a straight face? you can't just ignore a huge cost for a company and suddenly they are profitable. that's Enron level moronic. without constant R&D, the company gets beat by competitors that continue R&D. the product is not "good enough" to not continue improving.
if i ignored my major costs in my finances, i could retire, but i can't go to the grocery store and walk out with a basket of food while telling them that i'm ignoring this major cost in my life.
get real
A company that can pull in single digit billions in revenue for hundreds of billions in expenses just doesn't make sense.
> Most businesses are not using these systems t̶o̶ ̶d̶r̶a̶w̶ ̶a̶ ̶m̶a̶p̶.̶
FTFY
And no - while it might be obvious from the outside in that it probably won't happen, the continued existence of the business is still predicated on conversion to a for-profit. They don't just need the amount of money they've already "raised", they need too keep getting more money forever.
But this company is valued more than Netflix. The bar should not be this low.
This observation + sherlocking cursor suggests that perhaps sherlocking is the ideation strategy. Curious to see if they’re subsidizing token costs specifically to farm and Sherlock ideas
It hardly feels like a next generation release.
As a related anecdote (not saying that this is industry standard, just pointing out my own experience), the startup I work for launched their app four years ago, and, for all four of those years, we've had "Implement a Dark Mode design" sitting at the bottom of our own backlog. Higher priority feature requests are always pre-empting it.
PMs operating at this level ought to be bringing in some low cost UX improvements alongside major features. That simply isn't a sign that they've run ought of backlog. (That said, it is rather pathetic to paywall this)
A moment's consideration ought to show that Open AI has plenty of significant work they they can be doing, even if the core model never gets any better than this.