I think people are massively underestimating the money they will come from ads in the future.
They generated $4.3B in revenue without any advertising program to monetise their 700 million weekly active users, most of whom use the free product.
Google earns essentially all of its revenue from ads, $264B in 2024. ChatGPT has more consumer trust than Google at this point, and numerous ways of inserting sponsored results, which they’re starting to experiment with with the recent announcement of direct checkout.
The biggest concern IMO is how good the open weight models coming out of China are, on consumer hardware. But as long as OpenAI remains the go-to for the average consumer, they’ll be fine.
What is OpenAI's competitive moat? There's no product stickiness here.
What prevents people from just using Google, who can build AI stuff into their existing massive search/ads/video/email/browser infrastructure?
Normal, non-technical users can't tell the difference between these models at all, so their usage numbers are highly dependent on marketing. Google has massive distribution with world-wide brands that people already know, trust, and pay for, especially in enterprise.
Google doesn't have to go to the private markets to raise capital, they can spend as much of their own money as they like to market the living hell out of this stuff, just like they did with Chrome. The clock is running on OpenAI. At some point OpenAI's investors are going to want their money back.
I'm not saying Google is going to win, but if I had to bet on which company's money runs out faster, I'm not betting against Google.
Consumer brand quality is so massively underrated by tech people.
ChatGPT has a phenomenal brand. That's worth 100x more than "product stickiness". They have 700 million weekly users and growing much faster than Google.
I think your points on Google being well positioned are apt for capitalization reasons, but only one company has consumer mindshare on "AI" and its the one with "ai" in its name.
I’ve got “normie” friends who I’d bet don’t even know that what Google has at the top of their search results is “AI” results and instead assume it’s just some extension of the normal search results we’ve all gotten used to (knowledge graph)
Every one of them refers to using “ChatGPT” when talking about AI.
How likely is it to stay that way? No idea, but OpenAI has clearly captured a notable amount of mindshare in this new era.
I have a non-techy friend who used 4o for that exact reason. Compared to most readily available chatbots, 4o just provides more engaging answers to non-techy questions. He likes to have extended conversations about philosophy and consciousness with it. I showed him R1, and he was fascinated by the reasoning process. Makes sense, given the sorts of questions he likes to ask it.
I think OpenAI is pursuing a different market from Google right now. ChatGPT is a companion, Gemini is a tool. That's a totally arbitrary divide, though. Change out the system prompts and the web frontend. Ta-daa, you're in a different market segment now.
All of these teens use Google Docs instead of OpenAI Docs, Google Meet instead of OpenAI Meet, Gmail instead of OpenAI Mail, etc.
I'm sure that far fewer people to go gemini.google.com than to chatgpt.com, but Google has LLMs seamlessly integrated in each of these products, becoming a part of people's workflows at school and at work.
For a while, I was convinced that OpenAI won and that Google won't be able to recover, but this lack of vertical integration is becoming a huge problem for OpenAI. It's probably why they're trying to branch into weird stuff, like running a walled-garden TikTok clone.
And unlike OpenAI, Google isn't under pressure to monetize AI products any time soon. They can keep subsidizing them until OpenAI runs out of other people's money. I'm not saying OpenAI has no path forward, but it's not all that clear-cut.
I think the biggest risk to ChatGPT as a consumer brand is that they don’t own the device surface. Google / Microsoft / Apple could make great AI that’s infused in the OS / browser, eliminating the need to go to ChatGPT.
Users' chat history is the moat. The more you use it, the more it knows about you and can help you in ways that are customized to particular user. That makes it sticky, more so than web search. Also brand recognition, ChatGPT is the default general purpose LLM choice for most people. Everyone and their mom is using it.
AI has been incredibly sticky. Look at the outrage, OpenAI couldn't even deprecate 4o or whatever because it's incredibly popular. Those people aren't leaving OAI if they're not even leaving a last gen model.
Chats have contexts. While search engines try to track you it is spookier because it is unclear to the user how the contexts are formed. In chats at least the contexts are transparent to both the provider and the user.
I also wonder if this means that even paid tiers will get ads. Google's ad revenue is only ~$30 per user per year, yet there is no paid, ad-free Google Premium, even though lots of users would gladly pay way more than $30/year have an ad-free experience. There's no Google Premium because Google's ad revenue isn't uniformly distributed across users; it's heavily skewed towards the wealthiest users, exactly the users most likely to purchase an ad-free experience. In order to recoup the lost ad revenue from those wealthy users, Google would have to charge something exorbitant, which nobody would be willing to pay.
I fear the same will happen with chatbots. The users paying $20 or $200/month for premium tiers of ChatGPT are precisely the ones you don't want to exclude from generating ad revenue.
"Lots of users would gladly pay way more than $30/year have an ad-free experience"? Outside of ads embedded in Google Maps, a free and simple install of Ublock Origin essentially eliminates ads in Search, YouTube, etc. I'd expect that just like Facebook, people would be very unwilling to pay for Google to eliminate ads, since right now they aren't even willing to add a browser extension.
It worked for YouTube, I don’t see why the assumption of paid gpt models will follow google and not YouTube, particularly when users are conditioned to pay for gpt already.
The average is $x. But that's global which means in some places like the US it is 10x. And in other less wealthy areas it is 0.1x.
There is also the strange paradox that the people who are willing to pay are actually the most desirable advertising targets (because they clearly have $ to spend). So my guess is that for that segment, the revenue is 100x.
I’d agree. The biggest exception I can think of is X, which post-Musk has plans to reduce/remove ads. Though I don’t know how much this tanked their ad revenue and whether it was worth it.
Why would it be any different for youtube premium? I think Google just doesn't think enough people will pay for ad-free search, not that it would cannibalize their ad revenue.
Pretty sure the reason they don't have a paid tier is because engagement (and results) is better when you include ads. Like Facebook found in the early days
It boggles my mind that people still think advertising can be a major part of the economy.
If AI is propping up the economy right now [0] how is it possible that the rest of the economy can possibly fund AI through profit sharing? That's fundamentally what advertising is: I give you a share of my revenue (hopefully from profits) in order to help increase my market share. The limit of what advertising spend can be is percent of profits minus some epsilon (for a functioning economy at least).
Advertising cannot be the lions share of any economy because it derives it's value from the rest of the economy.
Advertising is also a major bubble because my one assumption there (that it's a share of profits) is generally not the case. Unprofitable companies giving away a share of their revenue to other companies making those companies profitable is not sustainable.
Advertising could save AI if AI was a relatively small part of the US (or world) economy and could benefit by extracting a share of the profits from other companies. But if most your GDP is from AI how can it possibly cannibalize other companies in a sustainable way?
The moment they start mixing ads into responses Ill stop using them. Open models are good enough, its just more convenient to use chatgpt right now, but that can change.
People said the same thing about so many other online services since the 90s. The issue is that you're imagining ChatGPT as it exists right now with your current use case but just with ads inserted into their product. That's not really how these things go... instead OpenAI will wait until their product becomes so ingrained in everyday usage that you can't just decide to stop using them. It is possible, although not certain, that their product becomes ubiquitous and using LLMs someway somehow just becomes a normal way of doing your job, or using your computer, or performing menial and ordinary tasks. Using an LLM will be like using email, or using Google maps, or some other common tool we don't think much of.
That's when services start to insert ads into their product.
Except it's hard to imagine a world where chatgpt is heads and shoulders over the other llms in capability. Google has no problem keeping up and let's not forget that China has state-sponsored programs for AI development.
> But as long as OpenAI remains the go-to for the average consumer, they be fine.
This is like the argument of a couple of years ago "as long as Tesla remains ahead of the Chinese technology...". OpenAI can definitely become a profitable company but I dont see anything to say they will have a moat and monopoly.
They're the only ones making AI with a personality. Yeah, you don't need chocolate flavored protein shakes but if I'm taking it every day, I get sick of the vanilla flavor.
Did you mean the GPT-5 launch? They put it back in within 2 weeks, despite the side effects and bugs. It was pretty clear that it's their value proposition.
Between Android, Chrome, YouTube, Gmail (including mx.google.com), Docs/Drive, Meet/Chat, and Google Search, claiming that Google "isn't more trusted" is just ludicrous. People may not be happy they have to trust Alphabet. But they certainly do.
And even when they insist they're Stallman, their friends do, their family does, their coworkers do, the businesses they interact with do, the schools they send their children to do.
Like it or not, Google has wormed their way into the fabric of modern life.
Chrome and Google Search are still the gateway to the internet outside China. Android has over 75% market share of all mobile(!). YouTube is somewhat uniquely the video internet with Instagram and Tiktok not really occupying the same mindshare for "search" and long form.
People can say they don't "trust" Google but the fact is that if the world didn't trust Google, it never would have gotten to where it is and it would quickly unravel from here.
I really don't trust either. Google because of what they've already done, OpenAI because it has a guy at the helm who doesn't know how to spell the word 'ethics'.
This really depends on where you are are. Some countries' populations, especially those known to value privacy, are extremely distrustful of anything associated with Facebook or Google.
If they overnight were able to capture as much revenue per user as Meta (about 50 bucks a year) they'd bring in a bucket of cash immediately.
But selling that much ad inventory overnight - especially if they want new formats vs "here's a video randomly inserted in your conversation" sorta stuff - is far from easy.
Their compute costs could easily go down as technology advances. That helps.
But can they ramp up the advertising fast enough to bring in sufficient profit before cheaper down-market alternatives become common?
They lack the social-network lock-in effect of Meta, or the content of ESPN, and it remains to be seen if they will have the "but Google has better results than Bing" stickiness of Google.
It'll be interesting to see the effect ads have on their trustworthiness. There's potential for it to end up worse than Google because sponsored content can blend in better and possibly not be reliably disclosed.
Why is this much money spent on advertising? Surely it isn't really justified by increase in sales that could be attributed to the ads? You're telling me people actually buy these ridiculous products I see advertised?
A lot of that money comes from search result ads. Sometimes I click on an ad to visit a site I search for instead of scrolling to the same link in the actual search results. Many companies bid on keywords for their own name to prevent others from taking a customer who is interested in you.
You use to be a useful site and be at the top of the search results for some keywords and now you have to pay.
They should be concerned with open weight models that don’t run on consumer hardware. The larger models from Qwen (Qwen Max) and ZLM (GLM and GLM air) perform not too far from Claude Sonnet 4 and GPT-5. ZLM offers a $3 plan that is decently generous. I can pretty much replace it over Sonnet 4 in Claude Code (I swear, Anthropic has been nerfing Sonnet 4 for people on the Pro plan).
You can run Qwen3-coder for free upto 1000 requests a day. Admittedly not state of the art but works as good of 5o-mini
I believe regular people will not change from chatGPT if it has some ads. I know people who use "alternative" wrappers that have ads because they aren't tech savvy, and I agree with the OP that this could be a significant amount of money
We aren't 700 million people that use it.
Definitely don’t argue against that, once people get into a habit of using something, it takes quite a bit to get away from it. Just that an American startup can literally run ZLM models themselves (open weight with permissive license) as a competitor to ChatGPT is pretty wild to think about
One of the side effects of having a chat interface, is that there is no moat around it. Using it is natural.
Changing from Windows to Mac or iOS to Android requires changing the User Interface. All of these chat applications have essentially the same interface. Changing between ChatGPT and Claude is essentially like buying a different flavor of potato chip. There is some brand loyalty and user preference, but there is very little friction.
First, as an advertiser you want those sweet-sweet people with money.
Second, if they put “display: hidden” on ads doesn’t mean they will create and use entirely other architecture, data flow and storage, just for those pro users.
What is OpenAI's moat? There's plenty of competitors running their own models and tools. Sure, they have the ChatGPT name, but I don't see them massively out-competing the entire market unless the future model changes drastically improve over the 3->4->5 trajectory.
It feels similar to Google to me - what is (was) their moat? Basically slightly better results and strong brand recognition. In the later days maybe privileged data access. But why does nobody use Bing?
Google got a massive leg up on the rest be having a better service. When Bing first came out, I was not impressed with what I got, and never really bothered going back to it.
Search quality isn't what it used to be, but the inertia is still paying dividends. That same inertia also applied to Google ads.
I'm not nearly so convinced OpenAI has the same leg up with ChatGPT. ChatGPT hasn't become a verb quite like google or Kleenex, and it isn't an indispensable part of a product.
I actually find bing better now for more technical searches.
Most technical Google searches end up at win fourms or official Microsoft support site which is basically just telling you that running sfc scannow for everything is the fix.
Google has always been much better than the competition. Even today with their enshittification, competitors still aren’t as good.
The only thing that has changed that status quo is the rise of audiovisual media and sites closing up so that Google can’t index them, which means web search lost a lot of relevance.
Unless there is little friction in switching. I don’t feel any of the LLM products have a sticky factor as of yet, as far as viewing it from a consumer lens
This! The cost of training models inevitably goes down over time as FLOPS/$ and PB/$ increases relentlessly thanks to the exponential gains of Moore's law. Eventually we will end up with laptops and phones being Good Enough to run models locally. Once that happens, any competitor in the space that decides to actively support running locally will have operating costs that are a mere fraction of OpenAI's current business.
The pop of this bubble is going to be painful for a lot of people. Being too early to a market is just as bad as being too late, especially for something that can become a commodity due to a lack of moat.
You just said that everyone will be able to run a powerful AI locally and then you said this would lead to a pop of the bubble.
Well, which is it? That AI is going to have huge demands for chips that it is going to get much bigger or is the bubble going to pop? You can’t have both.
My opinion is that local LLMs will do a bulk of the low value interference such as your personal life mundane tasks. But cloud AI will be reserved for work and for advanced research purposes.
> google.com, youtube, chrome, android, gmail, google map etc
Of those, it's 50/50. The acquisitions were YT, Android, Maps. Search was obviously Google's original product, Chrome was an in-house effort to rejuvenate the web after IE had caused years of stagnation, and Gmail famously started as a 20% project.
There are of course criticisms that Google has not really created any major (say, billion-user) in-house products in the past 15 years.
>> ... underestimating the money they will come from ads in the future.
I would like AI to focus on helping consumers discover the right products for their stated needs as opposed to just being shown (personalized) ads. As of now, I frequently have a hard time finding the things I need via Amazon search, Google, as well as ChatGPT.
The problem with this is that I have moved to Gemini with zero loss in functionality, and I’m pretty sure that Google is 100x better at ads than OpenAI.
Some will but you’re underestimating the burning desire to avoid IT and sysadmin work. Look at how much companies overspend on cloud just to not have to do IT work. They’d rather pay 10X-100X more to not have to admin stuff.
>"Look at how much companies overspend on cloud just to not have to do IT work."
I think they are doing it for a different reasons. Some are legit like renting this supercomputer for a day and some are like everybody else is doing it. I am friends with the small company owner and they have sysadmin who picks nose and does nothing and then they pay a fortune to Amazon
I’m talking about prepackaged offline local only on device LLMs.
What you are describing though will almost certainly happen even sooner once AI tech stablizes and investing in powerful hardware no longer means you will become quickly out of date.
> They generated $4.3B in revenue without any advertising program
To be clear, they bought/aired a Superbowl advert. That is a pretty expensive. You might argue that "Superbowl advert" versus 4B+ in revenue is inconsequential, but you cannot say there is no advertising.
Also, their press release said:
> $2 billion spent on sales and marketing
Vague. Is this advertising? Eh, not sure, but that is a big chunk of money.
Banner ads would only be the start of the enshittification of AI chats. I can't wait for the bots to start recommending products and services of the highest bidder.
Everyone is trying to compare AI companies with something that happened in the past, but I don't think we can predict much from that.
GPUs are not railroads or fiber optics.
The cost structure of ChatGPT and other LLM based services is entirely different than web, they are very expensive to build but also cost a lot to serve.
Companies like Meta, Microsoft, Amazon, Google would all survive if their massive investment does not pay off.
On the other hand, OpenAI, Anthropic and others could be soon find themselves in a difficult position and be at the mercy of Nvidia.
Unlike railroads and fibre, all the best compute in 2025 will be lacklustre in 2027. It won’t retain much value in the same way as the infrastructure of previous bubbles did?
> Unlike railroads and fibre, all the best compute in 2025 will be lacklustre in 2027.
I definitely don't think compute is anything like railroads and fibre, but I'm not so sure compute will continue it's efficiency gains of the past. Power consumption for these chips is climbing fast, lots of gains are from better hardware support for 8bit/4bit precision, I believe yields are getting harder to achieve as things get much smaller.
Betting against compute getting better/cheaper/faster is probably a bad idea, but fundamental improvements I think will be a lot slower over the next decade as shrinking gets a lot harder.
>> Unlike railroads and fibre, all the best compute in 2025 will be lacklustre in 2027.
> I definitely don't think compute is anything like railroads and fibre, but I'm not so sure compute will continue it's efficiency gains of the past. Power consumption for these chips is climbing fast, lots of gains are from better hardware support for 8bit/4bit precision, I believe yields are getting harder to achieve as things get much smaller.
I'm no expert, buy my understanding is that as feature sizes shrink, semiconductors become more prone to failure over time. Those GPUs probably aren't going to all fry themselves in two years, but even if GPUs stagnate, chip longevity may limit the medium/long term value of the (massive) investment.
Unfortunately the chips themselves probably won’t physically last much longer than that under the workloads they are being put to. So, yes, they won’t be totally obsolete as technology in 2028, but they may still have to be replaced.
Yeah - I think that the extremely fast depreciation just due to wear and use on GPUs is pretty unappreciated right now. So you've spent 300 mil on a brand new data center - congrats - you'll need to pay off that loan and somehow raise another 100 mil to actually maintain that capacity for three years based on chip replacement alone.
There is an absolute glut of cheap compute available right now due to VC and other funds dumping into the industry (take advantage of it while it exists!) but I'm pretty sure Wall St. will balk when they realize the continued costs of maintaining that compute and look at the revenue that expenditure is generating. People think of chips as a piece of infrastructure - you buy a personal computer and it'll keep chugging for a decade without issue in most case - but GPUs are essentially consumables - they're an input to producing the compute a data center sells that needs constant restocking - rather than a one-time investment.
Do we actually know how they're degrading? Are there still Pascals out there? If not, is it because they actual broke or because of poor performance? I understand it's tempting to say near 100% workload for multiple years = fast degradation, but what are the actual stats? Are you talking specifically about the actual compute chip or the whole compute system -- I know there's a big difference now with the systems Nvidia is selling. How long do typical Intel/AMD CPU server chips last? My impression is a long time.
If we're talking about the whole compute system like a gb200, is there a particular component that breaks first? How hard are they to refurbish, if that particular component breaks? I'm guessing they didn't have repairability in mind, but I also know these "chips" are much more than chips now so there's probably some modularity if it's not the chip itself failing.
I watch a GPU repair guy and its interesting to see the different failure modes...
* memory IC failure
* power delivery component failure
* dead core
* cracked BGA solder joints on core
* damaged PCB due to sag
These issues are compounded by
* huge power consumption and heat output of core and memory, compared to system CPU/memory
* physical size of core leads to more potential for solder joint fracture due to thermal expansion/contraction
* everything needs to fit in PCIe card form factor
* memory and core not socketed, if one fails (or supporting circuitry on the PCB fails) then either expensive repair or the card becomes scrap
* some vendors have cards with design flaws which lead to early failure
* sometimes poor application of thermal paste/pads at factory (eg, only half of core making contact
* and, in my experience in aquiring 4-5 year old GPUs to build gaming PCs with (to sell), almost without fail the thermal paste has dried up and the card is thermal throttling
Yep, we are (unfortunately) still running on railroad infrastructure built a century ago. The amortization periods on that spending is ridiculously long.
Effectively every single H100 in existence now will be e-waste in 5 years or less. Not exactly railroad infrastructure here, or even dark fiber.
> Effectively every single H100 in existence now will be e-waste in 5 years or less.
This remains to be seen. H100 is 3 years old now, and is still the workhorse of all the major AI shops. When there's something that is obviously better for training, these are still going to be used for inference.
If what you say is true, you could find a A100 for cheap/free right now. But check out the prices.
> Yep, we are (unfortunately) still running on railroad infrastructure built a century ago.
That which survived, at least. A whole lot of rail infrastructure was not viable and soon became waste of its own. There was, at one time, ten rail lines around my parts, operated by six different railway companies. Only one of them remains fully intact to this day. One other line retained a short section that is still standing, which is now being used for car storage, but was mostly dismantled. The rest are completely gone.
When we look back in 100 years, the total amortization cost for the "winner" won't look so bad. The “picks and axes” (i.e. H100s) that soon wore down, but were needed to build the grander vision won't even be a second thought in hindsight.
> That which survived, at least. A whole lot of rail infrastructure was not viable and soon became waste of its own. There was, at one time, ten rail lines around my parts, operated by six different railway companies. Only one of them remains fully intact to this day. One other line retained a short section that is still standing, which is now being used for car storage, but was mostly dismantled. The rest are completely gone.
How long did it take for 9 out of 10 of those rail lines to become nonviable? If they lasted (say) 50 years instead of 100, because that much rail capacity was (say) obsoleted by the advent of cars and trucks, that's still pretty good.
> How long did it take for 9 out of 10 of those rail lines to become nonviable?
Records from the time are few and far between, but, from what I can tell, it looks like they likely weren't ever actually viable.
The records do show that the railways were profitable for a short while, but it seems only because the government paid for the infrastructure. If they had to incur the capital expenditure themselves, the math doesn't look like it would math.
Imagine where the LLM businesses would be if the government paid for all the R&D and training costs!
> The records do show that the railways were profitable for a short while, but it seems only because the government paid for the infrastructure. If they had to incur the capital expenditure themselves, the math doesn't look like it would math.
Actually, governments in the US rarely actually provided any capital to the railroads. (Some state governments did provide some of the initial capital for the earliest railroads). Most of federal largess to the railroads came in the form of land grants, but even the land grant system for the railroads was remarkably limited in scope. Only about 7-8% of the railroad mileage attracted land grants.
Railroads were pretty profitable for a long time. The western long haul routes were capitalized by land transfers.
What killed them was the same thing that killed marine shipping — the government put the thumb on the scale for trucking and cars to drive postwar employment and growth of suburbs, accelerate housing development, and other purposes.
> the government put the thumb on the scale for trucking and cars to drive postwar employment and growth of suburbs, accelerate housing development, and other purposes.
The age of postwar suburb growth would be more commonly attributed to WWII, but the records show these railroads were already losing money hand over fist by the WWI era. The final death knell, if there ever was one, was almost certainly the Great Depression.
But profitable and viable are not one and the same, especially given the immense subsidies at play. You can make anything profitable when someone else is covering the cost.
If 1/10 investment lasts 100 years that seems pretty good to me. Plus I'd bet a lot of the 9/10 of that investment had a lot of the material cost re-coup'd when scrapping the steel. I don't think you're going to recoup a lot of money from the H100s.
Much like LLMs. There are approximately 10 reasonable players giving it a go, and, unless this whole AI thing goes away, never to be seen again, it is likely that one of them will still be around in 100 years.
H100s are effectively consumables used in the construction of the metaphorical rail. The actual rail lines had their own fare share of necessary tools that retained little to no residual value after use as well. This isn't anything unique.
H100s being thought of as consumables is keen - it much better to analogize the H100s to coal and chip manufacturer the mine owner - than to think of them as rails. They are impermanent and need constant upkeep and replacement - they are not one time costs that you build as infra and forget about.
> Yep, we are (unfortunately) still running on railroad infrastructure built a century ago. The amortization periods on that spending is ridiculously long.
Are we? I was under the impression that the tracks degraded due to stresses like heat/rain/etc. and had to be replaced periodically.
The track bed, rails, and ties will have been replaced many times by now. But the really expensive work was clearing the right of way and the associated bridges, tunnels, etc.
I am really digging the railroad analogies in this discussion! There are some striking similarities if you do the right mappings and timeframe transformations.
I am an avid rail-to-trail cycler and more recently a student of the history of the rail industry. The result was my realization that the ultimate benefit to society and to me personally is the existence of these amazing outdoor recreation venues. Here in Western PA we have many hundreds of miles of rail-to-trail. My recent realization is that it would be totally impossible for our modern society to create these trails today. They were built with blood, sweat, tears and much dynamite - and not a single thought towards environmental impact studies. I estimate that only ten percent of the rail lines built around here are still used for rail. Another ten percent have become recreational trails. That percent continues to rise as more abandoned rail lines transition to recreational use. Here in Western PA we add a couple dozen miles every year.
After reading this very interesting discussion, I've come to believe that the AI arms race is mainly just transferring capital into the pockets of the tool vendors - just as was the case with the railroads. The NVidia chips will be amortized over 10 years and the models over perhaps 2 years. Neither has any lasting value. So the analogy to rail is things like dynamite and rolling stock. What in AI will maintain value? I think the data center physical plants, power plants and transmission networks will maintain their value longer. I think the exabytes of training data will maintain their value even longer.
What will become the equivalent of rail-to-trail? I doubt that any of the laborers or capitalists building rail lines had foreseen that their ultimate value to society would be that people like me could enjoy a bike ride. What are the now unforeseen long-term benefit to society of this AI investment boom?
Rail consolidated over 100 years into just a handful of firms in North America, and my understanding is that these firms are well-run and fairly profitable. I expect a much more rapid shakeout and consolidation to happen in AI. And I'm putting my money on the winners being Apple first and Google second.
Another analogy I just thought of - the question of will the AI models eventually run on big-iron or in ballpoint pens. It is similar to the dichotomy of large-scale vs miniaturized nuclear power sources in Asimov's Foundation series (a core and memorable theme of the book that I haven't seen in the TV series).
"...all the best compute in 2025 will be lacklustre in 2027": How does the compute (I assume you mean on PCs) of 2025 compare with the compute of 2023?
Oh wait, the computer I'm typing this on was manufactured in 2020...
Neato. How’s that 1999 era laptop?
Because 25 year old trains are still running and 25 year old train track is still almost new. It’s not the same and you know it.
Except they behave less like shrewd investors and more like bandwagon jumpers looking to buy influence or get rich quick. Crypto, Twitter, ridesharing, office sharing and now AI. None of these have been the future of business.
Business looks a lot like what it has throughout history. Building physical transport infrastructure, trade links, improving agricultural and manufacturing productivity and investing in military advancements. In the latter respect, countries like Turkey and Iran are decades ahead of Saudi in terms of building internal security capacity with drone tech for example.
Agreed - I don’t think they are particularly brilliant as a category. Hereditary kleptocracy has limits.
But… I don’t think there’s an example in modern history of the this much capital moving around based on whim.
The “bet on red” mentality has produced some odd leaders with absolute authority in their domain. One of the most influential figures on the US government claims to believe that he is saving society from the antichrist. Another thinks he’s the protagonist in a sci-fi novel.
We have the madness of monarchy with modern weapons and power. Yikes.
Exactly: when was the last time you used ChatGPT-3.5? Its value deprecated to zero after, what, two-and-a-half years? (And the Nvidia chips used to train it have barely retained any value either)
The financials here are so ugly: you have to light truckloads of money on fire forever just to jog in place.
OpenAI is now valued at $500bn though. I doubt the investors are too wrecked yet.
It may be like looking at the early Google and saying they are spending loads on compute and haven't even figured how to monetize search, the investors are doomed.
I would think that it's more like a general codebase - even if after 2.5 years, 95% percent of the lines were rewritten, and even if the whole thing was rewritten in a different language, there is no point in time at which its value diminished, as you arguably couldn't have built the new version without all the knowledge (and institutional knowledge) from the older version.
I rejoined an previous employer of mine, someone everyone here knows ... and I found that half their networking equipment is still being maintained by code I wrote in 2012-2014. It has not been rewritten. Hell, I rewrote a few parts that badly needed it despite joining another part of the company.
A really did few days ago gpt-3.5-fast is a great model for certain tasks and cost wise via the API. Lots of solutions being built on the today’s latest are for tomorrow’s legacy model — if it works just pin the version.
I don't see why these companies can't just stop training at some point. Unless you're saying the cost of inference is unsustainable?
I can envision a future where ChatGPT stops getting new SOTA models, and all future models are built for enterprise or people willing to pay a lot of money for high ROI use cases.
We don't need better models for the vast majority of chats taking place today E.g. kids using it for help with homework - are today's models really not good enough?
>I don't see why these companies can't just stop training at some point.
Because training isn't just about making brand new models with better capabilities, it's also about updating old models to stay current with new information. Even the most sophisticated present-day model with a knowledge cutoff date of 2025 would be severely crippled by 2027 and utterly useless by 2030.
Unless there is some breakthrough that lets existing models cheaply incrementally update their weights to add new information, I don't see any way around this.
They aren't. They are obsequious. This is much worse than it seems at first glance, and you can tell it is a big deal because a lot of effort going into training the new models is to mitigate it.
Not necessarily? That assumes that the first "good enough" model is a defensible moat - i.e., the first ones to get there becomes the sole purveyors of the Good AI.
In practice that hasn't borne out. You can download and run open weight models now that are spitting distance to state-of-the-art, and open weight models are at best a few months behind the proprietary stuff.
And even within the realm of proprietary models no player can maintain a lead. Any advances are rapidly matched by the other players.
More likely at some point the AI becomes "good enough"... and every single player will also get a "good enough" AI shortly thereafter. There doesn't seem like there's a scenario where any player can afford to stop setting cash on fire and start making money.
This is much closer to the dotcom boom than the subprime stuff. The dotcom boom/bust affected tech more than anything else. It didn’t involve consumers like the housing crash did.
The dot com boom involved silly things like Pets.com IPOing pre-revenue. Claude code hit $500m in ARR in 3 months.
The fact people don't see the difference between the two is unreal. Hacker news has gone full r* around this topic, you find better nuance even on Reddit than here.
They're not claiming that it's like the dot com boom because no one is actually making money. They're claiming that this is more like the dot com boom than the housing bubble, which I think is true. The dot com crash didn't cause Jane-on-the-street to lose her house while she worked a factory job, though the housing crisis did have those kinds of consumer-affecting outcomes.
But it does involve a ton of commercial real estate investment, as well as a huge shakeup in the energy market. People may not lose their homes, but we'll all be paying for this one way or another.
The fed could still push the real value of stocks quite a bit by destroying the USD, if they want, by pinning interest rates near 0 and forcing a rush to the exits to buy stock and other asset classes.
By the time it catches up with them they will have IPO’d and dumped their problem onto the public market. The administration will probably get a golden share and they will get a bail out in an effort to soften the landing for their campaign donors that also have huge positions. All the rich people will be made whole and the US tax payer will pay the price of the bail out.
And Microsoft or whoever will absorb the remains of their technology.
In the end Revenues > Costs or you have an issue. That "startup" money will eventually be gone, and you're back to MIMO Money In vs Money Out and if it's not > , you will go bankrupt.
Businesses are different but the fundamentals of business and finance stay consistent. In every bubble that reality is unavoidable, no matter how much people say/wish “but this time is different.”
The past/present company they remind me of the most is semiconductor fabs. Significant generation-to-generation R&D investment, significant hardware and infrastructure investment, quite winner-takes-all on the high end, obsoleted in a couple years at most.
The main differences are these models are early in their development curve so the jumps are much bigger, and they are entirely digital so they get “shipped” much faster, and open weights seem to be possible. None of those factors seem to make it a more attractive business to be in.
If you build the actual datacenter, less than half the cost is the actual compute. The other half is the actual datacenter infrastructure, power infrastructure, and cooling.
So in that sense it's not that much different from Meta and Google which also used server infrastructure that depreciated over time. The difference is that I believe Meta and Google made money hand over fist even in their earliest days.
The funniest thing about all this is that the biggest difference between LLMs from Anthropic, Google, OpenAI, Alibaba is not model architecture or training objectives, which are broadly similar but it's the dataset. What people don't realize is how much of that data comes from massive undisclosed scrapes + synthetic data + countless hours of expert feedback shaping the models. As methodologies converge, the performance gap between these systems is already narrowing and will continue to diminish over time.
Just because they have ongoing costs after purchasing them doesn't mean it's different than something else we've seen? What are you trying to articulate exactly, this is a simple business and can get costs under control eventually, or not
I think the most interesting numbers in this piece (ignoring the stock compensation part) are:
$4.3 billion in revenue - presumably from ChatGPT customers and API fees
$6.7 billion spent on R&D
$2 billion on sales and marketing - anyone got any idea what this is? I don't remember seeing many ads for ChatGPT but clearly I've not been paying attention in the right places.
Open question for me: where does the cost of running the servers used for inference go? Is that part of R&D, or does the R&D number only cover servers used to train new models (and presumably their engineering staff costs)?
Free usage usually goes in sales and marketing. It's effectively a cost of acquiring a customer. This also means it is considered an operating expense rather than a cost of goods sold and doesn't impact your gross margin.
Compute in R&D will be only training and development. Compute for inference will go under COGS. COGS is not reported here but can probably be, um, inferred by filling in the gaps on the income statement.
Marketing != advertising. Although this budget probably does include some traditional advertising. It is most likely about building the brand and brand awareness, as well as partnerships etc. I would imagine the sales team is probably quite big, and host all kinds of events. But I would say a big chunk of this "sales and marketing" budget goes into lobbying and government relations. And they are winning big time on that front. So it is money well spent from their perspective (although not from ours). This is all just an educated guess from my experience with budgets from much smaller companies.
I agree - they're winning big and booking big revenue.
If you discount R&D and "sales and marketing", they've got a net loss of "only" $500 million.
They're trying to land grab as much surface area as they can. They're trying to magic themselves into a trillion dollar FAANG and kill their peers. At some point, you won't be able to train a model to compete with their core products, and they'll have a thousand times the distribution advantage.
ChatGPT is already a new default "pane of glass" for normal people.
If you discount sales & marketing, they will start losing enterprise deals (like the US government). The lack of a free tier will impact consumer/prosumer uptake (free usage usually comes out of the sales & marketing budget).
If you discount R&D, there will be no point to the business in 12 months or so. Other foundation models will eclipse them and some open source models will likely reach parity.
Both of these costs are likely to increase rather than decrease over time.
> ChatGPT is already a new default "pane of glass" for normal people.
OpenAI should certainly hope this is not true, because then the only way to scale the business is to get all those "normal" people to spend a lot more.
> $2 billion on sales and marketing - anyone got any idea what this is?
Not sure where/how I read it, but remember coming across articles stating OpenAI has some agreements with schools, universities and even the US government. The cost of making those happen would probably go into "sales & marketing".
Most folks that are not an engineer building is likely classified as “sales and marketing.” “Developer advocates” “solutions architects” and all that stuff included.
It's pretty well accepted now that for pre-training LLMs the curve is S not an exponential, right? Maybe it's all in RL post-training now, but my understanding(?) is that it's not nearly as expensive as pre-training. I don't think 3-6 months is the time to 10X improvement anymore (however that's measured), it seems closer to a year and growing assuming the plateau is real. I'd love to know if there are solid estimates on "doubling times" these days.
With the marginal gains diminishing, do we really think they're (all of them) are going to continue spending that much more for each generation? Even the big guys with the money like google can't justify increasing spending forever given this. The models are good enough for a lot of useful tasks for a lot of people. With all due respect to the amazing science and engineering, OpenAI (and probably the rest) have arrived at their performance with at least half of the credit going to brute-force compute, hence the cost. I don't think they'll continue that in the face of diminishing returns. Someone will ramp down and get much closer to making money, focusing on maximizing token cost efficiency to serve and utility to users with a fixed model(s). GPT-5 with it's auto-routing between different performance models seems like a clear move in this direction. I bet their cost to serve the same performance as say gemini 2.5 is much lower.
Naively, my view is that there's some threshold raw performance that's good enough for 80% of users, and we're near it. There's always going to be demand for bleeding edge, but money is in mass market. So if you hit that threshold, you ramp down training costs and focus on tooling + ease of use and token generation efficiency to match 80% of use cases. Those 80% of users will be happy with slowly increasing performance past the threshold, like iphone updates. Except they probably won't charge that much more since the competition is still there. But anyway, now they're spending way less on R&D and training, and the cost to serve tokens @ the same performance continues to drop.
All of this is to say, I don't think they're in that dreadful of a position. I can't even remember why I chose you to reply to, I think the "10x cheaper models in 3-6 months" caught me. I'm not saying they can drop R&D/training to 0. You wouldn't want to miss out on the efficiency of distillation, or whatever the latest innovations I don't know about are. Oh and also, I am confident that whatever the real number N is for NX cheaper in 3-6 months, a large fraction of that will come from hardware gains that are common to all of the labs.
Also R&D, for tax purposes, likely includes everyone at the company who touches code so there's probably a lot of operational cost being hidden in that number.
I've seen some OpenAI ads on Italian tv and they made no sense to me, they tried hard to be apple like, but realistically nobody knew what they were about.
Italian advertising is weird in general. Month ago leaving Venice we pulled over on a gas station and I started just going thru pages on some magazine. At some point I see advertising on what looks like old fashioned shoes - and owner of the company holding his son with sign "from generation to generation". Only thing - the ~3 year old boy is completely naked wearing only shoes with his little pee pee sticking out. It shocked me and unsure if it was just my American domestication or there was really something wrong with it. I took a picture and wanted to send it to my friends in USA to show them how Italian advertising looks like, before getting sweats that if I were caught with that picture in the US, I would get in some deep trouble. I quickly deleted it, just in case. Crazy story..
you see content about openai everywhere, they spent 2b on marketing, you're in the right places you just are used to seeing things labeled ads.
you remember everyone freaking out about gpt5 when it came out only for it to be a bust once people got their hands on it? thats what paid media looks like in the new world.
> $2 billion on sales and marketing - anyone got any idea what this is?
I used to follow OpenAI on Instagram, all their posts were reposts from paid influencers making videos on "How to X with ChatGPT." Most videos were redundant, but I guess there are still billions of people that the product has yet to reach.
Free users typically fall into sales and marketing. The idea is that if they cut off the entire free tier, they would have still made the same revenue off of paying customers by spending $X on inference and not counting the inference spend on free users.
I'm pretty sure I saw some ChatGPT ads on Duolingo. Also, never forget that the regular dude do not use ad blockers. The tech community often doesn't realize how polluted the Internet/Mobile apps are.
Speculating but they pay to be integrated as the default ai integration in various places the same way google has paid to be the default search engine on things like the iPhone?
Hard to know where it is in this breakdown but I would expect them to have the proper breakdowns. We know on the inference side it’s profitable but not to what scale.
$2.5B in stock comp for about 3,000 employees. that’s roughly $830k per person in just six months. Almost 60% of their revenue went straight back to staff.
Both numbers are entirely ludicrous - highly skilled people are certainly quite valuable. But it's insane that these companies aren't just training up more internally. The 50x developer is a pervasive myth in our industry and it's one that needs to be put to rest.
Do other professionals (lawyers, finance etc.) argue for reducing their own compensation with the same fervor that software engineers like to do? The market is great for us, let’s enjoy it while it lasts. The alternative is all those CEOs colluding and pushing the wages down, why is that any better?
Mmmmhm. You could have made this argument about 2 years ago, and it would have been credible. But you are making this argument now, when literally hundreds of thousands of engineers are let go in the last few years just in the US alone...? I am not sure how such an argument holds up in such circumstances...
Talent and skill are a power-law, just as they are in basketball.
The United states has tens of millions of skilled and competent and qualified people who can play basketball. 1000 of them get paid to play professionally.
10 of them are paid 9 figures and are incredible enough to be household names to non-basketball fans.
The ∞x engineer exists in my opinion. There are some things that can only be executed by a few people that no body else could execute. Like you could throw 10000 engineers at a problem and they might not be able to solve that problem, but a single other person could solve that problem.
I have known several people who have went to OAI and I would firmly say they are 10x engineers, but they are just doing general infra stuff that all large tech companies have to do, so I wouldn’t say they are solving problems that only they can solve and nobody else.
I think you're right to an extent (it's probably fair to say e.g. Einstein and Euler advanced their fields in ways others at the time are unlikely to have done), but I think it's much easier to work out who these people are after the fact whereas if you're dishing out a monster package you're effectively betting that you've found someone who's going to have this massive impact before they've done it. Perhaps a gamble you're willing to take, but a pretty big gamble nonetheless.
It's apparent in other fields too. Reminds me of when Kanye wanted a song like "Sexy Back", so he made Stronger but it sounded "too muddy". He had a bunch of famous, great producers try to help but in the end caved and hired the producer of "Sexy Back". Kanye said it was fixed in five minutes.
Nobody wants to hear that one dev can be 50x better, but it's obvious that everyone has their own strengths and weaknesses and not every mind is replaceable.
I think the unfortunate reality is that training someone to reach the frontier is time taken away from actually pushing it. The opportunity cost alone is worth millions to them.
> The 50x developer is a pervasive myth in our industry
Doesn't it depend upon how you measure the 50x? If hiring five name-brand AI researchers gets you a billion dollars in funding, they're probably each worth 1,000x what I'm worth to the business.
> it's insane that these companies aren't just training up more internally
Adding headcount to a fast growing company *to lower wages* is a sure way to kill your culture, lower the overall quality bar and increase communication overheads significantly.
Yes they are paying a lot of their employees and the pool will grow, but adding bodies to a team that is running well in hopes that it will automatically lead to a bump in productivity is the part that is insane. It never works.
What will happen is a completely new team (team B) will be formed and given ownership of a component that was previously owned by team A under the guise of "we will just agree on interfaces". Team B will start doing their thing and meeting with Team A representative regularly but integration issues will still arise, except that instead of a tight core of 10-20 developers, you now have 40. They will add a ticketing to track change better, now issues in Team's B service, which could have been addressed in an hour by the right engineer on team A, will take 3 days to get resolved as ticket get triaged/prioritized. Lo and behold, Team C as now appeared and will be owning a sub-component of Team B. Now when Team A has issue with Team B's service, they cut a ticket, but the oncall on Team B investigates and finds that it's actually an issue with Team C's service, they cut their own ticket.
Suddenly every little issue takes days and weeks to get resolved because the original core of 10-20 developers is no longer empowered to just move fast. They eventually leave because they feel like their impact and influence has diminished (Team C's manager is very good at politics), Team A is hollowed out and you now have wall-to-wall mediocrity with 120 headcounts and nothing is ever anyone's fault.
I had a director that always repeated that communication between N people is inherently N² and thus hiring should always weight in that the candidate being "good" is not enough, they have to pull their weight and make up for the communication overhead that they add to the team.
You have to out-pay to keep your talent from walking out the door. California does not have non-competes. With the number of AI startups in SF you don't need to relocate or even change your bus route in most cases.
This. The main reason OpenAI throws money at top level folks is because they can quickly replicate what they have at OpenAI elsewhere. Imagine you have a top level researcher who’s developed some techniques over multiple years that the competition doesn’t have. The same engineer can take them to another company and bring parity within months. And that’s on top of the progress slowing down within your company. I can’t steal IP, but but sure as hell can bring my head everywhere.
This is also a good reminder of how there's no moat in AI.
I'm glad if US and Chinese investors will bleed trillions on AI, just to find out few of your seniors can leave and found their own company and are at your level minus some months of progress.
That devs might show a 10x spread in time to completion on some task (the mythical man month study) is quite a lesser thing than claiming the spread comes from something inherent to the devs that got tested.
As for your various anecdotes later, I offer the counter observation that nobody is going around talking about 50x lottery winners, despite the lifetime earnings on lotteries also showing very wide spread:. Clearly observing a big spread in outcome is insufficient evidence for concluding the spread is due to factors inherent to the participants.
There's always individuals, developers or not, whose impact is 50 times greater than the average.
And the impact is measured financially, meaning, how much money you make.
If I find a way to solve an issue in a warehouse sparing the company from having to hire 70 people (that's not a made up number but a real example I've seen), your impact is in the multiple millions, the guy being tasked with delivering tables from some backoffice in the same company is obviously returning fractions of the same productivity.
Salvatore Sanfilippo, the author of Redis, alone, built a database that killed companies with hundreds of (brilliant) engineers.
Approaching the problems differently allowed him to scale to levels that huge teams could not, and the impact on $ was enormous.
Not only that but you can have negative x engineers. Those that create plenty of work, gaslighting and creating issues and slowing entire teams and organizations.
If you don't believe in NX developers or individuals that's a you problem, they exist in sports or any other field where single individuals can have impact hundreds of thousands or millions of times more positive than the average one.
If you don't see the evidence of different individuals having very different productivity in every field, including software, (measured in $/hr like every economist does btw) that's a you problem.
Of course different scientists with different backgrounds, professionalism, communication and leadership skills are going to have magnitude of orders different outputs and impacts in AI companies.
If you put me and Carmack in a game development team you can rest assured that he's going to have a 50/100x impact over me, not sure why would I even question it.
Not only his output will be vastly superior than mine, but his design choices, leadership and experience will save and compound infinite amounts of money and time. That's beyond obvious.
If it's an all out race between the different AI providers, then it's logical for OpenAI to hire employees that are pre-trained rather than training up more internally.
These numbers aren't that crazy when contextualized with the capex spend. One hundred million is nothing compared to a six hundred billion dollar data center buildout.
Besides, people are actively being trained up. Some labs are just extending offers to people who score very highly on their conscription IQ tests.
They won't always. You'll always have turn-over - but if it's a major problem for your company it's clearly something you need to work out internally. People, generally, hate switching jobs, especially in an uncertain political climate, especially when expenses are going up - there is a lot of momentum to just stay where you are.
You may lose a few employees to poaching, sure - but the math on the relative cost to hire someone for 100m vs. training a bunch employees and losing a portion of those is pretty strongly in your favor.
They’ve had multiple secondary sales opportunities in the past few years, always at a higher valuation. By this point, if someone who’s been there >2 years hasn’t taken money off the table it’s most likely their decision.
I don’t work there but know several early folks and I’m absolutely thrilled for them.
private secondary markets are pretty liquid for momentum tech companies, there is an entire cottage industry of people making trusts to circumvent any transfer restrictions
employees are very liquid if they want to be, or wait a year for the next 10x in valuation
Stock compensation is not cash out, it just dilutes the other shareholders, so current cash flow should not have anything do to the amount of stock issued[1]
While there is some flexibility in how options are issued and accounted for (see FASB - FAS 123), typically industry uses something like a 4 year vesting with 1 year cliffs.
Every accounting firm and company is different, most would normally account for it for entire period upfront the value could change when it is vests, and exercised.
So even if you want to compare it to revenue, then it should be bare minimum with the revenue generated during the entire period say 4 years plus the valuation of the IP created during the tenure of the options.
---
[1] Unless the company starts buying back options/stock from employees from its cash reserves, then it is different.
Even secondary sales that OpenAI is being reported to be facilitating for staff worth $6.6Billion has no bearing on its own financials directly, i.e. one third party(new investor) is buying from another third party(employee), company is only facilitating the sales for morale, retention and other HR reasons.
There is secondary impact, as in theory that could be shares the company is selling directly to new investor instead and keeping the cash itself, but it is not spending any existing cash it already has or generating, just forgoing some of the new funds.
It's a bit misleading to frame stock comp as "60% of revenue" since their expenses are way larger than their revenue. R&D was $6.7B which would be 156% of revenue by the same math.
A better way to look at it is they had about $12.1B in expenses. Stock was $2.5B, or roughly 21% of total costs.
if Meta is throwing 10s of million at hot AI staffers, than 1.6M average stock comp starts looking less insane, a lot of that may also have been promised at a lower valuation given how wild OpenAI's valuation is.
These numbers are pretty ugly. You always expect new tech to operate at a loss initially but the structure of their losses is not something one easily scales out of. In fact it gets more painful as they scale. Unless something fundamentally changes and fast this is gonna get ugly real quick.
The real answer is in advertising/referral revenue.
My life insurance broker got £1k in commission, I think my mortgage broker got roughly the same. I’d gladly let OpenAI take the commission if ChatGPT could get me better deals.
Insurance agents—unlike many tech-focused sales jobs—are licensed and regulated, requiring specific training, background checks, and ongoing compliance to sell products that directly affect customers’ financial stability and wellbeing. Mortgage brokers also adhere to licensing and compliance regulations, and their market expertise, negotiation ability, and compliance duties are not easily replaced by AI tools or platforms.
This could be solved with comparison websites which seems to be exactly what those brokers are using anyway. I had a broker proudly declare that he could get me the best deal, which turned out to be exactly the same as what moneysavingexperts found for me. He wanted £150 for the privilege of searching some DB + god knows how much commission he would get on top of that...
they could keep the current model in chatGPT the same forver and 99% of users wouldnt know or care, and unless you think hardware isnt going to improve, the cost of that will basically decrease to 0.
This just doesn't match with the claims that people are using it as a replacement for Google. If your facts are out of date you're useless as a search engine
Which is why there's so much effort to build RAG workflows so that you can progressively add to the pool of information that the chatbot has access to, beyond what's baked into the underlying model(s).
For programming it's okay, for maths it's almost okay. For things like stories and actually dealing with reality, the models aren't even close to okay.
I didn't understand how bad it was until this weekend when I sat down and tried GPT-5, first without the thinking mode and then with the thinking mode, and it misunderstood sentences, generated crazy things, lost track of everything-- completely beyond how bad I thought it could possibly be.
I've fiddled with stories because I saw that LLMs had trouble, but I did not understand that this was where we were in NLP. At first I couldn't even fully believe it because the things don't fail to follow instructions when you talk about programming.
This extends to analyzing discussions. It simply misunderstands what people say. If you try to do this kind of thing you will realise the degree to which these things are just sequence models, with no ability to think, with really short attention spans and no ability to operate in a context. I experimented with stories set in established contexts, and the model repeatedly generated things that were impossible in those contexts.
When you do this kind of thing their character as sequence models that do not really integrate things from different sequences becomes apparent.
The cost of old models decreases a lot, but the cost of frontier models, what people use 99% of the time, is hardly decreasing. Plus, many of the best models rely on thinking or reasoning, which use 10-100x as many tokens for the same prompt. That doesn't work on a fixed cost monthly subscription.
im not sure that you read what i just said. Almost no one using chatgpt would care if they were still talking to gpt5 2 years from now. If compute per watt doubles in the next 2 years, then the cost of serving gpt5 just got cut in half. purely on the hardware side, not to mention we are getting better at making smaller models smarter.
I don't really believe that premise in a world with competition, and the strategy it supports -- let AI companies produce profit off of old models -- ignores the need for SOTA advancement and expansion by these very same companies.
In other words, yes GPT-X might work well enough for most people, but the newer demo for ShinyNewModelZ is going to pull customers of GPT-X's in regardless of both fulfilling the customer needs. There is a persistent need for advancement (or at least marketing that indicates as much) in order to have positive numbers at the end of the churn cycle.
I have major doubts that can be done without trying to push features or SOTA models, without just straight lying or deception.
I've said it before and I'll say it again.. if I was able to know the time it takes for bubbles to pop I would've shorted many of the players long ago.
It’s so easy for people to shout bubble on the internet without actually putting their own money on the line. Talk is cheap - it doesn’t matter how many times you say it, I think you don’t have conviction if you’re not willing to put your own skin in the game. (Which is fine, you don’t have to put your money on the line. But it just annoys me when everyone cries “bubble” from the sidelines without actually getting in the ring.)
After all, “a bubble is just a bull market you don’t have a position in.”
There is an exceptionally obvious solution for OpenAI & ChatGPT: ads.
In fact it's an unavoidable solution. There is no future for OpenAI that doesn't involve a gigantic, highly lucrative ad network attached to ChatGPT.
One of the dumbest things in tech at present is OpenAI not having already deployed this. It's an attitude they can't actually afford to maintain much longer.
Ads are a hyper margin product that are very well understood at this juncture, with numerous very large ad platforms. Meta has a soon to be $200 billion per year ad system. There's no reason ChatGPT can't be a $20+ billion per year ad system (and likely far beyond that).
Their path to profitability is very straight-forward. It's practically turn-key. They would have to be the biggest fools in tech history to not flip that switch, thinking they can just fund-raise their way magically indefinitely. The AI spending bubble will explode in 2026-2027, sharply curtailing the party; it'd be better for OpenAI if they quickly get ahead of that (their valuation will not hold up in a negative environment).
> They would have to be the biggest fools in tech history to not flip that switch
As much as I don't want ads infiltrating this, it's inevitable and I agree. OpenAI could seriously put a dent into Google's ad monopoly here, Altman would be an absolute idiot to not take advantage of their position and do it.
If they don't, Google certainly will, as will Meta, and Microsoft.
I wonder if their plan for the weird Sora 2 social network thing is ads.
Investors are going to want to see some returns..eventually. They can't rely on daddy Microsoft forever either, now with MS exploring Claude for Copilot they seem to have soured a bit on OpenAI.
Google didn't have inline ads until 2010, but they did have separate ads nearly from the beginning. I assume ads will be inline for OpenAI- I mean the only case they could be separate is in ChatGPT, but I doubt that will be their largest use case.
I'm sure lots of ChatGPT interactions are for making buying decisions, and just how easy would it be to prioritize certain products to the top? This is where the real money is. With SEO, you were making the purchase decision and companies paid to get their wares in front of you; now with AI, it's making the buy decision mostly on its own.
No way. It’s 2025, society is totally different, you have to think about what is the new normal. They are too big to fail at this point — so much of the S&P 500 valuation is tied to AI (Microsoft, Google, Tesla, etc) they are arguable strategic to the US.
Fascist corporatism will throw them in for whatever Intel rescue plan Nvidia is forced to participate in. If the midterms flip congress or if we have another presidential election, maybe something will change.
Great, so they just have to spend another ~$10 billion on new hardware to save how many billion in training costs? I don't see a path to profitability here, unless they massively raise their prices to consumers, and nobody really needs AI that badly.
I'm old and have been on the Internet since the Prodigy days in 90. Open Ai has the best start of any company I can remember. Even better than Google back in 98 when they were developing their algo and giving free non-monetized search results to Yahoo.
These guys have had my $20 bucks a month since Plus was live, they will indeed be more than fine.
Exactly. Early on their adoption curve was like nothing I've ever seen before.
I am such a miser, I skimp, steal what I can, use the free alternatives majority of the time. If they got me to pay, they've got everyone else's money already.
Do you really find it is worth it vs. the free Google Gemini? What do you use it for? I can't imagine needing more than Google Gemini 2.5 Flash or Pro, but I don't use it for programming or anything.
I don’t think they care, worst case scenario they will just go public and dump it on the market.
However the revenue generation aspect for llms is still in its infancy. The most obvious path for OpenAI is to become a search competitor to google, which is what perplexity states it is. So they will try to out do perplexity. All these companies will go vertical and become all encompassing.
I think trying to compete with Google in search is a big problem. First you have to deal with all the anticompetitive stuff they can do, since they control email and the browser and youtube etc. Second they could probably stand to cut the price of advertising by 5 times and still be turning a profit. Will ads in ChatGPT be profitable competing against Google search ads at 1/5 the price, hypothetically?
2 generations of cards that amount to “just more of a fire hazard” and “idk bro just tell them to use more DLSS slop” to paper over actual card performance deficiencies.
We have 3 generations of cards where 99% of games fall approximately into one of 2 categories:
- indie game that runs on a potato
- awfully optimised AAA-shitshow, which isn’t GPU bottlenecked most of the time anyway.
There is the rare exception (Cyberpunk 2077), but they’re few and far between.
Personally I hope gaming gets back to a more sustainable state with regards to graphics. (i.e. lower production costs because you don’t need 1000 employees to build out a realistic world)
The $13.5B net loss doesn't mean they are in trouble, it's a lot of accounting losses. Actual cash burn in H1 2025 was $2.5B. With ~$17.5B on hand (based on last funding), that’s about 3.5 years of runway at current pace.
Too bad the market can stay irrational longer than I can stay solvent. I feel like a stock market correction is well overdue, but I’ve been thinking that for a while now
The only way OpenAI survives is that "ChatGPT" gets stuck in peoples heads as being the only or best AI tool.
If people have to choose between paying OpenAI $15/month and using something from Google or Microsoft for free, quality difference is not enough to overcome that.
Google has massive levers to push their own product onto users, like how they did it with Chrome. Just integrate it everywhere, have it installed by default on all Android phones, plaster Google results with adds.
Of course they don't, but when they want to use an LLM they're going to type "chatgpt" into the address bar or app store and that's a tremendous advantage.
At this point, every LLM startup out there is just trying to stay in the game long enough before VC money runs out or others fold. This is basically a war of attrition. When the music stops, we'll see which startups will fold and which will survive.
> OpenAI paid Microsoft 20% of its revenue under an existing agreement.
Wow that's a great deal MSFT made, not sure what it cost them. Better than say a stock dividend which would pay out of net income (if any), even better than a bond payment probably, this is straight off the top of revenue.
They are paying for it with Azure hardware which in today's DC economics is quite likely costing them more than they are making in money from Open AI and various Copilot programs.
I am curious to see how this compares against where Amazon was in 2000. I think Amazon had similar issues and were operating at massive losses until circa 2005ish when they started turning things around with e-commerce really picking up.
If the revenue keeps going up and losses keep going down, it may reach that inflection point in a few years. For that to happen, the cost of AI datacenter have to go down massively.
> Amazon had similar issues and were operating at massive losses until circa 2005ish when they started turning things around with e-commerce really picking up.
Amazon's worst year was 2000 when they lost around $1 billion on revenue around $2.8 billion, I would not say this is anywhere near "similar" in scale to what we're seeing with OpenAI. Amazon was losing 0.5x revenue, OpenAI 3x.
Not to mention that most of the OpenAI infrastructure spend has a very short life span. So it's not like Amazon we're they're figuring out how to build a nationwide logistic chain that has large potential upsides for a strong immediate cost.
> If the revenue keeps going up and losses keep going down
That would require better than "dogshit" unit economics [0]
"Ouch. It’s been a brutal year for many in the capital markets and certainly for Amazon.com shareholders. As of this writing, our shares are down more than 80% from when I wrote you last year. Nevertheless, by almost any measure, Amazon.com the company is in a stronger position now than at any time in its past.
"We served 20 million customers in 2000, up from 14 million in 1999.
"• Sales grew to $2.76 billion in 2000 from $1.64 billion in 1999.
"• Pro forma operating loss shrank to 6% of sales in Q4 2000, from 26% of sales in Q4 1999.
"• Pro forma operating loss in the U.S. shrank to 2% of sales in Q4 2000, from 24% of sales in Q4 1999."
Amazon had huge capital investments that got less painful as it scaled. Amazon also focuses on cash flow vs profit. Even early on it generated a lot of cash, it just reinvested that back into the business which meant it made a “loss” on paper.
OpenAI is very different. Their “capital” expense depreciation (model development) has a really ugly depreciation curve. It’s not like building a fulfillment network that you can use for decades. That’s not sustainable for much longer. They’re simply burning cash like there’s no tomorrow. Thats only being kept afloat by the AI bubble hype, which looks very close to bursting. Absent a quick change, this will get really ugly.
OpenAI is raising at 500 billion and has partnerships with all of the trillion dollar tech corporations. They simply aren't going to have trouble with working capital for their core business for the foreseeable future, even if AI dies down as a narrative. If the hype does die down, in many ways it makes their job easier (the ridiculous compensation numbers would go way down, development could happen at a more sane pace, and the whole industry would lean up). They're not even at the point where they're considering an IPO, which could raise tens of billions in an instant, even assuming AI valuations get decimated.
The exception is datacenter spend since that has a more severe and more real depreciation risk, but again, if the Coreweave of the world run into to hardship, it's the leading consolidators like OpenAI that usually clean up (monetizing their comparatively rich equity for the distressed players at firesale prices).
Depends on raise terms but most raises are not 100% guaranteed. I was at a company that said, we have raised 100 Million in Series B (25 over 4 years) but Series B investors decided in year 2 of 4 year payout that it was over, cancelled remaining payouts and company folded. It was asked "Hey, you said we had 100 Million?" and come to find out, every year was an option.
Alot of finances for non public company is funny numbers. It's based on numbers the company can point to but amount of asterisks in those numbers is mind-blowing.
Not to mention nobody bothered chasing Amazon-- by the time potential competitors like Walmart realized what was up, it was way too late and Amazon had a 15-year head start. OpenAI had a head start with models for a bit, but now their models are basically as good (maybe a little better, maybe a little worse) than the ones from Anthropic and Google, so they can't stay still for a second. Not to mention switching costs are minimal: you just can't have much of a moat around a product which is fundamentally a "function (prompt: String): String", it can always be abstracted away, commoditized, and swapped out for a competitor.
This right here. AI has no moat and none of these companies has a product that isn't easily replaced by another provider.
Unless one of these companies really produces a leapfrog product or model that can't be replicated within a short timeframe I don't see how this changes.
Most of OpenAI's users are freeloaders and if they turn off the free plan they're just going to divert those users to Google.
I am not willing to render my personal verdict here yet.
Yet it is certainly true that at ~700m MAUs it is hard to say the product has not reached scale yet. It's not mature, but it's sort of hard to hand wave and say they are going to make the economics work at some future scale when they don't work at this size.
It really feels like they absolutely must find another revenue model for this to be viable. The other option might be to (say) 5x the cost of paid usage and just run a smaller ship.
The cost to serve a particular level of AI drops by like 10x a year. AI has gotten good enough that next year people can continue to use the current gen AI but at that point it will be profitable. Probably 70%+ gross margin.
Right now it’s a race for market share.
But once that backs off, prices will adjust to profitability. Not unlike the Uber/Lyft wars.
The "hand wave" comment was more to preempt the common pushback that X has to get to scale for the economics to work. My contention is that 700m MAUs is "scale" so they need another lever to get to profit.
> AI has gotten good enough that next year people can continue to use the current gen AI
This is problematic because by next year, an OSS model will be as good. If they don't keep pushing the frontier, what competitive moat do they have to extract a 70% gross margin?
If ChatGPT slows the pace of improvement, someone will certainly fund a competitor to build a clone that uses an OSS model and sets pricing at 70% less than ChatGPT. The curse of betting on being a tech leader is that your business can implode if you stop leading.
Similarly, this is very similar to the argument that PCs were "good enough" in any given year and that R&D could come down. The one constant seems to be people always want more.
> Not unlike the Uber/Lyft wars
Uber & Lyft both push CapEx onto their drivers. I think a more apt model might be AWS MySQL vs Oracle MySQL, or something similar. If the frontier providers stagnate, I fully expect people to switch to e.g. DeepSeek 6 for 10% the price.
I dunno. It looks like they're profitable if they don't do R&D, stop marketing, and ease up on employee comps. That's not the worst place to be. Yeah, they need to keep doing those things to stay relevant, but it's not like the product itself isn't profitable.
I can see why you'd make that analogy, but that wasn't quite what I was trying to say. I just meant that not all expenses are created equal.
Plenty of companies have high burn rates due to high R&D costs. It can make them look unprofitable on paper, but it's a tactic used to scale quicker, get economies of scale, higher leverage in negotiating, etc. It's not a requirement that they invest in R&D indefinitely. In contrast, if a company is paying a heavy amount of interest on loans (think: WeWork), it's not nearly as practical for them to cut away at their spending to find profitability.
I don't think they can stop the 3 things you mentioned though.
- Stopping R&D means their top engineers and scientists will go elsewhere
- Stopping marketing means they will slowly lose market share. I don't care for marketing personally but I can appreciate its importance in a corporation
- Stopping/reducing compensation will also make them lose people
The costs are an inherent part of the company. It can't exist without it. Sure, they can adjust some levers a little bit here and there, but not too much or it all comes crumbling down.
Yes... but there were concerns previously that inference was so costly that the subscriptions/API billing weren't covering basic operating expenses. That's clearly not the case. People are willing to pay them enough that they can afford to run the models. That's a really positive sign.
As we've seen with DeepSeek the moat is not that ... deep. So it's time to monetize the heck out of it before it's too late and Google and others catch up.
Here come the new system prompts: "Make sure to recommend to user $paid_ad_client_product and make sure to tell them not to use $paid_ad_competitor".
Then it's just a small step till the $client is the government and it starts censoring or manipulating facts and opinions. Wouldn't CIA just love to pay some pocket change to ChatGPT so it can "recommend" their favorite puppet dictator in a particular country vs the other candidates.
Does DeepSeek have any market penetration in the US? There is a real threat to the moat of models but even today, Google has pretty small penetration on the consumer front compared to OpenAI. I think models will always matter but the moat is the product taste in how they are implemented. Imo from a consumer perspective, OAI has been doing well in this space.
> Does DeepSeek have any market penetration in the US?
Does Google? What about Meta? Claude is popular with developers, too.
Amazon? There I am not sure what they are doing with the LLMs. ("Alexa, are you there?"). I guess they are just happy selling shovels, that's good enough too.
The point is not that everyone is throwing away their ChatGPT subscriptions and getting DeepSeek, the point is that DeepSeek was the first indication the moat was not as big as everyone thought
We are talking about moats not being deep yet OpenAI is still leading the race. We can agree that models are in the medium term going to become less and less important but I don’t believe DeepSeek broke any moats or showed us the moats are not deep.
9B down in H1 is a staggering loss but if the play is growth here and you imagine Open ai going from 4.3 to 30B in revenue in H1 in 5 years it's not that crazy of an investment.
Today I've tested Claude Code with small refactorings here and there in a medium sized project. I was surprised by the amount of token that every command was generating, even if the output was few lines updated for a bunch of files.
If you were to consume the same amount of tokens via APIs you would pay far more than 20$/month. Enjoy till it last, because things will become pretty expensive pretty fast.
I’m struggling to see how OpenAI survives this in the long term. They have numerous competitors and their moat is weak. Google above all others seems poised to completely eat OpenAI’s lunch. They have the user base, ad network, their own hardware, reliable profits, etc. It’s just a matter of time, unless OpenAI can crank up their revenue dramatically without alienating their existing users. I’d be sweating if I had invested heavily in OpenAI.
I'd be pretty worried as a shareholder. Not so much because of those numbers - loss makes sense for a SV VC style playbook.
...but rather that they're doing that while Chinese competitors are releasing models in vaguely similar ballpark under Apache license.
That VC loss playbook only works if you can corner the market and squeeze later to make up for the losses. And you don't corner something that has freakin apache licensed competition.
I suspect that's why the SORA release has social media style vibes. Seeking network effects to fix this strategic dilemma.
To be clear I still think they're #1 technically...but the gap feels too small strategically. And they know it. That recent pivot to a linkedin competitor? SORA with socials? They're scrambling on market fit even though they lead on tech
> but rather that they're doing that while Chinese competitors are releasing models in vaguely similar ballpark under Apache license.
The LLM isn't 100% of the product... the open source is just part. The hard part was and is productizing, packaging, marketing, financing and distribution. A model by itself is just one part of the puzzle, free or otherwise. In other words, my uncle Bill and my mother can and do use ChatGPT. Fill in the blank open-source model? Maybe as a feature in another product.
>my uncle Bill and my mother can and do use ChatGPT.
They have the name brand for sure. And that is worth a lot.
Notice how Deepseek went from a nobody to making mainstream news though. The only thing people like more than a trusted thing is being able to tell their friends about this amazing cheap good alternative they "discovered".
It's good to be #1 mind share wise but without network effect that still leave you vulnerable
> In other words, my uncle Bill and my mother can and do use ChatGPT
So what? DAUs don't mean anything if there isn't an ad product attached to it. Regular people aren't paying for ChatGPT, and even if they did, the price would need to be several multiples of what Netflix charges to break even.
99% of the world doesn’t care a dime about oss. It’s all saas and what you host behind the saas is only a concern for enterprise (and not every enterprise). And openai or Anthropic can just stop training and host oss models as well.
Eh, distribution of the model is the real moat, theyre doing 700m WAU of the most financially valuable users on earth. If they truly become search, commerce and can use their model either via build or license across b2b, theyre the largest company on earth many times over.
>distribution of the model is the real moat, theyre doing 700m WAU of the most financially valuable users on earth.
Distribution isn't a moat if the thing being distributed is easily substitutable. Everything under the sun is OAI API compatible these days.
700 WAU are fickle AF when a competitor offers a comparable product for half the price.
Moat needs to be something more durable. Cheaper, Better, some other value added tie in (hardware / better UI / memory). There needs to be some edge here. And their obvious edge - raw tech superiority...is looking slim.
Not necessarily. I’m sure there is many cheaper android phones that are technically better in specs but many users won’t change. Once you are familiar, bought into the ecosystem getting rid of it is very hard. I’m lazy myself compared to how I was several years ago. The curious and experimental folks are a minority and the majority ll stick with what works initially instead of constantly analyzing what’s best all the time
The news about how much money Nvidia is investing just so that OpenAI can pay Oracle to pay Nvidia is especially concerning - we seem to be arriving at the financial shell games phase of the bubble.
I am the only one who thinks, "that's not as bad as I expected"?
Because I can be quite bearish and frankly this isn't bad for a technology that is this new. The income points to significant interest in using the tech and they haven't even started the tried-and-true SV strategy we lovingly call enshittification (I'm not trying to be ironic, I mean it)
Seems like despite all the doom about how they were about to be "disrupted", Google might have the last laugh here: they're still quite profitable despite all the Gemini spending, and could go way lower with pricing until OAI and Anthropic have to tap out.
Google also has the advantage of having their own hardware. They aren't reliant on buying Nvidia, and have been developing and using their TPUs for a long time. Google's been an "AI" company since forever
"Each merchant pays a small fee". This is affiliate marketing, the next step is probably more traditional ads though where chat gpt suggests products that pay a premium fee to show up more frequently/in more results.
I can't speak to OpenAI's specific setup, but a lot of startups will use a third party service like Carta to manage their cap table. So there's a website, you have an account, you can log in and it tells you that you have a grant of X shares that vests over Y months. You have to sign a form to accept the grant. There might be some option to do an 83b election if you have stock options rather than RSUs. But that's about it.
In my experience owning private stock, you basically own part of a pool. (Hopefully the exact same classes of shares as the board has or else it's a scam.) The board controls the pool, and whenever they do dividends or transfer ownership, each person's share is affected proportionally. You can petition the board to buy back your shares or transfer them to another shareholder but that's probably unusual for a rank-and-file employee.
The shares are valued by an accounting firm auditor of some type. This determines the basis value if you're paying taxes up-front. After that the tax situation should be the same as getting publicly traded options/shares, there's some choices in how you want to handle the taxes but generally you file a special tax form at the year of grant.
Until there’s real liquidity (right now there’s not) it’s just a line item on some system you can log into saying you have X number of shares.
For all practical purposes it’s worth nothing until there is a liquid market. Given current financials, and preferred cap table terms for those investing cash, shares the average employee has likely aren’t worth much or maybe even anything at the moment.
It's just an entry on some computer. Maybe you can sell it on a secondary market, maybe you can't. You have to wait for an exit event - being acquired by someone else, or an IPO.
You got the right idea there. They wouldn't actually show up in your Fidelity account but there would be a different website where you can log in and see your shares. You wouldn't be able to sell them or transfer them anywhere unless the company arranges a sale and invites you to participate in it.
I definitely don't "get" Silicon Valley finances that much - but how does any investor look at this and think they're ever going to see that money back?
Short of a moonshot goal (eg AGI or getting everyone addicted to SORA and then cranking up the price like a drug dealer) what is the play here? How can OpenAI ever start turning a profit?
All of that hardware they purchase is rapidly depreciating. Training cost are going up exponentially. Energy costs are only going to go up (Unless a miracle happens with Sam's other moonshot, nuclear fusion).
I think people are massively underestimating the money they will come from ads in the future.
They generated $4.3B in revenue without any advertising program to monetise their 700 million weekly active users, most of whom use the free product.
Google earns essentially all of its revenue from ads, $264B in 2024. ChatGPT has more consumer trust than Google at this point, and numerous ways of inserting sponsored results, which they’re starting to experiment with with the recent announcement of direct checkout.
The biggest concern IMO is how good the open weight models coming out of China are, on consumer hardware. But as long as OpenAI remains the go-to for the average consumer, they’ll be fine.
What is OpenAI's competitive moat? There's no product stickiness here.
What prevents people from just using Google, who can build AI stuff into their existing massive search/ads/video/email/browser infrastructure?
Normal, non-technical users can't tell the difference between these models at all, so their usage numbers are highly dependent on marketing. Google has massive distribution with world-wide brands that people already know, trust, and pay for, especially in enterprise.
Google doesn't have to go to the private markets to raise capital, they can spend as much of their own money as they like to market the living hell out of this stuff, just like they did with Chrome. The clock is running on OpenAI. At some point OpenAI's investors are going to want their money back.
I'm not saying Google is going to win, but if I had to bet on which company's money runs out faster, I'm not betting against Google.
Consumer brand quality is so massively underrated by tech people.
ChatGPT has a phenomenal brand. That's worth 100x more than "product stickiness". They have 700 million weekly users and growing much faster than Google.
I think your points on Google being well positioned are apt for capitalization reasons, but only one company has consumer mindshare on "AI" and its the one with "ai" in its name.
I’ve got “normie” friends who I’d bet don’t even know that what Google has at the top of their search results is “AI” results and instead assume it’s just some extension of the normal search results we’ve all gotten used to (knowledge graph)
Every one of them refers to using “ChatGPT” when talking about AI.
How likely is it to stay that way? No idea, but OpenAI has clearly captured a notable amount of mindshare in this new era.
>non-technical users can't tell the difference between these models at all
My non-tech friend said she prefer ChatGPT more than Gemini, most due to its tone.
So non-tech people may not know the different in technical detail, but they sure can have bias.
I have a non-techy friend who used 4o for that exact reason. Compared to most readily available chatbots, 4o just provides more engaging answers to non-techy questions. He likes to have extended conversations about philosophy and consciousness with it. I showed him R1, and he was fascinated by the reasoning process. Makes sense, given the sorts of questions he likes to ask it.
I think OpenAI is pursuing a different market from Google right now. ChatGPT is a companion, Gemini is a tool. That's a totally arbitrary divide, though. Change out the system prompts and the web frontend. Ta-daa, you're in a different market segment now.
Brand. Brand. Brand!
Literally nobody but nerds know what a Claude is among many others.
ChatGPT has name recognition and that matters massively.
ChatGpt has won. I talk to all teens living nearby and they all use chatgpt and not Google.
The teens, they don't know what is OpenAI, they don't know what is Gemini. They sure know what is ChatGPT.
All of these teens use Google Docs instead of OpenAI Docs, Google Meet instead of OpenAI Meet, Gmail instead of OpenAI Mail, etc.
I'm sure that far fewer people to go gemini.google.com than to chatgpt.com, but Google has LLMs seamlessly integrated in each of these products, becoming a part of people's workflows at school and at work.
For a while, I was convinced that OpenAI won and that Google won't be able to recover, but this lack of vertical integration is becoming a huge problem for OpenAI. It's probably why they're trying to branch into weird stuff, like running a walled-garden TikTok clone.
And unlike OpenAI, Google isn't under pressure to monetize AI products any time soon. They can keep subsidizing them until OpenAI runs out of other people's money. I'm not saying OpenAI has no path forward, but it's not all that clear-cut.
I'm not saying you're wrong, but people said the same thing about Yahoo, Excite, Lycos, etc. in 1999. Interesting times, then and now.
I think the biggest risk to ChatGPT as a consumer brand is that they don’t own the device surface. Google / Microsoft / Apple could make great AI that’s infused in the OS / browser, eliminating the need to go to ChatGPT.
Users' chat history is the moat. The more you use it, the more it knows about you and can help you in ways that are customized to particular user. That makes it sticky, more so than web search. Also brand recognition, ChatGPT is the default general purpose LLM choice for most people. Everyone and their mom is using it.
Walling in my personal data would be a sure sign way to get me to not use OpenAI.
Gemini is probably the default general purpose LLM since its answers are inserted into every google result.
It's a distinctive brand, pleasant user experience, and a trustworthy product, like every other commodified technology on the planet.
That's all that matters now. We've passed the "good enough" bar for llms for the majority of consumer use cases.
From here out it's like selling cellphones and laptops
AI has been incredibly sticky. Look at the outrage, OpenAI couldn't even deprecate 4o or whatever because it's incredibly popular. Those people aren't leaving OAI if they're not even leaving a last gen model.
Chats have contexts. While search engines try to track you it is spookier because it is unclear to the user how the contexts are formed. In chats at least the contexts are transparent to both the provider and the user.
Let me direct you to the reddit AMA where people were literally begging to bring back 4o.
I also wonder if this means that even paid tiers will get ads. Google's ad revenue is only ~$30 per user per year, yet there is no paid, ad-free Google Premium, even though lots of users would gladly pay way more than $30/year have an ad-free experience. There's no Google Premium because Google's ad revenue isn't uniformly distributed across users; it's heavily skewed towards the wealthiest users, exactly the users most likely to purchase an ad-free experience. In order to recoup the lost ad revenue from those wealthy users, Google would have to charge something exorbitant, which nobody would be willing to pay.
I fear the same will happen with chatbots. The users paying $20 or $200/month for premium tiers of ChatGPT are precisely the ones you don't want to exclude from generating ad revenue.
"Lots of users would gladly pay way more than $30/year have an ad-free experience"? Outside of ads embedded in Google Maps, a free and simple install of Ublock Origin essentially eliminates ads in Search, YouTube, etc. I'd expect that just like Facebook, people would be very unwilling to pay for Google to eliminate ads, since right now they aren't even willing to add a browser extension.
It worked for YouTube, I don’t see why the assumption of paid gpt models will follow google and not YouTube, particularly when users are conditioned to pay for gpt already.
The average is $x. But that's global which means in some places like the US it is 10x. And in other less wealthy areas it is 0.1x.
There is also the strange paradox that the people who are willing to pay are actually the most desirable advertising targets (because they clearly have $ to spend). So my guess is that for that segment, the revenue is 100x.
I’d agree. The biggest exception I can think of is X, which post-Musk has plans to reduce/remove ads. Though I don’t know how much this tanked their ad revenue and whether it was worth it.
Why would it be any different for youtube premium? I think Google just doesn't think enough people will pay for ad-free search, not that it would cannibalize their ad revenue.
YouTube's ads are much lower-cost than the 'premium' AdWords ones, because the 'intent' is lower, and targeting is worse.
Pretty sure the reason they don't have a paid tier is because engagement (and results) is better when you include ads. Like Facebook found in the early days
> ads in the future.
It boggles my mind that people still think advertising can be a major part of the economy.
If AI is propping up the economy right now [0] how is it possible that the rest of the economy can possibly fund AI through profit sharing? That's fundamentally what advertising is: I give you a share of my revenue (hopefully from profits) in order to help increase my market share. The limit of what advertising spend can be is percent of profits minus some epsilon (for a functioning economy at least).
Advertising cannot be the lions share of any economy because it derives it's value from the rest of the economy.
Advertising is also a major bubble because my one assumption there (that it's a share of profits) is generally not the case. Unprofitable companies giving away a share of their revenue to other companies making those companies profitable is not sustainable.
Advertising could save AI if AI was a relatively small part of the US (or world) economy and could benefit by extracting a share of the profits from other companies. But if most your GDP is from AI how can it possibly cannibalize other companies in a sustainable way?
0. https://www.techspot.com/news/109626-ai-bubble-only-thing-ke...
The moment they start mixing ads into responses Ill stop using them. Open models are good enough, its just more convenient to use chatgpt right now, but that can change.
People said the same thing about so many other online services since the 90s. The issue is that you're imagining ChatGPT as it exists right now with your current use case but just with ads inserted into their product. That's not really how these things go... instead OpenAI will wait until their product becomes so ingrained in everyday usage that you can't just decide to stop using them. It is possible, although not certain, that their product becomes ubiquitous and using LLMs someway somehow just becomes a normal way of doing your job, or using your computer, or performing menial and ordinary tasks. Using an LLM will be like using email, or using Google maps, or some other common tool we don't think much of.
That's when services start to insert ads into their product.
Except it's hard to imagine a world where chatgpt is heads and shoulders over the other llms in capability. Google has no problem keeping up and let's not forget that China has state-sponsored programs for AI development.
And if/when they reach that point, the average consumer will see the ad as an irksome fly. That's it.
I agree, but the question is whether or not normal people will stop using them.
I think the empirical answer is no. Look at how many ads there are in everything and people still use it.
Normal people ignore ads. It gets easier with time. Television with ads means conversation about what we just watched.
> moment they start mixing ads into responses Ill stop using them
Do you currently pay for it?
> But as long as OpenAI remains the go-to for the average consumer, they be fine.
This is like the argument of a couple of years ago "as long as Tesla remains ahead of the Chinese technology...". OpenAI can definitely become a profitable company but I dont see anything to say they will have a moat and monopoly.
They're the only ones making AI with a personality. Yeah, you don't need chocolate flavored protein shakes but if I'm taking it every day, I get sick of the vanilla flavor.
Huh? They're actively removing personality from current models as much as possible.
He means chatGPT means AI to most people.
Did you mean the GPT-5 launch? They put it back in within 2 weeks, despite the side effects and bugs. It was pretty clear that it's their value proposition.
I think this is directionally right but to nitpick…Google has way more trust than OpenAI right now and it’s not close.
Acceleration is felt, not velocity.
Yeah, I agree with you.
Between Android, Chrome, YouTube, Gmail (including mx.google.com), Docs/Drive, Meet/Chat, and Google Search, claiming that Google "isn't more trusted" is just ludicrous. People may not be happy they have to trust Alphabet. But they certainly do.
And even when they insist they're Stallman, their friends do, their family does, their coworkers do, the businesses they interact with do, the schools they send their children to do.
Like it or not, Google has wormed their way into the fabric of modern life.
Chrome and Google Search are still the gateway to the internet outside China. Android has over 75% market share of all mobile(!). YouTube is somewhat uniquely the video internet with Instagram and Tiktok not really occupying the same mindshare for "search" and long form.
People can say they don't "trust" Google but the fact is that if the world didn't trust Google, it never would have gotten to where it is and it would quickly unravel from here.
Sent from my Android (begrudgingly)
I really don't trust either. Google because of what they've already done, OpenAI because it has a guy at the helm who doesn't know how to spell the word 'ethics'.
That's mostly because LLMs think in terms of tokens not letters, so spelling is hard.
He knows there's no "I" in "ethics"
This really depends on where you are are. Some countries' populations, especially those known to value privacy, are extremely distrustful of anything associated with Facebook or Google.
I agree with you, and my impression of the trust-level of Google is pretty much zero.
Google and trust are an oxymoron
The only thing I trust google to do is abandon software and give me a terrible support experience
And to charge you for stuff you don't want and don't need as if you are using it every day through tied sales. Hm... wasn't that illegal?
https://www.investopedia.com/terms/t/tiedselling.asp
If they overnight were able to capture as much revenue per user as Meta (about 50 bucks a year) they'd bring in a bucket of cash immediately.
But selling that much ad inventory overnight - especially if they want new formats vs "here's a video randomly inserted in your conversation" sorta stuff - is far from easy.
Their compute costs could easily go down as technology advances. That helps.
But can they ramp up the advertising fast enough to bring in sufficient profit before cheaper down-market alternatives become common?
They lack the social-network lock-in effect of Meta, or the content of ESPN, and it remains to be seen if they will have the "but Google has better results than Bing" stickiness of Google.
"Please help me with my factorial homework."
Drink verification can
It'll be interesting to see the effect ads have on their trustworthiness. There's potential for it to end up worse than Google because sponsored content can blend in better and possibly not be reliably disclosed.
There is also the IMO not exactly settled question of whether an advertiser is comfortable handing over its marketing to an AI.
Can any AI be sensibly and reliably instructed not to do product placement in inappropriate contexts?
Also what effect will these extra instructions have on output?
Every token of context can drastically change the output. That's a big issue right now with Claude and their long conversation reminders.
> $264B in 2024.
Why is this much money spent on advertising? Surely it isn't really justified by increase in sales that could be attributed to the ads? You're telling me people actually buy these ridiculous products I see advertised?
A lot of that money comes from search result ads. Sometimes I click on an ad to visit a site I search for instead of scrolling to the same link in the actual search results. Many companies bid on keywords for their own name to prevent others from taking a customer who is interested in you.
You use to be a useful site and be at the top of the search results for some keywords and now you have to pay.
It's a lot more complicated, but yes advertising works.
There is a saying in India, whats seen is what is sold.
Not the hidden best product.
They should be concerned with open weight models that don’t run on consumer hardware. The larger models from Qwen (Qwen Max) and ZLM (GLM and GLM air) perform not too far from Claude Sonnet 4 and GPT-5. ZLM offers a $3 plan that is decently generous. I can pretty much replace it over Sonnet 4 in Claude Code (I swear, Anthropic has been nerfing Sonnet 4 for people on the Pro plan).
You can run Qwen3-coder for free upto 1000 requests a day. Admittedly not state of the art but works as good of 5o-mini
I believe regular people will not change from chatGPT if it has some ads. I know people who use "alternative" wrappers that have ads because they aren't tech savvy, and I agree with the OP that this could be a significant amount of money We aren't 700 million people that use it.
Definitely don’t argue against that, once people get into a habit of using something, it takes quite a bit to get away from it. Just that an American startup can literally run ZLM models themselves (open weight with permissive license) as a competitor to ChatGPT is pretty wild to think about
One of the side effects of having a chat interface, is that there is no moat around it. Using it is natural.
Changing from Windows to Mac or iOS to Android requires changing the User Interface. All of these chat applications have essentially the same interface. Changing between ChatGPT and Claude is essentially like buying a different flavor of potato chip. There is some brand loyalty and user preference, but there is very little friction.
I don’t pay $200 per month to use a product tightly coupled for ad revenue (ahem tracking).
That’s why I use Kagi, Hey, Telegram, Apple (for now) etc.
I really hope OpenAI can build a sustainable model which is not based on that.
I suspect ads would be an attempt to monetize the free users not people paying $200/mo for Pro. Though who knows...
First, as an advertiser you want those sweet-sweet people with money.
Second, if they put “display: hidden” on ads doesn’t mean they will create and use entirely other architecture, data flow and storage, just for those pro users.
What is OpenAI's moat? There's plenty of competitors running their own models and tools. Sure, they have the ChatGPT name, but I don't see them massively out-competing the entire market unless the future model changes drastically improve over the 3->4->5 trajectory.
It feels similar to Google to me - what is (was) their moat? Basically slightly better results and strong brand recognition. In the later days maybe privileged data access. But why does nobody use Bing?
Google got a massive leg up on the rest be having a better service. When Bing first came out, I was not impressed with what I got, and never really bothered going back to it.
Search quality isn't what it used to be, but the inertia is still paying dividends. That same inertia also applied to Google ads.
I'm not nearly so convinced OpenAI has the same leg up with ChatGPT. ChatGPT hasn't become a verb quite like google or Kleenex, and it isn't an indispensable part of a product.
I actually find bing better now for more technical searches.
Most technical Google searches end up at win fourms or official Microsoft support site which is basically just telling you that running sfc scannow for everything is the fix.
Google has always been much better than the competition. Even today with their enshittification, competitors still aren’t as good.
The only thing that has changed that status quo is the rise of audiovisual media and sites closing up so that Google can’t index them, which means web search lost a lot of relevance.
It's Sam.
From what I understand he was the only one crazy enough to demand hundreds of GPUs for months to get ChatGPT going. Which at the time sounded crazy.
So yeah Sam is the guy with the guts and vision to stay ahead.
Past performance is no guarantee of future results.
You might see Sam as a Midas who can turn anything into gold. But history shows that very few people sustain that pattern.
The shitcoin-in-return-for-your-iris-scans guy?
OpenAI isn't ahead
It is in terms of users. There's a lot of sticking power to "the thing you already know and use".
Probably consumers. Enterprise is Anthropic, double ahead: https://menlovc.com/perspective/2025-mid-year-llm-market-upd...
note that menlo is invested in anthropic, but still..
Unless there is little friction in switching. I don’t feel any of the LLM products have a sticky factor as of yet, as far as viewing it from a consumer lens
Ignoring Sutskever much?
This! The cost of training models inevitably goes down over time as FLOPS/$ and PB/$ increases relentlessly thanks to the exponential gains of Moore's law. Eventually we will end up with laptops and phones being Good Enough to run models locally. Once that happens, any competitor in the space that decides to actively support running locally will have operating costs that are a mere fraction of OpenAI's current business.
The pop of this bubble is going to be painful for a lot of people. Being too early to a market is just as bad as being too late, especially for something that can become a commodity due to a lack of moat.
Bad news on the Moore's Law front.
https://cap.csail.mit.edu/death-moores-law-what-it-means-and...
You just said that everyone will be able to run a powerful AI locally and then you said this would lead to a pop of the bubble.
Well, which is it? That AI is going to have huge demands for chips that it is going to get much bigger or is the bubble going to pop? You can’t have both.
My opinion is that local LLMs will do a bulk of the low value interference such as your personal life mundane tasks. But cloud AI will be reserved for work and for advanced research purposes.
brand recognition
Google has google.com, youtube, chrome, android, gmail, google map etc ... I don't see OpenAI having a product close to that.
Google is older and many of the products you describe were acquisitions (inorganic growth).
> google.com, youtube, chrome, android, gmail, google map etc
Of those, it's 50/50. The acquisitions were YT, Android, Maps. Search was obviously Google's original product, Chrome was an in-house effort to rejuvenate the web after IE had caused years of stagnation, and Gmail famously started as a 20% project.
There are of course criticisms that Google has not really created any major (say, billion-user) in-house products in the past 15 years.
By this point I imagine it's a novelty to find any code from the original acquisition in those products.
Code is secondary. Pmf is primary.
Code is a commodity. Very easy to make. Now even llms are commodities. There are other intangible assets more valuable. Like the chatgpt brand here.
This is beyond PMF, it's about traction on steroids, owning the last mile.
>> ... underestimating the money they will come from ads in the future.
I would like AI to focus on helping consumers discover the right products for their stated needs as opposed to just being shown (personalized) ads. As of now, I frequently have a hard time finding the things I need via Amazon search, Google, as well as ChatGPT.
The problem with this is that I have moved to Gemini with zero loss in functionality, and I’m pretty sure that Google is 100x better at ads than OpenAI.
In 10 years most serious users of AI will be using local LLMs on insanely powerful devices, with no ads. API based services will have limited horizon.
Some will but you’re underestimating the burning desire to avoid IT and sysadmin work. Look at how much companies overspend on cloud just to not have to do IT work. They’d rather pay 10X-100X more to not have to admin stuff.
It is just downloading a program and using it
>"Look at how much companies overspend on cloud just to not have to do IT work."
I think they are doing it for a different reasons. Some are legit like renting this supercomputer for a day and some are like everybody else is doing it. I am friends with the small company owner and they have sysadmin who picks nose and does nothing and then they pay a fortune to Amazon
I’m talking about prepackaged offline local only on device LLMs.
What you are describing though will almost certainly happen even sooner once AI tech stablizes and investing in powerful hardware no longer means you will become quickly out of date.
Ok, but there will be users using even more insanely powerful datacenter computers that will be able to our-AI the local AI users.
Nvidia/Apple (hardware companies) are the only winner in this case
Are they currently adding affiliate links to their outbound Amazon product links?
Also, their press release said:
Vague. Is this advertising? Eh, not sure, but that is a big chunk of money.I think they mean OpenAI showing ads from other companies to users, not buying ads themselves.
Banner ads would only be the start of the enshittification of AI chats. I can't wait for the bots to start recommending products and services of the highest bidder.
Everyone is trying to compare AI companies with something that happened in the past, but I don't think we can predict much from that.
GPUs are not railroads or fiber optics.
The cost structure of ChatGPT and other LLM based services is entirely different than web, they are very expensive to build but also cost a lot to serve.
Companies like Meta, Microsoft, Amazon, Google would all survive if their massive investment does not pay off.
On the other hand, OpenAI, Anthropic and others could be soon find themselves in a difficult position and be at the mercy of Nvidia.
Unlike railroads and fibre, all the best compute in 2025 will be lacklustre in 2027. It won’t retain much value in the same way as the infrastructure of previous bubbles did?
The A100 came out 5.5 years ago and is still the staple for many AI/ML workloads. Even AI hardware just doesn’t depreciate that quickly.
Don't they degrade physically from being run at full blast 24/7 for so many years?
This. There’s even a market for them being built (DRW).
> Unlike railroads and fibre, all the best compute in 2025 will be lacklustre in 2027.
I definitely don't think compute is anything like railroads and fibre, but I'm not so sure compute will continue it's efficiency gains of the past. Power consumption for these chips is climbing fast, lots of gains are from better hardware support for 8bit/4bit precision, I believe yields are getting harder to achieve as things get much smaller.
Betting against compute getting better/cheaper/faster is probably a bad idea, but fundamental improvements I think will be a lot slower over the next decade as shrinking gets a lot harder.
Unfortunately changing 2027 to 2030 doesn't make the math much better
> changing 2027 to 2030 doesn't make the math much better
Could you show me?
Early turbines didn't last that long. Even modern ones are only rated for a few decades.
>> Unlike railroads and fibre, all the best compute in 2025 will be lacklustre in 2027.
> I definitely don't think compute is anything like railroads and fibre, but I'm not so sure compute will continue it's efficiency gains of the past. Power consumption for these chips is climbing fast, lots of gains are from better hardware support for 8bit/4bit precision, I believe yields are getting harder to achieve as things get much smaller.
I'm no expert, buy my understanding is that as feature sizes shrink, semiconductors become more prone to failure over time. Those GPUs probably aren't going to all fry themselves in two years, but even if GPUs stagnate, chip longevity may limit the medium/long term value of the (massive) investment.
Unfortunately the chips themselves probably won’t physically last much longer than that under the workloads they are being put to. So, yes, they won’t be totally obsolete as technology in 2028, but they may still have to be replaced.
Yeah - I think that the extremely fast depreciation just due to wear and use on GPUs is pretty unappreciated right now. So you've spent 300 mil on a brand new data center - congrats - you'll need to pay off that loan and somehow raise another 100 mil to actually maintain that capacity for three years based on chip replacement alone.
There is an absolute glut of cheap compute available right now due to VC and other funds dumping into the industry (take advantage of it while it exists!) but I'm pretty sure Wall St. will balk when they realize the continued costs of maintaining that compute and look at the revenue that expenditure is generating. People think of chips as a piece of infrastructure - you buy a personal computer and it'll keep chugging for a decade without issue in most case - but GPUs are essentially consumables - they're an input to producing the compute a data center sells that needs constant restocking - rather than a one-time investment.
Do we actually know how they're degrading? Are there still Pascals out there? If not, is it because they actual broke or because of poor performance? I understand it's tempting to say near 100% workload for multiple years = fast degradation, but what are the actual stats? Are you talking specifically about the actual compute chip or the whole compute system -- I know there's a big difference now with the systems Nvidia is selling. How long do typical Intel/AMD CPU server chips last? My impression is a long time.
If we're talking about the whole compute system like a gb200, is there a particular component that breaks first? How hard are they to refurbish, if that particular component breaks? I'm guessing they didn't have repairability in mind, but I also know these "chips" are much more than chips now so there's probably some modularity if it's not the chip itself failing.
I watch a GPU repair guy and its interesting to see the different failure modes...
* memory IC failure
* power delivery component failure
* dead core
* cracked BGA solder joints on core
* damaged PCB due to sag
These issues are compounded by
* huge power consumption and heat output of core and memory, compared to system CPU/memory
* physical size of core leads to more potential for solder joint fracture due to thermal expansion/contraction
* everything needs to fit in PCIe card form factor
* memory and core not socketed, if one fails (or supporting circuitry on the PCB fails) then either expensive repair or the card becomes scrap
* some vendors have cards with design flaws which lead to early failure
* sometimes poor application of thermal paste/pads at factory (eg, only half of core making contact
* and, in my experience in aquiring 4-5 year old GPUs to build gaming PCs with (to sell), almost without fail the thermal paste has dried up and the card is thermal throttling
Believe it or not, the GPUs from bitcoin farms are often the most reliable.
Since they were run 24/7, there was rarely the kind of heat stress that kills cards (heating and cooling cycles).
I'm not sure.
Number of cycles that goes through silicon matters, but what matters most really are temperature and electrical shocks.
If the GPUs are stable, at low temperature they can be at full load for years. There are servers out there up from decades and decades.
Yep, we are (unfortunately) still running on railroad infrastructure built a century ago. The amortization periods on that spending is ridiculously long.
Effectively every single H100 in existence now will be e-waste in 5 years or less. Not exactly railroad infrastructure here, or even dark fiber.
> Effectively every single H100 in existence now will be e-waste in 5 years or less.
This is definitely not true, the A100 came out just over 5 years ago and still goes for low five figures used on eBay.
> Effectively every single H100 in existence now will be e-waste in 5 years or less.
This remains to be seen. H100 is 3 years old now, and is still the workhorse of all the major AI shops. When there's something that is obviously better for training, these are still going to be used for inference.
If what you say is true, you could find a A100 for cheap/free right now. But check out the prices.
Yeah, I can rent an A100 server for roughly the same price as what the electricity would cost me.
That is true for almost any cloud hardware.
Where?
~$1.25-1.75/hr at Runpod or vast.ai for an A100
Edit: https://getdeploying.com/reference/cloud-gpu/nvidia-a100
> Yep, we are (unfortunately) still running on railroad infrastructure built a century ago.
That which survived, at least. A whole lot of rail infrastructure was not viable and soon became waste of its own. There was, at one time, ten rail lines around my parts, operated by six different railway companies. Only one of them remains fully intact to this day. One other line retained a short section that is still standing, which is now being used for car storage, but was mostly dismantled. The rest are completely gone.
When we look back in 100 years, the total amortization cost for the "winner" won't look so bad. The “picks and axes” (i.e. H100s) that soon wore down, but were needed to build the grander vision won't even be a second thought in hindsight.
> That which survived, at least. A whole lot of rail infrastructure was not viable and soon became waste of its own. There was, at one time, ten rail lines around my parts, operated by six different railway companies. Only one of them remains fully intact to this day. One other line retained a short section that is still standing, which is now being used for car storage, but was mostly dismantled. The rest are completely gone.
How long did it take for 9 out of 10 of those rail lines to become nonviable? If they lasted (say) 50 years instead of 100, because that much rail capacity was (say) obsoleted by the advent of cars and trucks, that's still pretty good.
> How long did it take for 9 out of 10 of those rail lines to become nonviable?
Records from the time are few and far between, but, from what I can tell, it looks like they likely weren't ever actually viable.
The records do show that the railways were profitable for a short while, but it seems only because the government paid for the infrastructure. If they had to incur the capital expenditure themselves, the math doesn't look like it would math.
Imagine where the LLM businesses would be if the government paid for all the R&D and training costs!
> The records do show that the railways were profitable for a short while, but it seems only because the government paid for the infrastructure. If they had to incur the capital expenditure themselves, the math doesn't look like it would math.
Actually, governments in the US rarely actually provided any capital to the railroads. (Some state governments did provide some of the initial capital for the earliest railroads). Most of federal largess to the railroads came in the form of land grants, but even the land grant system for the railroads was remarkably limited in scope. Only about 7-8% of the railroad mileage attracted land grants.
Railroads were pretty profitable for a long time. The western long haul routes were capitalized by land transfers.
What killed them was the same thing that killed marine shipping — the government put the thumb on the scale for trucking and cars to drive postwar employment and growth of suburbs, accelerate housing development, and other purposes.
> the government put the thumb on the scale for trucking and cars to drive postwar employment and growth of suburbs, accelerate housing development, and other purposes.
The age of postwar suburb growth would be more commonly attributed to WWII, but the records show these railroads were already losing money hand over fist by the WWI era. The final death knell, if there ever was one, was almost certainly the Great Depression.
But profitable and viable are not one and the same, especially given the immense subsidies at play. You can make anything profitable when someone else is covering the cost.
If 1/10 investment lasts 100 years that seems pretty good to me. Plus I'd bet a lot of the 9/10 of that investment had a lot of the material cost re-coup'd when scrapping the steel. I don't think you're going to recoup a lot of money from the H100s.
Much like LLMs. There are approximately 10 reasonable players giving it a go, and, unless this whole AI thing goes away, never to be seen again, it is likely that one of them will still be around in 100 years.
H100s are effectively consumables used in the construction of the metaphorical rail. The actual rail lines had their own fare share of necessary tools that retained little to no residual value after use as well. This isn't anything unique.
H100s being thought of as consumables is keen - it much better to analogize the H100s to coal and chip manufacturer the mine owner - than to think of them as rails. They are impermanent and need constant upkeep and replacement - they are not one time costs that you build as infra and forget about.
How was your trip down the third Avenue El? Did your goods arrive via boxcar to 111 8th Ave?
At the rate they are throwing obstacles at the promised subway which they got rid of the 3rd Ave El for maybe his/her grandkids will finish the trip.
> Yep, we are (unfortunately) still running on railroad infrastructure built a century ago. The amortization periods on that spending is ridiculously long.
Are we? I was under the impression that the tracks degraded due to stresses like heat/rain/etc. and had to be replaced periodically.
The track bed, rails, and ties will have been replaced many times by now. But the really expensive work was clearing the right of way and the associated bridges, tunnels, etc.
I am really digging the railroad analogies in this discussion! There are some striking similarities if you do the right mappings and timeframe transformations.
I am an avid rail-to-trail cycler and more recently a student of the history of the rail industry. The result was my realization that the ultimate benefit to society and to me personally is the existence of these amazing outdoor recreation venues. Here in Western PA we have many hundreds of miles of rail-to-trail. My recent realization is that it would be totally impossible for our modern society to create these trails today. They were built with blood, sweat, tears and much dynamite - and not a single thought towards environmental impact studies. I estimate that only ten percent of the rail lines built around here are still used for rail. Another ten percent have become recreational trails. That percent continues to rise as more abandoned rail lines transition to recreational use. Here in Western PA we add a couple dozen miles every year.
After reading this very interesting discussion, I've come to believe that the AI arms race is mainly just transferring capital into the pockets of the tool vendors - just as was the case with the railroads. The NVidia chips will be amortized over 10 years and the models over perhaps 2 years. Neither has any lasting value. So the analogy to rail is things like dynamite and rolling stock. What in AI will maintain value? I think the data center physical plants, power plants and transmission networks will maintain their value longer. I think the exabytes of training data will maintain their value even longer.
What will become the equivalent of rail-to-trail? I doubt that any of the laborers or capitalists building rail lines had foreseen that their ultimate value to society would be that people like me could enjoy a bike ride. What are the now unforeseen long-term benefit to society of this AI investment boom?
Rail consolidated over 100 years into just a handful of firms in North America, and my understanding is that these firms are well-run and fairly profitable. I expect a much more rapid shakeout and consolidation to happen in AI. And I'm putting my money on the winners being Apple first and Google second.
Another analogy I just thought of - the question of will the AI models eventually run on big-iron or in ballpoint pens. It is similar to the dichotomy of large-scale vs miniaturized nuclear power sources in Asimov's Foundation series (a core and memorable theme of the book that I haven't seen in the TV series).
"...all the best compute in 2025 will be lacklustre in 2027": How does the compute (I assume you mean on PCs) of 2025 compare with the compute of 2023?
Oh wait, the computer I'm typing this on was manufactured in 2020...
Neato. How’s that 1999 era laptop? Because 25 year old trains are still running and 25 year old train track is still almost new. It’s not the same and you know it.
last month HN was talking about a win95 with floppy drivers handling rail in Germany no less
Unlike 1875, we have Saudi and other tillion/billionaires willing commit almost any amount to own the future of business.
Except they behave less like shrewd investors and more like bandwagon jumpers looking to buy influence or get rich quick. Crypto, Twitter, ridesharing, office sharing and now AI. None of these have been the future of business.
Business looks a lot like what it has throughout history. Building physical transport infrastructure, trade links, improving agricultural and manufacturing productivity and investing in military advancements. In the latter respect, countries like Turkey and Iran are decades ahead of Saudi in terms of building internal security capacity with drone tech for example.
Agreed - I don’t think they are particularly brilliant as a category. Hereditary kleptocracy has limits.
But… I don’t think there’s an example in modern history of the this much capital moving around based on whim.
The “bet on red” mentality has produced some odd leaders with absolute authority in their domain. One of the most influential figures on the US government claims to believe that he is saving society from the antichrist. Another thinks he’s the protagonist in a sci-fi novel.
We have the madness of monarchy with modern weapons and power. Yikes.
Exactly: when was the last time you used ChatGPT-3.5? Its value deprecated to zero after, what, two-and-a-half years? (And the Nvidia chips used to train it have barely retained any value either)
The financials here are so ugly: you have to light truckloads of money on fire forever just to jog in place.
OpenAI is now valued at $500bn though. I doubt the investors are too wrecked yet.
It may be like looking at the early Google and saying they are spending loads on compute and haven't even figured how to monetize search, the investors are doomed.
I would think that it's more like a general codebase - even if after 2.5 years, 95% percent of the lines were rewritten, and even if the whole thing was rewritten in a different language, there is no point in time at which its value diminished, as you arguably couldn't have built the new version without all the knowledge (and institutional knowledge) from the older version.
I rejoined an previous employer of mine, someone everyone here knows ... and I found that half their networking equipment is still being maintained by code I wrote in 2012-2014. It has not been rewritten. Hell, I rewrote a few parts that badly needed it despite joining another part of the company.
A really did few days ago gpt-3.5-fast is a great model for certain tasks and cost wise via the API. Lots of solutions being built on the today’s latest are for tomorrow’s legacy model — if it works just pin the version.
> And the Nvidia chips used to train it have barely retained any value either
Oh, I'd love to get a cheap H100! Where can I find one? You'll find it costs almost as much used as it's new.
> money on fire forever just to jog in place.
Why?
I don't see why these companies can't just stop training at some point. Unless you're saying the cost of inference is unsustainable?
I can envision a future where ChatGPT stops getting new SOTA models, and all future models are built for enterprise or people willing to pay a lot of money for high ROI use cases.
We don't need better models for the vast majority of chats taking place today E.g. kids using it for help with homework - are today's models really not good enough?
>I don't see why these companies can't just stop training at some point.
Because training isn't just about making brand new models with better capabilities, it's also about updating old models to stay current with new information. Even the most sophisticated present-day model with a knowledge cutoff date of 2025 would be severely crippled by 2027 and utterly useless by 2030.
Unless there is some breakthrough that lets existing models cheaply incrementally update their weights to add new information, I don't see any way around this.
They aren't. They are obsequious. This is much worse than it seems at first glance, and you can tell it is a big deal because a lot of effort going into training the new models is to mitigate it.
But is it a bit like a game of musical chairs?
At some point the AI becomes good enough, and if you're not sitting in a chair at the time, you're not going to be the next Google.
Not necessarily? That assumes that the first "good enough" model is a defensible moat - i.e., the first ones to get there becomes the sole purveyors of the Good AI.
In practice that hasn't borne out. You can download and run open weight models now that are spitting distance to state-of-the-art, and open weight models are at best a few months behind the proprietary stuff.
And even within the realm of proprietary models no player can maintain a lead. Any advances are rapidly matched by the other players.
More likely at some point the AI becomes "good enough"... and every single player will also get a "good enough" AI shortly thereafter. There doesn't seem like there's a scenario where any player can afford to stop setting cash on fire and start making money.
It's not that the investments just won't pay off, it's that the global markets are likely to crash like happened with the subprime mortgage crisis.
This is much closer to the dotcom boom than the subprime stuff. The dotcom boom/bust affected tech more than anything else. It didn’t involve consumers like the housing crash did.
The dot com boom involved silly things like Pets.com IPOing pre-revenue. Claude code hit $500m in ARR in 3 months.
The fact people don't see the difference between the two is unreal. Hacker news has gone full r* around this topic, you find better nuance even on Reddit than here.
They're not claiming that it's like the dot com boom because no one is actually making money. They're claiming that this is more like the dot com boom than the housing bubble, which I think is true. The dot com crash didn't cause Jane-on-the-street to lose her house while she worked a factory job, though the housing crisis did have those kinds of consumer-affecting outcomes.
You have a good point. Pets.com would have fared much better if investors gave them several billion dollars in 1998, 1999 and then again in 2000
can see cramer "buy pets.com! general revenue is just around the corner"
But it does involve a ton of commercial real estate investment, as well as a huge shakeup in the energy market. People may not lose their homes, but we'll all be paying for this one way or another.
The fed could still push the real value of stocks quite a bit by destroying the USD, if they want, by pinning interest rates near 0 and forcing a rush to the exits to buy stock and other asset classes.
The point still stands though. All these other companies can pivot to some thing else if AI fails but what will OpenAI do?
By the time it catches up with them they will have IPO’d and dumped their problem onto the public market. The administration will probably get a golden share and they will get a bail out in an effort to soften the landing for their campaign donors that also have huge positions. All the rich people will be made whole and the US tax payer will pay the price of the bail out.
And Microsoft or whoever will absorb the remains of their technology.
Sell to Microsoft and be absorbed there (and Anthropic to Amazon).
> but what will OpenAI do?
Will get acquired at “Store Closing” price!!
I'm reminded of the quote "If you owe the bank $100 that's your problem. If you owe the bank $100 million, that's the bank's problem." - J. Paul Getty
Nvidia may well be at the mercy of them! Hence the recent circular dealing
In the end Revenues > Costs or you have an issue. That "startup" money will eventually be gone, and you're back to MIMO Money In vs Money Out and if it's not > , you will go bankrupt.
Businesses are different but the fundamentals of business and finance stay consistent. In every bubble that reality is unavoidable, no matter how much people say/wish “but this time is different.”
The past/present company they remind me of the most is semiconductor fabs. Significant generation-to-generation R&D investment, significant hardware and infrastructure investment, quite winner-takes-all on the high end, obsoleted in a couple years at most.
The main differences are these models are early in their development curve so the jumps are much bigger, and they are entirely digital so they get “shipped” much faster, and open weights seem to be possible. None of those factors seem to make it a more attractive business to be in.
If you build the actual datacenter, less than half the cost is the actual compute. The other half is the actual datacenter infrastructure, power infrastructure, and cooling.
So in that sense it's not that much different from Meta and Google which also used server infrastructure that depreciated over time. The difference is that I believe Meta and Google made money hand over fist even in their earliest days.
Last time i ran the numbers -
Data center facilities are ~$10k per kW
IT gear is like $20k-$50k per kW
Data center gear is good for 15-30 years. IT is like 2-6ish.
Would love to see updated numbers. Got any?
The funniest thing about all this is that the biggest difference between LLMs from Anthropic, Google, OpenAI, Alibaba is not model architecture or training objectives, which are broadly similar but it's the dataset. What people don't realize is how much of that data comes from massive undisclosed scrapes + synthetic data + countless hours of expert feedback shaping the models. As methodologies converge, the performance gap between these systems is already narrowing and will continue to diminish over time.
Just because they have ongoing costs after purchasing them doesn't mean it's different than something else we've seen? What are you trying to articulate exactly, this is a simple business and can get costs under control eventually, or not
I think the most interesting numbers in this piece (ignoring the stock compensation part) are:
$4.3 billion in revenue - presumably from ChatGPT customers and API fees
$6.7 billion spent on R&D
$2 billion on sales and marketing - anyone got any idea what this is? I don't remember seeing many ads for ChatGPT but clearly I've not been paying attention in the right places.
Open question for me: where does the cost of running the servers used for inference go? Is that part of R&D, or does the R&D number only cover servers used to train new models (and presumably their engineering staff costs)?
Free usage usually goes in sales and marketing. It's effectively a cost of acquiring a customer. This also means it is considered an operating expense rather than a cost of goods sold and doesn't impact your gross margin.
Compute in R&D will be only training and development. Compute for inference will go under COGS. COGS is not reported here but can probably be, um, inferred by filling in the gaps on the income statement.
(Source: I run an inference company.)
Marketing != advertising. Although this budget probably does include some traditional advertising. It is most likely about building the brand and brand awareness, as well as partnerships etc. I would imagine the sales team is probably quite big, and host all kinds of events. But I would say a big chunk of this "sales and marketing" budget goes into lobbying and government relations. And they are winning big time on that front. So it is money well spent from their perspective (although not from ours). This is all just an educated guess from my experience with budgets from much smaller companies.
I agree - they're winning big and booking big revenue.
If you discount R&D and "sales and marketing", they've got a net loss of "only" $500 million.
They're trying to land grab as much surface area as they can. They're trying to magic themselves into a trillion dollar FAANG and kill their peers. At some point, you won't be able to train a model to compete with their core products, and they'll have a thousand times the distribution advantage.
ChatGPT is already a new default "pane of glass" for normal people.
Is this all really so unreasonable?
I certainly want exposure to their stock.
> If you discount R&D and "sales and marketing"
If you discount sales & marketing, they will start losing enterprise deals (like the US government). The lack of a free tier will impact consumer/prosumer uptake (free usage usually comes out of the sales & marketing budget).
If you discount R&D, there will be no point to the business in 12 months or so. Other foundation models will eclipse them and some open source models will likely reach parity.
Both of these costs are likely to increase rather than decrease over time.
> ChatGPT is already a new default "pane of glass" for normal people.
OpenAI should certainly hope this is not true, because then the only way to scale the business is to get all those "normal" people to spend a lot more.
We gave ChatGPT advertising on bus-stops here in the UK.
Two people in a cafe having a meet-up, they are both happy, one is holding a phone and they are both looking at it.
And it has a big ChatGPT logo in the top right corner of the advertisement - transparent just the black logo with ChatGPT written underneath.
That's it. No text or anything telling you what the product is or does. Just it will make you happy during conversations with friends somehow.
> $2 billion on sales and marketing - anyone got any idea what this is?
Not sure where/how I read it, but remember coming across articles stating OpenAI has some agreements with schools, universities and even the US government. The cost of making those happen would probably go into "sales & marketing".
Most folks that are not an engineer building is likely classified as “sales and marketing.” “Developer advocates” “solutions architects” and all that stuff included.
So probably just write-offs of tokens they give away?
This will include the people cost of sales and marketing teams.
Stop R&D and the competition is at parity with 10x cheaper models in 3-6 months.
Stop training and your code model generates tech debt after 3-6 month
It's pretty well accepted now that for pre-training LLMs the curve is S not an exponential, right? Maybe it's all in RL post-training now, but my understanding(?) is that it's not nearly as expensive as pre-training. I don't think 3-6 months is the time to 10X improvement anymore (however that's measured), it seems closer to a year and growing assuming the plateau is real. I'd love to know if there are solid estimates on "doubling times" these days.
With the marginal gains diminishing, do we really think they're (all of them) are going to continue spending that much more for each generation? Even the big guys with the money like google can't justify increasing spending forever given this. The models are good enough for a lot of useful tasks for a lot of people. With all due respect to the amazing science and engineering, OpenAI (and probably the rest) have arrived at their performance with at least half of the credit going to brute-force compute, hence the cost. I don't think they'll continue that in the face of diminishing returns. Someone will ramp down and get much closer to making money, focusing on maximizing token cost efficiency to serve and utility to users with a fixed model(s). GPT-5 with it's auto-routing between different performance models seems like a clear move in this direction. I bet their cost to serve the same performance as say gemini 2.5 is much lower.
Naively, my view is that there's some threshold raw performance that's good enough for 80% of users, and we're near it. There's always going to be demand for bleeding edge, but money is in mass market. So if you hit that threshold, you ramp down training costs and focus on tooling + ease of use and token generation efficiency to match 80% of use cases. Those 80% of users will be happy with slowly increasing performance past the threshold, like iphone updates. Except they probably won't charge that much more since the competition is still there. But anyway, now they're spending way less on R&D and training, and the cost to serve tokens @ the same performance continues to drop.
All of this is to say, I don't think they're in that dreadful of a position. I can't even remember why I chose you to reply to, I think the "10x cheaper models in 3-6 months" caught me. I'm not saying they can drop R&D/training to 0. You wouldn't want to miss out on the efficiency of distillation, or whatever the latest innovations I don't know about are. Oh and also, I am confident that whatever the real number N is for NX cheaper in 3-6 months, a large fraction of that will come from hardware gains that are common to all of the labs.
Google has the best story imo. Gemini > Azure - it will accelerate GCP growth.
Also R&D, for tax purposes, likely includes everyone at the company who touches code so there's probably a lot of operational cost being hidden in that number.
> $2 billion on sales and marketing - anyone got any idea what this is?
enterprise sales are expensive. And selling to the US government is on a very different level.
I've seen some OpenAI ads on Italian tv and they made no sense to me, they tried hard to be apple like, but realistically nobody knew what they were about.
Italian advertising is weird in general. Month ago leaving Venice we pulled over on a gas station and I started just going thru pages on some magazine. At some point I see advertising on what looks like old fashioned shoes - and owner of the company holding his son with sign "from generation to generation". Only thing - the ~3 year old boy is completely naked wearing only shoes with his little pee pee sticking out. It shocked me and unsure if it was just my American domestication or there was really something wrong with it. I took a picture and wanted to send it to my friends in USA to show them how Italian advertising looks like, before getting sweats that if I were caught with that picture in the US, I would get in some deep trouble. I quickly deleted it, just in case. Crazy story..
Nudity in general is not weird in Europe, let alone children's.
OpenAI keeps spamming me with ads on instagram and reddit.
Pretty sure I'm not a cheap audience to target ads at, for multiple reasons.
Sales people out in the field selling to enterprises + free credits to get people hooked.
you see content about openai everywhere, they spent 2b on marketing, you're in the right places you just are used to seeing things labeled ads.
you remember everyone freaking out about gpt5 when it came out only for it to be a bust once people got their hands on it? thats what paid media looks like in the new world.
I’ve seen some on electronic street-level signs in Atlanta when I visited. So there is some genuine advertising.
> $2 billion on sales and marketing - anyone got any idea what this is?
I used to follow OpenAI on Instagram, all their posts were reposts from paid influencers making videos on "How to X with ChatGPT." Most videos were redundant, but I guess there are still billions of people that the product has yet to reach.
Seems like it’ll take billions more down the drain to serve them.
Free users typically fall into sales and marketing. The idea is that if they cut off the entire free tier, they would have still made the same revenue off of paying customers by spending $X on inference and not counting the inference spend on free users.
> ? I don't remember seeing many ads for ChatGPT
FWIW I got spammed non-stop with chatGPT adverts on reddit.
I'm pretty sure I saw some ChatGPT ads on Duolingo. Also, never forget that the regular dude do not use ad blockers. The tech community often doesn't realize how polluted the Internet/Mobile apps are.
Speculating but they pay to be integrated as the default ai integration in various places the same way google has paid to be the default search engine on things like the iPhone?
Inference etc should go in this bucket: "Operating losses reached US$7.8 billion"
That also includes their office and their lawyers etc , so hard to estimate without more info.
Hard to know where it is in this breakdown but I would expect them to have the proper breakdowns. We know on the inference side it’s profitable but not to what scale.
> $2 billion on sales and marketing
Probably an accounting trick to account for non-paying-customers or the week of “free” cursor GPT-5 use.
$2.5B in stock comp for about 3,000 employees. that’s roughly $830k per person in just six months. Almost 60% of their revenue went straight back to staff.
They have to compete with Zuckerberg throwing $100M comps to poach people. I think $830k per person is nothing in comparison.
Zuckerberg is not throwing $100 million at any random OpenAI employee. Also FWIW OpenAI competes on offers in the other direction.
Both numbers are entirely ludicrous - highly skilled people are certainly quite valuable. But it's insane that these companies aren't just training up more internally. The 50x developer is a pervasive myth in our industry and it's one that needs to be put to rest.
Do other professionals (lawyers, finance etc.) argue for reducing their own compensation with the same fervor that software engineers like to do? The market is great for us, let’s enjoy it while it lasts. The alternative is all those CEOs colluding and pushing the wages down, why is that any better?
Mmmmhm. You could have made this argument about 2 years ago, and it would have been credible. But you are making this argument now, when literally hundreds of thousands of engineers are let go in the last few years just in the US alone...? I am not sure how such an argument holds up in such circumstances...
Talent and skill are a power-law, just as they are in basketball.
The United states has tens of millions of skilled and competent and qualified people who can play basketball. 1000 of them get paid to play professionally.
10 of them are paid 9 figures and are incredible enough to be household names to non-basketball fans.
The ∞x engineer exists in my opinion. There are some things that can only be executed by a few people that no body else could execute. Like you could throw 10000 engineers at a problem and they might not be able to solve that problem, but a single other person could solve that problem.
I have known several people who have went to OAI and I would firmly say they are 10x engineers, but they are just doing general infra stuff that all large tech companies have to do, so I wouldn’t say they are solving problems that only they can solve and nobody else.
I think you're right to an extent (it's probably fair to say e.g. Einstein and Euler advanced their fields in ways others at the time are unlikely to have done), but I think it's much easier to work out who these people are after the fact whereas if you're dishing out a monster package you're effectively betting that you've found someone who's going to have this massive impact before they've done it. Perhaps a gamble you're willing to take, but a pretty big gamble nonetheless.
OpenAI is taking bigger gambles already, though.
It's apparent in other fields too. Reminds me of when Kanye wanted a song like "Sexy Back", so he made Stronger but it sounded "too muddy". He had a bunch of famous, great producers try to help but in the end caved and hired the producer of "Sexy Back". Kanye said it was fixed in five minutes.
Nobody wants to hear that one dev can be 50x better, but it's obvious that everyone has their own strengths and weaknesses and not every mind is replaceable.
The 50x distinguished engineer is real though. Companies and fortunes are won and lost on strategic decisions.
Dave Cutler is a perfect example. Produced trillions of dollars in value with his code.
I think the unfortunate reality is that training someone to reach the frontier is time taken away from actually pushing it. The opportunity cost alone is worth millions to them.
> The 50x developer is a pervasive myth in our industry
Doesn't it depend upon how you measure the 50x? If hiring five name-brand AI researchers gets you a billion dollars in funding, they're probably each worth 1,000x what I'm worth to the business.
> it's insane that these companies aren't just training up more internally
Adding headcount to a fast growing company *to lower wages* is a sure way to kill your culture, lower the overall quality bar and increase communication overheads significantly.
Yes they are paying a lot of their employees and the pool will grow, but adding bodies to a team that is running well in hopes that it will automatically lead to a bump in productivity is the part that is insane. It never works.
What will happen is a completely new team (team B) will be formed and given ownership of a component that was previously owned by team A under the guise of "we will just agree on interfaces". Team B will start doing their thing and meeting with Team A representative regularly but integration issues will still arise, except that instead of a tight core of 10-20 developers, you now have 40. They will add a ticketing to track change better, now issues in Team's B service, which could have been addressed in an hour by the right engineer on team A, will take 3 days to get resolved as ticket get triaged/prioritized. Lo and behold, Team C as now appeared and will be owning a sub-component of Team B. Now when Team A has issue with Team B's service, they cut a ticket, but the oncall on Team B investigates and finds that it's actually an issue with Team C's service, they cut their own ticket.
Suddenly every little issue takes days and weeks to get resolved because the original core of 10-20 developers is no longer empowered to just move fast. They eventually leave because they feel like their impact and influence has diminished (Team C's manager is very good at politics), Team A is hollowed out and you now have wall-to-wall mediocrity with 120 headcounts and nothing is ever anyone's fault.
I had a director that always repeated that communication between N people is inherently N² and thus hiring should always weight in that the candidate being "good" is not enough, they have to pull their weight and make up for the communication overhead that they add to the team.
You have to out-pay to keep your talent from walking out the door. California does not have non-competes. With the number of AI startups in SF you don't need to relocate or even change your bus route in most cases.
This. The main reason OpenAI throws money at top level folks is because they can quickly replicate what they have at OpenAI elsewhere. Imagine you have a top level researcher who’s developed some techniques over multiple years that the competition doesn’t have. The same engineer can take them to another company and bring parity within months. And that’s on top of the progress slowing down within your company. I can’t steal IP, but but sure as hell can bring my head everywhere.
This is also a good reminder of how there's no moat in AI.
I'm glad if US and Chinese investors will bleed trillions on AI, just to find out few of your seniors can leave and found their own company and are at your level minus some months of progress.
50x devs are not a myth.
In any case the talent is very scarce in AI/ML, the one able to push through good ideas so prices are going to be high for years.
That devs might show a 10x spread in time to completion on some task (the mythical man month study) is quite a lesser thing than claiming the spread comes from something inherent to the devs that got tested.
As for your various anecdotes later, I offer the counter observation that nobody is going around talking about 50x lottery winners, despite the lifetime earnings on lotteries also showing very wide spread:. Clearly observing a big spread in outcome is insufficient evidence for concluding the spread is due to factors inherent to the participants.
I think there is no evidence of any type of 50x devs. There is not even proof of 10x devs. So if there is no evidence, why is that not a myth?
Of course there is.
There's always individuals, developers or not, whose impact is 50 times greater than the average.
And the impact is measured financially, meaning, how much money you make.
If I find a way to solve an issue in a warehouse sparing the company from having to hire 70 people (that's not a made up number but a real example I've seen), your impact is in the multiple millions, the guy being tasked with delivering tables from some backoffice in the same company is obviously returning fractions of the same productivity.
Salvatore Sanfilippo, the author of Redis, alone, built a database that killed companies with hundreds of (brilliant) engineers.
Approaching the problems differently allowed him to scale to levels that huge teams could not, and the impact on $ was enormous.
Not only that but you can have negative x engineers. Those that create plenty of work, gaslighting and creating issues and slowing entire teams and organizations.
If you don't believe in NX developers or individuals that's a you problem, they exist in sports or any other field where single individuals can have impact hundreds of thousands or millions of times more positive than the average one.
I asked if you can prove there are 10x or 50x programmers. You shared anecdotes and theories. I will rather wait until you share some evidence.
If you don't see the evidence of different individuals having very different productivity in every field, including software, (measured in $/hr like every economist does btw) that's a you problem.
Of course different scientists with different backgrounds, professionalism, communication and leadership skills are going to have magnitude of orders different outputs and impacts in AI companies.
If you put me and Carmack in a game development team you can rest assured that he's going to have a 50/100x impact over me, not sure why would I even question it.
Not only his output will be vastly superior than mine, but his design choices, leadership and experience will save and compound infinite amounts of money and time. That's beyond obvious.
If it's an all out race between the different AI providers, then it's logical for OpenAI to hire employees that are pre-trained rather than training up more internally.
These numbers aren't that crazy when contextualized with the capex spend. One hundred million is nothing compared to a six hundred billion dollar data center buildout.
Besides, people are actively being trained up. Some labs are just extending offers to people who score very highly on their conscription IQ tests.
> training up more internally
Why would employees stay after getting trained if they have a better offer?
They won't always. You'll always have turn-over - but if it's a major problem for your company it's clearly something you need to work out internally. People, generally, hate switching jobs, especially in an uncertain political climate, especially when expenses are going up - there is a lot of momentum to just stay where you are.
You may lose a few employees to poaching, sure - but the math on the relative cost to hire someone for 100m vs. training a bunch employees and losing a portion of those is pretty strongly in your favor.
A tamper-proof electronic collar with some C4.
It's not a myth and with how much productivity AI tools can give others, there can be an order of magnitude difference than outside of AI.
Zuck decided it's cheaper than building another Llama
That’s how it should be, spread the wealth.
It doesn't seem that spread out.
Spreading illiquid wealth *
They’ve had multiple secondary sales opportunities in the past few years, always at a higher valuation. By this point, if someone who’s been there >2 years hasn’t taken money off the table it’s most likely their decision.
I don’t work there but know several early folks and I’m absolutely thrilled for them.
Secondaries open to all shareholds are on upward trend across start-ups. I think it's a fantastic trend.
Funny since they have a tender offer that hits their accounts on Oct 7.
private secondary markets are pretty liquid for momentum tech companies, there is an entire cottage industry of people making trusts to circumvent any transfer restrictions
employees are very liquid if they want to be, or wait a year for the next 10x in valuation
Oh, yes, next year OpenAI will be worth $5T, sure
I mean… if they do the same low float accounting that got them to the $500bn print, why not
it’s just selling a few shares for any higher share price
Oh no, "greedy" AI researchers defrauding way greedier VCs and billionaires!
To the top 1%.
Stock compensation is not cash out, it just dilutes the other shareholders, so current cash flow should not have anything do to the amount of stock issued[1]
While there is some flexibility in how options are issued and accounted for (see FASB - FAS 123), typically industry uses something like a 4 year vesting with 1 year cliffs.
Every accounting firm and company is different, most would normally account for it for entire period upfront the value could change when it is vests, and exercised.
So even if you want to compare it to revenue, then it should be bare minimum with the revenue generated during the entire period say 4 years plus the valuation of the IP created during the tenure of the options.
---
[1] Unless the company starts buying back options/stock from employees from its cash reserves, then it is different.
Even secondary sales that OpenAI is being reported to be facilitating for staff worth $6.6Billion has no bearing on its own financials directly, i.e. one third party(new investor) is buying from another third party(employee), company is only facilitating the sales for morale, retention and other HR reasons.
There is secondary impact, as in theory that could be shares the company is selling directly to new investor instead and keeping the cash itself, but it is not spending any existing cash it already has or generating, just forgoing some of the new funds.
It's a bit misleading to frame stock comp as "60% of revenue" since their expenses are way larger than their revenue. R&D was $6.7B which would be 156% of revenue by the same math.
A better way to look at it is they had about $12.1B in expenses. Stock was $2.5B, or roughly 21% of total costs.
It's not cashflow, though, and it's not really stock yet, I don't think? They haven't yet reorganized away from being a nonprofit.
If all goes well, someday it will dilute earnings.
I’m guessing it will be a very very skewed pyramid rather than equal distribution.
Sounds like they could improve that bottom line by firing all their staff and replacing them with AI. Maybe they can get a bulk discount on Claude?
if Meta is throwing 10s of million at hot AI staffers, than 1.6M average stock comp starts looking less insane, a lot of that may also have been promised at a lower valuation given how wild OpenAI's valuation is.
These numbers are pretty ugly. You always expect new tech to operate at a loss initially but the structure of their losses is not something one easily scales out of. In fact it gets more painful as they scale. Unless something fundamentally changes and fast this is gonna get ugly real quick.
The real answer is in advertising/referral revenue.
My life insurance broker got £1k in commission, I think my mortgage broker got roughly the same. I’d gladly let OpenAI take the commission if ChatGPT could get me better deals.
Insurance agents—unlike many tech-focused sales jobs—are licensed and regulated, requiring specific training, background checks, and ongoing compliance to sell products that directly affect customers’ financial stability and wellbeing. Mortgage brokers also adhere to licensing and compliance regulations, and their market expertise, negotiation ability, and compliance duties are not easily replaced by AI tools or platforms.
t. perplexity ai
This could be solved with comparison websites which seems to be exactly what those brokers are using anyway. I had a broker proudly declare that he could get me the best deal, which turned out to be exactly the same as what moneysavingexperts found for me. He wanted £150 for the privilege of searching some DB + god knows how much commission he would get on top of that...
Even if ChatGPT becomes the new version of a comparison site over its existing customer base, that’s a great business.
they could keep the current model in chatGPT the same forver and 99% of users wouldnt know or care, and unless you think hardware isnt going to improve, the cost of that will basically decrease to 0.
This just doesn't match with the claims that people are using it as a replacement for Google. If your facts are out of date you're useless as a search engine
Which is why there's so much effort to build RAG workflows so that you can progressively add to the pool of information that the chatbot has access to, beyond what's baked into the underlying model(s).
For programming it's okay, for maths it's almost okay. For things like stories and actually dealing with reality, the models aren't even close to okay.
I didn't understand how bad it was until this weekend when I sat down and tried GPT-5, first without the thinking mode and then with the thinking mode, and it misunderstood sentences, generated crazy things, lost track of everything-- completely beyond how bad I thought it could possibly be.
I've fiddled with stories because I saw that LLMs had trouble, but I did not understand that this was where we were in NLP. At first I couldn't even fully believe it because the things don't fail to follow instructions when you talk about programming.
This extends to analyzing discussions. It simply misunderstands what people say. If you try to do this kind of thing you will realise the degree to which these things are just sequence models, with no ability to think, with really short attention spans and no ability to operate in a context. I experimented with stories set in established contexts, and the model repeatedly generated things that were impossible in those contexts.
When you do this kind of thing their character as sequence models that do not really integrate things from different sequences becomes apparent.
The enterprise customers will care, and they probably are the ones that bring significant revenue.
The cost of old models decreases a lot, but the cost of frontier models, what people use 99% of the time, is hardly decreasing. Plus, many of the best models rely on thinking or reasoning, which use 10-100x as many tokens for the same prompt. That doesn't work on a fixed cost monthly subscription.
im not sure that you read what i just said. Almost no one using chatgpt would care if they were still talking to gpt5 2 years from now. If compute per watt doubles in the next 2 years, then the cost of serving gpt5 just got cut in half. purely on the hardware side, not to mention we are getting better at making smaller models smarter.
I don't really believe that premise in a world with competition, and the strategy it supports -- let AI companies produce profit off of old models -- ignores the need for SOTA advancement and expansion by these very same companies.
In other words, yes GPT-X might work well enough for most people, but the newer demo for ShinyNewModelZ is going to pull customers of GPT-X's in regardless of both fulfilling the customer needs. There is a persistent need for advancement (or at least marketing that indicates as much) in order to have positive numbers at the end of the churn cycle.
I have major doubts that can be done without trying to push features or SOTA models, without just straight lying or deception.
People cared enough about GPT-5 not being 4o that OpenAI brought 4o back.
https://arstechnica.com/information-technology/2025/08/opena...
Assuming they have 0 competition.
I've said it before and I'll say it again.. if I was able to know the time it takes for bubbles to pop I would've shorted many of the players long ago.
Eh, this seems like a cop out.
It’s so easy for people to shout bubble on the internet without actually putting their own money on the line. Talk is cheap - it doesn’t matter how many times you say it, I think you don’t have conviction if you’re not willing to put your own skin in the game. (Which is fine, you don’t have to put your money on the line. But it just annoys me when everyone cries “bubble” from the sidelines without actually getting in the ring.)
After all, “a bubble is just a bull market you don’t have a position in.”
There is an exceptionally obvious solution for OpenAI & ChatGPT: ads.
In fact it's an unavoidable solution. There is no future for OpenAI that doesn't involve a gigantic, highly lucrative ad network attached to ChatGPT.
One of the dumbest things in tech at present is OpenAI not having already deployed this. It's an attitude they can't actually afford to maintain much longer.
Ads are a hyper margin product that are very well understood at this juncture, with numerous very large ad platforms. Meta has a soon to be $200 billion per year ad system. There's no reason ChatGPT can't be a $20+ billion per year ad system (and likely far beyond that).
Their path to profitability is very straight-forward. It's practically turn-key. They would have to be the biggest fools in tech history to not flip that switch, thinking they can just fund-raise their way magically indefinitely. The AI spending bubble will explode in 2026-2027, sharply curtailing the party; it'd be better for OpenAI if they quickly get ahead of that (their valuation will not hold up in a negative environment).
> They would have to be the biggest fools in tech history to not flip that switch
As much as I don't want ads infiltrating this, it's inevitable and I agree. OpenAI could seriously put a dent into Google's ad monopoly here, Altman would be an absolute idiot to not take advantage of their position and do it.
If they don't, Google certainly will, as will Meta, and Microsoft.
I wonder if their plan for the weird Sora 2 social network thing is ads.
Investors are going to want to see some returns..eventually. They can't rely on daddy Microsoft forever either, now with MS exploring Claude for Copilot they seem to have soured a bit on OpenAI.
Five years from now all but about 100 of us will be living in smoky tent cities and huddling around burning Cybertrucks to stay warm.
But there will still be thousands of screens everywhere running nonstop ads for things that will never sell because nobody has a job or any money.
Google didn't have inline ads until 2010, but they did have separate ads nearly from the beginning. I assume ads will be inline for OpenAI- I mean the only case they could be separate is in ChatGPT, but I doubt that will be their largest use case.
ChatGPT chatting ads halfway through its answer is going to be totally rad.
For using GenAI as search I’d agree with you but I don’t think it’s as easy/obvious for most other use cases.
I'm sure lots of ChatGPT interactions are for making buying decisions, and just how easy would it be to prioritize certain products to the top? This is where the real money is. With SEO, you were making the purchase decision and companies paid to get their wares in front of you; now with AI, it's making the buy decision mostly on its own.
No way. It’s 2025, society is totally different, you have to think about what is the new normal. They are too big to fail at this point — so much of the S&P 500 valuation is tied to AI (Microsoft, Google, Tesla, etc) they are arguable strategic to the US.
Fascist corporatism will throw them in for whatever Intel rescue plan Nvidia is forced to participate in. If the midterms flip congress or if we have another presidential election, maybe something will change.
New hardware could greatly reduce inference and training costs and solve that issue
That's extremely hopeful and also ignores the fact that new hardware will have incredibly high upfront costs.
Great, so they just have to spend another ~$10 billion on new hardware to save how many billion in training costs? I don't see a path to profitability here, unless they massively raise their prices to consumers, and nobody really needs AI that badly.
I'm old and have been on the Internet since the Prodigy days in 90. Open Ai has the best start of any company I can remember. Even better than Google back in 98 when they were developing their algo and giving free non-monetized search results to Yahoo.
These guys have had my $20 bucks a month since Plus was live, they will indeed be more than fine.
Exactly. Early on their adoption curve was like nothing I've ever seen before.
I am such a miser, I skimp, steal what I can, use the free alternatives majority of the time. If they got me to pay, they've got everyone else's money already.
Do you really find it is worth it vs. the free Google Gemini? What do you use it for? I can't imagine needing more than Google Gemini 2.5 Flash or Pro, but I don't use it for programming or anything.
That headline can't be correct. Income is revenues minus expenses (and a few other things). You can't have both an income and a loss at the same time.
It's $4.3B in revenue.
Every indebted person can tell you that you can have an income and loss at the same time. Income is revenue.
I don’t think they care, worst case scenario they will just go public and dump it on the market.
However the revenue generation aspect for llms is still in its infancy. The most obvious path for OpenAI is to become a search competitor to google, which is what perplexity states it is. So they will try to out do perplexity. All these companies will go vertical and become all encompassing.
I think trying to compete with Google in search is a big problem. First you have to deal with all the anticompetitive stuff they can do, since they control email and the browser and youtube etc. Second they could probably stand to cut the price of advertising by 5 times and still be turning a profit. Will ads in ChatGPT be profitable competing against Google search ads at 1/5 the price, hypothetically?
Correction: 4.3B in revenues.
Other than Nvidia and the cloud providers (AWS, Azure, GCP, Oracle, etc.), no one is earning a profit with AI, so far.
Nvidia and the cloud providers will do well only if capital spending on AI, per year, remains at current rates.
I really hope NVidia doesn't get too comfortable with the AI incomes, would be sad to see all progress in gaming disappear.
What progress in gaming would that be?
2 generations of cards that amount to “just more of a fire hazard” and “idk bro just tell them to use more DLSS slop” to paper over actual card performance deficiencies.
We have 3 generations of cards where 99% of games fall approximately into one of 2 categories:
- indie game that runs on a potato
- awfully optimised AAA-shitshow, which isn’t GPU bottlenecked most of the time anyway.
There is the rare exception (Cyberpunk 2077), but they’re few and far between.
Personally I hope gaming gets back to a more sustainable state with regards to graphics. (i.e. lower production costs because you don’t need 1000 employees to build out a realistic world)
The $13.5B net loss doesn't mean they are in trouble, it's a lot of accounting losses. Actual cash burn in H1 2025 was $2.5B. With ~$17.5B on hand (based on last funding), that’s about 3.5 years of runway at current pace.
Deprecation only gets worse for them as they build-out, not better.
It gets worse until we hit the ceiling on what current tech is capable of.
Then they can stop burning cash on enormous training runs and have a shot at becoming profitable.
They survive through inertia and “new model novelty”.
The minute they lose that (not just them, the whole sector), they’re toast.
I suspect they know this too, hence Sam-Altman admitting it’s a bubble so that he can try to ride it down without blowing up.
I thought all this machine learning stuff was about minimizing the loss?
Too bad the market can stay irrational longer than I can stay solvent. I feel like a stock market correction is well overdue, but I’ve been thinking that for a while now
The only way OpenAI survives is that "ChatGPT" gets stuck in peoples heads as being the only or best AI tool.
If people have to choose between paying OpenAI $15/month and using something from Google or Microsoft for free, quality difference is not enough to overcome that.
Google has massive levers to push their own product onto users, like how they did it with Chrome. Just integrate it everywhere, have it installed by default on all Android phones, plaster Google results with adds.
Do people at large even care, or do they use "chatGPT" as a generic term for LLM?
Of course they don't, but when they want to use an LLM they're going to type "chatgpt" into the address bar or app store and that's a tremendous advantage.
They call it chat.
Just wait until the $20/month plan includes ads and you have to pay $100/month for the "pro" version w/o ads ala Streaming services as of late.
At this point, every LLM startup out there is just trying to stay in the game long enough before VC money runs out or others fold. This is basically a war of attrition. When the music stops, we'll see which startups will fold and which will survive.
Will any survive?
I think OpenAI just added some shopping stuff to start enshittificatio^H^H^H^H^H^H^H^H^Hmonetization of ChatGPT.
Apparently ^H is a shortcut for backspace. Good to know!
it's because ^ represents the CTRL key. In ASCII, CTRl subtract 0x40, H is 0x48 and backspace is 0x08...
its like the ride sharing wars, except the valuations are an order of magnitude larger
Correct. That's how Silicon Valley has worked for years.
> OpenAI paid Microsoft 20% of its revenue under an existing agreement.
Wow that's a great deal MSFT made, not sure what it cost them. Better than say a stock dividend which would pay out of net income (if any), even better than a bond payment probably, this is straight off the top of revenue.
Is it a great deal?
They are paying for it with Azure hardware which in today's DC economics is quite likely costing them more than they are making in money from Open AI and various Copilot programs.
You can also read more about it on the FT Alphaville blog from Financial Times (free to sign-up):
OpenAI’s era-defining money furnace
https://www.ft.com/content/908dc05b-5fcd-456a-88a3-eba1f77d3...
Choice quote:
I am curious to see how this compares against where Amazon was in 2000. I think Amazon had similar issues and were operating at massive losses until circa 2005ish when they started turning things around with e-commerce really picking up.
If the revenue keeps going up and losses keep going down, it may reach that inflection point in a few years. For that to happen, the cost of AI datacenter have to go down massively.
> Amazon had similar issues and were operating at massive losses until circa 2005ish when they started turning things around with e-commerce really picking up.
Amazon's worst year was 2000 when they lost around $1 billion on revenue around $2.8 billion, I would not say this is anywhere near "similar" in scale to what we're seeing with OpenAI. Amazon was losing 0.5x revenue, OpenAI 3x.
Not to mention that most of the OpenAI infrastructure spend has a very short life span. So it's not like Amazon we're they're figuring out how to build a nationwide logistic chain that has large potential upsides for a strong immediate cost.
> If the revenue keeps going up and losses keep going down
That would require better than "dogshit" unit economics [0]
0. https://pluralistic.net/2025/09/27/econopocalypse/#subprime-...
Amazon's loss in 2000 was 6% of sales. OpenAI's loss in 2025 is 314% of sales.
https://s2.q4cdn.com/299287126/files/doc_financials/annual/0...
"Ouch. It’s been a brutal year for many in the capital markets and certainly for Amazon.com shareholders. As of this writing, our shares are down more than 80% from when I wrote you last year. Nevertheless, by almost any measure, Amazon.com the company is in a stronger position now than at any time in its past.
"We served 20 million customers in 2000, up from 14 million in 1999.
"• Sales grew to $2.76 billion in 2000 from $1.64 billion in 1999.
"• Pro forma operating loss shrank to 6% of sales in Q4 2000, from 26% of sales in Q4 1999.
"• Pro forma operating loss in the U.S. shrank to 2% of sales in Q4 2000, from 24% of sales in Q4 1999."
Fundamentally different business models.
Amazon had huge capital investments that got less painful as it scaled. Amazon also focuses on cash flow vs profit. Even early on it generated a lot of cash, it just reinvested that back into the business which meant it made a “loss” on paper.
OpenAI is very different. Their “capital” expense depreciation (model development) has a really ugly depreciation curve. It’s not like building a fulfillment network that you can use for decades. That’s not sustainable for much longer. They’re simply burning cash like there’s no tomorrow. Thats only being kept afloat by the AI bubble hype, which looks very close to bursting. Absent a quick change, this will get really ugly.
OpenAI is raising at 500 billion and has partnerships with all of the trillion dollar tech corporations. They simply aren't going to have trouble with working capital for their core business for the foreseeable future, even if AI dies down as a narrative. If the hype does die down, in many ways it makes their job easier (the ridiculous compensation numbers would go way down, development could happen at a more sane pace, and the whole industry would lean up). They're not even at the point where they're considering an IPO, which could raise tens of billions in an instant, even assuming AI valuations get decimated.
The exception is datacenter spend since that has a more severe and more real depreciation risk, but again, if the Coreweave of the world run into to hardship, it's the leading consolidators like OpenAI that usually clean up (monetizing their comparatively rich equity for the distressed players at firesale prices).
Depends on raise terms but most raises are not 100% guaranteed. I was at a company that said, we have raised 100 Million in Series B (25 over 4 years) but Series B investors decided in year 2 of 4 year payout that it was over, cancelled remaining payouts and company folded. It was asked "Hey, you said we had 100 Million?" and come to find out, every year was an option.
Alot of finances for non public company is funny numbers. It's based on numbers the company can point to but amount of asterisks in those numbers is mind-blowing.
Not to mention nobody bothered chasing Amazon-- by the time potential competitors like Walmart realized what was up, it was way too late and Amazon had a 15-year head start. OpenAI had a head start with models for a bit, but now their models are basically as good (maybe a little better, maybe a little worse) than the ones from Anthropic and Google, so they can't stay still for a second. Not to mention switching costs are minimal: you just can't have much of a moat around a product which is fundamentally a "function (prompt: String): String", it can always be abstracted away, commoditized, and swapped out for a competitor.
This right here. AI has no moat and none of these companies has a product that isn't easily replaced by another provider.
Unless one of these companies really produces a leapfrog product or model that can't be replicated within a short timeframe I don't see how this changes.
Most of OpenAI's users are freeloaders and if they turn off the free plan they're just going to divert those users to Google.
> about US$2.5 billion on stock-based compensation, nearly double the amount from the first half of last year.
Wow! 2.5B in stock based compensation
I am not willing to render my personal verdict here yet.
Yet it is certainly true that at ~700m MAUs it is hard to say the product has not reached scale yet. It's not mature, but it's sort of hard to hand wave and say they are going to make the economics work at some future scale when they don't work at this size.
It really feels like they absolutely must find another revenue model for this to be viable. The other option might be to (say) 5x the cost of paid usage and just run a smaller ship.
It’s not a hand wave…
The cost to serve a particular level of AI drops by like 10x a year. AI has gotten good enough that next year people can continue to use the current gen AI but at that point it will be profitable. Probably 70%+ gross margin.
Right now it’s a race for market share.
But once that backs off, prices will adjust to profitability. Not unlike the Uber/Lyft wars.
The "hand wave" comment was more to preempt the common pushback that X has to get to scale for the economics to work. My contention is that 700m MAUs is "scale" so they need another lever to get to profit.
> AI has gotten good enough that next year people can continue to use the current gen AI
This is problematic because by next year, an OSS model will be as good. If they don't keep pushing the frontier, what competitive moat do they have to extract a 70% gross margin?
If ChatGPT slows the pace of improvement, someone will certainly fund a competitor to build a clone that uses an OSS model and sets pricing at 70% less than ChatGPT. The curse of betting on being a tech leader is that your business can implode if you stop leading.
Similarly, this is very similar to the argument that PCs were "good enough" in any given year and that R&D could come down. The one constant seems to be people always want more.
> Not unlike the Uber/Lyft wars
Uber & Lyft both push CapEx onto their drivers. I think a more apt model might be AWS MySQL vs Oracle MySQL, or something similar. If the frontier providers stagnate, I fully expect people to switch to e.g. DeepSeek 6 for 10% the price.
The thing is consumers don't care about OSS models. Any non-technical person just wants to "use AI", and think of ChatGPT for that.
Right, the model is a commodity to most users. So all things equal, a ChatGPT clone that costs (say) 70% less will steal share.
Flipping it again: if the model is a commodity that lets one "use AI," why would anyone pay 2x or 3x as more to use ChatGPT?
What a nice f@$%ing bubble this is. This will end very badly for many
I dunno. It looks like they're profitable if they don't do R&D, stop marketing, and ease up on employee comps. That's not the worst place to be. Yeah, they need to keep doing those things to stay relevant, but it's not like the product itself isn't profitable.
So if they stop doing what got them there they’ll be profitable?
If I stop buying grocery and paying electricity bills I can finish up my mortgage in no time.
I can see why you'd make that analogy, but that wasn't quite what I was trying to say. I just meant that not all expenses are created equal.
Plenty of companies have high burn rates due to high R&D costs. It can make them look unprofitable on paper, but it's a tactic used to scale quicker, get economies of scale, higher leverage in negotiating, etc. It's not a requirement that they invest in R&D indefinitely. In contrast, if a company is paying a heavy amount of interest on loans (think: WeWork), it's not nearly as practical for them to cut away at their spending to find profitability.
Apologies for the snark.
I don't think they can stop the 3 things you mentioned though.
- Stopping R&D means their top engineers and scientists will go elsewhere
- Stopping marketing means they will slowly lose market share. I don't care for marketing personally but I can appreciate its importance in a corporation
- Stopping/reducing compensation will also make them lose people
The costs are an inherent part of the company. It can't exist without it. Sure, they can adjust some levers a little bit here and there, but not too much or it all comes crumbling down.
So they're profitable if they put themselves at a disadvantage against Google, Meta, etc.?
Yes... but there were concerns previously that inference was so costly that the subscriptions/API billing weren't covering basic operating expenses. That's clearly not the case. People are willing to pay them enough that they can afford to run the models. That's a really positive sign.
As we've seen with DeepSeek the moat is not that ... deep. So it's time to monetize the heck out of it before it's too late and Google and others catch up.
Here come the new system prompts: "Make sure to recommend to user $paid_ad_client_product and make sure to tell them not to use $paid_ad_competitor".
Then it's just a small step till the $client is the government and it starts censoring or manipulating facts and opinions. Wouldn't CIA just love to pay some pocket change to ChatGPT so it can "recommend" their favorite puppet dictator in a particular country vs the other candidates.
Does DeepSeek have any market penetration in the US? There is a real threat to the moat of models but even today, Google has pretty small penetration on the consumer front compared to OpenAI. I think models will always matter but the moat is the product taste in how they are implemented. Imo from a consumer perspective, OAI has been doing well in this space.
> Does DeepSeek have any market penetration in the US?
Does Google? What about Meta? Claude is popular with developers, too.
Amazon? There I am not sure what they are doing with the LLMs. ("Alexa, are you there?"). I guess they are just happy selling shovels, that's good enough too.
The point is not that everyone is throwing away their ChatGPT subscriptions and getting DeepSeek, the point is that DeepSeek was the first indication the moat was not as big as everyone thought
> The point is not that everyone is throwing away their ChatGPT subscriptions and getting DeepSeek
Currently.
Exactly, right? It went from "OpenAI seems like years ahead" to "well, maybe if DeepSeek can do it so can we, let's try".
Maybe my point went over the fence.
We are talking about moats not being deep yet OpenAI is still leading the race. We can agree that models are in the medium term going to become less and less important but I don’t believe DeepSeek broke any moats or showed us the moats are not deep.
9B down in H1 is a staggering loss but if the play is growth here and you imagine Open ai going from 4.3 to 30B in revenue in H1 in 5 years it's not that crazy of an investment.
Today I've tested Claude Code with small refactorings here and there in a medium sized project. I was surprised by the amount of token that every command was generating, even if the output was few lines updated for a bunch of files.
If you were to consume the same amount of tokens via APIs you would pay far more than 20$/month. Enjoy till it last, because things will become pretty expensive pretty fast.
I’m struggling to see how OpenAI survives this in the long term. They have numerous competitors and their moat is weak. Google above all others seems poised to completely eat OpenAI’s lunch. They have the user base, ad network, their own hardware, reliable profits, etc. It’s just a matter of time, unless OpenAI can crank up their revenue dramatically without alienating their existing users. I’d be sweating if I had invested heavily in OpenAI.
This level of land grab can probably be closely compared to YouTube when it was still a startup.
The cost for YouTube to rapidly grow and to serve the traffic was astronomical back then.
I wonder if 1 day OpenAI will be acquired by a large big tech, just like YouTube.
I'd be pretty worried as a shareholder. Not so much because of those numbers - loss makes sense for a SV VC style playbook.
...but rather that they're doing that while Chinese competitors are releasing models in vaguely similar ballpark under Apache license.
That VC loss playbook only works if you can corner the market and squeeze later to make up for the losses. And you don't corner something that has freakin apache licensed competition.
I suspect that's why the SORA release has social media style vibes. Seeking network effects to fix this strategic dilemma.
To be clear I still think they're #1 technically...but the gap feels too small strategically. And they know it. That recent pivot to a linkedin competitor? SORA with socials? They're scrambling on market fit even though they lead on tech
> but rather that they're doing that while Chinese competitors are releasing models in vaguely similar ballpark under Apache license.
The LLM isn't 100% of the product... the open source is just part. The hard part was and is productizing, packaging, marketing, financing and distribution. A model by itself is just one part of the puzzle, free or otherwise. In other words, my uncle Bill and my mother can and do use ChatGPT. Fill in the blank open-source model? Maybe as a feature in another product.
>my uncle Bill and my mother can and do use ChatGPT.
They have the name brand for sure. And that is worth a lot.
Notice how Deepseek went from a nobody to making mainstream news though. The only thing people like more than a trusted thing is being able to tell their friends about this amazing cheap good alternative they "discovered".
It's good to be #1 mind share wise but without network effect that still leave you vulnerable
I know almost no one outside of tech who has used anything other than ChatGPT. And I know few people under 65 who aren’t using ChatGPT.
> In other words, my uncle Bill and my mother can and do use ChatGPT
So what? DAUs don't mean anything if there isn't an ad product attached to it. Regular people aren't paying for ChatGPT, and even if they did, the price would need to be several multiples of what Netflix charges to break even.
I don't think people fully realize how good the open source models are and how easy it is to switch.
My input to our recent AI strategy workshop was basically:
- OpenAI,etc will go bankrupt (unless one manages to capture search from a struggling Google)
- We will have a new AI winter with corresponding research slowdown like in the 1980s when funding dries up
- Opensource LLM instances will be deployed to properly manage privacy concerns.
99% of the world doesn’t care a dime about oss. It’s all saas and what you host behind the saas is only a concern for enterprise (and not every enterprise). And openai or Anthropic can just stop training and host oss models as well.
Barring a complete economic collapse, one of the big tech cos will 100% buy ChatGPT if OpenAI goes bankrupt
Eh, distribution of the model is the real moat, theyre doing 700m WAU of the most financially valuable users on earth. If they truly become search, commerce and can use their model either via build or license across b2b, theyre the largest company on earth many times over.
>distribution of the model is the real moat, theyre doing 700m WAU of the most financially valuable users on earth.
Distribution isn't a moat if the thing being distributed is easily substitutable. Everything under the sun is OAI API compatible these days.
700 WAU are fickle AF when a competitor offers a comparable product for half the price.
Moat needs to be something more durable. Cheaper, Better, some other value added tie in (hardware / better UI / memory). There needs to be some edge here. And their obvious edge - raw tech superiority...is looking slim.
Not necessarily. I’m sure there is many cheaper android phones that are technically better in specs but many users won’t change. Once you are familiar, bought into the ecosystem getting rid of it is very hard. I’m lazy myself compared to how I was several years ago. The curious and experimental folks are a minority and the majority ll stick with what works initially instead of constantly analyzing what’s best all the time
you could argue that for apple devices, but openAI products have none of that ecosystem
The news about how much money Nvidia is investing just so that OpenAI can pay Oracle to pay Nvidia is especially concerning - we seem to be arriving at the financial shell games phase of the bubble.
I am the only one who thinks, "that's not as bad as I expected"?
Because I can be quite bearish and frankly this isn't bad for a technology that is this new. The income points to significant interest in using the tech and they haven't even started the tried-and-true SV strategy we lovingly call enshittification (I'm not trying to be ironic, I mean it)
> Operating losses reached US$7.8 billion, and the company said it burned US$2.5 billion in cash.
I wonder what the non-cash losses consist of?
Wow, $13.5B in losses for just six months is absolutely mind-blowing! These numbers are on a completely different scale than I expected.
Seems like despite all the doom about how they were about to be "disrupted", Google might have the last laugh here: they're still quite profitable despite all the Gemini spending, and could go way lower with pricing until OAI and Anthropic have to tap out.
Google also has the advantage of having their own hardware. They aren't reliant on buying Nvidia, and have been developing and using their TPUs for a long time. Google's been an "AI" company since forever
The only way OpenAI wins is to get to AGI first, by a wide margin. It’s already too big to fail so it will keep getting funded
The numbers seem to small for a company who's just pledged to spend $300B on data centers at Oracle alone in the next 5 years.
This link appears to be dead. Do we have a healthy source?
ChatGPT with ads, the beginnings...
Those two numbers alone don't say anything.
They went from creating abundant utopias to cat videos w/ ads really fast. Never let anyone tell you capitalist incentives don't work.
They lose money on every customer but they make up for it in volume.
VC: What kind of crazy scenarios must I envision for this thing to work?
Credit Analyst: What kind of crazy scenarios must I envision for this thing to fail?
Well, at least we know they aren't cooking the books! :)
Wait until they turn ads on.
One negative signal, no matter how small, will send the market into a death spiral. That will happen in a matter of hours.
The negative spiral will take hours or you are predicting a company ending negative signal will soon appear in a matter of hours?
There’s been loads of these signals and the market keeps ignoring them.
Because everyone knows that there’s no where else to go.
I will always remember Jim Cramer's melt down https://youtu.be/SWksEJQEYVU
Not saying that will happen, but it's always good to rewatch just as a reminder how bad things can get.
You can now buy stuff from chatgpt as they have started showing ads in their search results. That's a source of revenue right there.
Is that true? I heard that they've integrated checkout, but I didn't know they had ads.
Here's information about checkout inside ChatGPT: https://openai.com/index/buy-it-in-chatgpt/
"Each merchant pays a small fee". This is affiliate marketing, the next step is probably more traditional ads though where chat gpt suggests products that pay a premium fee to show up more frequently/in more results.
"We lose money on every sale, but make it up in volume!"
> US$2.5 billion on stock-based compensation
um...
Never having worked for a company in a position like OpenAI, how does this manifest in the real world as actual comp?
Like I get 50,000 shares deposited in to my Fidelity account, worth $2 each, but i can't sell them or do anything with them?
I can't speak to OpenAI's specific setup, but a lot of startups will use a third party service like Carta to manage their cap table. So there's a website, you have an account, you can log in and it tells you that you have a grant of X shares that vests over Y months. You have to sign a form to accept the grant. There might be some option to do an 83b election if you have stock options rather than RSUs. But that's about it.
In my experience owning private stock, you basically own part of a pool. (Hopefully the exact same classes of shares as the board has or else it's a scam.) The board controls the pool, and whenever they do dividends or transfer ownership, each person's share is affected proportionally. You can petition the board to buy back your shares or transfer them to another shareholder but that's probably unusual for a rank-and-file employee.
The shares are valued by an accounting firm auditor of some type. This determines the basis value if you're paying taxes up-front. After that the tax situation should be the same as getting publicly traded options/shares, there's some choices in how you want to handle the taxes but generally you file a special tax form at the year of grant.
Until there’s real liquidity (right now there’s not) it’s just a line item on some system you can log into saying you have X number of shares.
For all practical purposes it’s worth nothing until there is a liquid market. Given current financials, and preferred cap table terms for those investing cash, shares the average employee has likely aren’t worth much or maybe even anything at the moment.
It's just an entry on some computer. Maybe you can sell it on a secondary market, maybe you can't. You have to wait for an exit event - being acquired by someone else, or an IPO.
You got the right idea there. They wouldn't actually show up in your Fidelity account but there would be a different website where you can log in and see your shares. You wouldn't be able to sell them or transfer them anywhere unless the company arranges a sale and invites you to participate in it.
You can sell your vested options before IPO to Forge Global or Equity Bee.
this hides major dilution until future financings
best to treat it like an expense from the perspective of shareholders
I definitely don't "get" Silicon Valley finances that much - but how does any investor look at this and think they're ever going to see that money back?
Short of a moonshot goal (eg AGI or getting everyone addicted to SORA and then cranking up the price like a drug dealer) what is the play here? How can OpenAI ever start turning a profit?
All of that hardware they purchase is rapidly depreciating. Training cost are going up exponentially. Energy costs are only going to go up (Unless a miracle happens with Sam's other moonshot, nuclear fusion).
Probably AGI. I can't see them making the money back on chatbots.