Episode Transcript
[00:00:00] The reason I'm writing this today is that it feels like the tides are rapidly turning and multiple pale horses of the AI apocalypse have emerged. A big stupid magic trick in the form of OpenAI's rushed launch of its O1, codename Strawberry model. Rumored price increases for future OpenAI models and elsewhere, layoffs at scale AI and leaders fleeing OpenAI.
[00:00:29] These are all signs that things are beginning to collapse.
[00:00:35] Navigating the tools and the future of open source AI.
[00:00:39] This is AI Unchained.
[00:00:49] What is up, guys? Welcome back to AI Unchained. I am Guy Swan and this is where you will learn about and understand how to use the best in the open source AI world. And also where we will be making a little bit of sense out of what is happening in the industry and what is what the AI, the quote unquote technology of AI is in the first place. I want to reiterate to anybody who is new to the show, I suggest you go back and listen to the gentle introduction. I think this was the first read, actually we did the gentle introduction to LLMs and or large language models, how they work, because I think it's so important. I know it's a little bit technical, but I think it worked really well in audio actually. And even if you only pick up like 75% of it, I think it's so important to understand what these language models are so that you can understand. Because as I say towards the end of this piece, and I think this piece does a really good job of reiterating a bunch and is important context to understand going into this, is that what we think of as artificial intelligence is not artificial intelligence. And as I've said, AI is simply what it became called and. But I, but I don't think it was ever an accurate term. I never liked the name AI and I think Mark Maria actually put me on this article and it was a fantastic article. And I kept fighting between whether or not I should read the whole thing on the show because there were some.
[00:02:19] There were so many different pieces that I wanted to pull out and talk about, but it was so hard to just kind of like sum it up. I should have used an AI Boom.
[00:02:27] But.
[00:02:28] But in the end it was just so good. And I love a passionate rant and there's so much just like passion in this article and I don't agree with everything that he says. I don't agree with the entirety of what it is, but it's so much, it's. It's something that we've talked about with OpenAI and I've been kind of seeing the little breadcrumbs for the last year or two. That seems to be reiterating one of the opening themes of this show that we talked about and how OpenAI was likely not sustainable.
[00:03:05] But it's also been almost two years now and I admit that I was kind of having this growing like are they actually going to make this? And I hadn't looked at the hard numbers again and Ed Rezitrin does a fantastic job of pulling in a lot of the numbers and he has so many links to dig further into stuff. So I'm going to be verifying a lot of this stuff before we get into the next episode, which will mostly be a follow up to this, as well as other things that I've been covering on this. But I highly recommend you check out this piece go to a bunch of the links that he has. I obviously have him linked in the show notes and he has numerous other pieces that are really good with this. I just felt this was one of the best ones to talk about kind of the big picture and a lot of what's going on and how much could unravel in this industry, how much has actually been built on hype. And I have totally signed up for his newsletter and recommend you as well if you are interested in a really, really interesting take on a lot of the AI stuff and big tech in general.
[00:04:08] But with that I just want to go ahead and get into this piece. It's quite a long one as well, but trust me, stay to the end. This is a good one and I really think you're going to like it. So with that, let's get into today's article and it's titled the Subprime AI Crisis by Edward Zittran.
[00:04:30] None of what I write in this newsletter is about sowing doubt or hating, but a sober evaluation of where we are today and where we may end up on the current path. I believe that the artificial intelligence boom, which would be better described as a generative AI boom, is, as I've said before, unsustainable and will ultimately collapse.
[00:04:54] I also fear that said collapse could be ruinous to big tech, deeply damaging to the startup ecosystem, and will further sour public support for the tech industry.
[00:05:06] The reason I'm writing this today is that it feels like the tides are rapidly turning and multiple pale horses of the AI apocalypse have emerged. A big stupid magic trick in the form of OpenAI's rushed launch of its O1, codenamed Strawberry model. Rumored price increases for future OpenAI models and elsewhere, layoffs at scale AI and leaders fleeing OpenAI.
[00:05:32] These are all signs that things are beginning to collapse As a result. I think it's important to explain how precarious things are and why we are in a trough of magical thinking. I want to express my concerns about the fragility of this movement and the obsessiveness and directionlessness that brought us here. And I want some of us to do better.
[00:05:56] I also, and this perhaps is something I haven't focused on as much as I should have, want to highlight the potential human cost of an AI bubble implosion. Whether Microsoft and Google and the other big generative AI backers slowly wind down their positions or cannibalize their companies to keep OpenAI and Anthropic as well as their own generative AI efforts alive, I'm convinced that the end result will be the same. I fear tens of thousands of people will lose their jobs and and much of the tech industry will suffer as they realize that the only thing that can grow forever is cancer.
[00:06:38] There won't be much levity in this piece, I'm going to paint you a bleak picture, not just for the big AI players, but for tech more widely and for the people who work at tech companies, and tell you why I think the conclusion to this sordid saga, as brutal and as damaging as it will be, is coming sooner than you think.
[00:07:01] Let's begin.
[00:07:04] How does generative AI survive as we speak? OpenAI, the ostensible nonprofit that may soon become a for profit, is currently raising a funding round at a valuation of at least $150 billion and is expected to raise at least 6.5 billion, but potentially as much as 7 billion, led by Josh Kushner's Thrive Capital, with rumored participation from both Nvidia and Apple. As I've explained in agonizing detail in the past, OpenAI will have to continue to raise more money than any startup has ever raised in history in perpetuity to survive.
[00:07:49] Worse still, OpenAI also is trying to raise 5 billion in debt from banks, quote in the form of a revolving credit facility, according to Bloomberg. And the terms on revolving credit facilities tend to have higher rates of interest.
[00:08:06] The information also reports that OpenAI is in talks with MGX, a $100 billion investment fund backed by the United Arab Emirates, to invest primarily in AI and semiconductor companies, as well as potentially raising from the Abu Dhabi Investment Authority.
[00:08:24] This should be the biggest warning sign that things are going poorly, because absolutely nobody raises from the UAE or Saudis because they want to.
[00:08:35] They're the place you go if you need a lot of money and you're not confident anybody else will give it to you.
[00:08:43] Side note, as CNBC points out, one of the foundational partners of MGX, Mubadala, owns around $500 million in anthropic equity, which it acquired from the bankruptcy of FTX's assets. I'm sure Amazon and Google are thrilled about that conflict of interest As I discussed at the end of July, OpenAI needs to raise at least $3 billion but more like $10 billion to survive, as it is on course to lose $5 billion in 2024, a number that's likely to increase as more complex models demand more compute and more training data, with Anthropic CEO Dario Amadei predicting that future models may cost as much as $100 billion to train.
[00:09:34] As an aside, a $150 billion valuation in this case refers to how OpenAI is pricing shares in the company for investors, though shares is even a funny term in this case. For example, an investment of 1.5 billion at a $150 billion valuation in a normal company would get you 1% of the company, though as I'll get to in OpenAI's case, things are a lot weirder.
[00:10:08] OpenAI has already tried to raise at a $100 billion valuation earlier in the year, with some investors balking at the price tag in part because, and I paraphrase the Information's Kate Clark and Natasha Mascaneris of a growing concern over the overvaluation of generative AI companies.
[00:10:29] To get the round done, OpenAI may also convert from a nonprofit to a for profit entity. But the most confusing part of this story appears to be what investors are actually getting.
[00:10:43] Kate Clark of the Information reports that investors in the round are being told, and I quote, that they don't get traditional equity for their cash. Instead, they receive units that promise a slice of the company's profits once it starts to generate them. End quote.
[00:11:03] It's unclear whether a conversion to a for profit entity would fix this problem. And as OpenAI's bizarre nonprofit with a for profit arm corporate structure means that Microsoft has the rights to 75% of OpenAI's profits as part of its 2023 investment, though a shift to a for profit structure could theoretically include equity.
[00:11:27] Nevertheless, investing in OpenAI gives you profit participation units, or PPUs in in lieu of equity. And as Jack Raines wrote at Sherwood, if you own OpenAI, PPUs and the company never turns a profit and you don't sell them to someone who thinks OpenAI eventually will turn a profit. Your PPUs are worthless.
[00:11:56] Over the weekend, Reuters published a report that said any $150 billion valuation would be contingent on whether it's able to rework its entire corporate structure and in the process remove the cap on profits for investors, which are limited to 100x the original stake.
[00:12:15] This capped profit structure was introduced in 2019, with OpenAI saying any profits that exceed that ceiling would be returned to the nonprofit for the benefit of humanity.
[00:12:27] The company has in recent years altered that rule to allow for a 20% yearly increase in the cap starting from 20.
[00:12:36] Given OpenAI's existing profit sharing deals with Microsoft, not to mention its deep, deep unprofitability, any such returns are at the best theoretical. At the risk of sounding glib, 500% of nothing is still nothing.
[00:12:53] Reuters also added that any transition to a for profit structure, and thus a higher valuation than its most recent $80 billion, would also force OpenAI to renegotiate with existing investors who would see their stakes diluted.
[00:13:09] Separately, the Financial Times reports that investors have to sign up to an operating agreement that states it would be wise to view any investment in OpenAI's for profit subsidiary in the spirit of a donation and that OpenAI may never make a profit.
[00:13:29] A truly insane thing to sign that makes any investor in OpenAI fully deserving of any horrible fate that follows such a ridiculous investment.
[00:13:41] In effect, investors aren't getting a piece of OpenAI or any kind of control over OpenAI, but rather shares in the future profits of a company that loses over $5 billion a year and will likely lose more in 2025 if it makes it that far.
[00:14:04] OpenAI's models and products and we'll get into their utility in a bit are deeply unprofitable to operate with the information reporting that OpenAI is paying Microsoft an estimated $4 billion in 2024 to power ChatGPT and its underlying models. And that's with Microsoft giving it a discounted $1.30 per g, as opposed to the regular $3.40 to $4 that other customers pay.
[00:14:36] This means that OpenAI would likely be burning more like $6 billion a year on server costs if it wasn't so deeply wedded to Microsoft.
[00:14:49] And that's before you get into costs like staffing 1.5 billion a year and as I've discussed, training costs that are currently $3 billion for the year and will almost certainly increase.
[00:15:03] While the Information reported that OpenAI's revenue is 3.5 to 4.5 billion a year in July.
[00:15:11] The New York Times reported last week that OpenAI's annual revenues have now topped $2 billion, which would mean that the end of year numbers will likely trend toward the lower end of the estimate.
[00:15:25] In short, OpenAI is burning money, will only burn more money, and to continue burning more money, it will have to raise money from investors that are signing a document that says we may never make a profit.
[00:15:43] As I've written about previously, OpenAI's other problem is that generative AI, and by extension the model GPT and the product ChatGPT, doesn't solve the complex problems that would justify its massive costs.
[00:16:00] It has massive intractable problems as a result of these models being probabilistic, meaning that they don't know anything. They're just generating an answer or an image, or a translation or a summary based on training data, something that model developers are running out of at an incredible pace.
[00:16:23] Hallucinations, which occur when models authoritatively state something that isn't true, or in the case of an image or a video, makes something that looks wrong, are impossible to resolve without new branches of mathematics. And while you might be able to reduce or mitigate them, their existence makes it hard for business critical applications to truly rely on generative AI.
[00:16:50] And even if they did, it isn't clear whether generative AI actually provides much business value at all. The Information reported last week that customers of Microsoft's 365 suite, which includes things like Word, Excel, PowerPoint and Outlook, and more importantly a number of business focused software packages which in turn feed into consultancy services from Microsoft, are barely adopting its AI powered copilot products.
[00:17:18] With somewhere between 0.1% and 1% of its 440 million seats. That's 30 to $50 per person paying for the features.
[00:17:28] One firm testing the AI features is quoted as saying that most people don't find it that valuable right now, and others saying that many businesses haven't seen breakthroughs in productivity and other benefits and they're not sure when they will.
[00:17:45] And how much is Microsoft charging for these inessential features?
[00:17:50] A whopping $30 a seat on top of what you're already paying, or as much as $50 a month extra for copilots for sales.
[00:18:02] They're effectively asking customers to double their spend. And by the way, that's an annual commitment for products that don't seem to be helpful.
[00:18:12] As a note, Microsoft is such a clusterf that it will likely require its own newsletter in the future.
[00:18:20] This is the current state of generative AI, the literal leader in productivity and business software, cannot seem to find a product that people will pay for, in part because the results are so mediocre, and in part because the costs are so burdensome that it's hard to justify them.
[00:18:40] If Microsoft needs to charge this much, it's either because Satya Nadella wants to hit $500 billion in revenue by 2030. As revealed as part of a memo included in the public hearing over Microsoft's acquisition of Activision Blizzard, the costs are too high to charge much less, or some combination of the two.
[00:19:01] Yet the argument is almost always that the future of AI will blow us away, that the next generation of large language models are just around the corner, and they're going to be incredible.
[00:19:15] Last week we got our first real, definitive glimpse of what's around that corner, that future. And boy, was it underwhelming.
[00:19:26] A big, stupid magic trick OpenAI launched O1, codenamed Strawberry, on Thursday night with all the excitement of a dentist's appointment. Across a series of tweets, Sam Altman described O1 as OpenAI's most capable and aligned models yet, though he conceded that O1 was still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it.
[00:19:57] He promised it would deliver more accurate results when performing the kinds of activities which where there is a definitive right answer, like coding math problems or answering science questions.
[00:20:08] That by itself is incredibly revealing, but we'll expand on that later.
[00:20:14] First, let's talk about how it actually works. I'm going to introduce a bunch of new concepts here, but I promise I won't delve too deeply into the weeds. And if you actually want to read OpenAI's explanation, you can find it here.
[00:20:27] The link will be in the description of this podcast.
[00:20:30] When presented with a problem, O breaks it down into individual steps that hopefully would lead to a correct answer in a process called chain of thought. It's also a little easier if you consider O as two parts of the same model.
[00:20:46] On each step, one part of the model applies reinforcement learning, with the other one the model outputting stuff rewarded or punished based on the perceived correctness of their progress, the steps in its reasoning, and altering its strategies when punished. This is different to how other large language models work, in the sense that the model is generating outputs, then looking back at them, then ignoring or approving good steps to get to an answer, rather than just generating one and saying here you go.
[00:21:20] While this sounds like a seismic breakthrough or even another step towards the highly fetid artificial general intelligence or AGI. It isn't, and you can tell by the fact that OpenAI opted to release O1 as its own standalone product rather than a Number update to GPT.
[00:21:40] It's telling that the examples demonstrated by OpenAI, like math and science problems are, are ones where the answer can be known ahead of time and a solution is either correct or false, thus allowing the model to guide the chain of thought through each step.
[00:21:56] You'll notice OpenAI didn't show the O model trying to tackle complex problems, whether mathematical or otherwise, where the solution isn't known in advance. By its own admission, OpenAI has heard reports that O1 is more prone to hallucinations than GPT4 oh, and is less inclined to admit when it doesn't have the answer to a question when compared to other previous models. This is because despite there being a model that checks its work, the work checking part of the model is still capable of hallucinations, according to OpenAI. It's also, thanks to its chain of thought, more convincing to human users because O provides more detailed answers, people are more inclined to trust the outputs and even when they're completely wrong if you think that I'm being overly hard on OpenAI, consider the ways in which the company has marketed O1. It has described the reinforcement training process as thinking and reasoning, when in fact it's making guesses and then guessing on the correctness of those guesses at each step, where the end destination is often something that can be known in advance.
[00:23:09] It's an insult to people, actual human beings who when they think, are acting based on a complex tapestry of factors, from their experience to the knowledge they've accumulated across their lifetimes to their brain chemistry. While we may too guess about the correctness of each when reasoning through a complex problem, our guesses are based on something concrete rather than an inelegant mathematical flail, as with O1.
[00:23:37] And by God, it's expensive.
[00:23:42] Pricing for O1 Preview is $15 per million input tokens and $60 per million output tokens. In essence, it's three times as expensive as GPT4O for input and four times as expensive for output.
[00:24:00] There is, however, a hidden cost.
[00:24:03] Data scientist Max Wolff reports that OpenAI's reasoning tokens, the output it uses to get you your final answer, are not visible in the API, meaning that not only are O1's prices higher, the nature of the product requires it to charge you more often all of the things it generates to consider an answer. To be clear, this model is not thinking are also charged for making complex answers for things like coding likely incredibly expensive.
[00:24:37] So let's talk about accuracy.
[00:24:40] On Hacker News, the Reddit style site owned by Sam Altman's former alum Y Combinator, one person complained about O1 hallucinating libraries and functions when presented with a programming task, and making mistakes when asked questions where the answer isn't readily available on the Internet.
[00:24:59] On Twitter, Henrik Nyberg, a startup founder and former game developer, asked O1 to write a Python program that multiplied two numbers and then calculated the expected output of said program while it wrote the code correctly. Although said code could have been more succinct using one line instead of two, the actual result was wildly incorrect. Kartik Kanan, himself a founder of an AI company, tried a programming task on O, where it also hallucinated a non existent command for the API he was using.
[00:25:32] Another person, Sasha Yanxen, tried to play a game of chess with O1, and it hallucinated an entire piece onto the board and then lost because I'm a little shit. I also tried asking O1 to list the number of states with A in the name. After contemplating for 18 seconds, it provided the names of 37 states, including Mississippi. The correct number, by the way, is 36.
[00:26:01] When asked to list the states with the letter W in the name, it pondered for 11 seconds and included North Carolina and North Dakota. I also asked 01 to count the number of times the letter R appears in the word Strawberry, its pre release codename. It said 2openAI claims that O1 performs similarly to PhD students on challenging benchmark tasks in physics, chemistry and biology.
[00:26:28] Just not in geography, it seems, or basic elementary level English language tests, or math or programming. This is, I should note, the big stupid magic trick I predicted in a previous newsletter. OpenAI is shoving strawberry out of the door as a means of proving to investors and the greater public that the AI revolution is still here and what they have is a clunky, unexciting and expensive model.
[00:26:57] Worse still, it's kind of hard to explain why anybody should give a shit about 01, while Sam Altman will likely try and trump up its reasoning abilities.
[00:27:08] What people such as those with the money to keep bankrolling him will see is the 10 to 20 second wait time for answers, which still have issues with basic factual accuracy and the complete lack of of any exciting new features. Nobody gives a shit about better answers anymore. They want it to do something new. And I don't think that OpenAI has any idea how to make that happen. Altman's limp attempts to anthropomorphize 01 by making it think and use reasoning are obvious attempts to suggest that this is somehow part of the path to AGI. But even the most staunch AI advocates are having trouble getting excited.
[00:27:48] In fact, I'd argue that O1 shows that OpenAI is both desperate and out of ideas.
[00:27:56] The prices are not decreasing, the software is not becoming more useful, and the next generation model that we've been hearing about since last November has turned out to be a dud.
[00:28:07] These models are also desperate for training data, to the point that almost every large language model has ingested some sort of copyrighted material.
[00:28:16] This desperation led Runway, one of the largest generative video players, to create a company wide effort to collect thousands of YouTube videos and pirated content to train their models. And a federal lawsuit filed in August alleges that Nvidia did the same to numerous creators to train its Cosmos AI software.
[00:28:37] The legal strategy at this point is sheer force of will, hoping that none of these lawsuits reach the point where any legal precedent is set, or that might define training these models as a form of copyright infringement, which is exactly what a multidisciplinary study out of the Copyright Initiative recently found was the case.
[00:28:57] These lawsuits are progressing with a judge in August granting plaintiffs further claims of copyright infringement against stability AI and DeviantArt, which uses its models, as well as both copyright and trademark infringement claims against Midjourney. If any of these lawsuits prevail, it would be calamitous for OpenAI and Anthropic, and even more so for Google and Meta, whose Gemini and Llama models use data sets including millions of artists work, mostly because it's virtually impossible for AI models to forget training data, meaning that they'd need to be retrained from scratch, which would cost billions of dollars and dramatically reduce their efficacy at tasks they're already not particularly good at.
[00:29:48] I am deeply concerned that this entire industry is built on sand.
[00:29:54] Large language models at the scale of ChatGPT, Claude, Gemini and Llama are unsustainable and do not appear to have a path to profitability due to the compute intensive nature of generative AI.
[00:30:07] Training them necessitates spending hundreds of millions if not billions of dollars, and requires such a large amount of training data that these companies have effectively stolen from millions of artists and writers and hoped they'd get away with it.
[00:30:23] And even if you put these problems aside, Generative AI and its associated architectures do not appear to do anything revolutionary, and absolutely nothing about The Generative AI hype cycle has truly lived up to the term artificial intelligence.
[00:30:40] At best, Generative AI seems capable of generating some things correctly, sometimes summarizing documents or doing research at an intermediate level of faster.
[00:30:52] Microsoft's copilot for Microsoft 365 claims to have thousands of skills that give you infinite possibilities for enterprise, yet the examples it gives involve generating or summarizing emails, starting a presentation, using a prompt and querying Excel spreadsheets. Useful, perhaps, but hardly revolutionary.
[00:31:15] We're not in the early days since November 2022, Big Tech has spent over $150 billion in combined capex and investments into their own infrastructure and budding AI startups as well as in their own models. OpenAI has raised $13 billion and can effectively hire whoever they want as can Anthropic.
[00:31:43] The result of an industry wide Marshall Plan to get Generative AI off the ground has resulted in four or five near identical large language models, the world's least profitable startup, as well as thousands of overpriced and underwhelming integrations.
[00:32:02] Generative AI is being sold on multiple lies that it's artificial intelligence, that it's going to get better, that it will become artificial intelligence, that it is inevitable.
[00:32:17] Putting aside terms like performance as they are largely used as a means of generating things accurately or faster rather than being good at stuff, large language models have effectively plateaued.
[00:32:29] More powerful never seems to mean does more, and more powerful oftentimes means more expensive, meaning that you've just made something that doesn't do more but does cost more to run.
[00:32:42] If the combined forces of every venture capitalist and Big Tech hyperscaler have yet to come up with a meaningful use case that lots of people will actually pay for, there isn't one coming.
[00:32:54] Large language models, and yes, that's where all of these billions of dollars are going, are not going to magically sprout new capabilities because big tech and OpenAI sunk another $150 billion into them. Nobody is trying to make these things more efficient, or at the very least, nobody has succeeded in doing so. If they had, they'd be shouting it from the rooftops.
[00:33:20] What we have here is a shared delusion, a dead end technology that runs on copyright theft, one that requires a continual supply of capital to keep running as it provides services that are at best inessential, sold to us, dressed up as a kind of automation it doesn't provide that cost billions and billions of dollars and will continue to do so in perpetuity.
[00:33:47] Generative AI doesn't run on money or cloud credits so much as it does on faith. The problem is that faith, like investor capital, is a finite resource.
[00:33:59] My concern is that I believe we're in the midst of a subprime AI crisis where thousands of companies have integrated generative AI at prices that are far from stable and even further from profitable.
[00:34:13] Almost every AI powered startup that uses LLM features is based on some combination of GPT or claude. These models are built by two companies that are deeply unprofitable. Anthropics is on course to lose $2.7 billion this year and that have pricing designed to get more customers rather than make any kind of profit. OpenAI, as mentioned, is subsidized by Microsoft both in the cloud credits it received and and the preferential pricing Microsoft offers. And its pricing is entirely dependent on Microsoft's continued support, both as an investor and as a services provider, a problem that Anthropic faces with its deals with Amazon and Google.
[00:34:56] Based on how unprofitable they are, I hypothesize that if OpenAI or Anthropic charged prices closer to their actual costs, there would be a 10 to 100 times increase in the price of API calls, though it's impossible to say how much without the actual numbers.
[00:35:17] However, let's consider for a fact the numbers reported by the information estimate that OpenAI's server costs with Microsoft will be $4 billion in 2024, which I add are over two and a half times cheaper than what Microsoft charges others.
[00:35:34] And then consider that OpenAI still loses over $5 billion a year.
[00:35:41] OpenAI is more than likely charging only a small percentage of what it likely costs to run its models, and can only continue to do so if it is able to continually raise more venture funding than has ever been raised and continue to receive preferential pricing from Microsoft, a company that recently mentioned that it considers OpenAI a competitor. While I can't say that for certain, I would think it's reasonable to believe that Anthropic receives similarly preferential pricing from both Amazon Web Services and Google Cloud. Assuming that Microsoft gave OpenAI $10 billion of cloud credits and it spent $4 billion on server costs and let's say, $2 billion on training costs that are both sure to increase with the new 01 and Orion models.
[00:36:30] OpenAI will either need more credits or will have to start paying actual cash to Microsoft sometime in 2025.
[00:36:39] While it might be possible that Microsoft, Amazon and Google extend their preferred pricing indefinitely, the question is whether these transactions are profitable for them.
[00:36:49] As we saw following Microsoft's most recent quarterly earnings. There is growing investor concern about the capex spending required to build the infrastructure for generative AI, with many voicing skepticism about the potential profitability of the technology.
[00:37:04] And what we really don't know is how unprofitable generative AI is for the hyperscalers, because they bake those costs into other parts of their earnings. While we can't know for sure, I imagine if this stuff was in any way profitable, they'd be talking about the revenue they were receiving from it.
[00:37:23] They're not.
[00:37:25] The markets are extremely skeptical of the generative AI boom, and Nvidia CEO Jensen Huang had no real answers about AI's return on investment, leading to an historic $279 billion drop in Nvidia's market cap in a single day.
[00:37:43] This was the biggest rout in US markets history.
[00:37:48] The total value lost is the equivalent of nearly five Lehman Brothers at its peak value.
[00:37:55] And while the comparison ends there, Nvidia isn't even close to failing. And even if it did, the systemic impact wouldn't be anything near as damaging. It's still an insane amount of money and an indicator of the distorting power of AI on the markets.
[00:38:11] At the beginning of August, Microsoft, Amazon and Google all took a beating from the markets for their massive capital expenditures related to AI.
[00:38:21] And all three of them will face the wheel next quarter if they can't show a significant increase in revenue from the combined $150 billion, if not more, that they've put into new data centers and Nvidia GPUs.
[00:38:37] What's important to remember here is that other than AI, big tech is out of ideas.
[00:38:44] There are no more hyper growth markets left. And as firms like Microsoft and Amazon begin to show signs of declining growth, so too does their desperation to show the markets that they've still got it.
[00:38:58] Google, a company almost entirely sustained by multiple at risk monopolies in search and advertising, also needs something new and sexy to wave in front of the street. Except none of this is working because the products aren't useful enough, and it appears most of its revenue comes from companies trying out AI and then realizing that it isn't really worth it.
[00:39:20] At this point, there are two eventualities. One, Big Tech realizes that they've gotten in too deep and out of a deep fear of pissing off Wall street, chooses to reduce capital expenditures related to AI. And two, Big Tech, desperate to find a new growth pig, decides instead to cut costs to sustain their ruinous operations, laying off workers and reallocating capital from other operations as a means of sustaining the generative AI death march.
[00:39:49] It's unclear which will happen if Big Tech accepts that generative AI isn't the future.
[00:39:55] They don't really have anything else to wave at Wall street, but could do their own version of Meta's Year of Efficiency, reducing capital expenditures and laying off a bunch of people while also promising to slow down investment. By some degree, this is the most likely path for Amazon and Google, who, while desperate to make Wall street happy, still have their own profitable monopolies to point to, for now, at least.
[00:40:22] Nevertheless, there needs to be actual revenue growth from AI in the next few quarters, and it has to be material rather than some vague thing about how AI is a maturing market or annualized run rates, and said material contribution will have to be magnitudes higher if capital expenditures have increased along with it.
[00:40:47] I don't think it's going to be there. Whether or not it's Q3 or Q4, 2024, or even Q1, 2025, Wall street will begin punishing Big Tech for the sin of lust, and said punishment will savage these companies far more harshly than Nvidia, which despite Huang's bluster and empty platitudes, is the only company in the market that can actually point to how AI is increasing revenue.
[00:41:14] I worry somewhat that option 2 is far more that these companies are deeply committed to the idea that AI is the future, and their cultures are so thoroughly disconnected from the creation of software that solves the problems that real people face that they will burn the entire company to the ground.
[00:41:34] I deeply worry about the prospect of mass layoffs being used to fund the movement, and nothing about the last few years makes me think they'll do the right thing and walk away from AI.
[00:41:47] Big Tech has become thoroughly poisoned by management consultants. Amazon, Microsoft, and Google are all run by MBAs and in turn have surrounded them with similarly specious Googles. Like Google's Prabhakar Raghavin, who chased out the people that actually built Google Search so that he could run it.
[00:42:09] These people do not really face human problems and have created cultures dedicated to solving the imagined problems that software can fix.
[00:42:19] Generative AI must seem kind of magical when your entire life is either being in a meeting or reading an email, and I imagine the winning mindset of Satya Nadella mostly comes down to having the tech people sort it out.
[00:42:33] Sundar Pichai could have ended the entire generative AI boom in an instant if he simply looked at Microsoft's investment in OpenAI and laughed. But no, he had to follow because none of these men have any actual ideas, and these companies are not run by people that experience problems, let alone people that might actually know how to fix them.
[00:42:57] They're also desperate, and things have never gone like this for them before, other than when Meta burned billions on the metaverse. Yet this situation is so much bigger and uglier because they have put so much money and so thoroughly welded AI to their companies that removing it will be both embarrassing and damaging to their stocks and a tacit admission that all of this was a waste.
[00:43:25] All this could have been stopped earlier if the media had actually held them accountable.
[00:43:31] This narrative was sold through the same con as previous hype cycles, with the media assuming that these companies would just work it out despite the fact that it was blatantly obvious they wouldn't think I'm a doomer. Well, answer me this, what's the plan here?
[00:43:46] What does generative AI do next? Is your answer that they'll work it out, or that they have something behind the scenes? That's incredible. You're an unwitting participant in a marketing operation.
[00:44:00] No, really, we have to stop being conned by this shit. When Mark Zuckerberg claimed we are about to enter the metaverse, large swathes of the media, the New York Times, the Verge, CBS News and cnn, to name a few, humored an idea that was clearly flawed, one that looked like shit and was peddled using outright lies about the future. It was so obviously nothing other than a shitty VR world. Yet the Wall Street Journal was still over six months into the hype cycle when it was blatantly obvious the metaverse was bullshit, talking about it as the future vision for the Internet. And it happened with crypto web3 and NFTs too. The Verge, the New York Times, CNN, CBS News, all again participating in pumping technology that so clearly didn anything. Though I should add that when I say the Verge here, I really mean Casey Newton, who has a great reputation despite being on his third straight bag pumping, claiming in July that owning one of the most powerful LLMs could provide the company with a basis for all manner of money making products, while discussing a technology that only loses money and has yet to provide one truly useful, lasting product.
[00:45:09] I believe that at the very least, Microsoft will begin reducing costs in other areas of its business as a means of helping sustain the AI boom. In an email shared with me by a source from earlier this year, Microsoft's senior leadership team requested in a plan that was eventually scrapped reducing power requirements from multiple areas within the company as a means of freeing up power for GPUs, including moving other services compute to other countries as a means of freeing up capacity for AI.
[00:45:41] On the Microsoft section of anonymous social network Blind, we where you're required to verify that you have a corporate email of the company in question. One Microsoft worker complained in mid December 2023 of AI taking their money, saying that the cost of AI is so much that it is eating up pay raises and that things will not get better.
[00:46:02] In mid July, another shared their anxiety about how it was apparent to them that Microsoft had, quote, a borderline addiction to cut costs and in order to fund Nvidia's stock price with operational cash flows, and that doing so had damaged Microsoft's culture deeply. Another added that they believe that Copilot is going to ruin Microsoft in fiscal year 25, adding that the fiscal year 25 copilot focus is going to massively fail and that they knew of big Copilot deals in their country that have less than 20% usage after almost a year of proof of concept cuts and layoffs, adding that corporate risk too much and that Microsoft's huge AI investments are not going to be realized when Blind is anonymous it's hard to ignore the fact that there are many, many posts that tell a tale of cultural cancer in Redmond, with disconnected senior leadership that only funds projects if they have AI taped onto them.
[00:47:08] Multiple posts lament Satya Nadella's word salad approach and complain of a lack of bonuses or upward mobility in an organization focused on chasing an AI boom that may not exist. And at the very least, there's a deep cultural sadness to the company with the many posts I've seen oscillating between I don't like working here. I don't know why we're putting so much into AI and get used to it, because Satya doesn't care.
[00:47:36] Buried in the Informations article about the lack of adoption of Microsoft's Office AI features is one particularly worrying thought about the actual utilization of Microsoft's massive data center spend.
[00:47:52] Other signs back up those estimates. Around March of this year, Microsoft had set aside enough server capacity in its Data Centers for 365copilot to handle daily users of the AI Assistant in the low millions. According to someone with direct knowledge of those plans, it couldn't be learned how much of that capacity was in use at the time.
[00:48:13] Based on the Information's estimates, Microsoft has somewhere between 400,000 and 4 million users for its Office Copilot features, meaning that there's a decent chance that Microsoft has built out capacity that isn't getting used.
[00:48:29] Now, one could argue that they're building with the belief that this product category will grow.
[00:48:34] But here's another idea. What if it doesn't?
[00:48:38] What if, and this is a crazy idea, Microsoft, Google and Amazon built out these massive data centers to capture demand that may never arrive?
[00:48:49] I realize I sound a little crazy suggesting this, but back in March I made the point that I could find no companies that had integrated generative AI in a way that has truly benefited their bottom line.
[00:49:02] And just under six months later, I'm still looking the best that big companies appear to have is stapling AI functionality onto existing products and hoping that lets them sell more of them.
[00:49:16] Something that doesn't appear to be working for anybody or like Microsoft offering AI upgrades that don't seem to provide any real business value.
[00:49:25] While there may be companies integrating AI that are driving some degree of spend on Microsoft Azure, Amazon Web Services, in Google Cloud, I hypothesize a lot of this demand is driven by investor sentiment with companies investing in AI to keep the markets happy rather than any cost benefit analysis of actual utility.
[00:49:48] Nevertheless, these companies have spent a great deal of time and money baking in generative AI features to their products, and I hypothesize they will face one of a few different scenarios.
[00:50:00] 1. After developing and launching these features, they find customers don't want to pay for them, as Microsoft has found with its365 copilot. And if they can't find a way to make them pay for it now during the AI boom, they're going to be hard pressed to once people's bosses stop asking them to get in on AI.
[00:50:21] 2. After developing and launching these features, they can't seem to find a way to get users to pay for them and or at least pay extra for them. Which means that they'll have to bake AI into their existing products with no extra margin, effectively turning them into a parasite that eats away at revenue.
[00:50:39] This was a point raised by Jim Covello in Goldman Sachs Generative AI report that if AI's benefits are efficiency improvements such as being able to analyze documents faster, that's something that your competition can do too. If you look carefully, almost every generative AI integration is the same. Some sort of copilot that can answer questions for customers or internally Salesforce, Microsoft Box, Content Creation Box and IBM code generation Cognizant, GitHub, Copilot, and soon Agents the latest bullshit that will mean customizable chatbox that can connect to other parts of websites.
[00:51:19] This point also illustrates one of the biggest problems with generative AI that while powerful in some vague way said power mostly boils down to generates stuff based on other stuff rather than any intelligence. This is also why so many of the AI pages on companies websites see cognizance are gobbledygook. Their biggest sales pitch is you work it out.
[00:51:47] What I fear is a kind of cascade effect.
[00:51:50] I believe that a lot of businesses are trying AI at the moment, and once those trials end, gartner predicts that 30% of generative AI projects will be abandoned after their proof of concepts. By end of 2025, they'll likely stop paying for the extra features or stop integrating generative AI into their company's products.
[00:52:13] If this happens, it will reduce the already turgid revenue flowing to both hyperscalers providing cloud compute for generative AI applications and large language model providers like OpenAI and Anthropic, which could in turn create more price pressure on these companies as their already negative margins sour further.
[00:52:36] At that point, OpenAI and Anthropic will will almost certainly have to raise prices if they haven't already done so.
[00:52:44] Even though big tech could keep funding this boom, after all, they're pretty much entirely responsible for its existence. That won't help the startups that won't be able to afford to run their companies after getting used to discounted rates. Though there are cheaper alternatives like independent providers running Meta's llama models, it's hard to believe that they don't face exactly the same profitability problems and as the hyperscalers.
[00:53:10] It's also important to note that the hyperscalers are also terrified of pissing off Wall Street.
[00:53:17] While they could theoretically, as I fear, start doing layoffs and other cost cutting measures as a means of improving margins, these are short term solutions that only really work if they're able to somehow shake some money from the barren generative AI tree.
[00:53:34] In any case, it's time to accept that the money isn't there. It's time to stop and take stock of the fact that we're in the midst of the third delusional epoch of the tech industry.
[00:53:46] Yet unlike cryptocurrency and the metaverse, everybody has joined the party and decided to burn money pursuing an unsustainable, unreliable, unprofitable and environmentally destructive boondoggle sold to customers and businesses as artificial intelligence that will, quote, automate everything without it ever having a path to do so.
[00:54:10] So why does this keep happening?
[00:54:13] Why have we had movement after movement cryptocurrency the Metaverse, and now Generative AI that doesn't seem like it was actually made for a real person.
[00:54:23] Well, it's the natural result of a tech industry that's become entirely focused on on making each customer more valuable, rather than providing more value to the customer, or for that matter, actually understand who their customers are and what they need.
[00:54:43] The products you're being sold today almost certainly try to wed you to a particular ecosystem, one owned by Microsoft, Apple, Amazon or Google as a consumer at least, and in turn increase the burden of leaving said ecosystem. Even cryptocurrency, ostensibly a decentralized technology, quickly abandoned its freewheeling libertarian ideas and sought to consolidate users on one of the few big platforms like Coinbase, OpenSea, Blur or Uniswap, all backed by the same venture capital firms like Andries and Horowitz. Rather than being flag bearers for a new, radically independent online economic system, they were all only able to scale through the funds and connections of the same people that funded every other recent era of the Internet.
[00:55:32] The Metaverse, while a total sham, was an attempt by Mark Zuckerberg to own the next iteration of the Internet, one where the dominant platform is Horizon and I'll get to generative AI in a minute.
[00:55:46] Everything is about further monetization, about increasing the dollar per head value of each customer, be it through keeping them doing stuff on the platform to showing them more advertising, upselling them new features that are only kind of useful, or creating some new monopoly or oligopoly where only those with the massive war chest of big tech can really play, and very little is about delivering real value, real utility.
[00:56:17] Generative AI is so exciting, at least for a specific type of person, because big tech sees it as the next great monetizer, a means of creating a toll on every single product from consumer tech to the enterprise, because the majority of general COMPUTE goes through either OpenAI or Anthropic, which in turn flows back to either Microsoft, Amazon or Google, creating cloud compute revenue for them to continue expressing growth.
[00:56:47] The biggest innovation here isn't what generative AI does or can do, but rather the creation of an ecosystem that's hopelessly dependent upon a handful of hyperscalers and has no prospect of ever shaking its dependence.
[00:57:03] Generative AI may not be super useful, but it is really easy to integrate into stuff and make new things happen, creating all sorts of new things that a company could theoretically charge for, both for consumer apps and enterprise software companies that make millions or billions, upselling as many customers as possible.
[00:57:25] Sam Altman was smart enough to realize that the tech industry needed a new thing, a new technology that everybody could take a piece of and sell. And while he might not really understand technology, Altman understands the growth lust of the larger economy and productized transformer based architecture as something that everybody could sell, a magical tool that could plug into most things and make something different.
[00:57:53] The problem is that the desperation to integrate generative AI everywhere has shined a light on how disconnected these companies are from actual consumer needs or even running functioning businesses just doing new stuff has kind of worked for 20 years in the sense that simply launching new stuff and forcing salespeople to hock it was enough to keep growth going to the point that the tech industry's leaders have bought into a truly poisonous and deeply unprofitable business.
[00:58:25] The people running these companies, almost exclusively MBAs and management consultants that have never built a product or a technology company from the ground up, either don't understand or don't care that generative AI has no path to profitability, likely assuming it will naturally become profitable like Amazon Web Services, which took nine years to do so did. Despite these being two very, very different things, things have just worked out in the past, so why wouldn't they today?
[00:58:56] I mean, besides the fact that higher interest rates have dramatically reshaped the venture capital markets, reducing VC coffers and shrinking fund sizes, and the fact that sentiment towards tech has never been this negative, and the myriad of other factors, why 2024 is nothing like 2014, which are too lengthy to name in a newsletter that's already over 8,000 words long, those old chestnuts.
[00:59:21] The really worrying part is that other than AI, many of the companies don't seem to have any other new products.
[00:59:30] What else is there? What other things do they have to grow their companies?
[00:59:35] What else do they have?
[00:59:37] Nothing.
[00:59:38] They've got nothing.
[00:59:40] And that really is the problem.
[00:59:43] Because when it fails, the effects will invariably cascade down to other companies in the tech space.
[00:59:50] Every major tech player, both in the consumer and enterprise realm, is selling some sort of AI product integrating either one of the big large language models or their own invariably running cloud compute on one of the big tech players systems. On some level, every single one of these companies is dependent on big tech's willingness to subsidize the entire industry.
[01:00:15] I hypothesize a kind of subprime AI crisis is brewing where almost the entire tech industry has bought in on a technology sold at a vastly discounted rate, heavily centralized and subsidized by big tech. At some point, the incredible toxic burn rate of generative AI is going to catch up with them, which in turn will lead to price increases or companies releasing new products and features with wildly onerous rates like the egregious $2 a conversation rate for Salesforce's Agent Force product that will make even stalwart enterprise customers with budget to burn unable to justify the expense.
[01:00:59] What happens when the entire tech industry relies on the success of a kind of software that only loses money and doesn't create much value to begin with? And what happens when the heat gets too much and these AI products become impossible to reconcile with and these companies have nothing else to sell?
[01:01:18] I truthfully don't know.
[01:01:21] But the tech industry is building toward a grotesque reckoning with a lack of creativity enabled by an economy that rewards growth over innovation, monopolization over loyalty, and management over those who actually build things.