Generally speaking, you should be able to venture a good guess as to whether high profile companies are likely to rally or sell off post-earnings based on the magnitude of any beats or misses in the context of the current zeitgeist.
That latter bit — about contextualizing the scope of beats and misses by the prevailing market mood — is key. That’s why I italicized it.
If you’re a macro bellwether and everyone’s convinced a recession’s coming, an in-line guide might be good for a respectable post-earnings rally. An upside guide might lift the whole market. If, by contrast, you’re on the cutting edge of a technological revolution, or anywhere near the cutting edge, in-line won’t cut it. Guiding above the Street for the current quarter might not cut it either. In some cases, the bar might be so high that nothing will suffice to pacify the market.
With key AI-related names, we’re probably somewhere between upside guides not cutting it and nothing being good enough, which is to say yes, it’s still possible to clear the bar, but it’ll be very (very) hard. Cue an 8% adverse initial reaction to Micron’s current-quarter guide:
Forget whether that holds on Thursday (although I’ll quickly note that if it had, it would’ve been the largest percentage drop since March of 2022). The point is that the company didn’t really do anything “wrong.”
Consider this: Micron guided for between $7.4 billion and $7.8 billion in fiscal Q4 sales. Consensus was $7.58 billion. So, the midpoint’s bang in-line. The midpoint of the margin guide was likewise in-line, and the EPS outlook was easily better than expected (midpoint of $1.08 versus consensus $1.02). The company also beat on the top and bottom line for fiscal Q3 (i.e., last quarter).
But, coming back to what I said here at the outset, it was immediately obvious that wouldn’t be enough. Micron’s an AI craze beneficiary. Guiding consistent with the Street while nodding gingerly in the direction of some modest upside isn’t even close to sufficient in this environment. Micron’s added about $70 billion in market cap in 2024 so far courtesy of a rally that found the shares trading at twice their typical forward sales multiple.
Maybe the reaction was overdone, maybe it wasn’t. And maybe it’ll be “cured” by the end of the week. But coming as it did on the heels of a $430 billion three-day rout for Nvidia (the largest three-session value destruction event in US stock market history), it was a cautionary tale: Some of these AI bets are “overcooked,” as one Singapore-based strategist quoted by Bloomberg put it.
But don’t worry. This is a revolution. And Micron will be among the “biggest beneficiaries,” CEO Sanjay Mehrotra said on the call. “We are in the early innings of a multi-year race to enable artificial general intelligence, which will revolutionize all aspects of life,” he declared. “AI will also permeate to the edge via AI PCs and AI smartphones, as well as smart automobiles and intelligent industrial systems.”
Party on Huang. Party on Garth.



“If you’re going to spew, spew into this” holds up folded dixie cup
One of my favorite films
Everything has sped up. “Investors” now are demanding instant gratification on theme trades. While many guys hope for that when dating, it’s pretty silly when it comes to investing in long-term ideas, yes-no? .
Wish it was up ten but cant complain overall. Sanjay had some good propaganda in there but ive noticed over the years he has a degree of restraint in his stock pumping tactics. Theyll be ok.
In my mind, chips always become a commodity (remember SanDisk?) Once the data centers are built and equipped with chips, demand will plummet and software will take the baton back. Chip prices will plummet, and the next data centers will only be upgrading much more slowly than this initial compute build. And it’s possible a different chip designer will take meaningful market from NVDA.
Agree. For me, Nvidia is the camera and in the long run AI is all about the film = software.
Agree as well. It seems to me that the problem with betting on hardware demand is it’s possible to saturate the market, and, it’s harder for users to pivot when newer technology comes along.
MU is interesting. I bought it at pretty much the bottom, because when growth, earnings, EBIT, gross margin, capex cuts, everything are horrifically negative and getting worse, that’s when you buy memory semis.
You sell memory semis when everything is terrifically positive and getting better, so I took trims after the quick double and have been holding reduced positions, trying to figure out if the transcendence of AI will change the deep cyclical character of memory semis. Thinking I’m getting closer to a conclusion.
The next question, of course, is whether other semis that are less cyclical but still cyclical have had their characters changed.
Even with a seemingly powerful, logical and long-term tailwind, chips will remain somewhat cyclical. Similar to robotics & automation stocks.
Chips have the added burden of being incredibly capital-intensive relative to most other products. Expansion often leads to over-capacity in the sector.
It brings to mind the swings you see in the mining sector. It’s really no different than chips. You’ll always need copper/lithium etc but demand is NOT infinite, at least in the short run, which is all investors seem to care about these days.
How big will AI be, most certainly more than blockchain (if I recall early predictions were pretty phenomenal yet there is really only bitcoin to hang your hat on). Some form of AI will most certainly be used in every midsize and above company. Is it another nice tool like CRM or ERP, or will it be something magnitudes larger giving outsize returns for companies beyond the AI infrastructure providers.
Favorite pet topic of mine. I’m a big believer in the possibilities blockchain, but only because I actually understand it. The primary use case thing that’s so exciting to me about blockchain is not cryptocurrency, it’s the smart contracts and possibility of decentralized apps—being able to start the next Uber, eBay, or Amazon without needing to have the resources to pay for a server farm. At the moment, that’s way off in fantasyland, but only because the energy efficiency and computing horsepower aren’t there. Both of those will absolutely arrive someday. We are a very, very long way off from practical applications, but, I had a feeling as I got acquainted with smart contracts that I hadn’t had since the first time I saw a web browser in 1994. The possibilities are stunning.
Meanwhile, scalable blockchain-based decentralized apps are not going to be here for a long time, while AI has already filled the streets of my city with driverless cars. But, in the very long run, for me personally, it’s hard to guess which is going to be a more impactful technology. My gut says AI (actual usable AI, not the myth that’s being hyped, see my other comment on this page) is a major but predictable evolutionary advance, like web browsers and HTML, while blockchain, if they can overcome the practical obstacles, will be a revolutionary paradigm shift more comparable to the development of modern internet infrastructure itself.
Really curious to watch how this AI bubble goes. As with so many advanced technologies, obviously it does do incredible things, while the same time not even really doing a small fraction of what people are hyping it as doing now. Last night — representative recent experience — I had ChatGPT 4o fail almost 25 times in a row to correctly help me with a very simple CSS problem (refactoring a fairly simple animation to use compositor-only attributes, for you tech geeks out there). It got me 75% or 80% of the way there, and then failed abysmally, over and over, and I ended up with a completely meaningless answer that took way longer than it would’ve for me to simply Google to find out how to do it. I wonder how long it’s gonna take before the public at large realizes that that level of reliability is the norm, and what’s gonna happen to the AI investment thesis when they do.
Make no mistake, the capacity for it to help in very specialized areas is, I’d go so far as to say, breathtaking. But as far as broad general practical applications or anything even remotely approaching AGI, we are so far from that that it’s laughable. That is nothing like how the current systems work. PT Barnum would be hawking AI today.
An amusing “parlor game” is to ask people to explain just what AI actually is. What underpins it? How does it work? You will receive very few informed answers, outside of “I fooled around with it and generated a couple of really cool images!” Or “I heard it will replace every software coding job!”
Now, if the last one us even remotely correct (it’s not) why in heaven’s name are you buying Indian stocks??
I know even skilled developers who are insisting that it’s useful as a tool for programmers. However, I have yet to see one of them actually produce for me a prompt that gave them such practical results… I always make it a point to ask.
My favorite explanation is the famous tweet describing it as “spicy autocomplete”. If you’re not familiar, look it up, the guy really put his finger on it. Basically, he described the stages of getting to familiar with AI as: 1. Oh my God this is amazing. 2. Holy cow, this thing is going to put me out of a job. 3. Wait a minute, this thing is wrong a lot. 4. This is nothing but spicy autocomplete.
I guess the link is on my office PC, so I’ll have to produce an error-prone synopsis from memory. I came across an interview with the guy at AWS in charge of marketing AI tools to coders. He boasted how users were, on average, receiving around 15 suggestions per day from the AI tools.
What caught my eye was when he went on to boast that 36% were adopted. So, he was boasting of a 64% failure/reject rate? That’ll change the world for sure!
I’ll throw this out to add to your worry files or riducule bucket.
If LLMs are becoming capable of replacing many coding functions, don’t you suppose that it would be VERY enticining for Chinese, Russian and North Korean hackers to inject clever little back doors into the libraries underpiing the LLM- generated code? That might prove to be way more efficient than most “spear fishing” campaigns.
A silly worry I suppose – didn’t Sam Altman assure us that they all so very carefully screen their source databases for such things?
He needs to define “suggestions”. My experience with using Copilot in VS Code, for instance just for autocomplete suggestions of somewhere between two words and at absolute most three or four lines including comment text, is that it’s actually astoundingly good. And I’ve seen it do some pretty remarkable things, like telling it “change this conditional to handle the default case, and change all the output to lowercase“ and it’ll handle it perfectly, which is a very cool convenience. But asking the VS code Copilot chatbot to help with anything is uniformly an exercise in futility. Plus there’s a problem that on the rare occasions it produces something workable, I found that you build on it for a few days and only then discover a fundamental flaw that was too subtle to notice at first but winds up being significant enough to break the whole thing. You said AWS, not Microsoft, otherwise I’d assume by “accept” he’s talking about hitting the button to use the suggestion, but I don’t actually know what he means in the context of AWS. I think it’s safe to assume that his answer falls in its expected place in the hierarchy of lies, damned lies, and statistics, and that the actual rate of AI suggestions making it all the way through to production code is likely much lower than even that not terribly practical 36%. Here’s an example, I had copilot give me consecutively two mutually exclusive opposite answers about how JavaScript evaluates a certain expression. That means the accuracy I got from it on a simple question of fact, not even a question of it coming up with a working algorithm but a simple reference fact, was equivalent to a coin flip. His quoted 36% is even less accurate than a coin flip.
15 suggestions a day? I often get more than that in an hour, as with mounting frustration I try over and over again to get it to just solve the damn problem I asked without hallucinating or omitting basic facts.
People will realize. It’s going to be very interesting to watch this shake out.
Hmmmm. I’m not so worried about hackers putting anything in the code, it would be more of a concern about poisoning the models, I think. And a lot of it is open source, so it’ll get audited by geeks the world over.
The funny thing for me to think about is that the training corpii all come off of the Internet, which means, over time, the training corpii are going to contain much more machine-generated material. Where is that gonna lead? We’re gonna wind up with the semantic equivalent of that banjo-playing kid from “Deliverance”.
My totally uninformed guess is that a first attack vector will be via Python.
But the real gold mine for pro hackers will be if they can inflitrate Cobol and other more legacy languages which still run many things. Using programming languages which have fewer and fewer folks who worked in them day after day. It’ll be more and more tempting for less experinced coders to rely on LLM generated suggestions. Just “click it in there”!