Discussion about this post

User's avatar
Neural Foundry's avatar

The inference versus training split is the underappreciated part of this story. Meta's commitment to AMD chips at scale makes perfect economic sense when you're running billions of inference requests where every marginal dollar compounds. The 30-40% cost savigs on comparable performance is real money at their scale. AMD's timing with the software fixes was impeccable, catching the wave right as inference workloads are exploding.

Expand full comment
R L's avatar

Meta's public commitment to deploying AMD MI300 chips at scale validates your thesis that the software moat is eroding faster than the market realizes. The cost arbitrage you've outlined becomes exponentially more valuable when you consider that inference workloads are growing faster than training and will likely represent 80%+ of compute demand within 18 months. What's particularly astute is recognizing that 'good enough' plus 30-40% cost savings beats perfection in infrastructure economics, especiay when companies like Meta are running hundreds of millions of requests daily where marginal costs compound quickly. The timing of AMD's software fixes coinciding with the inference workload explosion is either incredibly lucky or brilliant positioning, but either way it's creating a genuine competitive dynamic that should compress NVIDIA's margins over the next few quarters.

Expand full comment

No posts