Feb. 28 at 5:00 PM
$NVDA $SPY $QQQ $GOOGL $MU
AMD, MU, Intel may come under pressure with the NEW AND IMPROVED inference chip
EXTREMELY BULLISH ROR NVDA
Reducing HBM footprint will improve costs
• Bottleneck: Slow decoding / token speed
Improvement : 5-10x faster (500+ tokens/sec)
Benefited: NVIDIA, OpenAI, hyperscalers
Hurt: AMD (MI300), Intel (Gaudi)
• Bottleneck: Memory bandwidth / latency wall
Improvement : On-chip SRAM, no HBM waits
Benefited: NVIDIA, cloud providers
Hurt: Google (TPU), Micron (HBM dips)
• Bottleneck: High power & energy waste
Improvement : Up to 90% less power
Benefited: NVIDIA, Meta, AWS
Hurt: Cerebras, SambaNova
• Bottleneck: Cost per token / supply chains
Improvement: 10x lower ops, less HBM reliance
Benefited: NVIDIA, all inference buyers
Hurt: AMD, Intel