Mar. 14 at 5:07 PM
The AI landscape is shifting from "Training" to "Inference," and the hardware battle is heating up. Amazon (
$AMZN) is now deploying Cerebras chips alongside its proprietary Trainium to maximize inference efficiency.
Cerebras Momentum: Following a massive
$10B+ deal with OpenAI in January, Cerebras is positioning itself as the premier NVIDIA alternative. CEO Andrew Feldman claims their wafer-scale chips run 20x faster than standard GPUs for specific workloads.
The Inference Pivot: As models move from development to massive-scale deployment, "Inference" (running the models) is becoming the primary cost driver.
Dominance Under Threat: While
$NVDA and
$AMD dominate training, specialized architectures (ASICs) like Trainium and Cerebras are optimized for the lower-latency, high-throughput needs of inference.