Gaudi 3 is not bad.That’s why I question folks proposing lower level standards - I can understand it for HPC problems, but not for GenAI inference. I think the big competitive battle going on right now is going to be about inference cost / power per token at the data center level, for every leading model. The good news is that Llama has been added to MLPerf 5.0. The bad news is that the focus is still on performance, so they aren’t looking at cost/power per token yet.
![]()
NVIDIA Blackwell & AMD MI325X Showdown In Latest MLPerf Inference Benchmarks: B200 Shatters Records, Instinct Fights Against Hopper
NVIDIA & AMD have submitted latest MLPerf Inference performance benchmarks of their latest GPUs, including Blackwell B200 & Instinct MI325X.wccftech.com