You are currently viewing SemiWiki as a guest which gives you limited access to the site. To view blog comments and experience other SemiWiki features you must be a registered member. Registration is fast, simple, and absolutely free so please, join our community today!
For more than 35 years, Arm has licensed its chip architecture and collected royalties on every processor made by customers like Apple, Nvidia, Google, Amazon, Intel and Microsoft. Now Arm is becoming their competitor, making physical silicon of its own for the first time, with Meta as the initial customer. CNBC got an exclusive first look at Arm's new AGI CPU that's "ruthlessly optimized" for running AI inference in data centers, and toured the $71 million lab in Austin, Texas, that Arm built for its new CPU.
I don't get it. These chips are just a bunch of Neoverse V3 cores. Nothing really special. I'm having trouble seeing advantages it has beyond the AWS Graviton 5. And then there's Ampere, also owned by SoftBank, which has custom Arm cores, which are supposedly superior to Neoverse V3 cores. Arm and SoftBank confuse me. Or am I missing something?
I am watching the livecast of Arm AGI CPU lunch,I think they want to join the competition of AI Server Architechture system,so they need CPU to provide total solutions.
I don't get it. These chips are just a bunch of Neoverse V3 cores. Nothing really special. I'm having trouble seeing advantages it has beyond the AWS Graviton 5. And then there's Ampere, also owned by SoftBank, which has custom Arm cores, which are supposedly superior to Neoverse V3 cores. Arm and SoftBank confuse me. Or am I missing something?
It's hard to get really good info on this ARM "AGI CPU", but it *looks* like the differences are:
- Graviton is more focused on general purpose, while ARM AGI is more biased towards Inference via improved inter-chip and off-chip bandwidth
- Graviton appears to be used exclusively by Amazon for AWS, where ARM AGI is the "public market offering" that anyone can use
It doesn't seem very special but it is at least a "general offering" that any server/cloud provider could buy/use.
.. and yes Ampere definitely has strong overlap with both Graviton and ARM AGI ..
It's hard to get really good info on this ARM "AGI CPU", but it *looks* like the differences are:
- Graviton is more focused on general purpose, while ARM AGI is more biased towards Inference via improved inter-chip and off-chip bandwidth
- Graviton appears to be used exclusively by Amazon for AWS, where ARM AGI is the "public market offering" that anyone can use
It doesn't seem very special but it is at least a "general offering" that any server/cloud provider could buy/use.
.. and yes Ampere definitely has strong overlap with both Graviton and ARM AGI ..
I think the marketing gimmick of tagging it as the "AGI CPU" is just silly. I can call myself young and good-looking, but as the Linda Ronstadt song goes, all it would take is "Just One Look" to know I'm full of you-know-what. Even Intel Xeon 6s have hardware accelerators which make it them better for AI than these "AGI" CPUs. I have trouble typing the AGI letters without chuckling. Arm, seriously, fire your marketing team.
I think the marketing gimmick of tagging it as the "AGI CPU" is just silly. I can call myself young and good-looking, but as the Linda Ronstadt song goes, all it would take is "Just One Look" to know I'm full of you-know-what. Even Intel Xeon 6s have hardware accelerators which make it them better for AI than these "AGI" CPUs. I have trouble typing the AGI letters without chuckling. Arm, seriously, fire your marketing team.
1. What do you think the competitive advantage between this ARM "AGI CPU" versus what nVidia is building internally that is also based on ARM architecture? Obviously NV's orchestration CPU is in production, but do you see ARM getting into NV's eco-system with the silicon in addition to IP play?
2. As for x86 comparison, we know Intel has been NV's partner lately, and I think from their latest earnings call, Intel is building and supplying their purposely built CPU into NV as well. So we can assume Intel at least understand very well, or as well as ARM, about NV's AI infrastructure needs. Assuming current Intel offering is sub-par compared to AGI CPU, what would it take for Intel (or AMD) to build one that's competitive or better that AGI CPU?
1. What do you think the competitive advantage between this ARM "AGI CPU" versus what nVidia is building internally that is also based on ARM architecture? Obviously NV's orchestration CPU is in production, but do you see ARM getting into NV's eco-system with the silicon in addition to IP play?
AGI versus Vera... to start, AGI uses a larger number of single-threaded cores. Vera uses a "spatial threading" strategy with two threads per core, but the threads have dedicated execution blocks, which is different than Intel's Simultaneous Multi-Threading (SMT). SMT designs use over-provisioned execution blocks which mostly allow the two threads to proceed without stalling. Mostly. All three strategies use shared caches to reduce cache coherency complexity. Without comparison traces of the same code on all three CPUs it is impossible to directly compare their performance.
It looks like getting into Nvidia's ecosystem is a matter of integrating NVLink port blocks onto the CPU dies and giving them cache-level access. Without that integration, scale-up link transfers will be essentially I/O operations, which are much higher overhead functions than direct cache access. Being in the coherent cache domain is said to be part of Vera's development. So it looks it to me like AGI is more likely to be in the "not Nvidia" market.
2. As for x86 comparison, we know Intel has been NV's partner lately, and I think from their latest earnings call, Intel is building and supplying their purposely built CPU into NV as well. So we can assume Intel at least understand very well, or as well as ARM, about NV's AI infrastructure needs. Assuming current Intel offering is sub-par compared to AGI CPU, what would it take for Intel (or AMD) to build one that's competitive or better that AGI CPU?
I would not assume Intel's Xeon approach is sub-par to Arm AGI, until we get data from direct comparison testing. That's probably going to take a while to be published. If Xeons integrate NVLink and AGIs don't, that right there is going to give Intel a substantial advantage. If anything is an advantage for AGI, it might be lower power consumption.
AGI versus Vera... to start, AGI uses a larger number of single-threaded cores. Vera uses a "spatial threading" strategy with two threads per core, but the threads have dedicated execution blocks, which is different than Intel's Simultaneous Multi-Threading (SMT). SMT designs use over-provisioned execution blocks which mostly allow the two threads to proceed without stalling. Mostly. All three strategies use shared caches to reduce cache coherency complexity. Without comparison traces of the same code on all three CPUs it is impossible to directly compare their performance.
It looks like getting into Nvidia's ecosystem is a matter of integrating NVLink port blocks onto the CPU dies and giving them cache-level access. Without that integration, scale-up link transfers will be essentially I/O operations, which are much higher overhead functions than direct cache access. Being in the coherent cache domain is said to be part of Vera's development. So it looks it to me like AGI is more likely to be in the "not Nvidia" market.
I would not assume Intel's Xeon approach is sub-par to Arm AGI, until we get data from direct comparison testing. That's probably going to take a while to be published. If Xeons integrate NVLink and AGIs don't, that right there is going to give Intel a substantial advantage. If anything is an advantage for AGI, it might be lower power consumption.