Array ( [content] => [params] => Array ( [0] => /forum/index.php?threads/intel-arc-b580.21735/ ) [addOns] => Array ( [DL6/MLTP] => 13 [Hampel/TimeZoneDebug] => 1000070 [SV/ChangePostDate] => 2010200 [SemiWiki/Newsletter] => 1000010 [SemiWiki/WPMenu] => 1000010 [SemiWiki/XPressExtend] => 1000010 [ThemeHouse/XLink] => 1000970 [ThemeHouse/XPress] => 1010570 [XF] => 2021770 [XFI] => 1050270 ) [wordpress] => /var/www/html )
Don't know as the GPU chip is made by TSMC not Intel's foundryI don’t think it’s very good for margins (they might be negative) but you don’t really worry about margins when breaking into a new segment…
Just curious - did you sell the 4090 now to get as much out of it as possible before the 5090 arrives?I sold my 4090 and purchased the B580. After some testing, I think it has potential. I used it with DaVinci Resolve 19 Pro to edit a 4K timeline, and it got the job done. However, its driver and software support need improvement to compete with NVIDIA's offerings. This should be achievable. Once the 24GB Battlemage model is available, I plan to purchase it and give my current one to my wife for video editing.
I roughly broke even with what I spent on the 4090. I won't be getting the 5090 since the leaked pricing seems way too high. However, I am planning to go for the Battlemage 24GB version insteadJust curious - did you sell the 4090 now to get as much out of it as possible before the 5090 arrives?
re: DaVinci Resolve -- does that app hit both the CPU and GPU when you're editing and rendering? and by "it got the job done" - were there any bugs with the B580 or just a bit slow?
This is a great information video XYang!Intel B580 GPU Review: Training a LLM/GPT Model in PyTorch (Follow-Up and vs. the Nvidia RTX 3090)
Thank you!This is a great information video XYang!
For the iGPU issue; you might try messaging Tom A Peterson on Twitter or some other source. "TAP" is their main outreach person for GPUs (he used to work at Nvidia) and if you can catch his eye it might go somewhere. Or alternatively open a bug report (/r/Intel might have some tips on how to do this properly).
I'm kinda curious what limits training performance. I know for running an actual AI (inference?) LLM, memory bandwidth is often more of a limiting factor than CPU or GPU performance. (This is why running Ollama models on a 3090 isn't necessarily slower than a 4090). I'll research this later. I liked that you provided 3 good data points here - B580, 14700K, and 3090.
For training models - could you buy 2 x B580s and get the benefit of more VRAM like you can for actually running models? (I think when you run models, if you get 2 cards, you get effectively 1.6X-1.8X the VRAM available).
Thanks for this video! (Subbed ).
Thanks for the articles -- the first link was really interesting with how it's sorta not parallel when training but you do get benefits of multiple GPUs. The extra bandwidth of NVLink makes a lot of sense for this.Another way to use multiple GPUs for training is by using a Granite Rapids/Sapphire Rapids-based workstation board. PCIe 5.0 lanes can facilitate inter-GPU communication. I am not sure the performance though.
PCI Express 6.0 Specification | PCI-SIG
pcisig.com
Very interesting - I need to play with Pytorch and TorchServe. So far I've just been using Ollama (under the Windows subsystem for Linux (WSL), and also on some dedicated Linux machines I have). It's able to effectively split workloads among multiple GPUs or split between GPU and CPU (i.e. you actually get some (small) speedup running a 70B model on a GPU+CPU vs CPU only). This is for inference and not training of course.I think you can use multiple GPUs to serve models relatively easily, as that does not require synchronization. However, I have not tried it yet: