Array
(
    [content] => 
    [params] => Array
        (
            [0] => /forum/index.php?threads/kaist-develops-next-generation-ultra-low-power-llm-accelerator-shames-nvidia.19794/
        )

    [addOns] => Array
        (
            [DL6/MLTP] => 13
            [Hampel/TimeZoneDebug] => 1000070
            [SV/ChangePostDate] => 2010200
            [SemiWiki/Newsletter] => 1000010
            [SemiWiki/WPMenu] => 1000010
            [SemiWiki/XPressExtend] => 1000010
            [ThemeHouse/XLink] => 1000970
            [ThemeHouse/XPress] => 1010570
            [XF] => 2021370
            [XFI] => 1050270
        )

    [wordpress] => /var/www/html
)

KAIST Develops Next-generation Ultra-low Power LLM Accelerator - Shames Nvidia!

Daniel Nenni

Admin
Staff member
Claim Samsung-fabbed chip is the first ultra-low power LLM processor.
1710116538081.png


A photo describing an artificial intelligence chip which processes a large language model with neuromorphic computing technology provided by the Ministry of Science and ICT on March 6, 2024 (Image courtesy of Yonhap)

SEOUL, Mar. 6 (Korea Bizwire) — A research team at the Korea Advanced Institute of Science and Technology (KAIST) has developed the world’s first artificial intelligence (AI) semiconductor capable of processing a large language model (LLM) with ultra-low power consumption, the science ministry said Wednesday.

The team, led by Professor Yoo Hoi-jun at the KAIST PIM Semiconductor Research Center, developed a “Complementary-Transformer” AI chip, which processes GPT-2 with an ultra-low power consumption of 400 milliwatts and a high speed of 0.4 seconds, according to the Ministry of Science and ICT.

The 4.5-mm-square chip, developed using Korean tech giant Samsung Electronics Co.’s 28 nanometer process, has 625 times less power consumption compared with global AI chip giant Nvidia’s A-100 GPU, which requires 250 watts of power to process LLMs, the ministry explained.

The chip is also 41 times smaller in area than the Nvidia model, enabling it to be used on devices like mobile phones.

The ministry said the utilization of neuromorphic computing technology, specifically spiking neural networks (SNNs), is essential to the achievement.

Previously, the technology was less accurate than deep neural networks (DNNs) and mainly capable of simple image classifications, but the research team succeeded in improving the accuracy of the technology to match that of DNNs to apply it to LLMs.

The team said its new AI chip optimizes computational energy consumption while maintaining accuracy by using unique neural network architecture that fuses DNNs and SNNs, and effectively compresses the large parameters of LLMs.

 
I have no doubt that this advance will be the first of a long line of advances ahead of us. The world has an increasing amount of collaboration, cooperation, communications and computational power to take on any challenge, we just have to learn to maximize it. No doubt formal, morphing structures will be built to achieve these ends of many types.
 
Everybody is gunning for NVIDIA, and rightly so. As I have said before, AI is a bubble, just like crypto, and it is getting ready to pop. My guess is 2025. I know of dozens of start-up companies making AI chips not to mention the big semiconductor companies. EDA/IP and TSMC will be the big benefactors, absolutely.
 
Back
Top