Array
(
    [content] => 
    [params] => Array
        (
            [0] => /forum/threads/qualcomm-and-cxmt-jointly-launch-a-standalone-npu-3d-dram-breaking-through-the-performance-bottleneck-of-mobile-ai.25035/
        )

    [addOns] => Array
        (
            [DL6/MLTP] => 13
            [Hampel/TimeZoneDebug] => 1000070
            [SV/ChangePostDate] => 2010200
            [SemiWiki/Newsletter] => 1000010
            [SemiWiki/WPMenu] => 1000010
            [SemiWiki/XPressExtend] => 1000010
            [ThemeHouse/XLink] => 1000970
            [ThemeHouse/XPress] => 1010570
            [XF] => 2031070
            [XFI] => 1060170
        )

    [wordpress] => /var/www/html
)

Qualcomm and CXMT jointly launch a standalone NPU + 3D DRAM, breaking through the performance bottleneck of mobile AI

Fred Chen

Moderator
Author: Su Ziyun | Publication Date: April 16, 2026, 12:13 PM

According to Wccftech, Qualcomm is collaborating with Chinese memory manufacturers Changxin Memory and GigaDevice to develop a standalone neural processing unit (NPU) that incorporates 3D DRAM, attempting to overcome the current bottleneck of mobile AI computing power being limited by the SoC architecture.

This solution aims to enhance long-term computing capabilities such as real-time translation and image generation by separating AI computing from the main chip and importing it into high-bandwidth memory. It is expected to be introduced into the mid-to-high-end mobile phone market in China as early as the end of 2026 to the beginning of 2027.

Currently, most smartphones integrate their NPUs into the SoC, which requires sharing power consumption and memory resources with the CPU and GPU, making it difficult for AI computing power to be output stably for extended periods.

The new solution adopts an independent architecture. According to Ming-Chi Kuo, an analyst at TF International Securities, it may be equipped with 4GB of customized 3D DRAM and significantly improve memory bandwidth through advanced packaging technologies such as TSV (Through Silicon Via) and Hybrid Bonding, surpassing the current LPDDR5X standard, thereby supporting more intensive AI tasks.

In terms of performance, this AI accelerator can provide a stable computing power of approximately 40 TOPS. In comparison, while current flagship phones boast a maximum of 100 TOPS, most can only achieve this briefly under specific conditions. If stable output can be maintained for an extended period, it has the potential to substantially improve the AI experience on mobile devices and even double the overall computing power.

However, cost pressures could be a key variable in its adoption. With rising memory prices, Chinese brands such as Xiaomi, OPPO, vivo, and Huawei are actively seeking to reduce DRAM and storage costs to maintain product profit margins. In this context, adding an additional dedicated NPU with 4GB of 3D DRAM will inevitably increase the overall BOM (Bill of Materials), potentially impacting manufacturers' willingness to adopt it.

Furthermore, the market demand for mobile AI remains unclear. TF International Securities analyst Ming-Chi Kuo points out that even with gradually maturing hardware, there is still a lack of application scenarios that can fully leverage the AI capabilities of devices. Consumers remain hesitant about paying a premium for this. While the combination of a dedicated NPU and 3D DRAM has the potential to reshape mobile AI architecture, its ultimate mainstream adoption depends on cost control and the maturity of the application ecosystem.

Furthermore, the market demand for mobile AI remains unclear. Ming-Chi Kuo pointed out that even as hardware conditions gradually mature, there is still a lack of application scenarios that can fully leverage the AI capabilities of devices, and consumers are still taking a wait-and-see approach regarding whether they are willing to pay a premium for it.


 
Last edited by a moderator:
Back
Top