Array
(
    [content] => 
    [params] => Array
        (
            [0] => /forum/threads/mram-getting-more-attention-at-smallest-nodes.18550/
        )

    [addOns] => Array
        (
            [DL6/MLTP] => 13
            [Hampel/TimeZoneDebug] => 1000070
            [SV/ChangePostDate] => 2010200
            [SemiWiki/Newsletter] => 1000010
            [SemiWiki/WPMenu] => 1000010
            [SemiWiki/XPressExtend] => 1000010
            [ThemeHouse/XLink] => 1000970
            [ThemeHouse/XPress] => 1010570
            [XF] => 2021770
            [XFI] => 1050270
        )

    [wordpress] => /var/www/html
)

MRAM Getting More Attention At Smallest Nodes

jms_embedded

Well-known member

Magneto-resistive RAM (MRAM) appears to be gaining traction at the most advanced nodes, in part because of recent improvements in the memory itself and in part because new markets require solutions for which MRAM may be uniquely qualified.

There are still plenty of skeptics when it comes to MRAM, and lots of potential competitors. That has limited MRAM to a niche role over the past couple decades, hampered by high costs, low density, and lower endurance. But the number of proponents is growing.

...

--------

I'm not familiar with the technology but I thought the article was fairly in-depth and a good read.
 
The density seems not a huge advance over SRAM, the article suggests double density. Well, if you use an optimized face-bonded SRAM chip like AMD's V-cache you almost double the density relative to the SRAM which is compiled to be compatible with a logic chip.

The other issue is speed. The signal from MRAM is weak so you need a good sense amp and time to integrate, which presumably translate to a bottleneck on access path per MB of storage. SRAM on a chip can provide tens of terabyte/sec throughput which is valuable for AI edge loads (not all the market, but almost all the new money looking for improved chips). Can MRAM come anywhere near that?
 
The density seems not a huge advance over SRAM, the article suggests double density. Well, if you use an optimized face-bonded SRAM chip like AMD's V-cache you almost double the density relative to the SRAM which is compiled to be compatible with a logic chip.
Yes, but you are still doubling your mm*2 of SRAM while doing that. You also need to hybrid bond the two dies which at least currently should be a yield hit, additionally the process does not seem to currently be running at a super high capacity (hopefully TSMC is working on expanding this alongside the rest of their advanced packaging expansion to make this cool tech more available). As you are no longer doing traditional packaging, packaging costs will increase potentially followed by thermals. The node you are stacking cache onto also needs to support TSVs, as well as said TSVs reducing area utilization of your base die.

Which isn’t to say I don’t think the technology has a place, because it seems the wind is blowing that way with the stagnation of SRAM scaling while logic keeps marching ahead. But it doesn’t seem like it will be the best choice for every application. Especially in this early stage while things aren’t mature yet.
 
Last edited:
It seems huge for us ASIC design houses
Yes it has its uses. I see it more as replacing NOR than replacing SRAM, since it can't match the bandwidth and locality of SRAM.
V-cache... you gotta go out to pads, then use 3D packaging, then the thermal issues and costs. Really?
Well, for sure it is only happening because there is no alternative which comes close to that performance.
 
additionally the process does not seem to currently be running at a super high capacity
Yep, advanced packaging is rapidly becoming a pain point. The problem is that the few pioneers like Sony imaging or AMD v-cache or YMTC Xtacking which prove it works, and works really well, open a flood of other companies that have similar needs but just were too small or too cautious to be pioneers. Now it is proven, they all want it, but capacity takes time to build.
But it doesn’t seem like it will be the best choice for every application. Especially in this early stage while things aren’t mature yet.
If you don't really need SRAM performance and overall bandwidth that you can get from distributed embedded SRAM, then you have other choices. But, not if SRAM really is what you need.

It is mature enough. It just isn't cheap, and capacity is not available for all who want it.
 
Yes it has its uses. I see it more as replacing NOR than replacing SRAM, since it can't match the bandwidth and locality of SRAM.
Again, this is for where the foundry embeds some macro of NVM onto your SoC or microcontroller. They are really offering nothing (in terms of e-Flash) past 28nm node.

For standalone NOR (only NVM on chip), the market is very small, not much to say there. But in terms of increasing capacity upwards of Gb, it's going to run into 3D NAND eventually, especially as NAND has been touted as NOR replacement.
 
Last edited:
So for those of us designing ASICs with many cores on the edge with sane NRE budgets, can I assume that L1 cache uses SRAM, L2 will use a flavor of MRAM, and inductors must be avoided?
 
So for those of us designing ASICs with many cores on the edge with sane NRE budgets, can I assume that L1 cache uses SRAM, L2 will use a flavor of MRAM, and inductors must be avoided?
Foundries have mainly developed MRAM as a Flash macro replacement not SRAM replacement. The cell designs are quite different. Only Intel has come close with publishing a targeted last level cache application, obviously not in their chips (not yet at least).
 
Last edited:
Can MRAM come anywhere near that?

No, but MRAM use cases don't really need that.

SRAM on lower nodes is quite leaky, and needs all power saving tricks from big chips to realise the benefits from lower nodes.

MRAM on the other hand can be completely powered off, no gating tricks needed as such.

The core can be either suspended with untouched register/cache content, or powered off completely.

Off-chip toggle MRAM still has no alternative for high rate logging purposes, due to write cycles far surpassing NOR.
 
No, but MRAM use cases don't really need that.

SRAM on lower nodes is quite leaky, and needs all power saving tricks from big chips to realise the benefits from lower nodes.

MRAM on the other hand can be completely powered off, no gating tricks needed as such.

The core can be either suspended with untouched register/cache content, or powered off completely.

Off-chip toggle MRAM still has no alternative for high rate logging purposes, due to write cycles far surpassing NOR.
RAM+Flash, or battery-backed RAM in some cases.
 
Not quite there for SRAM, not a sure case against RRAM or NOR either. High-retention current accelerating breakdown (limiting endurance) is the issue.
Screenshot_20230827_132409_Chrome.jpg


TMRC 2023 paper C3 https://sites.google.com/umn.edu/tmrc2023/
 
Last edited:
SSD's are getting pretty scary. >500 MB/s.
Unfortunately, to access SSDs you need to do OS I/O operations, which are comparatively very slow and inefficient, even with the NVMe protocol, compared to memory accesses. The only currently-defined fix that comes to mind for individual SSDs would be RDMA support, which hasn't appeared due to cost and power challenges. NVMe over Fabrics supports RDMA, but for now this is only implemented by storage system servers, which only use RDMA in NVMe over Fabrics between the storage systems and the application servers, and even this use model for RDMA is not widely deployed. The storage industry also seems to be investing in the far slower NVMe over TCP protocol, because it's cheaper to implement in storage system servers and doesn't need RDMA-capable networking adapters.
 
Back
Top