Array
(
    [content] => 
    [params] => Array
        (
            [0] => /forum/index.php?threads/sram-scaling-isnt-dead-after-all-%E2%80%94-tsmcs-2nm-process-tech-claims-major-improvements.21414/
        )

    [addOns] => Array
        (
            [DL6/MLTP] => 13
            [Hampel/TimeZoneDebug] => 1000070
            [SV/ChangePostDate] => 2010200
            [SemiWiki/Newsletter] => 1000010
            [SemiWiki/WPMenu] => 1000010
            [SemiWiki/XPressExtend] => 1000010
            [ThemeHouse/XLink] => 1000970
            [ThemeHouse/XPress] => 1010570
            [XF] => 2021370
            [XFI] => 1050270
        )

    [wordpress] => /var/www/html
)

SRAM scaling isn't dead after all — TSMC's 2nm process tech claims major improvements

Daniel Nenni

Admin
Staff member
TSMC

(Image credit: TSMC)

SRAM scaling came to a screeching halt with the last round of new process nodes, portending a dark future where on-chip memories would become increasingly expensive. However, contrary to what we've seen in the past, SRAM scaling apparently isn't dead after all.

TSMC has announced that its N2 process technology (2nm-class) offers substantial improvements in performance, power efficiency, and area (PPA) compared to previous-generation nodes. However, there is one more thing that TSMC hasn't yet publicly discussed: considerably smaller SRAM cells and higher SRAM density (38 Mb/mm^2), which will have an impact on the costs and performance of next-generation CPUs, GPUs, and system-on-chips.

TSMC's upcoming N2 node will debut with gate-all-around (GAA) nanosheet transistors, promising a significant power reduction and a boost in performance and transistor density. Compared to the N3E fabrication technology, chips built on N2 are expected to reduce power usage by 25% to 30% (at equivalent transistor count and frequency), boost performance by 10% to 15% (with the same transistor count and power), and achieve a 15% increase in transistor density (maintaining the same speed and power).

But a noteworthy aspect of TSMC's N2 is that this production node also shrinks HD SRAM bit cell size to around 0.0175 µm^2 (enabling SRAM density of 38 Mb/mm^2), down from 0.021 µm^2 in the case of N3 and N5, according to a paper that TSMC will present at the upcoming IEDM conference this December.

This is a major breakthrough as SRAM has become particularly hard to scale in recent years. For example, TSMC's N3B (1st Generation 3nm-class technology) provided little advantage over N5 (a 5nm-class node) in this regard, while the HD SRAM bit cell size of N3E (2nd Generation 3nm process) is 0.021 µm^2 and offers no advantages in terms of SRAM scaling compared to N5. With N2, TSMC has managed to finally shrink HD SRAM bit cell size and, therefore, increase SRAM density.

TSMC's GAA nanosheet transistors appear to be the main enabler of smaller HD SRAM bit cell sizes. GAA transistors offer improved electrostatic control over the channel by completely surrounding it with the gate material, which helps reduce leakage and allows transistors to scale down in size while maintaining performance. This enables better scaling of transistor dimensions, which is crucial for reducing the size of individual components like SRAM cells. Also, GAA structures allow for more precise threshold voltage tuning, which is essential for the reliable operation of transistors overall, and SRAM cells in particular, making it possible to further shrink their sizes.

Modern CPU, GPU, and SoC designs are very SRAM-intensive as these processors rely heavily on SRAM for numerous caches to handle large amounts of data efficiently. Accessing data from memory is both performance-draining and power-intensive, making ample SRAM crucial for optimal performance. Looking ahead, demand for caches and SRAM is set to keep growing, so TSMC's achievement with SRAM cell size represents a very important one.

Earlier this year, TSMC said that N2's gate-all-around nanosheet transistors were delivering over 90% of their target performance, and yields for 256 Mb (32 MB) SRAM devices were surpassing 80% in certain batches. As of March 2024, the average yield for 256 Mb SRAM had reached approximately 70%, up significantly from around 35% in April 2023. Device performance has also shown steady improvement, with higher frequencies achieved without increasing power consumption.

 
IEDM TSMC Paper.jpg


TSMC’s New, Industry-Leading 2nm CMOS Logic Platform: In a late-news paper, TSMC researchers will unveil the world’s most advanced logic technology. It is the company’s forthcoming 2nm CMOS (i.e., N2), platform, designed for energy-efficient computing in AI, mobile, and HPC applications. It offers a 15% speed gain (or 30% power reduction) at >1.15x chip density versus the most advanced logic technology currently in production, TSMC’s own 3nm CMOS (N3) platform, introduced in late 2022.

The new N2 platform features GAA nanosheet transistors; middle-/back-end-of-line interconnects with the densest SRAM macro ever reported (~38Mb/mm2); and a holistic, system-technology co-optimized (STCO) architecture offering great design flexibility. That architecture includes a scalable copper-based redistribution layer and a flat passivation layer (for better performance, robust CPI, and seamless 3D integration); and through-silicon vias, or TSVs (for power/signal with F2F/F2B stacking). The researchers say the N2 platform is currently in risk production and scheduled for mass production in 2H’ 25. N2P (5% speed enhanced version of N2) targets to complete qualification in 2025 and mass production in 2026.

At left above is a graph showing that the new N2 high-density cells gain 14~15% speed@power vs. N3E FinFlex 2-1 fin cells across the Vdd range; a 35% power savings at higher voltage; and a 24% power savings at lower voltage.


At right above, the cross-sectional image shows that the N2 platform’s Cu redistribution layer (RDL) and passivation provide seamless integration with 3D technologies.

(Paper #2.1, “2nm Platform Technology Featuring Energy-Efficient Nanosheet Transistors and Interconnects Co-Optimized with 3DIC for AI, HPC and Mobile SoC Applications,” G. Yeap et al, TSMC)

 
This is great news!. Do we think SRAM scaling can continue for at least another few nodes via GAAFET or is this a one time benefit?
 
This is great news!. Do we think SRAM scaling can continue for at least another few nodes via GAAFET or is this a one time benefit?

We will be covering the TSMC paper at IEDM but the answer is yes. TSMC works closely with Synopsys and Apple on SRAM optimization. SRAM is critical for SoCs. There is a lot you can do with FinFlex and NanoFlex DCO optimizations.
 
View attachment 2431

TSMC’s New, Industry-Leading 2nm CMOS Logic Platform: In a late-news paper, TSMC researchers will unveil the world’s most advanced logic technology. It is the company’s forthcoming 2nm CMOS (i.e., N2), platform, designed for energy-efficient computing in AI, mobile, and HPC applications. It offers a 15% speed gain (or 30% power reduction) at >1.15x chip density versus the most advanced logic technology currently in production, TSMC’s own 3nm CMOS (N3) platform, introduced in late 2022.

The new N2 platform features GAA nanosheet transistors; middle-/back-end-of-line interconnects with the densest SRAM macro ever reported (~38Mb/mm2); and a holistic, system-technology co-optimized (STCO) architecture offering great design flexibility.
The SRAM macro shrinking is not super surprising since TSMC is always good about finding clever design tricks to get more density, but the bitcell size shrink in spite of the suspected lack of CD shrink is more so. I have only seen the bitcell number in the Tom's article. Is that bitcell value what TSMC has quoted or is it just people reverse engineering that bitcell shrink based on the SRAM macro density improvement? If that bitcell value is correct, I hope TSMC elaborates how they got there. But I suspect in true TSMC fashion they won't, and I will just have to wait until fall 2026 to look at A20 Pro teardowns :rolleyes:.
That architecture includes a scalable copper-based redistribution layer and a flat passivation layer (for better performance, robust CPI, and seamless 3D integration); and through-silicon vias, or TSVs (for power/signal with F2F/F2B stacking). The researchers say the N2 platform is currently in risk production and scheduled for mass production in 2H’ 25. N2P (5% speed enhanced version of N2) targets to complete qualification in 2025 and mass production in 2026.
I am kind of shocked that TSMC is finally doing Cu top metal (not because it is hard just because they haven't for decades). It makes me genuinely curious as to why NOW?! This is early 2000s technology. My understanding is wirebond packages need Al top metal, but not every die TSMC makes goes into a wirebond package. For all of the guys out there using BGA's you would think the small cost adder for Cu would be a no brainier with the extra performance it gives. But alas, be it Apple, NVIDIA, or AMD you would see Al top metal even on N3/N3E. So it again begs the question, if M4 couldn't justify Cu for the top metal, why is Cu now the default?
At right above, the cross-sectional image shows that the N2 platform’s Cu redistribution layer (RDL) and passivation provide seamless integration with 3D technologies.

(Paper #2.1, “2nm Platform Technology Featuring Energy-Efficient Nanosheet Transistors and Interconnects Co-Optimized with 3DIC for AI, HPC and Mobile SoC Applications,” G. Yeap et al, TSMC)

It may have taken 6 years, but huzzah, they have finally matched the 10nm Super Fin MIM cap! I was wondering if their SHP-MiM would be TSMC's first 3D MIM cap on an active die (as they showed off 3D MIM caps for passive COWS-interconnect bridges). But that would be a no. I do wonder if A14 (or whatever they end up calling it) finally integrates a 3D MIM cap?
This is great news!. Do we think SRAM scaling can continue for at least another few nodes via GAAFET or is this a one time benefit?
That is a complex question, and it deserves a complex answer! So all else being equal HNS should reduce SRAM density rather than increasing it (assuming a bit cell drive degradation is unacceptable). uHD SRAM bitcells use 1 fin transistors. By nature, a HNS Xtor is basically a finFET fin but sliced into horizontal sheets. As a result it has less drive area in the same lateral space as a one fin finFET device. The real benefit of HNS is the slightly better electrostatic control, not being forced to increase transistor width in discrete intervals, and that you can get the drive area of multiple fins in a lesser area (on account of the space between the fins now being one continuous sheet of varrying width).

In the specific case of N2 SRAM bitcell being denser than N3 and N3E, it is strongly suspected that the better channel control is not being used to scale the poly pitch (cell width). There are two opportunities I can think of to enable this scaling. One N3 has a dielectric wall that is formed in every space that is larger than one fin pitch. For uHD SRAM (and to a lesser degree the logic cells) this blows out the cell height way larger than N5 despite the aggressive pitch scaling (see fig 1). N3E got rid of this wall and went back to cutting the metal gate to isolate the devices from each other (hence why N3E is not less dense than N3 despite having a relaxed poly pitch). It wouldn't shock me if with N2 they further cleaned this up and for a lack of better words reclaimed some space they weren't able to squeeze out on N3E. Theory two is that they are moving the spacing between diffusions (ie N-P) from 2 fin pitches to 1 fin pitch. Intel showed it could be done on intel 4, and I would be shocked if TSMC wasn't far behind in reclaiming this "dead" space. Bonus round: while it wouldn't account for the whole 20% bitcell density boost, but I suspect that N2 just has the optical shrink built in rather than being an optional N2P feature. Originally TSMC said N2 would be a >10% density boost vs N3E, and that N3P would be a 4% boost over N3E (likely due to a 2% optical shrink like they did with N5->N4). Then after a little while they said N2 would be a 15% boost vs N3E. Recently they also opened up that N2P would not have any density uplift vs N2. To me that reads as N2 was going to unwind the optical shrink of N3P and that N2P would reintroduce the optional optical shrink. However N2 was healthy enough that reintroducing the optical shrink wouldn't have negatively impacted TSMC's ability to hit Apple's schedule to launch N2 products in 2026. Reintroducing the optical shrink would also pull double duty with addressing customer concerns over the small PPA uplift/cost per FET increase over N3P (assuming customers "cross shopping" with N3P are willing to take whatever the D0 hit on N3P vs N2 with optical shrink is).

As for further scaling opportunities beyond N2, the rules stay the same. More lateral shrink or go 3D. Post N2 I would assume TSMC will start scaling poly pitch again. The question is if they will continue with metal gate cut and find a different way to do their contacts in a MGC friendly way, or if they will go back to self aligned contacts with self aligned gate endcaps/dielectric wall. Since GAA is by definition less of an improvement over finFET (from the perspective of electrostatic control) the headroom for additional Lg shrink is not likely to be particularly impressive. IMO the main avenue for further SRAM and logic shrink is by reducing cell height. While yes there is the obvious avenue of shrinking feature sizes (at the cost of worse performance, cap, and wafer cost). So finding avenues to squeeze your existing transistors closer togheter seem like the better bet. The most common way to go about that is reducing the number of metal tracks needed for a given function. TSMC would get a large one time benefit if on A14 they commit to only using BSPDN going forward as they can remove the spacing that is dedicated for FS power rails from their cell heights. From there the next obvious route for improvement is for moving some of the signal lines to the BS (if my understanding of SRAM bitcell design is correct moving the word line with your VSS/VDD would provide another big one time scaling benefit). BS signaling would also become an issue of greater importance once logic makes the jump to 3D.

While doing all of this cell height scaling alot of work would need to be done to shave off every last bit of parasitic capacitance (as devices would now be closer together and you will see power AND performance degradations if things remain otherwise unchanged) and alot of work would be needed for maximizing drive per unit area. On N2 TSMC will be able to have thick nanosheets, but to make use of all of that BEOL cell height reduction the device width will also need to shrink. In theory IMEC's forksheet concept allows you to further shrink the N-P allowing for a wider nanosheet in a given area. But that comes at the cost of some of your electrostatic control. I don't think I have seen TSMC write any papers on forksheets yet, so it would seem they are trying to rush to CFET and skip over forksheet-FET. If TSMC insists on continuing to have two TD teams as development times extend from 2-3 years to 5-7 years (per Dr. Y. J. Mii); that probably makes sense. Assuming intel does hit a 2 year cadence between "full" nodes post 18A, then TSMC will need to make sure that their average full node PPA uplift is at least 1.5-2x of intel's average full node PPA uplift as TSMC formalizes their cadence of 3-4 years between full nodes (assuming their aspiration is ensuring that they consistently stay ahead of intel).


fig 1.
1731002774132.png


Edit:
Another Idea just came to me for how TSMC could get such a large SRAM bitcell shrink despite not having a feature size reduction and HNS-FET on paper requiring Xtors that are wider than a single fin device to deliver the same drive as said single fin device. My off the wall idea is that TSMC's uHD SRAM could be going from 6 transistors to 4. In theory it can be done, but my very rudimentary understanding is that those extra 2 transistors help minimize bitcell leakage and increase important figures of merit for a memory like retention time. Maybe TSMC has gotten their SRAM Vmin and transistor leakage so low that they can get away with a 4T SRAM bitcell having acceptable characteristics? However I would file this idea as less likely than the above theories. 6T bitcells have been the norm for many decades at this point because it hits a nice sweet spot for about as dense as you can go without having a bitcell that isn't a very good bitcell. For kicks and giggles, after doing some back of the envelope calculations to reverse engineer the reported values under the assumptions of 4T SRAM and a wider N2 device create values that at least seem potentially plausible.

Some VERY ROUGH napkin math that shouldn't be taken with any degree of seriousness:
N2 bitcell area * 1.2 = N3E bitcell area
1.2*(4*A2) = 6*A3
A2/A3 = 1.25
In other words an N2 uHD SRAM device could be 25% wider than a 1 fin N3E Xtor while still being allowing for an SRAM bitcell that is 0.83x the size. Of course this is a very simplistic approximation and doesn't account for any potential BEOL scaling bottlenecks.
 
Last edited:
The bit cell size is not mentioned in the press sheet, it doesn't look like it's supposed to be the inverse of the macro density.

TSMC’s New, Industry-Leading 2nm CMOS Logic Platform: In a late-news paper, TSMC researchers will unveil the world’s most advanced logic technology. It is the company’s forthcoming 2nm CMOS (i.e., N2), platform, designed for energy-efficient computing in AI, mobile, and HPC applications. It offers a 15% speed gain (or 30% power reduction) at >1.15x chip density versus the most advanced logic technology currently in production, TSMC’s own 3nm CMOS (N3) platform, introduced in late 2022.

The new N2 platform features GAA nanosheet transistors; middle-/back-end-of-line interconnects with the densest SRAM macro ever reported (~38Mb/mm2); and a holistic, system-technology co-optimized (STCO) architecture offering great design flexibility. That architecture includes a scalable copper-based redistribution layer and a flat passivation layer (for better performance, robust CPI, and seamless 3D integration); and through-silicon vias, or TSVs (for power/signal with F2F/F2B stacking). The researchers say the N2 platform is currently in risk production and scheduled for mass production in 2H’ 25. N2P (5% speed enhanced version of N2) targets to complete qualification in 2025 and mass production in 2026.
 
Last edited:
The SRAM macro shrinking is not super surprising since TSMC is always good about finding clever design tricks to get more density, but the bitcell size shrink in spite of the suspected lack of CD shrink is more so. I have only seen the bitcell number in the Tom's article. Is that bitcell value what TSMC has quoted or is it just people reverse engineering that bitcell shrink based on the SRAM macro density improvement? If that bitcell value is correct, I hope TSMC elaborates how they got there. But I suspect in true TSMC fashion they won't, and I will just have to wait until fall 2026 to look at A20 Pro teardowns :rolleyes:.

I am kind of shocked that TSMC is finally doing Cu top metal (not because it is hard just because they haven't for decades). It makes me genuinely curious as to why NOW?! This is early 2000s technology. My understanding is wirebond packages need Al top metal, but not every die TSMC makes goes into a wirebond package. For all of the guys out there using BGA's you would think the small cost adder for Cu would be a no brainier with the extra performance it gives. But alas, be it Apple, NVIDIA, or AMD you would see Al top metal even on N3/N3E. So it again begs the question, if M4 couldn't justify Cu for the top metal, why is Cu now the default?

It may have taken 6 years, but huzzah, they have finally matched the 10nm Super Fin MIM cap! I was wondering if their SHP-MiM would be TSMC's first 3D MIM cap on an active die (as they showed off 3D MIM caps for passive COWS-interconnect bridges). But that would be a no. I do wonder if A14 (or whatever they end up calling it) finally integrates a 3D MIM cap?

That is a complex question, and it deserves a complex answer! So all else being equal HNS should reduce SRAM density rather than increasing it (assuming a bit cell drive degradation is unacceptable). uHD SRAM bitcells use 1 fin transistors. By nature, a HNS Xtor is basically a finFET fin but sliced into horizontal sheets. As a result it has less drive area in the same lateral space as a one fin finFET device. The real benefit of HNS is the slightly better electrostatic control, not being forced to increase transistor width in discrete intervals, and that you can get the drive area of multiple fins in a lesser area (on account of the space between the fins now being one continuous sheet of varrying width).

In the specific case of N2 SRAM bitcell being denser than N3 and N3E, it is strongly suspected that the better channel control is not being used to scale the poly pitch (cell width). There are two opportunities I can think of to enable this scaling. One N3 has a dielectric wall that is formed in every space that is larger than one fin pitch. For uHD SRAM (and to a lesser degree the logic cells) this blows out the cell height way larger than N5 despite the aggressive pitch scaling (see fig 1). N3E got rid of this wall and went back to cutting the metal gate to isolate the devices from each other (hence why N3E is not less dense than N3 despite having a relaxed poly pitch). It wouldn't shock me if with N2 they further cleaned this up and for a lack of better words reclaimed some space they weren't able to squeeze out on N3E. Theory two is that they are moving the spacing between diffusions (ie N-P) from 2 fin pitches to 1 fin pitch. Intel showed it could be done on intel 4, and I would be shocked if TSMC wasn't far behind in reclaiming this "dead" space. Bonus round: while it wouldn't account for the whole 20% bitcell density boost, but I suspect that N2 just has the optical shrink built in rather than being an optional N2P feature. Originally TSMC said N2 would be a >10% density boost vs N3E, and that N3P would be a 4% boost over N3E (likely due to a 2% optical shrink like they did with N5->N4). Then after a little while they said N2 would be a 15% boost vs N3E. Recently they also opened up that N2P would not have any density uplift vs N2. To me that reads as N2 was going to unwind the optical shrink of N3P and that N2P would reintroduce the optional optical shrink. However N2 was healthy enough that reintroducing the optical shrink wouldn't have negatively impacted TSMC's ability to hit Apple's schedule to launch N2 products in 2026. Reintroducing the optical shrink would also pull double duty with addressing customer concerns over the small PPA uplift/cost per FET increase over N3P (assuming customers "cross shopping" with N3P are willing to take whatever the D0 hit on N3P vs N2 with optical shrink is).

As for further scaling opportunities beyond N2, the rules stay the same. More lateral shrink or go 3D. Post N2 I would assume TSMC will start scaling poly pitch again. The question is if they will continue with metal gate cut and find a different way to do their contacts in a MGC friendly way, or if they will go back to self aligned contacts with self aligned gate endcaps/dielectric wall. Since GAA is by definition less of an improvement over finFET (from the perspective of electrostatic control) the headroom for additional Lg shrink is not likely to be particularly impressive. IMO the main avenue for further SRAM and logic shrink is by reducing cell height. While yes there is the obvious avenue of shrinking feature sizes (at the cost of worse performance, cap, and wafer cost). So finding avenues to squeeze your existing transistors closer togheter seem like the better bet. The most common way to go about that is reducing the number of metal tracks needed for a given function. TSMC would get a large one time benefit if on A14 they commit to only using BSPDN going forward as they can remove the spacing that is dedicated for FS power rails from their cell heights. From there the next obvious route for improvement is for moving some of the signal lines to the BS (if my understanding of SRAM bitcell design is correct moving the word line with your VSS/VDD would provide another big one time scaling benefit). BS signaling would also become an issue of greater importance once logic makes the jump to 3D.

While doing all of this cell height scaling alot of work would need to be done to shave off every last bit of parasitic capacitance (as devices would now be closer together and you will see power AND performance degradations if things remain otherwise unchanged) and alot of work would be needed for maximizing drive per unit area. On N2 TSMC will be able to have thick nanosheets, but to make use of all of that BEOL cell height reduction the device width will also need to shrink. In theory IMEC's forksheet concept allows you to further shrink the N-P allowing for a wider nanosheet in a given area. But that comes at the cost of some of your electrostatic control. I don't think I have seen TSMC write any papers on forksheets yet, so it would seem they are trying to rush to CFET and skip over forksheet-FET. If TSMC insists on continuing to have two TD teams as development times extend from 2-3 years to 5-7 years (per Dr. Y. J. Mii); that probably makes sense. Assuming intel does hit a 2 year cadence between "full" nodes post 18A, then TSMC will need to make sure that their average full node PPA uplift is at least 1.5-2x of intel's average full node PPA uplift as TSMC formalizes their cadence of 3-4 years between full nodes (assuming their aspiration is ensuring that they consistently stay ahead of intel).


fig 1.
View attachment 2434

Edit:
Another Idea just came to me for how TSMC could get such a large SRAM bitcell shrink despite not having a feature size reduction and HNS-FET on paper requiring Xtors that are wider than a single fin device to deliver the same drive as said single fin device. My off the wall idea is that TSMC's uHD SRAM could be going from 6 transistors to 4. In theory it can be done, but my very rudimentary understanding is that those extra 2 transistors help minimize bitcell leakage and increase important figures of merit for a memory like retention time. Maybe TSMC has gotten their SRAM Vmin and transistor leakage so low that they can get away with a 4T SRAM bitcell having acceptable characteristics? However I would file this idea as less likely than the above theories. 6T bitcells have been the norm for many decades at this point because it hits a nice sweet spot for about as dense as you can go without having a bitcell that isn't a very good bitcell. For kicks and giggles, after doing some back of the envelope calculations to reverse engineer the reported values under the assumptions of 4T SRAM and a wider N2 device create values that at least seem potentially plausible.

Some VERY ROUGH napkin math that shouldn't be taken with any degree of seriousness:
N2 bitcell area * 1.2 = N3E bitcell area
1.2*(4*A2) = 6*A3
A2/A3 = 1.25
In other words an N2 uHD SRAM device could be 25% wider than a 1 fin N3E Xtor while still being allowing for an SRAM bitcell that is 0.83x the size. Of course this is a very simplistic approximation and doesn't account for any potential BEOL scaling bottlenecks.
Must N2 be a pitch shrink from N3?
 
Last edited:
N2 must be a pitch shrink from N3?
You never know but I would place it in the unlikely category. The chip level density improvement is too anemic for it to be likely. If there was a pitch shrink (and the resultant cost and cycle time increase) plus all of the extra processing steps for GAA then the cost per wafer cost increase would be so much larger than the density increase that I suspect N2 would be too cost prohibitive for any application beyond data center chips when compared to N3P/X.

There is also the historical angle to consider. 16FF was TSMC'S first finFET process and it completely reused the 20nm BEOL to help minimize the risk to their program. As you could have expected, the density improvement for 16FF was minimal even by today's more relaxed standards of what a full node density uplift is. All of the indications that TSMC has been giving point to N2 following the same playbook as they used on 16FF. FWIW IanD also has asserted that N2 and N3 use the same BEOL/design rules.

There is also the practicality angle. I would assume reusing the BEOL from N3 makes it easier for designers to get started on N2. That way all of their focus can be on getting the most out of the better tunability and worse PMOS performance on GAA nodes. Then once they get a handle on that they can dip their toes in on A16 BSPDNs without having to fully commit (as A16 has IP compatibility with N2 and allows for a partial BSP implementation).
 
You never know but I would place it in the unlikely category. The chip level density improvement is too anemic for it to be likely. If there was a pitch shrink (and the resultant cost and cycle time increase) plus all of the extra processing steps for GAA then the cost per wafer cost increase would be so much larger than the density increase that I suspect N2 would be too cost prohibitive for any application beyond data center chips when compared to N3P/X.

There is also the historical angle to consider. 16FF was TSMC'S first finFET process and it completely reused the 20nm BEOL to help minimize the risk to their program. As you could have expected, the density improvement for 16FF was minimal even by today's more relaxed standards of what a full node density uplift is. All of the indications that TSMC has been giving point to N2 following the same playbook as they used on 16FF. FWIW IanD also has asserted that N2 and N3 use the same BEOL/design rules.

There is also the practicality angle. I would assume reusing the BEOL from N3 makes it easier for designers to get started on N2. That way all of their focus can be on getting the most out of the better tunability and worse PMOS performance on GAA nodes. Then once they get a handle on that they can dip their toes in on A16 BSPDNs without having to fully commit (as A16 has IP compatibility with N2 and allows for a partial BSP implementation).
The pitch shrink anemia is partly because they still use an SRAM circuit that made sense in the planar era and seem never to have changed that. The planar design depended upon variable strength transistors and when translated into FinFET ended up with a design that was 2 rails high and needed a physical cut between cells for one of the channels (to modulate transistor strength) resulting in a design with empty space at the heart of the cell, and difficult to shrink in the sameway the rest of the logic cells do.

But if you rethink the circuit to use a normal logic transmission gate instead of the lagacy approach of two weak transmission gates, the need to modulate the transistor strength goes away and the layout fits in a one-rail design with no cut and no unused space. This is especially nice in ribbon/GAA FET design where the SRAM cell is just like the normal logic cells and will scale the same way.
 
The pitch shrink anemia is partly because they still use an SRAM circuit that made sense in the planar era and seem never to have changed that. The planar design depended upon variable strength transistors and when translated into FinFET ended up with a design that was 2 rails high and needed a physical cut between cells for one of the channels (to modulate transistor strength) resulting in a design with empty space at the heart of the cell, and difficult to shrink in the sameway the rest of the logic cells do.

But if you rethink the circuit to use a normal logic transmission gate instead of the lagacy approach of two weak transmission gates, the need to have some weaker transistors goes away and the layout fits in a one-rail design with no channel cut and no unused space. This is especially nice in ribbon/GAA FET design where the SRAM cell is just like the normal logic cells and will scale the same way.
 
Back
Top