100X800 Banner (1)

2022 Semiconductor Supercycle and 2023 Crash Scenario

2022 Semiconductor Supercycle and 2023 Crash Scenario
by Daniel Nenni on 08-26-2022 at 6:00 am

2021 Semiconductor Forecast Summary

Charles Shi, semiconductor analyst at Needham & Company, an US-based investment bank and asset management firm, hosted an expert call on semiconductor cycles with Malcolm Penn, Founder and CEO of Future Horizons on 18 August 2022, with over 100 financial analysts in attendance.  The following bulletin is a summary of the discussions.

Charles:  In January 2021, you forecasted 2021 semiconductor growth was going to be 18 percent as the base case and 24 percent as the bull case. Around the same time, Gartner, your former employer, was forecasting 11.6 percent of growth. Now we know the actual number was 26.3 percent and you were probably the closet. But at the same time when you gave people that strong double-digit forecast, here is what you said: “Enjoy the super cycle for now, because it will crash in 2023!”

Being a financial analyst myself, I always find it both refreshing and frightening to hear the level of conviction in your statement about the future. Also “crash” would not be the word I use in my research reports, but we will touch upon this later. Having said that, you were spot on about the industry downturn that is just getting started, even when your statement was really made 18 months ago. So now, my first question: What’s your forecasting methodology that gives you this level of confidence? And what are the key metrics that you look at?

Malcolm:  Yes, I’m actually very proud of this achievement, but both forecasts were very scary and I was publicly derided for being ‘ever-optimistic’ on the impending double-digit super-cycle (the glory days are over) and ‘off my rocker’ re the 2023 crash (the cycles were dead, it’s different this time).

I spent a lot of sleepless nights checking and double checking my assumptions and scenarios but kept coming back to the same answers and conclusions.  In the end I decided … “If that’s what my research shows then it’s my professional responsibility to publish and be damned”.

I’ve always believed in the need for integrity and accountability so one of the tenants of my forecasting methodology has always been transparency … to explain the ‘how and why’ behind the forecasts and then be judged by what actually transpired.

I also see it my role to provide timely and insightful information at the start of each year, not when the outcome’s painfully obvious, also not to follow the fashion or consensus just because these are the safe options … anyone can do this, but it’s doing the industry a disservice.  Being independent helps … I don’t have a corporate position to defend, which means I can take a contrarian view if needed and I don’t need ‘wiggle room’ … I can live or die on my own integrity.

A third fundamental tenant is not to keep constantly changing the goal posts … a forecast, if properly considered, ought to be good for the whole year with just interim ‘progress report’ updates on the way.  The maths will affect the final forecast number but never the assumptions.

I also don’t believe in ‘models’ or ‘secret sauce’ rather a ‘balance of probabilities’ methodology based on data analysis and experience.  I see pictures not numbers, a bit like a jigsaw puzzle but with half the pieces missing … It won’t show all the detail but I can still see the picture.

I look at four key data trends, the global economy (which sets the overall appetite for demand), unit sales (how many parts are sold), capacity (how many parts could be made) and ASPs (what price they were sold at).  There’s lots of good economic data from e.g. the IMF (my prime ‘go-to’ source); WSTS provides good data on unit sales and ASPs and SEMI provides good data on semiconductor equipment capital expenditure.  Unfortunately, since 2011 capacity data is no longer published by SICAS, but SEMI’s CapEx spend data is a very good proxy to it.

I liken these metrics to the biblical four horsemen of the apocalypse because, whilst they are all inter-related (but independent), their impact on the market could be either off-setting or supporting.  For example, over-investment and excess capacity in a strong economy will be quickly absorbed whereas excess capacity in a weak economy and slowing demand will have a stronger and longer-lasting impact.

Charles:  Now I want to double click on each of the metrics that you look at. First, the economy. You said semiconductor industry can have a bad year in a good economy but not a good year in a bad economy. Is it though? Maybe this is a bit U.S.-centric, but in 2020, the U.S. GDP declined by 3.5 percent but it was a healthy growth year for the semiconductor industry. Is that the exception to the rule?

Malcolm:  That’s a really good question.  The GDP took a hit in 2020 because the world went into lockdown in March of that year, but it was more like someone hit the pause button on the world, not a ‘real’ economic downturn, and we hadn’t all suddenly gone back to living in caves … we still had all of our creature comforts, governments all over the world were pumping cash into their economies and firms were furloughing staff or paying their employees to work from home.

My original forecast for 2020 was for 10 percent growth, against a tight capacity background.  In the end it grew 6.6 percent, slowed down by the Q2 economic downturn, but it quickly rebounded driven by a spurt in home office and home entertainment demand.

Charles:  So we talked about the economy, Covid and 2020. What is the chip shortage issue all about that were in the headlines through 2021? What caused that? More specifically, how did you manage to correctly call 2021 a strong double-digit growth year as early as October 2020?

Malcolm:  I had been cautioning for some time that the chip industry was too finely tuned, running on empty with no slack in the system, meaning any slight increase in demand would very quickly escalate into shortage.  A cyclical market upturn was always inevitable and would have hit home in 2020 had Covid not initially reined back demand in Q2.

That push back was short-lived though and by October wafer fab capacity was maxed out against a historically low level of CapEx investment.  With inventory levels low, shortages were already starting to bite and lead-times were extending.  Three out of four of my matrices were fundamentally sound with only the economic outlook fragile, hostage to the ongoing Covid uncertainty, but with a lot of fiscal policy support to do ‘whatever it takes’ to avoid a recession.

With no tight capacity relief possible before 2022 at the earliest, based on the industry momentum and the systemic supply-side tightness, it was impossible for 2021 to show anything less than strong double-digit growth, hence my 18 percent number, with a 24 percent upside.

Charles:  Second, let me ask you about the unit growth rate. You said long-term unit growth has always been 8 percent. Right now people are talking about electric vehicle and autonomous driving, and expect the semiconductor content in electric vehicles are going to six times more than internal combustion engine. How is this not bending the long-term growth curve to the upside?

Malcolm:  The industry is currently shipping around 250 billion semiconductor devices every quarter – one trillion units per year – and these parts both increase their penetration in existing markets and constantly create new markets, such as the ones you mentioned, markets that were either impossible or too expensive using the previous levels of technology.  That’s the real economic impact of Moore’s Law.

Without these new markets, the long-term growth in the industry would falter, but all these do is keep the industry on its 8 percent growth track; offsetting other applications that are slowing down or falling by the wayside, they are not altering the overall status quo.  The long-term average IC unit growth rate has been averaging 8 percent since records are available so I believe it’s a very reliable statistic.

Charles:  Next I want to ask you about capacity. We may have just touched upon it a bit. Through this semiconductor super cycle, you hear foundries claiming that they are sold out through 2022, 2023. Some even stick their neck out and said that supply is no longer unlimited and supply will chase demand going forward. What’s your response to these statements, as you just said that the supply/demand balance has always been bumping between under- and over-supply?

Malcolm:  Balancing supply with demand has always been the industry’s Achille’s heel, driven by the inherent mismatch between short term volatility in demand and the long (one to two-year) lead time for new capacity exacerbated by the long (four month typically) production cycle time.

In the past, the industry used to forecast what capacity was needed two years out, an impossible task to get right, resulting in systemic over capacity but that all changed when TSMC elected to stop building this speculative capacity.  Instead, it would only build new capacity against firm customer commitment.

Right now, given the shortages, supply is now no longer unlimited but as new capacity comes on stream, that status quo will flip.  It’s fundamentally inherent in the economics … economists call it the hog cycle.  That’s never going to change until the industry as a whole takes collective responsibility for the supply chain, which inevitable means the end customers taking more fiscal responsibility which in turn is going to impact their bottom line.  There is absolutely no indication any end customer, their shareholders, or the financial community, has any stomach for that.

Charles:  Lastly, before we move onto forecasting, I want to ask you about ASP. You said the long-term IC ASP growth is zero. But through 2021 and 2022, we saw foundries raising prices by as much as 20 percent, and even OSATs, the packaging houses, are also raising prices. I’ll give you an example… Even wire bonding, the most legacy packaging service, the price of it is now 20 percent higher than where it was in 2020. Some say those ASP increases are here to stay. What do you say?

Malcolm:  This is one of the more controversial and counter-intuitive industry statistics but it’s the culmination of a lot of associated facts.  For example, the average revenue per square inch of silicon processed has been constant since the beginning of time; likewise, the average number of ICs per wafer area.  It’s the corollary of Moore’s Law … for every new more complex IC we design, another existing one gets cost reduced or shrunk.  Moore’s Law enables us to either get more bang for the same buck or pay less bucks for the same bang.

If ASPs are 20 percent higher now than in 2020, you can virtually stake your life on them being 20 percent lower down the road.  Interestingly, and something that’s never grabbed the world’s attention, Gordon Moore had a second law, namely “The long-term IC ASP trend is $1.”  My data analysis has shown this observation both true and equally prophetic, but largely denied.

Charles: Now let’s look at the upcoming downturn more closely. Malcolm, I’ll be honest with you. Wally Rhines, the former CEO of Mentor Graphics, Daniel Nenni, Co-Founder of SemiWiki, and I met got together at an industry event here in California back in May and we talked about your forecast of a semiconductor crash as early as Q4 2022. Back then, most of the semiconductor companies have reported their Q1 results and guided their Q2. The numbers were not bad. Three of us were talking behind your back that hey Malcolm might be too pessimistic. But hey, over the last month, we’ve got Micron, Intel, AMD, Nvidia, and Micron again, massively reduced their outlook for the second half of this year. Earlier on this call, I said you warned people that “enjoy the super cycle, but it’ll crash”. Where numbers are trending do look like a “crash”. So allow me to ask, is the worst news in? Are we going to have a soft landing this time?

Malcolm:  The short answer to both questions is ‘no’ and ‘no’.  June’s WSTS data was just the start of the bad news, I fully expect Q3 data to be worse, it’s inevitable, and the forward-looking company statements are already paving the ground.  Once the correction starts, it quickly escalates into a downward spiral.  Orders start to soften, some get delayed or even cancelled, lead times start to shrink and bloated inventories get purged softening fab demand even more and so on and so forth.  Unit shipments inevitably decline.

Faced with depleted backlogs and spare fab capacity, ‘bomb the price … any order’s a good order’ is the name of the day, meaning ASPs quickly plummet soon after.  The combination of reduced unit demand and falling ASPs is a lethal combination for negative sales growth.

Charles:  So you don’t think there will be a soft landing. What’s your forecast for 2023?

Malcolm:  Well, let’s just say if there is, it will be an industry first.  Once you get into the situation of severe product shortages and 50-week long lead times, customers have no option but to increase buffer stocks and inventory over and above what they actually need.  Stockpiling is inevitable and orders, as seen by the supplier, are inflated by this excess demand.

Once supply catches up with this now inflated level of demand, products become more available, lead times start to shrink and customers pare back their orders and start to liquidate their bloated inventories.  This frees up even more supply availability, reducing lead times further, triggering even more inventory liquidation and paring back of orders.  This positive feedback fuels a downward spiral which only stops once lead times reach a sustainably low level (typically 4-6 weeks) and customer inventory levels have been purged.

Simultaneous with this slowdown in unit shipments, ASPs fall, hence the plummeting nature of sales value.  It’s a deluge, soft landings just don’t happen.  This process started in June 2022 and it typically lasts four quarters.  Our forecast was based on negative quarter on quarter growth for the second half of 2022 and first half of 2023, with the market bottoming in Q2-2023 and recovering from Q3-2023 onwards.  Rolling these quarterly growth numbers forward led to a 22 percent year-on-year decline, 2023 vs. 2022.

Charles:  Wow 22 percent decline in 2023. How do we go from here to a 22 percent decline in 2023? Maybe let me ask from a different angle… You said there is no soft landing. Is the path forward more like a V, or a U, or a K?

Malcolm:  As mentioned before, the market collapse typically last four quarters and the ensuing recover spans a three to five-year period, so it’s more the shape of a hockey stick on its side.  There is nothing to indicate why this current market correction and recovery will be different.

Charles:  Based on your current forecast, on a quarterly basis, when will the rate of decline be the greatest? And when will we start to recover?

Malcolm:  When I made the forecast back in January 2022, it was based on a Q3-2022 decline of 3.2 percent followed by 10 percent declines in Q4 2022 and Q1-2023, both being, in addition to the correction, seasonally weak quarters, before bottoming out in Q2-2023, with a further 8 percent decline, followed by 2.5 percent growth in Q3 and a 2.5 percent seasonal decline in Q4.

Granted, a 22 percent decline sounds pretty bad but it’s really just a compensation adjustment from 2021’s 26.6 percent growth.  Also, when you look historically back at the at the magnitude of previous swings, from the end of the boom to the start of the bust, you see swings ranging from 69 to 21 percentage points, so going from plus 6 or even 10 percent growth in 2022 to minus 22 percent in 2023 is only a swing of 28-32 percentage points, at the historical low end of the range.  Corrections, when they do happen are always very steep.

Charles:  I believe there are good number of people on the call who want to push back on your very grim view about the road ahead. Let me ask my follow-up questions first.  My first follow-up… Many foundries this time have signed LTAs with their customers. Could LTAs smooth the curve a bit there?

Malcolm:  Well, that’s certainly the hope, but I think their impact will be disappointing, a negotiating tool at best.  They can never compensate for the fact the customers no longer need the product on order so, at best, suppliers may manage to enforce a price penalty and offset the early impact on ASPs but this will only kick the can down the road and drag out the downturn.  You cannot force customers to take product they do not need so the impact on unit shipments will be negligible and fab capacity will be underutilized which will trigger an ASP decline.

Firms will also need to think about the long-term marketing impact on their supplier-customer relationship.  Forcing their customers too hard now might just result in a lost customer tomorrow.

Charles:  My second follow-up… you touched upon this a bit… the equipment demand remains very high and equipment supply is still very tight… we hear lead time for equipment is now 12-18 months. Now we are seeing some cancellation and pushouts from the memory companies, but not so much on the non-memory side. This elevated level of capacity investment… what do you think is going to do to the upcoming downcycle if nobody cancels equipment orders?

Malcolm:  No firm ever sees the downturn coming yet, after it happens, it’s always seen has having been inevitable!  Push outs and cancellations will eventually follow through, but there’s always a ‘two quarter lag’, up and down.  Memory is always the fastest to react, but the other sectors will also soon feel the pain, except at the extreme leading-edge where capacity is always tight.

If firms are part way through a capacity expansion, there is always a point of no return whereby they have no choice but to finish what’s been started, a bit like V1 speed for an aircraft, even though it will exacerbate the excess capacity situation.  The spend thus continues after the market has collapsed but eventually capacity expansion plans are adjusted to the new demand reality and which means push outs, cancellations and delays, the bull-whip effect is inevitable.

Charles:  My last follow-up… CHIPS Act. The CHIPS Act is now signed into law. What’s your general thought on this government subsidized on-shoring? The funding is going to get allocated and get into capacity investments. What’s your view on that?

Malcolm:  This is one area where my European location really comes to the fore and my first piece of advice is to look carefully at the European experience … copy the bits that have worked well and learn from the areas that were less successful.

It’s good that governments around the world have woken up, at least temporarily, to the strategic nature of semiconductors but in general, these initiatives fail to grasp the cause of the problem, namely why did the off-shoring occur in the first place?

What inevitably happens in these situations is the proposed programmes start with where we currently are and focus on essentially ‘band aid’ solutions … patching up the current system, rather than starting the other way around, focusing on what do we want the industry to be like in 20 or so years’ time in the face of all the financial and geo-political issues that need to be addressed.

None of the current industry and political thinking or initiatives even start to address these issues and, whilst government investment in the cash-hungry chip industry is to be welcomed, it will have only limited long-term success unless the reasons chip firms off-shored production in the first place are addressed, especially those once IDM firms who elected to go fabless.  A sizable amount of offshored chip production is from firms who have no desire whatsoever to own and operate fabs … anywhere.

In addition, none of these programmes have the support of the end market customer, they are all bottom-up initiatives.  From Apple down, there is not a single end-customer championing any on-shoring initiative, nor any customer willing to pay a potential price premium for on-shored product.

Industry needs to collectively address this problem and then tell its respective governments what supportive financial and policy incentives are needed to encourage this top-down demand-driven industry market pull.

For a monthly updates, see Future Horizons’ Semiconductor Monthly Update Report:

https://www.futurehorizons.com/page/137/

For the latest industry outlook, join Future Horizons’ 13 September 2022 Webinar

https://us02web.zoom.us/webinar/register/9216604665338/WN_4vgIAPA9TW6co4Wl-z5zRw

Also read:

EDA Product Mix Changes as Hardware-Assisted Verification Gains Momentum

WEBINAR: A Revolution in Prototyping and Emulation

An EDA AI Master Class by Synopsys CEO Aart de Geus


Automotive Semiconductor Shortage Over?

Automotive Semiconductor Shortage Over?
by Bill Jewell on 08-25-2022 at 5:00 pm

Global Light Vehicle Production 2022

Several signs point to an easing of the shortages of semiconductors for automotive applications. However, the production of light vehicles will likely remain below full potential through at least 2023. LMC Automotive’s July forecast of light vehicle production called for 81.7 million units in 2022, up 2% from 2021. LMC’s January forecast was for 13% unit growth in 2021, over 4 million more units than the current forecast. The July forecast calls for production growth of 5% in 2023 and 7% in 2024. LMC’s April projection for 2023 and 2024 was 4 million units higher each year than the July projection. 2024 production of 91.7 million units should still be below the pre-pandemic level of 94 million units in 2018, five years earlier.

The key reasons for the shortages of automotive semiconductors are:

  • Automotive manufacturers cut back on semiconductor orders severely at the beginning of the COVID-19 pandemic in early 2020. The auto companies were fearful of being stuck with excess inventories of cars if demand fell significantly due to the pandemic. When the automakers tried to increase orders, they had lost their place in line and were behind other industries such as PCs and smartphones.
  • Many automakers used a just-in-time ordering system to avoid excess inventories. This left them with almost no buffer inventories. Also, most semiconductors used in automotive are bought by the companies supplying the systems (engine controls, dashboard electronics, etc.) rather than the automakers, leading to a more complex supply chain.
  • Semiconductors used in automobiles have a long design-in cycle and must be qualification standards. Thus, it is difficult for an automaker to change suppliers in the short term.
  • Due to the long design cycles and long product life, semiconductors used in automotive applications use older process nodes than most other applications. As shown in the table below, McKinsey estimated 72% of the semiconductor wafers for automotive in 2021 used process nodes of 90 nanometer or greater compared to 52% of all applications. Only 6% of automotive demand is for 14 nanometer and less process nodes, compared to 21% for all applications. Semiconductor manufacturers concentrate their capital spending on the more advanced process nodes, with only modest expansion of capacity in older nodes. TSMC, the dominant wafer foundry, makes 65% of its revenue from advanced process nodes and only 12% of its revenue from nodes of 90 nanometers or greater. Only 5% TSMC’s revenue is from automotive, compared to 38% from smartphones. Thus, automakers are generally a lower priority for foundries.

Given all the above factors, it will take time for all the supply issues to be resolved. Recent comments from major automakers reveal mixed trends in resolving the semiconductor shortages.

  • Toyota – shortages at least through 3Q 2022
  • Volkswagen – shortages easing
  • Hyundai – shortages easing
  • General Motors – shortage impact into 2023
  • Stellantis – shortages through the second half of 2022
  • Honda – outlook uncertain due to shortages
  • Nissan – recovery in next few months
  • Ford – shortages still an issue
  • Mercedes-Benz – no significant supply issues
  • BMW – no production delays due to shortages
  • Volvo – back at full supply
  • Bosch (parts supplier) – shortages into 2023

The table below list the major suppliers of semiconductors to the automotive market. The top ten suppliers accounted for 46% of the total automotive semiconductor market of $69 billion in 2021 (according to WSTS). Automotive is a significant portion of total semiconductor revenues for each of these companies, ranging from 17% to 50%.

The five largest suppliers of automotive semiconductors also had varied outlooks on the shortages in their recent financial reports for 2Q 2022:

  • Infineon Technologies – gradual easing of shortages throughout 2022
  • NXP Semiconductors – demand will still exceed supply in 3Q 2022
  • Renesas Electronics – inventory back to planned levels at end of 2Q 2022.
  • Texas Instruments – inventories still below desired levels
  • STMicroelectronics – capacity sold out through 2023

Shortages of automotive semiconductors will likely remain through at least the year 2023. Although a few automakers indicate they are back at full production, most report continuing shortages. The shortages will prevent automakers from producing enough vehicles to meet demand in 2022 and 2023, resulting in continued high prices for most vehicles.

Automakers and semiconductor suppliers are working to try to prevent such severe shortages in the future. Automakers are adjusting their just-in-time inventory models. Automakers are also working more closely with semiconductor suppliers to communicate their short-term and long-term needs. Semiconductors will become even more crucial to automakers as trends toward electric vehicles and driver-assist technologies continue.

Also read:

Electronics is Slowing

Semiconductors Weakening in 2022

Semiconductor CapEx Warning


Enhanced X-NAND flash memory architecture promises faster, denser memories

Enhanced X-NAND flash memory architecture promises faster, denser memories
by Dave Bursky on 08-25-2022 at 10:00 am

x nand timing vs slc

Although the high-performance X-NAND memory cell and architecture were first introduced in 2020 by Neo Semiconductor, designers at Neo haven’t rested on that accomplishment and recently updated the cell and the architecture in a second-generation implementation to achieve 20X the performance of conventional quad-level-cell (QLC) 3D NAND memories. The Gen2 X-NAND, unveiled at this month’s Flash Memory Summit, achieves that improvement through an enhanced architecture that allows the 3D NAND flash programming (i.e., data writes) to occur in parallel using fewer planes. Able to deliver a 2X performance improvement over the first generation X-NAND technology, the Gen2 architecture remains compatible with current manufacturing technologies and processes thus giving adopters of the technology a competitive advantage over existing 3D NAND flash memory products.

According to Andy Hsu, Neo’s CEO, the Gen2 technology incorporates zero-impact architectural and design changes that do not increase manufacturing costs while offering improvements in throughput and reductions in latency. The X-NAND technology can be implemented with all flash cell structures – SLC, MLC, TLC, QLC, and PLC – while delivering higher performance at comparable manufacturing costs. Neo is currently looking to partner with memory manufacturers who will license the X-NAND technology and then design and manufacture the high-performance memory chips.

NAND-flash memories based on the QLC cell and 3D stacking have been widely adopted in many applications thanks to the high storage capacities possible and their relatively low cost per bit. The one drawback is their slow write speed. The X-NAND technology overcomes that limitation and improves the QLC NAND read and write speeds threefold and the sequential read/write throughput by 15 to 30X. Further improvements in the Gen2 technology let memories deliver SLC-like (single-level cell) performance but with the higher capacity and lower cost of QLC implementations. Transfer rates of up to 3.2 Gbytes/s are possible with the Gen2 X-NAND technology.

In Neo’s Gen1 design the company employs a unique SLC/QLC parallel programming scheme that allows the data to be programmed to QLC pages at SLC speed for the entire memory capacity. This also solves the conventional NAND’s SLC cache-full problem (see the figure). When the conventional SLC cache is full for traditional NAND, the data will be directly written to QLC cells, and the write speed will be reduced to below 12%. X-NAND solves this problem, explains Hsu, and it also provides an excellent solution for heavy-write systems such as data centers with NAS systems.

Furthermore, the X-NAND’s 16-64 plane architecture provides parallelism at the chip level. When compared to a conventional NAND that uses 2 to 4 planes, one X-NAND chip can provide the same parallelism of 4 to 16 NAND chips. This would allow small-form-factor packaging such as in M.2 and eMMC memory modules. Additionally, X-NAND’s bit-line capacitance is only 1/4 – 1/16 that of a conventional NAND and thus the bit line’s power consumption for read and write operations can be reduced to about 1/4 – 1/16 (or by about 25% – 90%). This significantly increases the battery life for smartphones, tablets, and IoT devices.

www.neosemic.com

Also read:

WEBINAR: A Revolution in Prototyping and Emulation

ARC Processor Summit 2022 Your embedded edge starts here!

WEBINAR: Design and Verify State-of-the-Art RFICs using Synopsys / Ansys Custom Design Flow


Getting Ahead with Semiconductor Manufacturing Equipment and Related Plasma Reactors

Getting Ahead with Semiconductor Manufacturing Equipment and Related Plasma Reactors
by Kalar Rajendiran on 08-25-2022 at 6:00 am

Figure 1 Dry Etching Process Classification

Advanced semiconductor fabrication technology is what makes it possible to pack more and more transistors into a sq.mm of a wafer. The rapidly increasing demand for advanced-process-based chips has created huge market opportunities for semiconductor manufacturing equipment vendors. According to SEMI, worldwide sales of semiconductor manufacturing equipment in 2021 rose 44% to an all-time record of $102.6 billion. While the opportunities are big, delivering cost-effective equipment optimized for mass production involves overcoming a number of challenges.

In June 2022, SemiWiki published an article titled “Leveraging Simulation to Accelerate the Design of Plasma Reactors for Semiconductor Etching Processes.” That article was a brief introduction, leading up to a webinar on simulation techniques to accelerate design of plasma reactors for semiconductor etching processes. The webinar was presented by Richard Cousin of Dassault Systèmes and is now available on-demand for viewing. This article covers some salient points from that webinar.

Benefits of Simulation When Designing Plasma Reactors

Simulations help understand how a device will behave even before the device is actually manufactured. Whether the device being designed is a chip or a plasma reactor, the savings in terms of time and money is worth investing in a good simulation tool.

In the case of plasma reactors, the design has to balance many different parameters to accommodate various requirements and plasma characteristics. The plasma characteristics include the density profiles, the ionization rate, the effect of the pressure and the type of gas used as well as the influence of the geometry to prevent damages. Simulation can, for example, help experiment with different numerical and physical settings for a plasma reactor to improve uniformity of ion density profiles. It can also help perform thermal coupling analysis with the aim of eliminating vulnerabilities to reactor damage. It can also allow experimenting with different types of gases, pressures and power levels.

The benefits of simulation fall into three categories.

  • Predicting and explaining experimental results, particularly when no diagnostics are available in advance
  • Reducing cost of development by optimizing performance and reliability before actually manufacturing the plasma reactor (or modifying an existing reactor)
  • Accelerating device validation and decision making on right manufacturing processes

Focus of Dassault Systèmes’ SIMULIA Tools

Depending on the pressure levels deployed, dry etching process would be a physical process or a chemical process. Refer to the Figure 1 below for the dry etching classification spectrum and the focus range of the spectrum for SIMULIA tools.

Figure 1: Dry Etching Process Classification as a function of the Pressure level

The anisotropic etching process is well suited for nanoscale size features. Also, more physical parameters can be controlled to characterize the plasma, such as the input power as well as the pressure of the neutral gas which controls the plasma density.

Simulation Techniques Deployed by SIMULIA Tools

The SIMULIA tools help analyze not only the steady-state model after the plasma is formed but also the transient model on how/when the plasma is formed. SIMULIA can use a microscopic approach or a macroscopic approach for the simulations.

Microscopic Approach

Under the microscopic approach, it uses a time-domain kinetic approach with a Poisson based Particle-In-Cell code to analyze space charge interactions. Several particle interactions are taken into account in a global Monte-Carlo Collision (MCC) type model. The ionization of the neutral gas, its excitation and the elastic collisions are considered simultaneously to compute the plasma kinetics.

Macroscopic Approach

Under the macroscopic approach, SIMULIA tools treat the plasma as bulk for RF plasma analysis and matching network optimization. Both linear and non-linear Drude dispersion models are available as options. This is well suited for use when designing Capacitive Coupled Plasma (CCP) type of reactors. With the application of a Bias Magnetic field, an Electric Gyrotropic Dispersion model is available for use when designing Inductive Coupled Plasma (ICP) type of reactors, for example.

SIMULIA Tool Use Cases

The SIMULIA tool can be used to simulate various types of plasma reactors. The following are three examples that were presented.

The DC Magnetron Sputtering Reactor Example

This design setup is as follows:

Pressure range: 1 to 5 mTorr

Target voltage: -400V

Target materials: Al, Cu or Ti

Target thickness: 1 mm to 3 mm

Goal: Estimating the target erosion profile for understanding long term sputtering efficiency

Figure 2 below shows the high correlation between the simulated results and measured results, as it relates to target erosion profile prediction.

Figure 2: Very good agreement between the predicted results and the experimental data for the Target erosion profile in a DC Magnetron Sputtering Device example

GEC CCP Reactor (Capacitive Coupled Plasma) Example

The design setup is as follows:

Pressure: 200 mTorr

Temperature: 300K

Gas: Argon neutral gas

RF-Voltage: 60V Peak-to-Peak

Discharge over 13.56MHz RF-Voltage

Goal: Control of plasma homogeneity

Figure 3 below shows the ion density profile as predicted by the simulator. This compares very well to density profiles presented in the technical literature for CCP reactors.

Figure 3: Well-known GEC Cell CCP Reactor. Ion Density Profile in good agreement with the published results for this Device 

The VHF ICP Reactor (Inductive Coupled Plasma) Example

The design setup is as follows:

Pressure: 30 mTorr

Gas: Argon neutral gas

Input power: 300 W

EM-Field distribution: 13.56MHz

Goal: Characterize the physical parameters, understand the physical principles and identify potential issues and damages

Figure 4 below shows the plasma homogeneity issue and potential for damage from energy enhancement.

Figure 4: Electron Energy Enhancement localized which affects the plasma homogeneity and could lead to potential damage of the ICP Reactor

 


Automating and Optimizing an ADC with Layout Generators

Automating and Optimizing an ADC with Layout Generators
by Daniel Payne on 08-24-2022 at 10:00 am

Layout Geneator tool flow min

I first got involved with layout generators back in 1982 while at Intel, and about 10% of a GPU was automatically generated using some code that I wrote. It was an easy task for one engineer to complete, because the circuits were digital, and no optimization was required. In an IEEE paper from the 2022 18th International Conference on Synthesis, Modeling, Analysis and Simulation Methods and Applications to Circuit Design, there was an article authored by experts from Fraunhofer IIS/EAS, MunEDA, IMST Gmbh and Dresden University of Technology. I’ll share what I learned from their article, “A Multi-Level Analog IC Design Flow for Fast Performance Estimation Using Template-based Layout Generators and Structural Models.”

Analog designs like an ADC require that you start with a transistor-level schematic, do an initial IC layout, extract the parasitics, simulate, then measure the performance to compare against the specifications. This manual process is well understood, yet it requires iterations that can take weeks to complete, so there has to be a better approach. In the paper they describe a more automated approach, built upon three combined techniques:

  • Template-based generator
  • Parasitic estimation from object-oriented templates
  • Fast model-based simulation

The following diagram shows the interaction and flow between optimization, performance estimation and layout generators for an ADC:

Generator Template, Model, Performance Estimation

There’s a SystemC AMS model of the pipeline ADC circuit, and the parameters define things like the number of device rows, while the model defines the behavior of non-ideal capacitors and OpAmp offsets. The flow is designed to be executed, and when optimized it reaches an acceptable performance criteria.

The inner loop makes an estimate of the layout parasitics in about 5 seconds, and then the ADC is optimized and a layout generated in about 1 minute. The layout generator uses the best parameter set, and generates the capacitor structures. Layout capacitance values for device and wires were pre-characterized to enable fast estimates in the template approach. The optimization step is using estimated parasitics, not extracted parasitics, saving time.

A SystemC AMS model of a pipeline ADC has both behavioral and structural details, so that engineers can trade off accuracy versus runtime. Using an analytical model enables a thousand runs in just a few minutes. The outer loops adds the ADC model, and that run takes about 50 seconds to complete.

This generator template approach even estimates layout parasitics, capacitor variation and device mismatch. Both global and local process variations were taken into account.

Results

Starting from transistor-level schematics for the ADC, a parameterized model was built. Having a model enabled fast simulation and optimization, with the goals of:

  • Reduced layout area
  • Specific layout aspect ratio
  • Minimal error in the effective capacitor ratio
  • Robustness against process variations and mismatch

Layout-level optimization used an EDA tool from MunEDA called WiCkeD, which has a simulator in a loop approach, and the template was the simulator:

Optimization with WiCkeD

As the optimizer needs to find the performance for a set of design parameters, it asks the template to evaluate them. The optimizer finds the direction to change the design parameters that improve the layout. Evaluating a template takes under 5 seconds, so the optimization can quickly reach an optimal set of layout parameters.

To get the best capacitor array aspect ratio they selected an input parameter range of W and L for the unit devices, and the number of rows in the array. Next, they simulated the worst-case performance, including offsets in under two hours with 114 individual parameterizations. The worst-case transfer functions of the ADC model, based on numbers of rows in the capacitive array and various W over L values are shown below, where the ideal curve is dashed:

Worst-case transfer functions

Summary

Analog design and optimization is more difficult than digital design, because there are more interdependencies and trade-offs involved. A new approach with layout generators, template-based layout estimates and optimization has been demonstrated successfully for an ADC circuit, and it uses optimization technology from MunEDA, called WiCkeD. Instead of taking days to weeks, this approach met specifications for an ADC in just minutes.

Related Blogs


Hazard Detection Using Petri Nets. Innovation in Verification

Hazard Detection Using Petri Nets. Innovation in Verification
by Bernard Murphy on 08-24-2022 at 6:00 am

Innovation New

Modeling and verifying asynchronous systems is a constant challenge. Petri net models may provide an answer. Paul Cunningham (GM, Verification at Cadence), Raúl Camposano (Silicon Catalyst, entrepreneur, former Synopsys CTO and now Silvaco CTO) and I continue our series on research ideas. As always, feedback welcome.

The Innovation

This month’s pick is Hazard Detection in a GALS Wrapper: a case study. The paper published in the 2005 International Conference on Application of Concurrency to System Design. The authors are/were from Humboldt-Universität zu Berlin and IHP Microelectronics, Frankfurt.

This is an old paper yet interesting for use of Petri nets in modeling asynchronous processes, a problem domain that continues to crop up in modern designs. The authors apply their analysis to a GALS wrapper for a baseband block. Hazards can occur in handoff between a local clock driving the pipeline in the absence of external activity and the clock associated with an external input.

The method constructs Petri net patterns for a range of elementary gates, as state transition graphs (STGs) based on input signal states and edges. These nets are composable into a full circuit. In modeling, if an output edge with more than one marking is reachable from a starting state or might lead to such a marking, that represents a hazard. This is a widely studied problem and the basis of the method. Through a series of abstractions, the authors were able to detect several hazards and a potential deadlock.

Paul’s view

This month’s paper was a total blast from the past for me, bringing back memories from my own PhD in the formal verification of asynchronous circuits. The paper describes a method for detecting hazards (the potential for wires to glitch) in control circuits for a “globally asynchronous locally synchronous” (GALS) design style, but the star of the show for me is the authors’ use of a wonderful notation called a Petri net.

A Petri net looks like a state machine diagram, but it overlays the traditional node and arc notation with the concept of a “token”, literally drawn as a black dot that can go inside nodes. These tokens enable Petri nets to elegantly visualize concurrent interactions between state machines. The notation is not new and dates to the 1960s. Much of what we do in modern SOC verification involves wrestling with concurrency and this paper reminded me that Petri nets may warrant some fresh attention to improve complex protocol verification (CXL, USB4, etc.), cache coherency verification, or EDA tools based on the new Portable Stimulus language for transaction level randomization like Cadence’s Perspec.

As a further side comment, we all know about the duality between state machines and regular expressions. Likewise with Petri nets, there is an extension of regular expression syntax called “Trace Expressions” that introduces a special concurrency construct called the “weave” operator enabling a similar duality between Petri nets and Trace Expressions…but now we’re getting way too close to my PhD for comfort.

OK, back to the paper: the authors develop a system to represent every gate in a circuit with a Petri net fragment. They then compose these Petri net fragments together to form a giant Petri net for their GALS control circuits. Hazards can be formally expressed as certain states of this giant Petri net, and so proving that a circuit is hazard-free reduces to proving that these states are unreachable, which can be done with standard off-the-shelf model checking tools.

It’s a tight paper, and the authors’ use of Petri nets to formalize hazards is beautiful.  But unfortunately, the scalability of the underlying model checking problem generated by their method is very poor, and in the end the authors have to make some significant simplifying assumptions to get their checker to converge. That said, even with these assumptions they did find some real hazards in the control circuits that needed fixing.

Raúl’s view

The authors show how to model a chip that implements baseband processing compliant with IEEE802.11a. The implementation is globally asynchronous locally synchronous (GALS). The asynchronous part, which is of interest for the paper, is implemented as a wrapper around synchronous parts. It consists of 5 blocks and ensures that a synchronous pipeline functions properly. This they build up from elementary gate models, each represented by a Petri Net model. Once built, they make a key observation: that a hazard can occur when an edge place c carries more than one token. The problem of determining if a particular token pattern in a Petri-Net can occur is called the reachability problem.

They aim to solve this problem with a model checker, e.g., LoLA [10] or SMV[6]. The initial model results in a Petri-Net of 288 places and 526 transitions, too complex to be solved. The model is simplified applying hazard preserving abstraction techniques such as merging gates, abstracting structures. Also considering each of the 5 blocks separately. And they use VHDL simulations to confirm if a “dangerous” state transition could really occur. Several hazards could be detected with the simplified model. The authors conclude that “Fortunately, in the usual application scenario for GALS blocks no reachable state generates any of the detected potential hazards. Nevertheless, in order to increase reliability of the asynchronous wrapper we have fixed all those potential hazards in the system. Without the formal verification this would not have been feasible.”

I think this is an interesting modeling technique which allows formal verification. The modeling part is very clear, I found it elegant to model both signal values and edges. Detecting hazards when multiple edges can occur through solving reachability in Petri-Nets. The formal verification part however applies several techniques which are cumbersome and difficult to follow. Concluding thoughts:  Petri-Nets were popular in the 80s and 90s, particularly to model asynchronous systems. Asynchronous systems never became ubiquitous. Almost all digital systems are synchronous and just communicate, necessarily, with the outside world in a well understood asynchronous manner. There was no follow up to this paper. But even if not practical, it is a stimulating contribution to a hard problem.

My view

Despite the theoretical elegance of Petri Nets, I struggled to find papers on this topic in our domain. I did find several related papers based on different methods which all commented on scalability problems for Petri Nets. Perhaps this realization accounts in part for the lack of papers.


EDA Product Mix Changes as Hardware-Assisted Verification Gains Momentum

EDA Product Mix Changes as Hardware-Assisted Verification Gains Momentum
by Lauro Rizzatti on 08-23-2022 at 10:00 am

Semiwiki Hero Image Lauro Rizzatti

The Design Automation Conference, as always, is a good barometer on the state of EDA and my area of interest, verification. The recent DAC offered plenty of opportunities to check on trends and the status quo.

Remarkably, exhibitors and attendees were upbeat about the chip design landscape despite concerns about supply chain shortages and the impending semiconductor downturn. Based on my informal analysis, the hardware-assisted verification (HAV) segment and the future for chip design verification looks exceedingly bright. The big three players (Cadence, Siemens EDA and Synopsys) all hint at big gains in HAV tool adoption and point to emerging market segments for this growth.

Three executive-level presentations from three financial and market analysts that were part of the DAC program confirmed what I think and helped put into perspective the outlook for the EDA and semiconductor industry.

Sunday night, Charles Shi, Principal, Senior Analyst at Needham & Company, talked about “EDA to Power Through Semiconductor Cycles.” Shi reminded us that the semiconductor industry is cyclical and appears to be on the brink of a downturn. More positively, he outlined the reasons why the EDA industry will power through semiconductor cycles and emerge stronger on the other side.

Shi pointed to the slowdown of Moore’s Law driving secular growth of EDA and reminded attendees that EDA is part of an interdependent ecosystem that includes fabless, IP, foundry, equipment and materials. He added that chiplets are creating greater design complexity as the industry moves from 2D to 2.5/3D ICs, forcing stronger EDA and IP collaboration and tightly coupled package/chip co-design for faster design convergence. According to Shi, this means design and verification are getting harder and system-level design and analysis will increase EDA spending.

His last slide predicted that EDA is going to be okay if there is a recession and a semiconductor downcycle. Foundry, EDA and IP growth will continue to outperform the semiconductor industry, Shi concluded.

From my perspective, this is good news and portends many more opportunities for HAV as we move into more diverse applications areas and help manage chip functionality.

Rich Wawrzyniak, Senior Market Analyst from Semico Research, seemed to concur with Shi. He presented a well-balanced talk on “Semiconductor Market Trends: Tying it all Together for the Big Picture” Monday at the DAC Pavilion, beginning with a current assessment of the landscape. While semiconductor sales are high –– 26% growth in 2021, 6.3% projected for 2022 –– there’s a potential for a -0.9% forecast in 2023. Mitigating factors could create a downturn despite ongoing demand, a tight labor market and new application areas, he cautioned.

Much like Shi, Wawrzyniak pointed to advanced SoCs with elevated complexity levels and gate counts as drivers for increased functionality and performance. Wawrzyniak further pointed to the growing number of IP blocks per chip rises with each new process node and a reason for rising design costs.

As he wrapped up, Wawrzyniak predicted continued EDA revenue growth with more designs and more complex design starts. EDA is starting to move into new areas for growth and diversification.

Again, I concluded that more functionality and performance in a chip combined with more chip applications demand HAV solutions.

Jay Vleeschhouwer, Managing Director of Griffin Securities, presented “The State of EDA: A View from Wall Street” Tuesday in the DAC Pavilion. The most analytical of the three presentations, it nonetheless painted a positive view of EDA. For example, he noted that the EDA industry has grown each year for more than a decade with revenue surpassing $10 billion in 2021. The growth, he said, has been sustained across multiple product categories, a result of semiconductor and system engineering requirements in design, process and system complexity.

One the most important product mix changes over the past five to 10 years, according to Vleeschhouwer, has been the growth of HAV, both hardware emulation and FPGA prototyping. While he didn’t break out HAV revenue specifically, I estimate it’s approaching $1 billion, a huge revenue increase from hovering around $300-$500 million for almost two decades or since 2000.

Vleeschhouwer identified two areas where semiconductor company are investing –– software development and silicon development. He added that investments in silicon development remains where the majority of EDA revenue comes from. Software development is welcome news for HAV suppliers since it’s the only verification solution to support both software and silicon.

Given we’re more than halfway through 2022, we should expect more good news through yearend. Next year’s DAC could have a vibrant exhibit floor with attendees flocking to see the latest product news. I look forward to it and think you will, too.

Also Read:

An EDA AI Master Class by Synopsys CEO Aart de Geus

ARC Processor Summit 2022 Your embedded edge starts here!

WEBINAR: Design and Verify State-of-the-Art RFICs using Synopsys / Ansys Custom Design Flow


WEBINAR: A Revolution in Prototyping and Emulation

WEBINAR: A Revolution in Prototyping and Emulation
by Daniel Nenni on 08-23-2022 at 6:00 am

MimicPro Picture

This webinar will introduce to you a revolutionary new way to do prototyping and emulation at best-in-class performance, productivity, and pricing by unifying the hardware and a new software stack so one system is capable of prototyping and delivering essential emulation functionality.

Register Here

The speed of Moore’s law has slowed. However, with plans for 3nm production silicon in the second half of 2022 and with 2nm and 1nm in the works, the amount of IP being put on a single SOC is exploding. Also, the ability to verify IP both stand-alone and in the overall context of the SOC is becoming a greater challenge for the verification community at large.

Take the area of consumer electronics and we all know what our cell phones can do and deliver in terms of productivity and entertainment. Another device that is representative of massive amounts of IP placed on a SOC is the Drone market. Drones have upwards of DSLR sensors and record 5k+ video, requiring incredibly high-speed data interfaces. They leverage GPS multiple networks acquiring signals from well over 20 satellites to maintain a perfectly still position even on a breezy day. Even with Moore’s law slowing down, it is still driving us toward lower nodes and packing more and more IP into a single SOC creating a massive verification and software development challenge.

The same can be said at the enterprise level with networking bandwidth exploding and customized silicon now found at the desktop compute level. Data centers are all looking for faster and faster networking, and these complex SOCs need to be verified with firmware developed before silicon back from the FAB. The need for a faster and lower-cost platform for emulation and prototyping is apparent and Corigine’s EDA product line is aimed directly at those requirements.

Today, companies invest millions in costly emulation systems that need special power and cooling requirements and perform sub-par in order to run verification. In this session we will introduce you to Corigine’s approach to solving these issues using their unified platform for prototyping and essential emulation with a new level of performance.

Traditionally, costly emulators have been the tools deployed in top fabless companies around the world. With VC funding on the rise for fabless start-ups, these players need similar capabilities and Corigine with the only unified platform for prototyping and emulation functionality is answering these market requirements.

Corigine’s goal was to create an FPGA-based system bridging the gap between prototyping and major emulation functions. Doing so would reduce the investment in the key emulation functionality and require only a single platform for both teams meaning much lower overall infrastructure support costs. In addition, scalability and automation were key elements and with push button, automated timing optimized partitioning, fully automated handling of complex clock domains and the elimination of time-consuming recompiles, Corigine’s MimicPro product line is this revolutionary step forward for both Prototyping and Emulation.

Register Here

Also read:

Bringing Prototyping to the Desktop

A Next-Generation Prototyping System for ASIC and Pre-Silicon Software Development


A clear VectorPath when AI inference models are uncertain

A clear VectorPath when AI inference models are uncertain
by Don Dingee on 08-22-2022 at 10:00 am

Achronix VectorPath Accelerator Card with Speedster 7t1500 FPGA for running AI inference models and more

The chase to add artificial intelligence (AI) into many complex applications is surfacing a new trend. There’s a sense these applications need a lot of AI inference operations, but very few architects can say precisely what those operations will do. Self-driving may be the best example, where improved AI model research and discovery are on a frantic pace. What should a compute environment look like when AI inference models are uncertain?

Software adds AI inference flexibility but at a cost

A familiar reflex in the face of uncertainty is opting for software programmability. This dynamic has dominated large-scale CPU core development for generations. Faster processors debut, programmers write more software until it takes all the new-found capacity, then another round of even faster processors appears. But there’s a mismatch between a bulked-up CPU core and the fine-grained parallelized workloads in AI inference, and inefficiency becomes overwhelming.

Then GPUs showed up at the AI inference party with many smaller, parallelized cores and multithreading. On the surface, scaling up a software-programmable field of fast GPU cores seems a better fit for the fine-grained inference workload. If one has room for a rack of GPU-based hardware, it’s possible to pack a lot of TOPS in a system. But bigger GPUs start presenting other issues for AI inference in the form of sub-optimal interconnects and memory access. Hardware utilization isn’t great, and determinism and latency are suspect. Power consumption and cooling also head in the wrong direction.

Hardware could optimize around a known workload

If this is starting to sound like the case for a workload optimized custom SoC, that’s because it is. Design high-performance execution units, optimize memory access and interconnects, and organize them around running an AI inference model.

We’re seeing off-the-shelf AI inference SoCs popping up all over – primarily targeting one specific class of AI inference problem. There are SoCs designed to run YOLO models on facial recognition. Others optimize for driver assistance functions like lane adherence or emergency braking. AI inference is starting to get traction in areas like pharmaceutical research. If the AI inference models are well-defined, optimizing the workload in hardware is achievable.

But different AI inference models do not map onto layers or execution units the same way. Optimizing hardware around one model can be utterly inefficient for running another model. Making optimization matters worse, some of these more complex problems call for running different types of AI inference models concurrently on separate parts of the problem.

Niching down a custom SoC too tightly can result in lock-in, possibly preventing an enhanced AI inference model from running efficiently without a hardware redesign. That’s terrible news for a long life cycle project where the breakthrough AI inference innovations are yet to happen. It’s also not healthy for return on investment if volumes on a custom SoC are too low.

If only there were fast, programmable AI inference hardware

Several IP vendors are working on the specifics of reconfigurable AI inference engines with higher utilization and efficiency. Most are after the premise of co-design, where one looks at the AI inference models at hand and then decides how to configure the engine to run them best.

Recapping, we don’t know what the best hardware solution looks like when we start the project. We need a platform to explore combinations of IP quickly and change the design accordingly, maybe many times during development. And we must respond promptly and keep pace with state-of-the-art AI inference methods from new research. Also, if we’re going to a custom SoC, we need an inexpensive platform for software development before silicon is available.

Before thinking about designing a workload-optimized SoC, or even thinking about one at all if volumes are low, we should be thinking about an FPGA-based solution. The fact that an application may depend on AI inference models currently in flux reinforces that choice.

On that backdrop comes the Achronix VectorPath Accelerator Card, jointly designed with BittWare and now in general availability. It carries the Achronix Speedster 7t1500 FPGA, with its unique multi-fracturable MAC array matched to high performance LRAM and BRAM. Much of the attention on this design focuses on its blazing Ethernet connectivity for applications like high-frequency trading. It’s also an 86 TOPS engine with a 2-dimensional NoC for optimizing IP interconnects, plus 4 Tbps bandwidth to GDDR6 memory. Sensor data can come via those Ethernet ports or MCIO lanes at PCIe Gen5 data rates or on legacy interfaces over GPIO.

In short, it’s a powerful platform for AI inference, whether starting with third-party IP or designing it in-house. It drops into a host system easily with its PCIe form factor. More importantly, it allows designers to cope with projects starting when AI inference models are uncertain. We expect AI inference software and IP vendors to adopt Achronix into their ecosystems soon, and we’ll watch for future developments.

Visit the Achronix VectorPath Accelerator Card page for videos, datasheets, white papers, and more information on how this can help your AI inference project.


Protecting Critical IP With Next-Generation Zero Trust Security

Protecting Critical IP With Next-Generation Zero Trust Security
by Kalar Rajendiran on 08-22-2022 at 6:00 am

Securing Access Controls

While semiconductors are the enablers for high-tech solutions, the semiconductor industry was not at the forefront of Cloud adoption. There were many valid concerns behind the slow adoption, a primary reason being the threat to intellectual property (IP) security. IP in this context refers to not just chip building blocks but rather everything including tools, design methodologies and operational processes. Naturally, cloud providers took the IP security concern seriously and developed strong security mechanisms. They invested heavily into developing a secure infrastructure with tight security protocols and procedures for data at rest and data in motion.

With this increased level of cloud security, the chip industry started migrating to cloud computing. But accessing corporate data was still via on-prem equipment or via company supplied laptop computers and other mobile devices. The company’s IT department typically installed security software to protect against viruses and malware. As many other industries started accommodating bring your own devices (BYOD) policies with their employees, semiconductor industry was still lagging on this front.

Evolving Workplace

The semiconductor work environment has changed a lot over the recent years. Data and tools are nowadays accessible using both personal and company supplied devices. Access is made from within the firewall as well as from outside. When the Covid-19 pandemic hit, businesses were pushed into a work from home (WFH) model with predominant data access from outside the firewall. Of course, the primary mode for remote accessing data and tools was via a virtual private network (VPN).

VPNs are insecure because once hacked into, hackers can potentially access the entire network. Once an attacker has breached the network through a compromised device, hackers can try to steal corporate IP, customer or financial information, or launch a ransomware attack.

Security Challenges

In spite of elaborate security measures put in place, corporate networks are still hacked into on a regular basis. The usual hacks are made possible through a combination of phishing and breaking into a VPN client. With employees, contractors, vendors, supply-chain partners and customers needing different levels of access to various data, the threat of break-in is ever present.

Just last week, Cisco announced that their corporate network was hacked into.

Following is an excerpt from a blog post by Cisco Talos Intelligence Group, one of the largest commercial threat intelligence teams in the world.

  • The attacker conducted a series of sophisticated voice phishing attacks under the guise of various trusted organizations attempting to convince the victim to accept multi-factor authentication (MFA) push notifications initiated by the attacker. The attacker ultimately succeeded in achieving an MFA push acceptance, granting them access to VPN in the context of the targeted user.
  • Initial access to the Cisco VPN was achieved via the successful compromise of a Cisco employee’s personal Google account.

Following is an excerpt from SecurityWeek.com as it relates to the above Cisco news.

  • The attacker managed to enroll new devices for MFA and authenticated to the Cisco VPN. Once that was achieved, they started dropping remote access and post-exploitation tools. The hackers escalated their privileges, created backdoors for persistence, and moved to other systems in the environment, including Citrix servers and domain controllers.

Security risks are only growing over time, prompting companies to favor a Zero Trust security model over a VPN-based model. As per a report on helpnetsecurity.com, attacks against VPNs were up nearly 2,000% in 2021. According to a VPN Risk Report based on Zscaler’s survey of over 350 cybersecurity professionals:

  • 72% of organizations are concerned that VPN may jeopardize IT’s ability to keep their environments secure
  • 67% of enterprises are considering remote access alternative to a traditional VPN
  • 72% of companies are prioritizing the adoption of a Zero Trust model

Zero Trust Security (ZTS)

Even before Covid-19 impacted the workplace forever, ZTS was already in motion as a solution for addressing security threats. The ZTS model takes the approach of verifying everyone and every device that is seeking access to confirm they are who they claim to be. It does not matter whether the access is being sought from within the network or from outside.

Fundamental Requirements for Enterprise ZTS

  • Access visibility – Full visibility of all access to critical applications and data
  • Access control – Authenticate and control of who can access what resources, applications, and data
  • Protocol control – Whitelisted software and protocol controls to block unapproved or malicious traffic
  • Universal enforcement – Unified method across enterprise on-prem, in the cloud, or remote
  • Business agility – Dynamic project-based controls, no infrastructure “rip and replace”

A primary challenge of Zero Trust approach is ensuring security without slowing things down. Access to sensitive data is needed on an ongoing basis to communicate and collaborate on projects. Team members’ roles and access levels vary and change on a frequent basis. ZTS implementation should be such that people don’t lose access, leading to productivity loss.

The Figure below highlights where data leaks typically happen and how attacks propagate through a network.

Zentera CoIP® Platform and Technology

Zentera has developed the next generation ZTS solution using its patented technology. Its solution takes a differentiated approach of in-line ZTS. The Zentera CoIP Platform implements a Zero Trust security fabric using an overlay technology architecture. The platform authenticates users using a company’s existing identity providers and endpoints and applications using certificates and fingerprints. This method enables a role-based access to specific applications and resources. Accordingly, secure virtual zones can be created, managed and torn down rapidly as the requirements call for. Refer to the Figure below for the solution concept.

The CoIP platform integrates the following three core engines to implement ZTS.

The CoIP Chamber

The chamber implements the concept of a secure virtual development environment wherein applications are cloaked and lateral network access is eliminated. Inbound and outbound flow control are allowed only for authorized users, apps and services.

The CoIP Network

The CoIP network connects applications rather than connect networks across silos. The applications’ connectivity is based on application identity. IP collisions across environments are avoided through the use of overlay IP addressing.

CoIP Access

This engine authorizes each access based on an user’s role and privileges. The users and devices are authenticated through MFA. This engine also sets up non-step and one-stop access between specific applications instead of connecting entire networks together.

CoIP Technology in Action

Whether it is the employee, contractor, vendor or customer, they gain access to the remote desktop chamber via the CoIP platform. From there, they gain access to the application chamber and other applications based on CoIP access controls setup. Even if someone wants access from an on-prem workstation, they still have to go through the remote desktop chamber before they can access applications and data. Refer to the Figure below.

In the particular use-case shown in the above Figure, blanket access to the internet has been prohibited. For use-cases where internet access is required for specific purposes, the technology supports those requirements. For example, let’s say that access to a company’s remote data server is required, that specific connectivity can be authenticated and setup.

Also Read:

Post-quantum cryptography steps on the field

CEO Interview: Jaushin Lee of Zentera Systems, Inc.

Memory Security Relies on Ultra High-Performance AES-XTS Encryption/Decryption