100X800 Banner (1)

High Efficiency Edge Vision Processing Based on Dynamically Reconfigurable TPU Technology

High Efficiency Edge Vision Processing Based on Dynamically Reconfigurable TPU Technology
by Kalar Rajendiran on 05-02-2022 at 6:00 am

Fast model evolution Flexibility is key

While many tough problems relating to computing have been solved over the years, vision processing is still challenging in many ways. Cheng Wang, Co-Founder and CTO of FlexLogix Technologies gave a talk on the topic of edge vision processing at Linley’s Spring 2022 conference. During that talk he references how Gerald Sussman took the early steps of computer vision processing way back in 1966. Gerald, a first-year undergraduate student under the guidance of MIT AI Lab co-founder Marvin Minsky tried to link a camera to a computer. Much progress has happened since then. Of course, the requirements and the markets for computer vision haven’t stayed static during this time.

The early era of computer vision processing focused on industrial grade computing equipment that tolerated large form factors and high costs of the solutions. Fast forward to the most recent decade, neural network models and GPUs have played critical roles in advancing vision processing capabilities. But delivering solutions in smaller form factors and at low costs is still a challenge. In his talk, Cheng discusses the reasons behind these challenges and FlexLogix’s solution to edge vision processing based on dynamically reconfigurable TPU technology. The following are some excerpts from his presentation.

Performance, Efficiency and Flexibility

Edge computer vision requires extreme amount of processing at Teraops rates. And the vision solutions need to demonstrate high accuracy at low latencies, operate at low power and be available at low cost points. While GPUs can deliver the performance, they are large, expensive and power hungry and thus not a good match for edge compute devices. And GPUs count on a huge amount of memory bandwidth via DDR type interfaces.  On top of these challenges, the neural models are also fast evolving. Not only are new models emerging at a rapid rate, even the same models undergo incremental changes at a frequent rate. Refer to Figure below to see how frequently the popular model YOLOv5 is going through changes.

The processing of neural network models is very different from general purpose processing when it comes to compute work load and memory access patterns. Each layer may require vary computational loads relative to the memory bandwidth that layer requires. And this changes dynamically as different layers are processed. So, an optimal approach to solving the challenges counts on memory efficiency and future proofing for changing models. Graph streaming will help reduce DRAM requirements but bandwidth matching on a varying load is difficult.

FlexLogix’s Dynamic TPU

FlexLogix’s Dynamic TPU offers a flexible, load-balanced, memory-efficient solution for edge vision processing applications.

The Dynamic TPU is implemented using Tensor Processor Arrays (ALUs) and EFLX logic. The architecture enables very efficient layer processing across multiple Tensor Processor Arrays that communicate via FlexLogix’s XFLX InterConnect and access L2 SRAM for memory efficiency. As the TPU uses EFLX cores, the control and data paths are future proofed for changes in activation functions and operator changes. By streaming data at a sub-graph level, more efficient bandwidth matching is made possible. Refer to Figure below.

While a GPU-based edge vision processing solution may consume power in the 75W-300W range , a Dynamic TPU based solution will consume in the 6W-10W range. Whereas a GPU-based solution predominantly relies on GDDR, a Dynamic TPU-based solution relies on local connections, XFLX connections, flexible L2 memories and LPDDR.

The FlexLogix solution includes the InferX SDK which directly converts a TensorFlow graph model to dynamic InferX hardware instance. A Dynamic TPU-based solution will yield a much higher efficiency on the Inference/Watt and Inference/$ metrics compared to a GPU or CPU based solution. All in all, a superior performance with software flexibility and future proofing versus ASIC solutions.

On-Demand Access to Cheng’s talk and presentation

You can listen to Cheng’s talk from here, under Session 5.  You will find his presentation slides here, under Day 2- PM Sessions.

Also read:

A Flexible and Efficient Edge-AI Solution Using InferX X1 and InferX SDK

Flex Logix and Socionext are Revolutionizing 5G Platform Design

Using eFPGA to Dynamically Adapt to Changing Workloads

 


ITSA – Not So Intelligent Transportation

ITSA – Not So Intelligent Transportation
by Roger C. Lanctot on 05-01-2022 at 10:00 am

ITSA Not So Intelligent Transportation

The Infrastructure Investment and Jobs Act (IIJA) passed last year in the U.S. earmarks billions of dollars that can be used for the deployment of potentially life-saving C-V2X car connectivity technology. The U.S. Department of Transportation and state DOTs are poised to commence that spending, but one thing stands in the way of car maker or state DOT willingness to proceed – a lawsuit by the Intelligent Transportation Systems of America (ITSA) and the American Association of State Highway Transportation Officials (AASHTO).

ITSA and AASHTO are seeking to reverse the Federal Communication Commission’s (FCC) re-allocation of 45MHz of spectrum in the 5.9GHz band – previously preserved for dedicated short range communication (DSRC) use by automobiles – for unlicensed Wi-Fi use. ITSA and AASHTO want the 45MHz restored.

The judge is unlikely and probably does not have the authority to reverse the FCC’s unanimous decision. The only possible path to success for the ITSA and AASHTO would be for the judge to find the FCC’s decision-making process somehow flawed. This is highly unlikely – which means that the legal action is a waste of time and money and maybe…lives.

Ironically, ITSA and AASHTO wave the bloody flag in their efforts to preserve the prior spectrum allocation – claiming their efforts are intended to save lives the best way possible with connected car technology. Their efforts are actually further delaying the prospect of the adoption of connected car technology in the near term.

ITSA CEO Laura Chase comments in an opinion piece in the latest ITSA magazine:

“While there is no magic bullet to reduce crashes and fatalities, we have a responsibility to use all the tools at our disposal to save lives. The best tool we currently have is connected vehicle technologies – but without wide-scale deployment, we can’t hope to move the needle on reducing traffic fatalities.”

Well, Laura, we can’t move that needle as long as ITSA and AASHTO continue to inject uncertainty into the regulatory process. Car makers need clarify, alignment, and commitment. State and Federal contracting authorities, too, need clarity. ITSA and AASHTO have muddied the waters and put a bullet in the head of potentially life-saving infrastructure projects incorporating C-V2X technology.

Even stranger, Chase says the ITSA is simultaneously working on re-imbursement for stranded DSRC deployments – something that is already provided for in the IIJA. The only good news from ITSA is that the organization appears to be “accepting” the FCC’s overt endorsement of cellular-based C-V2X technology over DSRC. Thank goodness for small things.

The legal action by ITSA and AASHTO means that car companies or state DOTs are frozen. Proposals can’t be written and money cannot be allocated until the case is resolved.

Multiple car companies and DOTs have applied for waivers from the FCC to proceed with their projects – but the FCC has not even posted the waiver requests, which are subject to public comment. ITSA and AASHTO have gummed up the very process for which they have worked for more than 20 years – to bring V2X technology to the market.

A senior General Motors executive speaking as part of a 5.9GHz forum at the recent ITSA event in Charlotte, N.C., said, of this legal action: “You lost half of the dedicated 5.9GHz spectrum because you did nothing with it for 20 years. If you don’t do something with it now you’re likely to lose the rest of it.”

Worse, though, is the reality that the legal action by ITSA and AASHTO is actually costing the U.S. valuable time in the race to compete with China. China long ago abandoned DSRC as the primary connected car technology in favor of C-V2X.

As many as 13 auto makers in China have either already introduced C-V2X-equipped vehicles or have announced plans to do so. In the U.S., Ford Motor Company, Audi of America, and Jaguar Land Rover and multiple state DOTs have submitted waiver requests to introduce the technology – and, so, they wait.

ITSA and AASHTO are on the wrong side of history. These organizations are wasting time, money, and lives in the interest of turning back the clock. The FCC has spoken. The spectrum has been allocated. The billions of dollars have been approved. It’s time for ITSA and AASHTO to simply get out of the way.

Also read:

OnStar: Getting Connectivity Wrong

Tesla: Canary in the Coal Mine

ISO 26262: Feeling Safe in Your Self-Driving Car


Has KLA lost its way?

Has KLA lost its way?
by Robert Maire on 05-01-2022 at 6:00 am

KLA SPIE 2022

-KLA has another great QTR in face of overwhelming demand
-Supply chain issues obliterated by backlog
-Longer term technology leadership concerns are increasing
-We see limited upside near term & remain cyclically cautious

Another great quarter- demand remains super strong

KLA’s performance remains great as does overall semiconductor equipment demand. KLA reported revenues of $2.3B versus expectations of $2.2B and non GAAP EPS of $5.13 versus street of $4.82. Guidance was for revenues of $2.3B to $2.55B versus current expectations of $2.36B with earnings in the range of $4.93 to $6.03 versus current expectations of $5.30

KLA can “dial in” numbers given the huge backlog

Historically KLA has almost always been able to accurately dial in numbers for the next quarter given the huge and long backlog they have. The current backlog is out the door and down the street and not likely to shorten any time soon.

This results in the ability to both guide and deliver numbers wherever they wish. Some segments remain a little bit lumpy due to high selling prices or mix shifts. With most deliveries running at a year or more and over $8B in solid orders we don’t see a lot of risk to the backlog right now. However we have seen backlog deflate in prior cycles but we never with the level we currently have.

Supply chain issues remain but not very impactful

Supply chain issues remain “fluid” but the large backlog clearly mitigates most if not all of that instability. As compared to other companies in the industry that typically run a more turns business KLA can modulate to adapt to shortages.

Yield management continues to be a crucial market segment

Growth in all things yield management remain very strong especially in emerging markets that have a lot more to learn when it comes to semiconductor manufacturing. This means China which remains a huge market for semiconductor tool makers including KLA.

This adds perhaps a bit more risk but so far we don’t think the US government is in a mood to upset the Chinese by turning up the heat in trade restrictions given the Ukraine situation.

Has KLA lost its way?

KLA’s first product was reticle inspection back in 1975 and it has been one of the two pillars of the company along with wafer inspection for the entire life of the company. We think the reticle inspection pillar of the company is weakening, though perhaps not completely, it has certainly lost the technology lead and along with it the future business.

A former upstart pimple on KLA, Lasertec, has clearly taken the technology lead and with it, the most profitable as well as future of the reticle inspection market. Lasertec likely has the dominant share of leading edge reticle inspection revenue as well and will likely expand that lead.

Lasertec’s recent quarter

Lasertec recently announced its quarter and along with it projections for the next twelve months business which they expect will come in at over $2B versus KLA’s just reported $611M in the quarter in “patterning” which likely does not represent pure 100% reticle inspection tools.

More importantly, Lasertec is the only game in town in EUV actinic inspection. We believe KLA’s actinic tool has been further delayed by issues with several hardware subsystems let alone the fact that the noble gas Xenon, which the system runs on, has skyrocketed recently from $10 a liter to over $200 a liter if you can get it…and still climbing.

We have heard that KLA’s E-Beam reticle inspection (the 8XX) tool has not been popular customers and public data shows that E-Beam is just way too slow. But right now customers may settle for a slower 8XX tool as actinic is years away from either Lasertec or maybe eventually KLA.

KLA may point to “print check” (AKA print and pray) which uses a wafer inspection tool to look at what has been printed from the reticle on the wafer but its not a direct (only inferred) system that is useless in a mask shop anyway. Actinic is clearly the gold standard and only Lasertec has it.

Data points from SPIE

We recently attended SPIE (a conference about all things lithography) and a talk given by a major chipmaker who is first in line for High NA EUV tools spoke about High NA reticle inspection and showed a picture of a system…and it wasn’t KLA.

So currently where the industry is and is going in reticle inspection is not KLA and KLA may not have time to catch up given the delays. We have seen this movie before as ASML was around 10-15 years delayed in getting EUV scanners to market. KLA’s multi year self imposed halt in the program certainly made things even worse.

KLA still does a great business in older technology reticle inspection for all those second and third tier fabs in China but that’s not saying a lot.

Weakness in E Beam wafer tools

While reticle inspection may already be a fait acompli we are also starting to get more concerned about wafer inspection. ASML recently announced a 5X5, 25 multibeam (not multicolumn…there is a difference) E Beam wafer inspection tool in the Hermes division. ASML has been winning in wafer defect inspection while AMAT has been exploding in the E Beam wafer metrology market. KLA still dominates in optical, which is about 4 times the size of E Beam, but clearly needs to catch up to ASML and AMAT in E Beam.

The stock

The results and financials are great…as always. Demand remains super strong. We certainly are not concerned about the near term but have questions about the longer term especially when the market eventually slows.

Right now customers are desperate for tools and anything that will help the yields of ever more complex process so KLA is in a good seat. Perhaps not as good as ASML but second best.

Much of the current success is due to momentum, size and desperation not necessarily technology leadership. This makes us more concerned about the longer term issues .

While 2022 seems almost “in the bag” we are more concerned about where things go when the tide goes out and exposes issues in the longer term.
From a valuation perspective its hard to fight the negative tape in chip stocks and much of the strong performance including a strong second half is already baked into the numbers and expectations.

We don’t see a lot of upside headroom in the stock and see more longer term potential downside at this point which would make us avoid putting more money to work here.

Also read:

LRCX weak miss results and guide Supply chain worse than expected and longer to fix

Chip Enabler and Bottleneck ASML

DUV, EUV now PUV Next gen Litho and Materials Shortages worsen supply chain


Podcast EP75: Getting There is Half the Fun – Connecting the Digital World with Alphawave IP

Podcast EP75: Getting There is Half the Fun – Connecting the Digital World with Alphawave IP
by Daniel Nenni on 04-29-2022 at 10:00 am

Dan is joined by Tony Pialis, the co-founder and CEO of Alphawave, a global leader in high-speed connectivity IP enabling industries such as AI, autonomous vehicles, 5G, hyperscale data centers, and more. Tony and Dan discuss the requirements for data connectivity across many high-growth markets and what is required for successful deployment.

Tony is the former VP of Analog and Mixed-Signal IP at Intel and has co-founded three semiconductor IP companies, including Snowbush Microelectronics Inc (sold to Gennum/Semtech and now part of Rambus) and V Semiconductor Inc (acquired by Intel).

The views, thoughts, and opinions expressed in these podcasts belong solely to the speaker, and not to the speaker’s employer, organization, committee or any other group or individual.


CEO Interview: Dr. Robert Giterman of RAAAM Memory Technologies

CEO Interview: Dr. Robert Giterman of RAAAM Memory Technologies
by Daniel Nenni on 04-29-2022 at 6:00 am

RAAM Memory Group Photo SemiWiki

Dr. Robert Giterman is Co-Founder and CEO of RAAAM Memory Technologies Ltd, and has over nine-years of experience with the research and development of GCRAM technology, which is being commercialized by RAAAM. Dr. Giterman obtained his PhD from the Emerging Nanoscaled Circuits and Systems Labs Research Center in Bar-Ilan University. Following the completion of his PhD in 2018, he joined the Telecommunications Circuits Laboratory in the Ecole Polytechnique Federale de Lausanne, Switzerland, as a post-doctoral researcher. As part of his research, he has led the front-end and physical implementations of multiple ASICs, and mentored numerous PhD thesis and MSc projects in the field of VLSI embedded memories. Dr. Giterman has authored over 40 scientific papers and holds 10 patents.

First, please tell me about RAAAM?
RAAAM Memory Technologies Ltd. is an innovative embedded memory solutions provider, that delivers the most cost-effective on-chip memory technology in the semiconductor industry. RAAAM’s silicon-proven Gain-Cell RAM (GCRAM) technology combines the density advantages of embedded DRAM with SRAM performance, without any modifications to the standard CMOS process available from multiple foundries.

RAAAM’s patented GCRAM technology can be used by semiconductor companies as a drop-in replacement for SRAM in their SoCs, allowing to significantly reduce fabrication costs through a significant die size reduction. Alternatively, increasing the on-chip memory capacity in the same die size enables a dramatic reduction in the off-chip data movement to resolve the memory bottleneck. This increase in on-chip memory capacity will enable additional features that can enable industry growth for applications in the areas of AR/VR, Machine Learning (ML), Internet-of-Things (IoT), and Automotive.

What problem are you solving?
Important industry growth drivers, such as ML, IoT, Automotive and AR/VR, operate on ever-growing amounts of data that is typically stored off-chip in an external DRAM. Unfortunately, off-chip memory accesses are up-to 1000x more costly in latency and power compared to on-chip data movement. This limits the bandwidth and power efficiency of modern systems. In order to reduce these off-chip data movements, almost all SoCs incorporate large amounts of on-chip embedded memory caches that are typically implemented with SRAM and often constitute over 50% of the silicon area. This memory bottleneck is further aggravated since SRAM scaling has been increasingly difficult in recent nodes, shrinking only at a rate of 20%-25% compared to almost 50% scaling for logic.

Can you tell us more about GCRAM technology?
GCRAM technology relies on a high-density bitcell that requires only 2-3 transistors (depending on priorities on area or performance). This structure offers up-to 2X area reduction over high-density 6T SRAM designs. The bitcell is composed of decoupled write and read ports, providing native two ported operation, with a parasitic storage node capacitor keeping the data. Unlike conventional 1T-1C eDRAM, GCRAM does not rely on delicate charge sharing to read the data. Instead, our GCRAM provides an active read transistor that provides an amplified bit-line current, offering low-latency non-destructive readout without the need for large storage capacitors. As a result, GCRAM does not require any changes or additional costs to the standard CMOS fabrication process and scales with technology when properly designed.

While the concept of 2T/3T memory cells has been tried in the past, reduction of the parasitic storage capacitor and concerns about increasing leakage currents has so far discouraged its application beyond 65nm. RAAAM’s patented innovations comprise clever circuit design at both memory bitcell and periphery levels, resulting in significantly reduced bitcell leakage and enhanced data retention times, as well as specialized refresh algorithms optimized for various applications, ensuring very high memory availability even under the most extreme operating conditions. In fact, we had demonstrated the successful scaling of GCRAM technology across process nodes of various foundries (e.g., TSMC, ST, Samsung, UMC), including recent silicon demonstrators in 28nm (Bulk and FD-SOI) and 16nm FinFET technologies implementing up-to 1Mbit of GCRAM memory macros.

Can you share details about your team at RAAAM and what has been done to validate the GCRAM technology?
RAAAM founders, including Robert Giterman, Andreas Burg, Alexander Fish, Adam Teman and Danny Biran, bring over 100+ combined years of semiconductor experience. In fact, RAAAM is built on a decade of world-leading research in the area of embedded memories, and GCRAM in particular. Our work on GCRAM technology has been demonstrated on 10 silicon prototypes of leading semiconductor foundries in a wide range of process nodes ranging from 16nm to 180nm, including bulk CMOS, FD-SOI and FinFET processes. Our work on GCRAM is documented by more than 30 peer-reviewed scientific publications in books, journals, and conference proceedings, and is protected by 10 patents.

Who is going to use RAAAM’s technology and what will they gain?
RAAAM’s GCRAM technology enables a significant chip fabrication cost reduction or highly improved performance, resolving the memory bottleneck for semiconductor companies in various application fields. Since GCRAM is directly compatible with any standard CMOS process and uses an SRAM-like interface, it can easily be integrated into existing SoC designs.

As an example for potential system benefits, we can look at the Machine Learning accelerators domain using a 7nm AI processor integrating 900MB of SRAM on a single die. In this case, the SRAM area constitutes over 50% of the overall die size. Replacing SRAM with RAAAM’s GCRAM technology can provide a reduction of up-to 25% of the overall die size, resulting in up-to $35 savings per die.

Alternatively, for memory-bandwidth limited systems, increasing the on-chip memory capacity can bring substantial performance and power improvements. In fact, the required DRAM bandwidth is often inversely proportional to the on-chip memory capacity. With off-chip memory accesses being up-to 1000x more costly in power and latency compared to on-chip data movement, replacing SRAM with 2X more GCRAM capacity at the same area footprint significantly reduces the off-chip bandwidth requirements and enables RAAAM’s customers to gain a competitive advantage in the power consumption of their chip.

What is RAAAM’s engagement model?
RAAAM follows an IP vendor licensing model. Semiconductor companies can license RAAAM’s GCRAM technology for a fee and production unit royalties RAAAM implements the front-end memory controller and GCRAM-based hard memory macros according to the customer specifications and delivers a soft RTL wrapper (using a standard SRAM interface), which instantiates the GCRAM hard

macros (GDS) and the soft refresh control (RTL). Additionally, the customer receives a characterization report of the hard memory macro and a behavioral model for system-level verification. At present,

RAAAM is working on the implementation and qualification of a GCRAM-based memory compiler, which will enable RAAAM’s customers to automatically generate the complete front and back-end views of GCRAM IP and corresponding characterization reports according to customer specifications.

Can you tell us about your recent achievements?
RAAAM has made very exciting progress recently. First, we have been evaluating the benefits of our technology for leading semiconductor companies, which has confirmed our projected substantial improvements from a performance and cost perspective over existing solutions based on SRAM. In fact, we have recently engaged with a very large semiconductor company on a long-term, co-development project and we continue running customer evaluations for various application fields and process nodes. We see growing interest in our technology in a variety of applications, both in very advanced process (7nm and beyond) nodes and in less advanced ones (16nm and higher). Finally, we are extremely pleased to have joined the Silicon Catalyst Incubator, allowing us to gain access to their comprehensive ecosystem of In-Kind Partners, Advisors, and Corporate VC and institutional investor network.

What is on the horizon for RAAAM?
Our product development roadmap includes full memory qualification in selected nodes of leading semiconductor foundries, based on customer demand. In addition, we have on-going discussions with numerous foundries for further technology migration to their next generation process nodes. Furthermore, we are looking to expand our embedded memory platform and introduce design flow automation based on our memory compiler development efforts. To this end, we are in the process of raising Seed funding to fully qualify our GCRAM technology and to accelerate our company’s overall business growth.

A preliminary GCRAM product brief is available upon request, please send an email to info@raaam-tech.com. Additional information can be found at: https://raaam-tech.com/technology  https://www.linkedin.com/company/raaam

Also read:

CEO Interview: Dr. Esko Mikkola of Alphacore

CEO Interview: Kelly Peng of Kura Technologies

CEO Interview: Aki Fujimura of D2S


Freemium Business Model Applied to Analog IC Layout Automation

Freemium Business Model Applied to Analog IC Layout Automation
by Daniel Payne on 04-28-2022 at 10:00 am

animate preview min

Freemium is the two words “free” and “premium” combined together, and many of us have enjoyed using freemium apps on our phones, tablets and desktop devices over the years. The concept is quite simple, you find an app that is useful, and download the free version, mostly to see if it operates as advertised, and then decide if there’s enough promise to warrant buying the fully-featured version. But wait, is there actually any EDA vendor offering a freemium business model?

Yes, about a year ago, the UK-based company Pulsic introduced their Animate Preview tool to the EDA world as a free download. The only requirement is that you are using Cadence Virtuoso IC6.1.6, IC6.1.7 or IC6.1.8 software. I had a Zoom call with three Pulsic folks this month to better understand this freemium model:


Mark Williams, CEO

  • Mark Waller, Director of User Enablement
  • Otger Perich, Digital Marketing

Q: Why a freemium model?

A: The typical EDA evaluation cycle for a new EDA tool is way too long. Often requiring an NDA to be agreed, terms and conditions to be negotiated, and time and resources for a formal evaluation. It can take many weeks before potential customers can start to really get to know the product’s capabilities.

We wanted to find a way to shortcut this process and remove all of the barriers to entry. With the freemium model, any interested engineer can quickly and directly download a free version and get started in minutes instead of weeks.

To make the freemium model work, we made Animate easy to use with a very simple UI, easy to learn and operate.

Q: What does Animate Preview do?

A: Animate Preview works within their Cadence Virtuoso schematic editor, where a circuit designer can quickly see the automatically created initial layout of their analog cells in minutes. The designer can see the effect of their circuit design decisions in the layout and get accurate area estimates. The free version contains all the features of the premium product, the user can do everything that can be done in the paid version, but they can only then save the design outline and IO pins.

The paid version is called Preview Plus, and with that version, you can save the automatically created initial layouts into OpenAccess. The saved layout includes all the detailed placement information and is a great starting point for creating the final analog block layout.

Animate Preview inside the schematic editor

Q: How long does it take to learn Animate Preview?

A: It’s fast; from downloading the app to seeing the first circuit layout can happen in as little as 20 minutes because it’s a simple process of filling out a form and opening the link in an email to get started. Anyone with a Cadence Virtuoso environment for schematics can use Animate Preview on their analog cells. We’re using a cloud-based license, so you don’t need to think about licensing.

Q: Does the Pulsic tool come with any design examples?

A: Yes, we ship with a Pulsic PDK with example designs in that technology, plus there’s a set of videos to get you started. It’s all designed to just run out of the box. As well as the getting started videos, there is a series of 2-minute tutorials, with 22 tutorials available.

Animate Preview runs in the background when you open a schematic in Virtuoso, which you use just like you normally would. The layouts appear automatically and are updated when circuit changes are made, all without the user needing to create any constraints. Just install and then see the auto-generated IC layouts based on schematics.

Q: What process technology is supported for analog IC layout generation?

A: Our focus has been to ensure that Animate creates great results for TSMC processes from 180nm down to 22nm. However, Animate will work with any planar process on OpenAccess with Cadence P-Cells. We have customers using Animate on many other processes from several fabs. We’re also starting to support some FD-SOI technology, but no FinFET yet.

Q: Is the generated IC layout always DRC clean?

A: Yes, the generated IC layout should be DRC clean, especially for TSMC processes. For other processes, if the rules are in the OA tech file, Animate will obey them. Most customers get good results out of the box, but if a user has any issues, they can contact Pulsic for better support.

Animate Preview generated layout

Q: So, who is using Animate for analog IC cell layout automation?

A: One company that we can talk about is Silicon Labs, out of Austin, Texas; back in 2019, when they were using an early version of the Animate technology, they said, “In our initial evaluation of Animate, we needed to achieve both efficiency and quality for our analog IC layouts, and Animate provided excellent results equal to using traditional approaches but in far less time,” said Stretch Young, Director of Layout at Silicon Labs.  “Collaborating with Pulsic, we see opportunities to improve the quality of our layout, which will increase productivity and save design time.”

Q: How many downloads so far of Animate Preview from your web site?

A: About 360 engineers have downloaded Animate so far. About 100 of these downloaders have created IC layouts, and we’ve followed up with 10s of engagements.

Q: What are some of the benefits of offering a freemium model for EDA tools?

A: With the freemium model, there is less pressure. We see that the users like the free download experience, and then we support them when they have follow-up questions. Users can see the benefits of analog automation within days without the hassle and pressure of the usual EDA sales process. Only if they like what they see and want to save the placement do they need to talk to us.

Launching a new product in COVID times was always going to be a challenge, but a big benefit for us was that we didn’t have to travel to do prospecting because it’s been all online evaluations. So we were able to reach the target audience much quicker.

Q: What types of IC design end markets are attracted to analog IC layout automation?

A: The IoT market has been the most significant sweet spot so far because of the need to be quick to market cheaply and the ability to iterate quickly.  Automotive and general analog IP providers also see great results from our tool.

Q: What are the limitations of Animate Preview as an EDA tool?

A: Animate Preview is designed for core analog cells. The tool is always-on inside the Cadence Virtuoso Schematic Editor and continually updates as you change the schematic. So you just leave it on all of the time, but it will warn you if it cannot apply the technology to a cell. A built-in circuit suitability check will warn you when the circuit is not suitable for Animate, e.g., a hierarchy that is too large or a digital block. Animate Preview will automatically create a layout for analog blocks with up to 100 schematic symbols. With Preview Plus, the user can create a layout for larger analog blocks; it might take a few minutes instead of seconds to produce a result.

Q: Will your company be attending DAC in SFO this summer?

A: Yes, look for our booth, and there will be a theatre setup to show the benefits of analog IC layout automation.

Q: How does Animate Preview work, under the hood?

A: Animate is radically different from other IC layout automation because it has a PolyMorphic approach in a virtual space, producing optimal IC layouts. It really is a unique architecture. The polymorphic engine is patented, but we don’t talk about how it works.

Related Blogs


The Path Towards Automation of Analog Design

The Path Towards Automation of Analog Design
by Tom Simon on 04-28-2022 at 6:00 am

Early parasitics estimation for analog design

You may have noticed that I have been writing a lot more about analog design lately. This is no accident. Analog and custom blocks are increasingly important because of the critical role they play in enabling many classes of systems, such as automotive, networking, wireless, mobile, cloud, etc.  Many of the SoCs needed for these markets are developed on advanced nodes, including FinFET. However, new design rules and other design complexities at these advanced nodes are making analog design more difficult and challenging.

Synopsys has a presentation at this year’s CICC that is titled “Has the Time for Analog Design Automation Finally Come?”, authored by Dave Reed and Avina Verma, which offers a close examination of methods for accelerating and improving how analog design is done. I had a chance to talk with them recently about their presentation. Automation of analog design is a laudable goal but has proven elusive. In part Dave and Avina attribute this to the fact that it’s more difficult for analog designers to provide a concise set of constraints to describe their design objectives. Asking analog engineers to create extensive text-based rules to drive the automation tools often results in just as much work as doing the physical design in the first place.  They also say that tool designers need to ensure design tools match the way designers want to work.

Their point is that each stage of the design process has a preferred creation and editing method that any tools for automation should accommodate. They believe that encouraging iterative design is better than asking for a big up-front investment to specify the results. There are several key goals for an automation flow. Faster layout should be possible with automated correct device level placement and device level routing. Design closure requires consideration of resistance, capacitance and electromigration issues. Designers want to get early insight into parasitics. Lastly, design reuse, if done right, can offer a huge productivity boost.

As one example of using graphical methods, they point to the way Custom Compiler uses a symbolic graphical palette to pre-define placement patterns for devices. Along with this it provides a real-time display of the actual layout visible at the same time. Visual feedback is provided with color coded device visualizations and a centroid display. It also provides an easy way add dummies and guard rings.

Device routing automatically connects large device arrays while ensuring matched R/C routes. Interconnect with user-controlled hand-crafted quality is created with greater ease than with manual methods. Just as with placement, Custom Compiler provides a graphical palette of predefined routing patterns that designers can choose from. Users can drive the router by guiding it using their cursor on the layout. It comes with automatic connection cloning, pin taping and via generation. There is also interactive DRC and obstruction feedback.

The key to moving from schematic to layout design closure is understanding layout parasitics quickly and accurately. Without this, rework effort can become considerable. Instead of having to wait until the design is LVS clean to run LPE and RC extraction, Custom Compiler’s schematic driven layout (SDL) flow gives layout engineers parasitics throughout the layout process. Before nets are routed, estimates are used. As nets are incrementally hooked up actual extracted parasitics are inserted for each one.

Early parasitics estimation for analog design

Even though the fully extracted design is not available until the end, enough information is available early in the process to help provide useful feedback. This is vastly preferable to waiting until the end of the layout process to get physical parasitic information. Synopsys has also been working on using machine learning to help improve prediction of parasitics for even better estimates earlier in the process.

I mentioned above that templates can be used to help drive placement. Dave and Avina talked about how existing designs can be mined to easily produce templates for device placement. Dave said that this is a favorite feature for a lot of users.

With the added complexity of advanced nodes, specifically with new complex design rules and the need to place or modify arrays of FinFET devices, automation of the analog layout process promises big gains in productivity and design quality. Dave and Avina argue that the time has finally come for the automation of analog designs. They understand that this will never be a “push the big red button” sort of thing but will instead be made up from numerous discrete capabilities that are easy for designers to integrate into their workflow.

More information is available through CICC in their educational session archives and also on the Synopsys web page for their custom design platform.

Also read:

Design to Layout Collaboration Mixed Signal

Synopsys Tutorial on Dependable System Design

Synopsys Announces FlexEDA for the Cloud!


Semiconductor CapEx Warning

Semiconductor CapEx Warning
by Bill Jewell on 04-27-2022 at 4:00 pm

CAPEX Growth 2022

Semiconductor makers are planning strong capital expenditure (CapEx) growth in 2022. According to IC Insights, 13 companies plan to increase CapEx in 2022 by over 40% from 2021. The largest CapEx for 2022 will be from TSMC at $42 billion, up 40%, and Intel at $27 billion, up 44%. IC Insights is forecasting total semiconductor industry CapEx at $190 billion in 2022, up 24% from $154 billion in 2021. 2021 CapEx was up 36% from $113 billion in 2020.

Could this large increase in CapEx lead to overcapacity and a downturn in the semiconductor market? Our analysis at Semiconductor Intelligence has identified points where significant increases in CapEx result in a downturn or significant slowdown in the semiconductor market in the following year or two. The chart below shows the annual change in semiconductor CapEx (green line on the left scale) and the annual change in the semiconductor market (blue line on the right scale). The CapEx data is from Gartner from 1984 to 2007 and from IC Insights from 2008 to 2022. The semiconductor market data is from WSTS. In the last 38 years, semiconductor CapEx growth has exceeded 56% six times (red “danger” line). In each of those six cases, semiconductor market growth has decelerated significantly (greater than 20 percentage points) in the following year. In three of the six cases the market declined the following year. In three of the years from 1984 through 2017, CapEx has exceeded 27% growth (yellow “warning” line) but been less than 56%. In each of these three years (1994, 2006 and 2017) the semiconductor market declined two years later.

2021 CapEx growth of 36% puts it above the warning ling but below the danger line. IC Insights current forecast of 24% CapEx growth in 2022 is close to the warning line. Increases in 2022 CapEx plans could put growth over the 27% warning line but is very unlikely to approach the 56% danger line. So, are we in for a market downturn in 2023?

A few factors may come into play to avoid the overcapacity/downturn cycle this time. Previous large jumps in CapEx have resulted from semiconductor companies chasing fast growing emerging markets. In 1984 it was PCs. In 2000 it was internet infrastructure. In 2010 it was smartphones. In each of these cases, the end market either declined the following year (PCs and internet infrastructure) or slowed (smartphones). In the current situation, semiconductor companies are trying to alleviate shortages, especially in the automotive market. Increasing semiconductor content in vehicles is driving demand for semiconductors. Automotive companies fell behind in semiconductor procurement when they cut production during the pandemic beginning in 2020. In the current case, the demand for automotive semiconductors is not likely to weaken anytime soon.

Another factor is most of the current growth is coming from non-memory companies. In previous cycles, memory companies have been a major driver of CapEx growth. With DRAMs and flash memory primarily commodity products, they are more prone to over-supply and price declines in downturns. In 2021, memory companies grew CapEx 33%, similar to the 38% growth for non-memory companies. In 2022, memory companies are more cautious; we estimate 7% growth in CapEx. With this estimate, non-memory companies CapEx growth would be 36% in 2022. Most of the non-memory products are non-commodity and the companies are more closely linked to their end market customers.

CapEx growth should not be looked at in isolation. Absolute levels of CapEx relative to the semiconductor market give an indication whether CapEx is too high. The graph below shows semiconductor CapEx as a percentage of the semiconductor market on an annual and five-year average basis. Over the last 38 years, from 1984 to 2021, CapEx has averaged 23% of the semiconductor market. The five-year average ratio has ranged from 18% to 28%. The ratio has been on an uptrend for the last several years, with the five-year average reaching 27% in 2022 based on forecasts from IC Insights and WSTS. This data indicates the ratio may be close to a peak, indicating lower CapEx in the near future.

Our conclusion is the increase in CapEx should lead to caution, but not to panic. There is no indication of an end demand bubble, such as with the PC and internet infrastructure. Most of the growth is driven by non-memory companies, which largely produce non-commodity products. But the CapEx growth in 2021 and 2022 should be of concern based on historical trends. Our current forecast for the semiconductor market is 15% growth in 2022 and 5% to 9% in 2023. At the low end, 5% growth in 2023 would be a 21 point drop from 26.2% growth in 2021. This would fit the model, with the 36% CapEx growth in 2021 above the 27% warning line and leading to an over 20-point growth rate decline two years later in 2023.

Also read:

Electronics, COVID-19, and Ukraine

Semiconductor Growth Moderating

COVID Still Impacting Electronics


Podcast EP74: A Tour of the DAC Engineering Tracks with Dr. Ambar Sarkar

Podcast EP74: A Tour of the DAC Engineering Tracks with Dr. Ambar Sarkar
by Daniel Nenni on 04-27-2022 at 10:00 am

Dan is joined by Dr. Ambar Sarkar, a member of the Design Automation Conference (DAC) Executive Committee and platform architect at Nvidia. Ambar and Dan explore the new Engineering Tracks at DAC – their purpose and noteworthy content. Topics such as the cloud, global supply chain and silent hardware errors are discussed, along with details of the popular Poster Gladiator competition.

The views, thoughts, and opinions expressed in these podcasts belong solely to the speaker, and not to the speaker’s employer, organization, committee or any other group or individual.


ML-Based Coverage Refinement. Innovation in Verification

ML-Based Coverage Refinement. Innovation in Verification
by Bernard Murphy on 04-27-2022 at 6:00 am

Innovation New

We’re always looking for ways to leverage machine-learning (ML) in coverage refinement. Here is an intriguing approach proposed by Google Research. Paul Cunningham (GM, Verification at Cadence), Raúl Camposano (Silicon Catalyst, entrepreneur, former Synopsys CTO and now Silvaco CTO) and I continue our series on research ideas. As always, feedback welcome.

The Innovation

This month’s pick is Learning Semantic Representations to Verify Hardware Designs. This paper published in the 2021 NeurIPS. The authors are from Google and Google Research.

The research uses simulation data as training input to learn a representation for the currently covered subset of a circuit state transition graph. In inference, the method uses this representation to predict whether a newly defined test can meet new cover points, much faster than running the corresponding simulation. The architecture of the reported tool, Design2Vec, is based on a blending of Graph Neural Network (GNN) reasoning about the RTL CDFG structure and RNN reasoning about sequential evolution through the STG.

The paper positions Design2Vec as an augment to a constrained-random (CR) vector generation process. The method generates CR vectors as usual, then ranks these using a gradient ascent algorithm to maximize the probability of covering target cover points. The simulator then runs tests with highest predicted coverage.

The authors detail evaluations across a couple of RISC-V-based designs, also the Google TPU, and show compelling results in improving coverage over constrained random methods alone..

Paul’s view

This is a great paper, on a center stage topic in commercial EDA today. The paper studies two very practical opportunities to use ML in mainstream digital verification. First using ML as a rapid low-cost way to predict the coverage a test will achieve. And second using ML to automatically tune test parameters to maximize coverage.

On the first, the paper eloquently demonstrates that predicting coverage without understanding anything about the design (where the design is a black box) doesn’t work very well (50% accuracy across 3 testcases). However, if features derived from the design’s control-dataflow-graph (CDFG) are also fed into the predictor then it can work quite well (80-90% accuracy across the same testcases).

The way the CDFG is modeled in their neural network is very slick, building incrementally on other published work for modeling software program control flow in a neural network using a softmax function.

On the second opportunity, they compare their CDFG-based neural network with another tool that uses an entirely black box algorithm using Bayesian optimization. Here the results are less conclusive, showing data for only 1 testcase, and for this case showing only marginal benefit using the CDFG-based neural network over Bayesian optimization

Stepping back for a moment, I believe there are huge opportunities to use ML to improve coverage and productivity in digital verification. We are investing heavily in this area at Cadence. I applaud the Google authors of this paper for investing and sharing their insights. Thank you!

Raúl’s view

The authors address the problem of coverage, hard-to-cover branches and generating tests to cover them. Their approach is through training to predict whether a cover point is activated by an input test vector. The CDFG architecture is captured by 4 different graph neural networks, out of which an enhanced IPA-GNN (Instruction Pointer Attention Graph NN [6]) called RTL IPA-GNN works marginally best.

Design2Vec is also used for test generation for given cover points. The method uses predicted probability in a gradient-based search to maximize detection probability. Tests generated are run through an RTL simulator to get the actual coverage. Results comparing to Vizier [18], a Google tool using Bayesian optimization are not conclusively superior.

They ran coverage prediction experiments on two small RISC-V cores and a TPU. They look at points covered by10-90% of random tests to exclude trivial cases. The authors compare results against 3 methods: statistical frequency random patterns; a multi-layer perceptron (MLP) treating the design as a black box; an MLP with node sequence embedding, allowing generalization across cover points.

Design2Vec beats other approaches by 20% (against statistical frequency) to about 3% (against node sequence embedding). Notably MLP black box approach does worse than statistical frequency for the TPU (large design). In the words of the authors “the MLP performs catastrophically poorly on the test data”. For me, the main insight is that embedding the design in the architecture is key in building a semantic representation.

The authors stress the “potential of deep learning to make a quantum leap in progress in the area of verification”. Their results back this up.

My view

If you download this paper, you may notice that it is missing some appendices. The appendices are useful though not essential for full understanding. You might find this live presentation will bridge that gap.

Also read:

Cadence and DesignCon – Workflows and SI/PI Analysis

Symbolic Trojan Detection. Innovation in Verification

Leveraging Virtual Platforms to Shift-Left Software Development and System Verification