RVN! 26 Banner revised (800 x 100 px) (600 x 100 px)

How to Grow with Poise and Grace, a Tale of Scalability from ClioSoft

How to Grow with Poise and Grace, a Tale of Scalability from ClioSoft
by Mike Gianfagna on 06-23-2020 at 10:00 am

Screen Shot 2020 05 10 at 1.06.27 PM

ClioSoft published a white paper recently entitled Best Practices are the Foundations of a Startup. The piece discusses the needs and challenges associated with building a scalable infrastructure to support growth.

Before I get into more details on ClioSoft’s white paper, I would offer my own experience on this topic – the need to build a chip company with scale in mind is absolutely critical. I will draw on my experience at eSilicon.  I was one of the first employees at the company, so I had a front-row seat to watch the company’s growth.  And grow we did, from a modest mainstream ASIC suppler to an advanced 2.5D FinFET ASIC supplier, building some of the most complex chips in the world.

There are many parts of this story. I will focus briefly on just one – the compute infrastructure for chip design and tapeout. In the early days, eSilicon operated its own compute farm, first on-site and then at a co-located facility not far from its headquarters in the Bay Area. We owned the computers and managed the whole thing in house. For the types of chips we were doing, this strategy was predictable and effective.

As we began to grow and move from mainstream designs to cutting-edge FinFET and 2.5D designs, it became clear the “owner/operator” model was going to break. The advanced chips we were contemplating required at least 10X more compute resources, often a lot more than that near tapeout. We couldn’t afford to buy all that gear. And even if we could, hiring enough people to manage a facility like that would be daunting.

The company attacked the problem in two parts.  First, we outsourced our data center operation and the management of it to a large service provider who did that kind of thing all the time.  Our un-manageable capital and manpower problem was now a manageable operating expense problem. Later on, we saw the additional benefits of moving to the cloud, so we did that to further manage expense and allow massive on-demand bursts of compute during tapeout. By staying ahead of the need, our infrastructure was able to scale with the company. I’d like to give a shout-out to the long-time CIO at eSilicon who had the foresight to stay ahead of the curve – Naidu Annamaneni.

Back to the ClioSoft white paper. This discussion treats compute infrastructure scalability as well as design methodology scalability.  The reasons to adopt best practices are explained well. The piece also spends some time on the design management aspects of the problem.  This one has multiple dimensions. Storing and managing design data are part of it of course.

The white paper makes a compelling case for getting collaboration tools such as design data management correct at the beginning of a company’s life. The need for simplicity and agility are also addressed. To whet your appetite, here are some of the topics covered:

  • Naming conventions
  • Data storage and backup conventions and processes
  • Design flows and handoffs
  • Design management tools and methodology
  • Issue tracking and other collaboration tools
  • Project and schedule management

You can access the new ClioSoft white paper here. Happy reading.

ClioSoft was launched in 1997 by Srinath Anantharaman as a self-funded company, with the SOS design collaboration platform as its first product. The objective then, was to help manage front end flows for SoC designs.

The SOS platform was later extended to incorporate analog and mixed-signal design flows wherever Cadence Virtuoso® was predominantly used. SOS is currently integrated with tools from Cadence®, Synopsys®, Mentor and Keysight Technologies®. ClioSoft also provides an enterprise IP management platform for design companies to easily create, publish and reuse their design IPs called designHUB.

Today ClioSoft, driven by the experience and innovation of a number of engineers, is the market leader for design data and IP management solutions and the #1 choice for analog and mixed-signal designers.

Also Read

How to Modify, Release and Update IP in 30 Minutes or Less

Best Practices for IP Reuse

WEBINAR REPLAY: AWS (Amazon) and ClioSoft Describe Best Cloud Practices


Design Technology Co-Optimization (DTCO) for sub-5nm Process Nodes

Design Technology Co-Optimization (DTCO) for sub-5nm Process Nodes
by Tom Dillinger on 06-23-2020 at 6:00 am

scaled metal resistance

Summary
Design Technology Co-Optimization (DTCO) analysis was pursued for library cell PPA estimates for gate-all-around (GAA) devices and new metallurgy options.  The cell design and process recommendations are a bit surprising.

Introduction
During the “golden years” of silicon technology evolution that applied Dennard scaling, the tasks of fabrication process development and library circuit design were rather disjoint.  Physical and electrical models for scaled devices and interconnects were derived and IP design progressed while the process was being qualified.  The development of “contactless” local FEOL metallization and the transition to damascene Cu BEOL interconnects introduced some additional considerations – yet, process bring-up and IP design were still relatively distinct.

Several factors led to the need for a much closer collaboration between process and library development, a partnership which has been described as “design technology co-optimization” (DTCO):

  • The end of Dennard scaling (around 2006) required that the physical layout design rules for device and interconnect fabrication were each individually optimized. The drawn-to-electrical device dimension bias became an integral part of establishing circuit density and performance targets.
  • The slowing of supply voltage scaling meant that the reductions in device current and device power were not keeping up with the physical density improvements. (The increasing contribution of device sub-threshold leakage currents exacerbated the problem.)  Circuit library design needed to accommodate power distribution network (PDN) optimizations that addressed I*R and L*di/dt voltage drop concerns.
  • The quantization of device dimensions associated with the introduction of FinFET technology further complicated library IP physical and electrical design. Cell dimensions were established to provide for a discrete number of nFET and pFET fins.  Additionally, diverging markets results in the bifurcation of IP library design, into high-performance and high-density variants.

The figure below illustrates a “typical” 6-track (6T) cell definition, with four internal first metal signal wiring tracks and 2X-wide power and ground rails shared between abutting (flipped) cell rows for the PDN.

Figure 1.  Cross-section of 6T cell, in a FinFET technology.  (Source:  Synopsys)

In all these cases, DTCO was necessary for process and IP development (and for the subsequent introduction of enhanced process variants at the same physical node).

Recent DTCO Analysis

Fast-forwarding to the present…

At the recent VLSI 2020 Symposium on Circuits and Technology, a prevalent theme was the transition from FinFET to gate-all-around (GAA) “nanosheet” devices.

A team from Synopsys and Applied Materials provided an invited talk with compelling data from their DTCO analysis of GAA-based cell library design for sub-5nm nodes. [1]  I found their analysis to be extremely interesting, and their conclusion a bit surprising.

Specifically, in support of additional first metal dimension reduction, the team offered the following insights:

  • Cu wire scaling will be impeded by the need for a (relatively large) damascene barrier/seed layer.
  • New metallurgy will need to be considered for the first metal layer, either Cobalt (also damascene, thinner barrier layer) or Molybdenum (subtractive etch; or, Ruthenium).
  • Alternatively, wider Cu could continue to be used (specifically, for power/ground rails).
  • Via resistance will be subject to similar issues with current Cu metallurgy. New metal options will also need to be investigated.

The figures below illustrate the relative comparison between these metals, as the dimensions are scaled with advancing process lithography.

Figure 2.  Comparison of the resistivity of different metals with dimensional scaling, as a function of the cross-sectional area. 

Note that damascene-based patterning at these dimensions will result in more impactful metal grain boundary influence on the electron mobility.  Yet, Cu still outperforms the other metal options evaluated for large cross-sectional area.

Figure 3.  Comparison of the via resistance of different metals.

The DTCO team analyzed several different GAA-based cell topologies, evaluating their PPA characteristics.  (Models for a three-layer GAA device with 13nm channel length and varying widths were developed by the team – these models were then applied for the cell design analysis.)  The figure below depicts the six topologies considered.

Figure 4.  Cell design alternatives.

The DTCO experiment parameters included:

  • Evaluating Cu vs. Mo for signal wires (10nm wide, 20nm pitch)
  • Adjusting GAA nanosheet device widths
  • Exploring different cell track height options (5T, 5.5T, 6T, 6.5T, 20nm pitch)
  • Defining multiple PDN options:
    • Wide Cu power rails (larger cell size, less second metal stitching density required, measured as a multiple of the contacted poly pitch)
    • Narrow Cu power rails (much tighter second metal stitching required)
    • Thicker Cu power rails w/Mo signal wires (requiring a unique process to combine multiple metallurgies and extend the signal wire metal vertically)
    • Introducing Mo buried power rails, with (optional) connectivity to smaller Cu power rails in the cell

The figure below details the four options A, B, C, and D.  Options A and B continue to use Cu as the metallurgy.  Designs C and D use thicker Cu for power rails, with additional processing to introduce a different metal for signal wires.

Figure 5.  DTCO analysis cell design options, maintaining Cu as the power rail metallurgy.  C and D use a “deeper” Cu power rail.  The figure includes an additional VLSI 2020 Symposium reference that describes the process for incorporating composite interconnects for power (Cu) and signal wires (Co), as in options C and D.

The figure below illustrates the cell design incorporating a buried power rail (BPR), with Mo (or Ru) as the metallurgy.

Figure 6.  DTCO analysis design options incorporating a buried power rail (Mo or Ru metallurgy).

An important facet to this analysis is the need to pursue the physical implementation of a large, complex block design, to evaluate:  local pin access issues, overall routability (with the requisite PDN stitching density), power, and performance.  The DTCO team completed the design of a GPU core for the six cell library options.  The PPA results are shown below (as iso-power) – the 6T cell design serves as the baseline (“B”, 120nm tall, Cu signal wires, dense 30nm Cu power rails).

Figure 7. DTCO results for the six cell design options.

The “champion” of this PPA optimization analysis is configuration “C” – a 6T cell (120nm high) with a “deep” Cu first metal power rail and an alternative metallurgy for signal wires.

Note that the cost of the additional steps to provide for a composite metallurgy at first metal is not included in the analysis, nor were the steps to introduce the buried power rails.

I found the result to be of interest, as it demonstrates the significance of DTCO analysis.  At advanced process node dimensions, the choice of (damascene or subtractive etch) metallurgies, the requisite patterning, and the GAA device design width are all strongly interrelated.

Of particular note were the results of the BPR cell designs, “E” and “F”.  Note that the GPU block design area was minimal for these choices, suggesting that this may be the best option for embedded SRAM arrays.  (Indeed, another presentation at VLSI 2020 provided a compelling argument for the bit cell area reductions afforded by using BPR for ground distribution, with VDD remaining on first metal.)

It will be intriguing to see how the foundries and their customers address the varied material and patterning DTCO options for sub-5nm nodes, for logic, SRAM, I/O’s, and mixed-signal IP.  Of course, cost will be another critical factor in the ultimate selection.  For more information on the Synopsys DTCO flows and these recent results, please follow this link.

-chipguy

References
[1]  Moroz, V., et al., “Can We Ever Get to a 100nm Tall Library?  Power Rail Design for 1nm Technology Node”, VLSI Symposium on Circuits and Technology 2020, paper JFS3.2.

Images supplied by the VLSI Symposium on Technology & Circuits 2020.

Also Read:

Webinar: Optimize SoC Glitch Power with Accurate Analysis from RTL to Signoff

The Problem with Reset Domain Crossings

What’s New in CDC Analysis?


CEO Interview: Deepak Kumar Tala of SmartDV

CEO Interview: Deepak Kumar Tala of SmartDV
by Daniel Nenni on 06-22-2020 at 10:00 am

SmartDV CEO Interview 2020

SMARTDV is one of the biggest small EDA companies in the industry today in regards to products, customers and number of licenses in use, absolutely. They have a portfolio of more than 600 Design & Verification Solutions, everything from Design & Verification IP to Formal Verification IP, Post-Silicon Verification IP & Synthesizable Transactors.

Semiwiki first encountered SmartDV in 2015 and in my experience verification and IP go together like peas and carrots so let’s get a quick update. You can also spend quality time with SmartDV at this year’s Virtual DAC, sign up HERE. It is free if you love DAC!

Who is SmartDV?
We are a group of more than 250 experienced ASIC and SoC design and verification engineers based in Bangalore, India, dedicated to offering the largest selection of high-quality Design and Verification IP solutions. Our U.S. headquarters is in San Jose, Calif.

When did you form SmartDV?
We formed in 2007 and are celebrating our 13th anniversary this month. While our product portfolio expanded beyond providing verification IP only into a complete line of design and verification solutions, our goal to be a committed and trustworthy IP vendor who provides on-demand support remains the same.

Why did you start SmartDV?  What opportunities drove you?
We saw an opportunity to move a strictly transactional business to a service-oriented business where one size IP does not fit all SoC design groups. Servicing users is an important part of our business model since many project groups are in need of tailored, customized IP.

Initially, we served an unfulfilled need for quality verification IP, an essential piece of any verification strategy, given that verification consumes approximately 70% of a project schedule.

Our projections proved to be accurate. The IP provider business has matured into a large, viable segment of the electronics supply chain and a multi-billion dollar business. The revenue, according to the most recent ESD Alliance Market Statistics Service (MSS) news release, totaled $900.6 million in Q4 2019, a 4 percent increase compared to Q4 201. The four-quarters moving average increased 10.1 percent.

What is SmartDV’s core strength?
When I think of our core strengths, I think first and foremost about our exceptionally talented engineering group with expertise that spans design and verification. That expertise, commitment and responsiveness to our user community as well as our broad product portfolio is what makes our customer service and support stand out from our competition. As a result, we have a proven track record with a large, repeat user base and our solutions are used in hundreds of design projects throughout the global electronics industry.

Our engineering group continually delivers up-to-date design and verification IP products that gives us the largest, most extensive solutions offering that has grown over the last 13 years. Our products range today from Design and Verification IP, Formal (assertion) Verification IP, Post-Silicon Verification IP to SimXL, a portfolio of synthesizable transactors for accelerating system-level, SoC testing on hardware emulators or FPGA prototyping platforms.

A point of pride for us is our proprietary automated compiler-based technology for rapid development and deployment of new design and verification IP to support new industry standard protocols. The compiler gives us the ability to customize any of our design and verification solutions to meet specific customer design needs. This ensures quick delivery of products compliant with standard protocol specifications for new or evolving networking, storage, automotive, bus, MIPI, display and defense and aerospace applications. Consequently, we often deliver first-to-market design and verification solutions simultaneous to a new industry protocol standard’s availability.

It’s for these reasons that SmartDV Technologies has the reputation for being the Proven and Trusted choice for Smart Design and Verification Solutions.

What’s new with SmartDV?
We reached a milestone last month with availability of more than 600 Design and Verification solutions, an achievement that gives all of us great satisfaction. I credit our experienced and tireless development group, our active participation in standards organizations and our proprietary compiler. It gives us the competitive advantage to quickly produce and market a host of products to support the chip design and verification process.

Our most recent product news is delivery of a series of video, imaging and entertainment system design IP compliant with a variety of standard protocol specifications, including:

  • V-by-One, a high-speed serial video interface for HDTV
  • VESA DSC (Display Stream Compression), a video compression and decompression standard
  • HDCP 2.3 (High-Bandwidth Digital Content Protection) used to encrypt and authenticate digital signals for copyright-protected media, including movies, TV shows and audio
  • HDMI CEC (Consumer Electronics Control), a feature of HDMI that allows devices connected to HDMI to be controlled by just one remote
  • HDMI eARC (enhanced Audio Return Channel), an HDMI feature that enables high-quality digital audio to be sent back from the TV via HDMI
  • CXP (CoaXpress), a high-speed imaging standard for serial transmission of video and still images
  • SLVS-EC (Scalable Low Voltage Signaling with Embedded Clock), a high-speed serial interface scheme for image data transmission

What markets and applications does SmartDV support?
Our design and verification solutions are used in hundreds of audio, video and multimedia, automotive, communications, computing, networking, security, storage, memory, wireless and mobile chip projects.

Instead of applications, I’ll focus on differentiation. We offer a distinct advantage and unique capabilities unmatched by in-house resources. Our engineering group can verify the correct functionality and compliance with an industry standard of the production-proven core through a full test suite of functional coverage models. Each Verification IP block for emulation and FFGA prototyping, for example, comes as synthesizable RTL code and full API compatibility to move designs from simulation to emulation. Our experience from working with other users on previous tapeouts gives us the know how to find hard-to-find design bugs. Our various support options and customization can help meet a verification engineering group’s needs.

What trends is SmartDV tracking?
As you might expect, we actively track new industry standards or updates to existing industry standards so we can be first to market with our IP.

RISC-V is interesting to us, as it is to everyone else in the semiconductor industry and it will bring change to the verification landscape. What will change is unknown, which is why we’re tracking it so carefully.

Also Read:

Fractal CEO Update 2020

CEO Interview: Johnny Shen of Alchip

Tortuga Logic CEO Update 2020


Seeing is Believing, the Benefits of Delta’s Low-Resolution Vision Chip

Seeing is Believing, the Benefits of Delta’s Low-Resolution Vision Chip
by Mike Gianfagna on 06-22-2020 at 6:00 am

Screen Shot 2020 06 15 at 10.16.13 AM

Presto Engineering recently held a webinar discussing vision chip technology – what a vision chip is, what are the applications and how can you optimize its use.  Samer Ismail, a design engineer at Presto Engineering with deep domain expertise in vision chip technology was the presenter.  Samer takes you on a very informative journey about image processing and where vision chips fit.  At first glance, a “low resolution” vision chip sounds like a way to compromise a design. In fact, it is a way to optimize machine vision applications.

I will take you through some of the insights offered by Samer during the webinar. I highly recommend you view Delta’s entire low resolution vision chip webinar here, the entire event is 40 minutes with an excellent Q&A session – you will learn a lot.

The topics covered in this webinar are as follows. I’ll cover a bit of detail on each one to whet your appetite.

  • What is a vision chip?
  • Why low-resolution vision chip?
  • Vision chips in standard CMOS process
  • Working principle
  • Vision algorithms
  • Applications
  • Q&A

First, what is a vision chip? It’s NOT just a CMOS image sensor. Rather, a vision chip has the ability to capture an image (with a CMOS image sensor typically) and also perform analysis on that image with a combination of analog and digital circuits to extract information about the image. This reminded me of edge vs. cloud processing. Getting the processing closer to the data source has some significant advantages and that’s what is going on with a vision chip. Information on the ubiquity of vision chips was surprising to me. You’ll need to watch the webinar to judge for yourself.

Why use a low-resolution vision chip?  Simply put, latency, power, cost and area all benefit from using a low-resolution device. Think analyzing a 64 X 64-pixel image vs. a one-megapixel image. The benefits can be substantial if your application can fit a low-resolution profile. The types of applications that benefit are discussed during the webinar.

Typical high-resolution image sensors use a specialized process, one that produces bigger and more expensive sensor dies. These processes typically don’t support non-volatile memory. So, if you plan to capture an image and use an embedded processor on the same die to analyze it, this is going to be difficult without embedded memory. If you are in the low-resolution domain, these problems go away since you can use standard, lower cost manufacturing processes. There are several other benefits of a lower cost process as well, described in the webinar.

An architectural overview of Presto’s new Heimdal 2 vision chip is then provided. The elements of the architecture, its flexibility, capabilities and features and how to apply the device to various vision tasks are all discussed. Samer goes into substantial detail here. The Heimdal 2 is available on an evaluation board and an example application using the device is presented. A methodology to use Heimdel 2 to develop vision applications is also reviewed by Samer, using well-known algorithms as examples. This part of the webinar is a very good tutorial on image processing algorithms and how to implement them in hardware, with a special focus on the low-resolution applications and benefits.

Samer ends his presentation with two use case examples of a low-resolution vision chip – finding a vacant parking space and monitoring shopping behavior in a store. The webinar concludes with a short, but useful Q&A session. If machine vision is of interest, I highly recommend you watch this webinar.


Embedded MRAM for High-Performance Applications

Embedded MRAM for High-Performance Applications
by Tom Dillinger on 06-21-2020 at 10:00 am

embedded memory requirements

Summary
A novel spin-transfer torque magnetoresistive memory (STT-MRAM) IP offering provides an attractive alternative for demanding high-performance embedded applications.

Introduction
There is a strong need for embedded non-volatile memory IP across a wide range of applications, as depicted in the figure below.

The future scaling of embedded non-volatile flash memory IP is ineffective at more advanced nodes.  Several alternative memory technologies have been pursued as a “flash replacement” – e.g., phase change memory (PCM) materials, resistive change memory (RRAM), spin-transfer torque magnetoresistive memory (STT-MRAM).  These technologies offer dense bit cells (“1T1R”) and operate by changing the static electrical resistance of the cell as a result of the “Write1” and “Write0” pulse current and magnitude through the material.  A read operation senses the resistance magnitude when the cell is accessed, with much reduced cell current.  The ratio between the two resistances is ideally very high, to accelerate the read operation.

As a replacement for embedded flash, these technologies are evaluated against a number of criteria:

  • Non-volatility measures – i.e., operating temperature range, data retention (very temperature-dependent)
  • Bit density
  • Bit cell resistance ratio
  • Write access time, read access time
  • Array write granularity
  • Low power
  • Endurance (reflected as the # of R/W cycles before a bit error rate threshold is exceeded)
  • Additional fabrication complexity (i.e., cost)

The embedded memory technology that has the fastest adoption rate is currently STT-MRAM.  A cross-section of the “magnetic tunnel junction” (MTJ) of the bit cell is illustrated below.  The cell consists of two ferromagnetic layers separated by a thin tunnel oxide.  The magnetic polarization of the “free layer” is altered by the direction and magnitude of the write current.  The electrical resistance through the layers differs greatly, whether the free layer polarization is “parallel” or “anti-parallel” to the reference layer.  Previous semiwiki articles have described the operation of the STT-MRAM in detail. [ Refs. 1, 2]

The magnetic and electron tunnel layers for the STT-MRAM are readily fabricated and lithographically patterned.  The MTJ satisfies the typical embedded flash requirements, as listed in the first figure, with the great benefit of highly granular addressability.

For the set of embedded memory applications listed above demanding very high performance and endurance, these attractive characteristics of STT-MRAM technology will require ongoing R&D investment.

STT-MRAM for High Performance and Endurance

At the recent VLSI 2020 Symposium, a team from GLOBALFOUNDRIES introduced a novel high-performance STT-MRAM offering. [3]   A new MTJ material layer stack was developed, to optimize the read access time and concurrently (and significantly) extend the number of endurance cycles.  (Additional process engineering focus was also given to tighter CD lithography.)

The overall specifications for this new STT-MRAM IP are given in the table below:

Note that the endurance target for high-performance applications is defined using a bit error rate (BER) limit of 1E-06 (1 ppm) with a 10nsec write pulse.

The engineering tradeoff for this high-performance offering is that the retention property is reduced to 10sec @ 125 degrees C, as the MTJ energy barrier for this new high-performance cell is much lower compared to an eFlash-like replacement design.  (The retention of 10sec @ 125C equates to ~1 week @ 85C.)  This will necessitate a low-overhead refresh cycle.

A couple of interesting engineering optimizations were added to the array implementation by the GLOBALFOUNDRIES team:

  • An adaptive operating voltage is applied to the MTJ array.
  • The read sense amplifier is “trimmed” for optimal performance.

For example, the voltage bias for both the Write1 and Write0 current directions is adapted to respond to an internal temperature sensor.  The required Vop is higher at lower temperature – e.g., +10% at -40C (and -16% at 125C) compared to 25C.  This is due to the higher “coercive” magnetic field at low temperature that has to be overcome to alter the polarization.

The following figures highlight the technology qualification data that GLOBALFOUNDRIES presented at the symposium.

The first figure shows the median BER for two different MTJ material stacks with different write pulse widths, as a function of Vop.  Stack “C” was optimized for a single write pulse of 10nsec.  (Note that longer current pulses and/or multiple pulses, potentially including an intermediate read-verify operation, improves the BER.)

The figure below illustrates the BER for a read access cycle for stack “C” at 125C, with sense amp trimming.

STT-MRAM Reliability

Reliability evaluations were undertaken to ensure no adjacent bit cell “disturb” fails.

The endurance specification for the STT-MRAM array required development of an MTJ lifetime model, using BER data taken at accelerated voltage and temperature conditions.  (The time required to exercise sufficient array data would be prohibitive, necessitating the accelerated stress technique commonly used for other failure mechanism models.)  The GLOBALFOUNDRIES team noted that there is extensive model history for the time-dependent dielectric breakdown (TDDB) of conventional device gate oxides, but as yet, little modeling history of MTJ lifetime breakdown mechanisms.

The result of the endurance data (for 1ppm BER) is the reliability model illustrated below:

The figure illustrates the model extrapolation to >1E12 endurance cycles at -40C, using the most aggressive write cycle pulse of 10nsec.  For higher temperatures than -40C (lower Vop) and greater write pulse widths, the number of endurance cycles for this optimized MTJ stack will be much greater.

With extensive R&D engineering, the GLOBALFOUNDRIES team has demonstrated a novel MTJ materials stack, providing a “high performance” variants of an STT-MRAM array.  Whereas the initial IP offerings of this technology provide an attractive replacement for non-volatile eFlash, this new technology pushes STT-MRAM into an extremely competitive position for the NVRAM applications described in the opening figure.

For more info on the STT-MRAM technology from GLOBALFOUNDRIES, please follow this link.

-chipguy

References

[1]  https://semiwiki.com/semiconductor-manufacturers/tsmc/283868-tsmc-32mb-embedded-stt-mram-at-isscc2020/

[2]  https://semiwiki.com/semiconductor-manufacturers/samsung-foundry/5960-stt-mram-coming-soon-to-an-soc-near-you/

[3]  Lee, T.Y., et al., “Fast Switching of STT-MRAM to Realize High Speed Applications”, VLSI 2020 Symposium, paper TM3.3.

[4[ Naik, V. “A Reliable TDDB Lifetime Projection Model Verified Using 40Mb STT-MRAM Macro at Sub-ppm Failure Rate to Realize Unlimited Endurance for Cache Applications”, VLSI 2020 Symposium, paper TM3.4.

Images supplied by the VLSI Symposium on Technology & Circuits 2020.

Also Read:

Webinar on eNVM Choices at 28nm and below by Globalfoundries

GLOBALFOUNDRIES Sets a New Bar for Advanced Non-Volatile Memory Technology

Specialized Accelerators Needed for Cloud Based ML Training


Uber: Pariah to Paragon

Uber: Pariah to Paragon
by Roger C. Lanctot on 06-21-2020 at 8:00 am

Uber Pariah to Paragon

For years, the lords of Lyft and Uber have declaimed their intention to vanquish car ownership and displace public transportation. It really was as simple and as blunt as that. For sure there would be collateral damage including rental car companies and taxi operators and millions of under-compensated drivers – but the bottom line was crush, kill, destroy.

It was a simpler time when taxi operators hated Uber and Lyft (and their ilk) as did airports, cities, and rental car companies. All the while, Uber soaked up billions in fares and force fed those fees into software development focused on optimizing routing and logistics resulting in a trove of patent filings and, in a remarkable turning point for the company, the signing, this week, of a public transport agreement with Marin County in California.

In about a decade’s time Uber has transformed from transportation industry pariah to potential panacea. Uber has made peace with airport operators who have created massive dedicated ride hailing pickup areas transforming traffic flows at airports around the world. And Uber has co-opted taxi operators as sub-contractors in multiple markets where regulations required this shift.

Cities, too, have turned to Uber (and other operators) to serve areas under-served by existing mass transit options either chronologically or temporally – the buses can’t run all night long, after all. In fact, Uber has recast its messaging around transit painting itself as a supportive partner.

From Uber’s Website: https://www.uber.com/us/en/community/supporting-cities/transit/

“Ridesharing can not only help people get to their nearest transit stop but also:

  1. Fill in gaps in existing public transit service
  2. Provide access to underserved communities
  3. Alleviate the demand for parking
  4. Reduce costs of underused routes or services”

This is a very different song for Uber. It raises the stakes. In fact, it changes the game.

The Uber app’s reach, familiarity, and ease of use for both booking and paying are able to overcome multiple pain points for public transit operators. Uber has integrated public transportation options within its app in multiple cities around the world but the Marin County deal is unique.

Uber (and other ride hailing operators) has built their market capital on their ability to rapidly consolidate local ad hoc transportation activity in the form of millions of rides and billions of dollars creating overwhelming clout with adjacent service providers – such as public transit and airports.

In Marin County, Uber will manage rides for public minibuses – leveraging the Uber platform’s ability to match riders travelling in the same direction. Rides will cost $4 per mile, or $3 for those with disabilities or other mobility issues, with the fee going directly to Marin Transit. Uber will not collect a commission, instead charging the authority a flat monthly rate for the next two years, totaling no more than $80,000 over that period, according to terms reported by the Financial Times.

This first deal appears to be a camel’s-nose-under-the-tent proposition and marks a critical turning point as cities around the world emerge from lockdown fearful that their public transportation systems will be permanently and negatively impacted by COVID-19 concerns among normally reliable riders. Transit operators have tried, with mixed success, to foster their own app-based transit access.

Integration with Uber – or other regional ride hailing companies – now appears to be an essential element to expanding ridership – while re-opening and recovering from the COVID-19 outbreak.

Uber, of course, has taken matters further by taking on the back end routing and logistics in Marin County – not normally a strength of local transit operators. Uber is uniquely adept at managing the integration of ad hoc and scheduled transportation services.

Uber can put to work its network effect to help cities optimize public and private transit options. There are multiple ironies in the effort as it will simultaneously draw transit users to the Uber app and Uber users to transit. In all likelihood both Uber and transit operators will benefit.

The Marin County deal has the potential to upend Uber’s original nihilistic approach of laying waste to all non-Uber transportation alternatives. It also has the effect of signaling to Google, HERE, Intel/Mobileye (which recently acquired transit app Moovit), and regionally dominant ride hailing operators – Grab, Gojek, Gett, DiDi Chuxing, and Yandex – that this is the new path forward.

It further sets the stage for Uber and its like to capture the mobility-as-a-service flag with integrated transportation options, payment, logistics, and, as in Marin County, the back-end systems.  All those years that taxi operators, rental car companies, airports, and cities were crabbing and complaining about Uber’s disruptive and sometimes illegal activities, the company was laying the foundation to fundamentally transform and take charge of transit networks – maybe.

It’s no wonder that Uber is fighting strenuously against Los Angeles’ Mobility Data Sharing initiative. It’s clear now that Uber’s crown jewels are its patents and logistical knowhow – and its data. It’s also clear that the key to managing urban transit lies in data – and ride hailing operators around the world are decoding cityscapes every day.

Uber is battling MDS on privacy grounds, but the battle for ownership and control of data is more existential for cities and for ride hailing operators. What is shaping up is a war over the hearts, minds, and wallets of the travelling public. Uber brings a tattered reputation to the battlefield, but it is important to understand that the one-time pariah is increasingly seen as a paragon of transit virtue.


DVCon 2020 Virtual Follow-Up Conference!

DVCon 2020 Virtual Follow-Up Conference!
by Daniel Nenni on 06-19-2020 at 6:00 am

DVCon 2020 Logo SemiWiki

As most of you know DVCon 2020 was our first conference to be cut short by the Pandemic. SemiWiki bloggers Bernard Murphy, Mike Gianfagna, and I were there with full schedules but at the last minute it was called off. It really was an eerie feeling, the emptiness of it all.

The rest of our EDA live events followed suit and went virtual which is the new normal. A nice thing about virtual conferences, like webinars, is that you can go back and watch the replays as time permits and that is exactly what DVCon has done for all interested parties (for a limited time).

The program has been available to registered attendees exclusively from May 26 – June 16 and will open to the public from June 17 – August 14.

 “I’m thrilled that we are continuing the 2020 program online with many of the presentations that were unable to be presented during the live conference,” stated Vanessa Cooper, DVCon U.S. Technical Program Committee Chair. “Due to the rapidly evolving environment at the onset of the conference, we were able to pivot the live program quickly and compress if from four days to three, but that still left many without the benefit of the informative, technical material. I’d like to thank the technical program committee for continuing its efforts and coordinating with the presenters to record some of their papers, posters, tutorials and short workshops for this on-demand experience.”

Each session provided registered attendees with the ability to post publicly viewable questions to a forum that were answered by presenters throughout the first three weeks of the online experience. The forums are no longer active but you can still see the Q&A, which quite frequently is the best part, absolutely.

DVCon is sponsored by Accellera Systems Initiative (Accellera) which has a dedicated landing page on SemiWiki HERE.

Please visit the DVCon U.S. 2020 virtual conference website for program details and to access the virtual conference.

Save the date:  DVCon U.S. 2021 will be held March 1-4.

About DVCon
DVCon is the premier conference for discussion of the functional design and verification of electronic systems. DVCon is sponsored by Accellera Systems Initiative, an independent, not-for-profit organization dedicated to creating design and verification standards required by systems, semiconductor, intellectual property (IP) and electronic design automation (EDA) companies. For more information about Accellera, please visit www.accellera.org. For more information about DVCon U.S., please visit www.dvcon.org. Follow DVCon on Facebook https://www.facebook.com/DVCon or @dvcon_us on Twitter or to comment, please use #dvcon_us.

Also Read:

Accellera Tackles Functional Safety, Mixed-Signal

Functional Safety Comes to EDA and IP

Accellera IP Security Standard: A Start


Talking Sense with Moortec: Staying on the right side in worst case conditions – Power (Part 1)

Talking Sense with Moortec: Staying on the right side in worst case conditions – Power (Part 1)
by Tim Penhale-Jones on 06-18-2020 at 10:00 am

Tim Penhale Jones

In this first part of a 2-part blog series, we look at defining worst case conditions, focusing specifically on device power.

With great power, comes great responsibility…

With each new technology node especially FinFET, the dynamic conditions within a chip are changing and becoming more complex in terms of process speeds, thermal activity and supply variation. Dennard Scaling brought about the ability for power to be scaled down with each successive node so that power per unit area stayed roughly constant. However, as highlighted by John Hennessy at last year’s AI Hardware summit, since the mid-2000s this is no longer the case and we have seen the steady increase in power density per unit silicon area. Hennessy made the point that with Dennard scaling ending and Moore’s Law slowing down, transistor power and costs were no longer heading in the right direction and there’s no free ride for future performance just from process developments.

Worst case is getting worse!

What this means, is that chips have the propensity to run hotter and in-chip voltage drops are getting bigger. These two factors of increased process variation and the end of Dennard Scaling combine to mean the worst case is definitely getting worse! In addition to worst case performance which we will cover in the second part of this blog, SoC designers are being forced to focus on worst case power and voltage drop scenarios. To address these issues, it is no coincidence that the majority of FinFET SoC designs include a fabric of sensors for in-chip process, voltage and temperature monitoring.

Worst case power is not just about the maximum power dissipation although that is naturally a good starting point.  It is also about bursts of activity which cause temperature cycling and power differences which cause temperature gradients across the chip. FinFET processes require particular attention for potential hotspots as not only do they offer fantastic logic densities with the associated increased power per unit area, but their 3D fin type structures are not great at dissipating heat. Ideallystrategies need to be implemented to reduce maximum hotpot Tj (junction temperatures), as these impact lifetime and leakage current, they are also needed to reduce temperature gradients and cycling which impacts reliability. The trend with very large FinFET SoCs is to embed tens of temperature sensors to monitor potential hotspots around the chip, or alternatively, to use the recently launched Distributed Thermal Sensor (DTS) from Moortec.

Strategies employed for thermal management range from simple thermal cutoff where some, or worst case, all of the circuitry is switched off or ramped down if a certain temperature is reached, to more sophisticated DFS and DVFS schemes where the operating point and power in terms of clock frequency and supply voltage can be controlled and dropped to a lower level. Thermal load balancing involves allocating up-coming tasks to processors based on the level of their free processing capacity and their temperature. In all these cases an accurate temperature sensor provides the benefit of delaying the point at which action needs to be taken and therefore ensuring maximum processing power is maintained as long as possible. Less accurate temperature sensors require a larger temperature guard band (Check out our previous blog to learn more) which means for AI chips the processors will be switched off or to a slower throughput mode, at an earlier time and that’s not good for AI.

Associated with worst case power are worst case currents which cause IR voltage drops on chip. Particularly difficult to predict in advance are changes in voltage drops due to step changes in workload. The large SoCs are invariably software driven, but how the end customer will program these chips and how their worst case workload profile looks, is not always clear.  Including voltage and temperature monitors onchip especially for critical blocks gives visibility of the on-chip conditions and how these change with different workload profiles.

Multiple potential hotspots & temperature gradients ?

SoC development teams are faced not just with resolving traditional worst case timing issues but also worst case power. The latter can lead to multiple potential hotspots, temperature gradients and also difficult to predict voltage drops across large SoCs. Embedding a fabric of accurate in-chip monitors on SoCs provides excellent visibility of on-chip conditions.

This is seen as an essential tool for bring up, characterization and optimization on a per die basis especially for SoC development teams who are pushing the limits on advanced FinFET nodes but who want to stay on the right side in worst case conditions. As the old saying goes…’with great power, comes great responsibility’ and this is certainly the case when it comes to managing power conditions on advanced node devices.

Look out for the second part of this blog series entitled “Staying on the right side in worst case conditions – Performance” which will be available early Julywhere we will look at defining the worst case in terms of chip performance where timing analysis is key!


The Moving Target Known as UPF

The Moving Target Known as UPF
by Tom Simon on 06-18-2020 at 10:00 am

UPF hierarchy

As if engineers did not have enough difficulty just getting everything right so that their designs are implemented functionally correct, the demands of lowering power consumption require changes that can affect functionality and verification. Techniques such as power gating, clock gating, mixed supply voltage, voltage and clock scaling are commonly used to reduce dynamic and static power consumption of ICs. Each of these methods require changes to the design in ways that can affect functionality. Thus power reduction methods must be something that can be specified and even more importantly verified, both in the context of whether they work correctly but also such that they do not adversely affect chip operation.

Starting back in 2006 the IEEE took on the challenge of providing a standard to help the process of reducing power consumption with their IEEE 1801 specification. The first version of IEEE 1801 also known as the Unified Power Format (UPF) specification was published in 2007. It would be nice to say that this specification worked perfectly and remains little changed to this day. Sadly, the industry has encountered a steep learning curve regarding what makes an optimal power specification format. Commensurate with this, adoption of 1801 has been chronically low. To deal with issues in the 1.0 version, the 2.0 version followed fairly quickly in 2009.

Each newer versions of IEEE 1801 has worked to improve the specification. Usability, clarity and ambiguity have all be improved. The latest version is 3.1. It differs from the previous versions in both syntax and semantics. The challenge for designers today is that IEEE 1801 is used in design specification, power analysis and implementation tools, as well as in functional implementation and verification tools. In short, there are a lot of touch points for IEEE 1801. Things get messy when among the tools used in the flow engineers encounter support for differing versions of the spec.

But it gets worse, legacy designs may have used earlier versions, muddying the question of which version to use across a new design. Likewise, externally sourced IP, which may be a black box, might lack visibility or insight so that IEEE 1801 can be used effectively. Optimistic designers might hopefully wish that UPF directives such as “upf_version” will help to sort things out. However, “upf_version” is really only a suggestion to tools indicating what syntax was used to write the directives. It by no means is a get out of jail card for designs built with earlier versions.

There is some help in the form of useful guidance in a white paper written by Madhur Bhargava from Mentor, a Siemens business, where many of the key differences between the versions are discussed and strategies for dealing with them are offered. The white paper, titled “UPF 1.0, UPF 2.0, UPF 2.1, UPF 3.0, and now UPF 3.1: Which is the right design standard for my design?”, helps by offering some clarity to the changes over time.

First off the paper gives its own summary of the evolution of UPF. There is a discussion on the newest UPF 3.1 that goes over the most important changes and provides historical context regarding the reason for the changes and previous handing, if any, of the issue being addressed. The paper has a detailed section on backwards compatibility, which necessarily covers both syntax and semantics – and they do need to be addressed both individually and together. While this is not light reading, and one could hope that in the end there was an easier story to tell, it is fortunate that major progress has been made and much learned. Anyone contemplating using UPF should read through this paper, as it offers a lot of useful information in one place in a well-organized format. If you are interested, the paper can be downloaded from Mentor’s website.


Cruise Controlling Its Destiny

Cruise Controlling Its Destiny
by Roger C. Lanctot on 06-18-2020 at 6:00 am

Cruise Controlling Its Destiny

The Information tells us that General Motors’ Cruise Automation self-driving car unit has acquired automotive radar maker Astyx. The move can be seen as a simple defensive move to preserve access to valuable radar technology along the path to realizing Cruise’s vision of a robotaxi infused future.

There are a variety of of wrinkles to this acquisition – with a veritable Shar Pei of implications for GM, for Cruise, and for the autonomous vehicle world. There are three key elements underlying the financial aspects of the deal – the specifics of which have yet to be acknowledged, disclosed or confirmed – cash on hand, valuation, and market cap.

Thanks to investors such as Honda and Softbank, Cruise has $5B-$6B to burn – and burn it will at upwards of $250M/quarter. Cruise’s valuation, following the latest round of investment, stands at $19B. This compares to GM’s market cap of $39B, Waymo’s valuation of $30B, and Alphabet’s market cap of $1T.

It is amusing to ponder the prospect of Cruise’s valuation potentially exceeding GM’s market cap. I have an image in my mind of Cruise founder and president Kyle Vogt, just back from running seven marathons on seven continents in seven days (look it up), strolling into GM CEO Mary Barra’s office and “telling her how it’s going to be” at GM now that he owns the place.

Don’t hold your breath for that.

Cruise is playing a high stakes game of chicken in its pursuit of the autonomous driving Holy Grail. It must maintain its valuation, while steadily making progress toward an actual value-creating proposition, while burning massive amounts of cash…but not TOO much. To that end, Cruise announced layoffs recently which were thought to include personnel from the prior acquisition of Strobe – a Lidar supplier.

The most notable aspect of the Astyx acquisition is that it is a purchase that GM itself would have never made – in a million years. Car makers typically prefer not to take possession of their suppliers. Car makers prefer to have smaller suppliers, such as Astyx, partner with existing so-called “Tier 1” suppliers. In fact, Astyx has existing relationships with ZF, Hella, Autoliv, and Nvidia.

In the race to autonomy, though, acquisitions of this sort are common as AV developers seek to own and protect their “software stack” – the core of their AV solution. Typically, though, these acquisitions have occurred at the Tier 1 level with acquirers such as Continental, Bosch, and Aptix, while car makers have made select “investments” or engaged in partnerships.

While the Cruise move on Astyx may bave been defensive, such acquisitions have an offensive nature as well. The Cruise/Astyx deal may disrupt some competing AV developers that have been working with Astyx. They may be forced to rethink those relationships. For Cruise, though, taking possession of Astyx’s development team might help accelerate Cruise’s progress toward autonomy.

In spite of Mary Barra’s frequent claims that Cruise is on track to meet its goals and hitting its milestones – the reality appears to be something quite different. The latest announcements surrounding the much-heralded Cruise Origin robotaxi entering production in 2022 suggests a timeline shifting inexorably into the future – as has happened at most other OEMs and independent developers working in the space.

The odds are as long as the stakes are high in the game that Cruise is playing. I still remember running into Waymo CEO John Krafcik on the floor of the Geneva International Motor Show three years ago where he noted that Cruise was racking up hundreds of thousands of miles on city streets while Waymo was notching millions of miles on highway, surface, and city streets.

The issue is critical to determining which approach to autonomy will scale the most rapidly. Waymo has made clear its long-term intention to offer an autonomous driving “kit” to makers of all types of vehicles operating in a wide range of environments while continuing to operate its own nascent all-purpose autonomous driving solution. Cruise appears to be committed exclusively to the robotaxi path.

The outlook for robotaxi development is grim. Even assuming Cruise is successful, at the end of the day it will have mastered San Francisco after years of development work. How much time, effort, and money will it take to tackle Los Angeles? New York? Shanghai?

The acquisition of Astyx could set the stage for a Cruise shift toward a more Waymo-like operational capability encompassing urban and highway operation. If so, it probably would make sense to abandon the Cruise Origin entirely.

If Cruise is bent on taking on highway driving, though, it will run headlong into GM’s in-house Super Cruise development efforts which are proceeding on an entirely parallel path with little or no cross pollination. GM’s Warren Tech Center team and Cruise are pursuing completely separate hardware and software stacks with very different visions of value creation. Only the GM in-house team has crossed the threshold into mass production.

Whether offensive or defensive, the Astyx acquisition does not appear to measurably advance the cause of vehicle automation in the short-term, nor does it appear to alter Cruise’s strategic plans either to expand its operational domain or to shift its go to market strategy to a Waymo-style AV kit approach. But if Cruise intends to maintain or enhance its current valuation it needs to show measurable progress in 2020.

The Astyx acquisition may simply reflect Cruise taking advantage of diminished post-COVID-19 valuations, and a minority Astyx shareholder’s (ZF) need for cash, to add to its development team and provide its own internal stimulus. It won’t fix a flawed business model. So Cruise may be in control, but it won’t wag the dog.