webinar banner2025 (1)

Is ARM TechCon Really Worth it?

Is ARM TechCon Really Worth it?
by Daniel Nenni on 11-05-2018 at 6:00 am

The ARM TechCon organizers are asking me what I thought of this year’s conference so here it goes. As you know I am a big fan of ARM TechCon and feel it is one of the better conferences for SoC design. This year however I noticed a big change in demographics. Maybe the location change had something to do with it but I definitely saw different faces this year. I also saw fewer faces. I’m sure the conference organizers will come out with some big numbers but in my opinion it was a down year in regards to qualified attendees, especially the ones I met in the exhibit hall.

I signed more than 200 Prototypical books in the S2C booth, mostly the first day but maybe 50 or so on day two. When I sign books I get to talk to the people and I always ask where they work and what they do. This year it was overwhelmingly software development versus chip design. In looking at the 200+ badges that were scanned by S2C you will only find a handful of actual chip designers. In fact, the number of people off the street just collecting give away stuff was inordinately high. I attribute that to the location, downtown San Jose versus Santa Clara, the high tech hub of Silicon Valley. Location location location.

According to the S2C sales person who covered the booth, only about 10 of the 200+ people who got books were qualified ASIC prototyping leads. In the other booths I talked to qualified leads were in the single digits so the book give away gave a bit of an edge. The cost of the booth and related fixings was about $10,000 which gets you to about $1,000 per lead. Of course, getting out there and being seen counts but it is very hard to financially justify.

That being said, my constructive criticism is to move ARM TechCon back to Santa Clara and reduce the cost of participation for emerging companies. The EDA and IP ecosystem has changed due to acquisitions. There are now a handful of very large and medium sized companies with many, many small companies. The landscape is changing yet the conferences fight to stay the same, not good.

One of the reasons we started SemiWiki was to give emerging companies a fighting chance against the bigger EDA and IP companies by providing a level playing field for branding and demand creation. Of course the big companies were quick to join and they do drive a significant amount of traffic for the greater good of SemiWiki. We are, however, still focused on helping emerging companies in all aspects of business (funding, sales, marketing, business development, acquisition, etc…) and will continue to do so, absolutely.

Last week we were at the Linley Processor Conference. Linley does a very nice job, in fact, I would say they do the best regional conference in regards to content and cost. The first keynote was standing room only at the Santa Clara Hyatt (my favorite location). Linley Gwennap did a great presentation on “Breaking Bottlenecks in Processor Design” which we will blog about later.


Connected Cars High Tech Low Demand

Connected Cars High Tech Low Demand
by Roger C. Lanctot on 11-04-2018 at 6:01 am

The latest reliability report from Consumer Reports which dumped on domestics and rained glory upon Asian imports highlighted a conundrum facing car makers. The more effort car makers put into innovation, the greater the risk of consumer confusion, system failure and lousy reliability scores.

The picture is even worse if one takes into account key areas of innovation sweeping the industry including: electrification, connectivity, autonomous operation and car sharing. The one attribute all of these value propositions have in common is low consumer demand.

Strategy Analytics surveys of consumer car buying priorities typically show an interest in reliability, fuel efficiency and safety while the industry responds with complexity. The technology laggards in Japan and South Korea walk away with the bouquets year after year.

Factor in the prevalence of smartphones and the onset of digital assistants and the bulk of high tech auto gear is regarded as irrelevant or prohibitively expensive by consumers. This means that all of the billion dollar investments that the auto industry is making in startups, acquisitions and in-house development are targeting low demand, high cost applications.

Two applications, in particular, stand out: connectivity and car sharing. Car makers feel compelled to embrace connectivity for safety reasons (eCall mandate in Europe for automatic crash notification), cybersecurity protection, software updates and vehicle diagnostics. But consumers are less than enthusiastic.

Given the high cost of adding an embedded wireless connection, auto makers are desperate to recover their investments with subscription-based services or marketing platforms. The most egregious example of this effort by car makers to directly monetize vehicle connections is GM’s Marketplace – a privacy violating, driver distraction quatch being foisted on unsuspecting consumers.

Marketplace is redundant and irrelevant. The auto industry is poised on the threshold of an app-less future. Driver and passenger interactions with on-board and off-board content, services and applications will be driven by AI-infused voice interfaces not on-screen icons.

Connectivity remains a hard sell because car makers continue to demand a price rather than including it in the cost of the vehicle. Even wireless carriers – such as Deutsche Telekom and Verizon – are losing interest in the subscription-based model and abandoning the connected car market to the likes of AT&T, Vodafone, KDDI and T-Mobile.

The other application with a weak market pulse is car sharing. Car makers and some rental companies have tapped into rivulets of consumer interest, but a true groundswell remains elusive. The muted response is readily attributable to the fact that the market leaders are ruled by their legacy car rental and vehicle sales operations.

There is a lot of talk of the rise of mobility services and transportation-as-a-service platforms – but market fragmentation and an attention deficit means crucial advertising and marketing dollars are directed toward selling and renting cars, not toward sharing cars. Where’s Clay Christensen when you need him?

The only creative destruction underway in transportation is coming from pure play ride hailing companies like Gett, Uber, Grab, Lyft and others. A simple message of a cheap, convenient transportation alternative translates into rapid adoption and market penetration.


After more than 20 years of car sharing (some might say 50 years) only 260,000 cars globally are available as part of car sharing programs large and small, free floating and station based. Is the sector ripe for transformation and rapid growth? Perhaps.

What could change the demand scenario for connected and shared cars? Two things. Car companies need to build in connectivity and make connectivity an intrinsic element of owning, sharing, renting or simply operating a car. The onset of 5G technology will help.

As for car sharing? It may take the rise of pure play car sharing startups like Mevo (Australia), Zazcar (Brazil) or Car2Go (Israel) to shake things up. Car buyers are telling car makers they’re pretty content with the status quo. It’s going to take a marketing jolt or a fundamental business model shift to alter the underlying market dynamics. If legacy car makers can’t deliver, startups (like Tesla) will step in – rather, are stepping in.


Q4 Hiring remains strong Outlook for Q1 2019

Q4 Hiring remains strong Outlook for Q1 2019
by Mark Gilbert on 11-04-2018 at 6:00 am

Things are going quite well in the Valley, ESPECIALLY if traffic is any indication. I still remember the BUST with the empty office buildings and light traffic, back in the early 2000’s…while I wish for traffic like that always, I certainly prefer the busy times, guess you cannot have it all. The more traffic, the more jobs; they invariably go hand in hand as a sign.

As for ARM Techcon, it seemed busy enough, though with the new layout and location made it somewhat difficult to compare to previous years. This year, the conference moved to the San Jose Convention Center, which is a much larger footprint than the Santa Clara CC, where it has been previously held, for many years. The show did seem larger and attendance was robust with most of the folks I chatted with telling me, they were having a good show. Daniel Nenni, famed top dog at SemiWiki, signed so many of his new book (his 6[SUP]th[/SUP] I think), that he would not even give me one for fear of running out…so I just took one when he wasn’t looking. I will make him sign it next time and then sell it on EBAY.

Needless to say that IOT devices, Software, Hardware and all the functionality and issues associated with them, are going through the roof. Security is essential, and we picked up a new client doing just that, and needing new people. Obviously, DAC is far more in my wheelhouse (with my 21[SUP]st[/SUP] coming up), but I must say that we are moving far quicker into IOT, than I could have imagined and quite happy about the add-on business. The reality is, there are several connections between EDA/Semi and IOT and not only do some of the skills often blend, but many people that left EDA, have moved into IOT. Most of the skill-sets are obviously different, but fortunately for me, I have had several friends who have moved over and given me the opportunity to help them find new people. Now that I have been doing IOT placements for the last several years, and having numerous placements under my belt, I understand the domain so much better and am learning more about this new area, every day. I think the same can be said for domains like AI and Machine Learning.

Hiring is still strong and the H-1 Visa situations (and there are many) have complicated hiring. Engineers on H-1s and thinking about changing jobs, have more to consider in this climate of uncertainty.

As for hiring, the reality is that right now, we are on the high side of what I call the PENDULAM SWING, and that swing changes every several years. I have seen many! The PENDULAM SWING is when either the momentum, advantage is on the Company or Candidates side…right now, the pendulum is at its highest point, with companies needing people and the candidates having more control of their destiny, without the pressure or fear that the opportunity in front of them, is their only option. Companies must take note of this and make sure they do not lose a viable candidate in hopes of finding someone better…it is quite the risk to take today with no guarantees if they will or how long that can take.

I was swamped in back to back meetings preparing for next year. By all indications, next year will start strong, as strong as this year is ending. I for one certainly hope so.


UMC the Next China Chip Embargo?

UMC the Next China Chip Embargo?
by Robert Maire on 11-03-2018 at 7:00 am

The DOJ last week released a previously sealed indictment against UMC, Jinhua and three individuals for stealing Micron trade secrets. The indictment is very damning of UMC with very specific semiconductor technology files that were stolen by UMC related people then passed on to Jinhua through its joint development with UMC.

DOJ UMC indictment link

The indictment clearly lays out that UMC and individuals were responsible for the actual theft and that Jinhua was the recipient of the stolen goods. Usually the thief is treated more harshly than the receiver of stolen goods. The only thing that has likely kept UMC off the embargo list for now is that they are Taiwan based and are probably cooperating as fast as they can for fear of being put out of business.

To be clear Jinhua/China is far from innocent as UMC was just their henchmen.

UMC is clearly frightened as they have immediately severed all ties with Jinhua and tried to distance themselves. They canceled the DRAM joint development. Continuing the joint development would be pointless anyway now that Jinhua can’t get tools needed from the US to get a fab to work.

We would not be surprised if the stolen Micron information finds its way to other Chinese fabs such as Yangtze. That would likely result in an immediate embargo of the new receiver of stolen goods.

Its also obvious that if UMC gets embargoed that they would be forced out of business at some point as it would be difficult to keep their fleet of US made tools operating in their fab.

However we would not be surprised to see UMC get added to the embargo as it makes logical sense to punish the thief as well as the receiver. After the US initiates the embargo, there will likely be a quick settlement similar to ZTE where UMC will pay a huge fine and be forced to change out management just like ZTE did.

In the meantime, shipments to UMC will be halted just like shipments to ZTE were halted for several months. Just like in the ZTE case, US companies will be hurt as they won’t be able to do business with UMC, so it will be in the interests of the US to settle as well after US equipment companies scream.

Even though UMC is not a big spender it could be yet another down tick of overall Capex going into 2019. We think the embargo weapon could be used after the election when the fallout doesn’t matter, so we would stay tuned.

The stocks
This is obviously a very, very big potential negative for UMC. Its a positive for Micron, and a positive for TSMC. Its obviously a negative for semiconductor equipment companies in the US and ASML as well. We remained concerned as to what China’s response will be in the escalating tit for tat, and then how the US will respond in turn. We don’t think we have seen the last Asian company to be embargoed by the US, we expect more to follow.


Playing the Long Game with 56G SerDes

Playing the Long Game with 56G SerDes
by Daniel Nenni on 11-02-2018 at 12:00 pm

IP has always been a hot topic on SemiWiki with quite a bit of our readers doing the “make versus buy” analysis. SerDes is one of the more difficult “makes” so it is mostly a “buy” analysis made by networking class ASIC architects and a handful of other applications that demand high performance and throughput. Based on the recent traffic patterns, data centers and the coming 5G infrastructure are spiking SerDes interest, absolutely.

The new SerDes twist is “long reach”, a term I was not completely familiar with so I got on the phone with my goomba Mike Gianfagna at eSilicon. He put me in touch with Hugh Durdan, Vice President, Strategy and Products at eSilicon and Matt Burns, Technical Marketing Manager at Samtec Inc. Samtec is a privately held electronic interconnect solutions company that enables the SerDes “reach beyond the rack” implementation – high-speed serial communication over an eye-popping five meters of copper cable.

“eSilicon will demonstrate true long-reach capability for the most demanding system environments in next-generation 25 and 50 Tb/s switches with Samtec Twinax Flyoverä Cable Assemblies. Developers can accelerate their time to market with a high-performance, flexible, easy-to-configure SerDes that targets all of the common Ethernet, Fibre Channel and CPRI standards as well as proprietary rates common in data center and 5G
wireless infrastructure applications.”


5-meter prototype Samtec EBCM Series ExaMAX® Backplane Cable Assembly

Q: What’s different about the SerDes you are announcing?

A: from eSilicon. There are several things that set this device apart. It is a full DSP (transmit and receive) 56G SerDes in 7nm with true long-reach performance. Many other so-called DSP-based SerDes are in reality a hybrid with a digital transmit path but an analog receive path, which impacts performance. We also offer exceptional control and configurability through a graphical user interface that talks to the firmware running on an embedded SiFive processor built into the SerDes. From the GUI, users can access all monitoring features such as non-destructive eye diagrams, SNR & BER, bathtubs, histograms and power measurements.

Q: Regarding this “reach beyond the rack” demo, why is this significant?

A: from eSilicon. Customers want to leverage existing infrastructure and drive higher and higher data rates through legacy backplanes. With our ability to achieve low bit error rates over very long channels we enable customers to extend the life of their existing equipment vs. having to do a “fork-lift upgrade”. Data center customers also want extreme flexibility in connecting systems to each other. Thanks to the highly robust design of our SerDes and Samtec’s cable products, we are able to offer 56G operation over five meters of copper cable.

A: from Samtec. At 56G data rates, trace lengths on a PCB get really short. There are numerous design techniques engineers can use to combat this: advanced PCB laminates, additional PCB layers or clock and data recovery and retimers. Samtec’s solution is to get the signal out of the noisy and lossy PCB into low-noise, ultra-low-skew twinax copper cable as soon as possible. Long length copper cable also enables a paradigm shift in system architecture in the data center, whether inside the rack or beyond that rack.

Q: Tell me a little abou Samtec as my readers may not be familiar with your company. What do you make, where does it fit in the end product?

A: from Samtec. Samtec has been the service leader in the connector industry since our founding in 1976. At our core, we continuously support our legacy flexible stacking headers and sockets. Additionally, we are expanding our high-performance interconnect portfolio into backplane, board-to-board, precision RF, optical engines and high-speed cable assemblies based on our Twinax Flyover™ technology.

Q: Tell me about the long cables. What’s involved in manufacturing them? Why are they so accurate?

A: from Samtec. One key factor is our manufacturing process. Our twinax cables are co-extruded with low Dk FEP dielectric using real-time closed-loop control. This high-precision process eliminates the inconsistencies of individually extruded dielectric twinax cabling. Our cables are grounded via twinax copper alloy, eliminating the need for a drain wire. When combined with our high-performance connectors, the results are 56G data rates over the longest backplane cable assemblies in the industry.

The demonstration will be conducted in the Samtec booth #521 at Supercompute in Dallas from Nov. 12 – 15. Sign up now for a private briefing HERE. And Matt Burns has a nice blog series HERE.

AbouteSilicon
eSilicon provides complex FinFET ASICs, market-specific IP platforms and advanced 2.5D packaging solutions. Our ASIC-proven, differentiating IP includes highly configurable 7nm 56G/112G SerDes plus networking-optimized 16/14/7nm FinFET IP platforms featuring HBM2 PHY, TCAM, specialized memory compilers and I/O libraries. Our neuASIC™ platform provides AI-specific IP and a modular design methodology to create adaptable, highly efficient AI ASICs. eSilicon serves the high-bandwidth networking, high-performance computing, artificial intelligence (AI) and 5G infrastructure markets. www.esilicon.com

About Samtec
Much more than just another connector company, Samtec puts people first with a commitment to exceptional service, quality products, and convenient design tools. We believe that people matter, and taking care of our customers and our employees is paramount in how we approach our business. This belief is deeply ingrained throughout the organization, and means that you can expect exceptional service coupled with technologies that take the industry further faster. www.samtec.com


Basic Techniques for Managing EDA Business

Basic Techniques for Managing EDA Business
by Daniel Nenni on 11-02-2018 at 7:00 am

The “20 Questions with Wally Rhines” series finale

During the period from 1997, I had the privilege to work with Greg Hinckley, a superb “out of the box” thinker and excellent operational manager. We had such a common view of the world that we almost always agreed upon the best approach to problems and opportunities and hardly needed to confer, although we regularly did discuss almost all significant issues. One of our common attributes was a contrarian attitude. If everyone else was headed in the same direction, opportunity must lie in a different direction. That “Do what others aren’t doing” served as a guidepost for our moves at Mentor.

Following are some of the approaches that Greg or I brought to Mentor, or that evolved from combining our previous experience through interaction together

  • Listen for silence
    • Divisions, product groups, people and managers are always vying for the attention of top management and will usually let you know the good things they are doing
    • If you never hear from, or about, an entity in your organization, there are two possibilities:
      • They are doing nothing worthwhile and make no difference to the company, in which case they should be eliminated, or
      • They are trying to hide something, in which case they should be investigated

       

     

  • Life never gets any easier
    • Ignoring the problems of today only forces you to deal with bigger problems tomorrow
    • Never hope that excess inventory will be any easier to sell in the future than it is today—-write it off now
    • Never assume that an overdue account will find a way to pay you in the future—-it’s even less likely than the present

     

  • What you do infrequently, you do poorly
    • Send an annual invoice, not a quarterly one, for your customers to renew their maintenance contract — the person who deals with it will have to depend upon you to tell him how to handle it
    • Restrict the unfavorable options in your contracts so that customers only deal with them at renewal time. If they have to deal with options regularly, they will become very good at it.
    • Don’t expect your own organization to become efficient at things they rarely deal with—try to automate those things or consolidate them with someone who deals with similar issues regularly

     

  • Making rational decisions based upon data is more important than being right all the time
    • Can you provide a rational basis for why you are making the decision?
    • If you turn out to be wrong, you can retest your rationale and make a better decision next time
    • You may be wrong but at least you’re rational

     

  • Budgeting starts at the top
    • Asking organizations to determine how much they need to spend in the coming year (or quarter, month, day, etc.) is a worthless exercise and a waste of company resources—–the answer will almost always be more than the company can afford
    • Budgeting must start with an objective sales and revenue decision, heavily influenced by those who will have to achieve the sales. Budgets can then be developed that will fit into that sales and revenue plan.
    • Never be tempted to increase the revenue forecast when the cost budgeting pressures appear insurmountable. It will be even more difficult to reduce the budget later when the revenue comes in at (or below) the realistic level that is lower than the optimistic revised forecast.

     

  • In a chip company, management needs to focus first on what is NOT working. In a software company, management should focus first on what IS working.
    • Problems in a semiconductor company (low yields, rapidly falling prices, excess inventory, etc.) can bankrupt you quickly; fixing them has immediate, measurable benefit.
    • In a software company, inventory is unimportant and incremental revenue nearly equals incremental pre-tax profit. Costs are largely fixed because they are mostly people-related and people should not be treated as a variable expense. Incremental revenue is everything so maximum effort should be focused upon using existing resources to produce additional revenue.

     

  • Spend your time doing what others are NOT doing
    • Strategies that are similar to what the other mainstream companies are doing will rarely develop any unique advantage for your customers
    • Needs that are not being addressed offer more opportunity than those that are receiving lots of attention

     

  • For an employee, it is a great success for management to adopt your idea and claim it as its own, even if you are not acknowledged as the creator. For management, it is a great success to see employees act upon your idea and claim it as their own, even if they do not acknowledge you as the source.
  • In technical software, users rarely change the product they are using until it becomes incapable of completing the required task—not when it becomes, slow, inefficient, hard to use or unpopular.
    • Benchmarking with competitive products results primarily in price competition for the customer to secure a lower price to continue using the same software
    • Companies that offer a new and improved product would be better off to use their time by:
      • Participating in benchmarks of new applications that can’t be handled by the older, competitive tool, or
      • Finding new applications for the tool where there is not already an incumbent supplier.

       

     

  • When you acquire an EDA competitor, you don’t eliminate competition—-you create it. In EDA, people are the competitors and they keep doing what they are good at doing by creating new companies whenever the one they are working for is acquired.
  • A new product or service is not really profitable until you receive your first order from a customer with whom you’ve never interacted
    • As long as orders come only from direct effort and interaction by your own organization, you are expending too much money for the revenue you are achieving
    • When word of mouth and third party support start generating orders, you’ve gone viral and have a winning, profitable product

     

  • If you only play defense, your competitor will eventually score.
    • You must counterattack to make your competitor play defense
    • Selling your new, innovative products to your existing customers may be a starting point but success comes only when you penetrate new customers who are usually already served by your competitor
    • Statistically, your competitor will eventually get lucky if he keeps attacking your accounts and is not burdened with defense of his own accounts if you don’t attack them.

     

  • When your company is not number one is a specific market, always compare yourself to the leader(s). When, or if, you become number one, never compare yourself to anyone.
    • Market communication is all about consideration to purchase. You want to be considered when purchasing decisions are made so the association with the leader enhances the possibility of customer comparison of the leader’s product or service with yours.
    • If you are number one in a market, it’s counterproductive to provide your customer with a list of competitors to evaluate, even if you think your product is superior.

     

This is not meant to be a complete formula for successful EDA management. It’s a collection of ideas that aren’t normally discussed in business management training. Yet I believe that these observations, combined with other standard good management practices and an unwavering commitment to ethical treatment of customers, employees and shareholders, helped Mentor to recover from a difficult period and ultimately demonstrate leadership and success in a difficult high technology business.

The 20 Questions with Wally Rhines Series


Designing Integrated ADAS Domain Controller SoCs with ISO 26262 Certified IP

Designing Integrated ADAS Domain Controller SoCs with ISO 26262 Certified IP
by Camille Kokozaki on 11-01-2018 at 12:00 pm

As new automotive Advanced Driver Assistance System (ADAS) based product releases intensifies while a more stringent set of safety requirements are mandated, it is not surprising that subsystem and electronic suppliers are looking for pre-designed and ISO 26262 certified IP that can address both imperatives of schedule and safety when being integrated into SoCs. Add to that the necessity of lower power and higher performance, you begin to look at newer architectures that minimize footprint, risk, and effort and maximize the performance at the same time.

Traditionally, the electronic control units (ECUs) for individual ADAS applications have been placed throughout the car: the forward collision avoidance ECU located in the windshield, park assist ultrasonic sensors, and the processor in the rear. ECUs integrate the multiple ADAS applications into centralized domains to combine multiple ADAS functions. The new class of integrated domain controller ECUs utilizes data transferred from the car’s remote sensors such as cameras, LIDARs, radar, ultrasonic, and other sensors to the integrated domain controller for processing by a high-performance ADAS system-on-chip (SoC).

The IP in the integrated ADAS domain controller SoC must also meet the highest Automotive Safety Integrity Levels (ASILs), must be designed and tested for grade 1 and 2 temperatures, and must fully adhere to the automotive quality management process. In addition, to meet the power and performance requirements of the new integrated ADAS domain controller SoC architecture, designers are moving to more stringent process technologies, such as FinFETs, making it even more important to use automotive-certified IP in advanced foundry processes.

A Shift to Integrated ADAS Domain Controller SoC Architectures

According to the August 2016 Traffic Safety Facts Research Noteby the National Highway Traffic Safety Administration (NHTSA), “the nation lost 35,092 people in crashes on U.S. roadways during 2015, a 7.2% increase which is the largest increase in nearly 50 years.” It was analyzed that about 94% of those accidents were caused by human error, and the rest by the environment and mechanical failures.

The opportunity to reduce car accidents is making automotive ADAS even more critical. Automatic emergency braking, pedestrian detection, surround view, park assist, driver drowsiness detection, and gaze detection are among the many ADAS applications that assist drivers with safety-critical functionalities to reduce car accidents. Figure 1 shows an integrated ADAS domain controller SoC with a centralized ECU where data from numerous sensors travels to a central ECU and is then processed via an ADAS processor.


Figure 1: Data from sensors travel to a central ECU and are processed via a vision processor

High volumes of data are driving the adoption of 64-bit processors for automotive ADAS applications. The shift from a distributed architecture to a more centralized ECU is more prevalent, and since the ECUs are integrated, the ADAS SoCs are becoming very complex, requiring the latest semiconductor features, semiconductor process technologies, along with other technologies for ADAS domain controller SoCs:

  • Ethernet manages high data volume including time-sensitive data and reduces point-to-point wiring
  • LPDDR4/4x operates at data rates of up to 3200 megabits per second and beyond, which speeds up the DRAM operations in automotive-grade SoCs
  • MIPI standards like MIPI Camera Serial Interface and Display Serial Interface provide high-performance connectivity in imaging and display applications
  • PCI Express has high-reliability processor-to-processor connectivity for 4G radios or the future 5G radios and external SSDs
  • 5G and IEEE standards, like 802.11p, help provide real-time updates of maps or images to and from the Cloud, and vehicle-to-vehicle or vehicle-to-infrastructure communications
  • Security protocols in hardware and software for data protection to and from connectivity via USB, WiFi or Bluetooth
  • Sensor and control subsystems offload the host processor and fuse sensor data to manage the different type of sensor data provided by the sensors
  • More advanced manufacturing process technologies from the traditional 90-nanometer (nm), 65-nm and 40-nm to more advanced 16-nm, 14-nm, and even 7-nm FinFET processes

Safety-critical applications are significantly increasing the adoption of ADAS SoCs. However, it is required that the ADAS SoC along with all semiconductor components including the IP that is integrated into the SoC meet the ISO 26262 functional safety standard.

Meeting ISO 26262 Functional Safety Standard Requirements

ISO 26262 is a standard that defines the impacts of failures in automotive systems at four different Automotive Safety Integrity Levels (ASILs): A, B, C, and D; ASIL D is the highest level of functional safety. The ISO 26262 standard defines all the processes, development efforts, and standards that automotive development organizations must implement and comply with when developing products for safety-critical systems. A key objective of ISO 26262 standard is to minimize the susceptibility to all types of random hardware failures, including permanent failures or transient failures, by:

  • Defining the functional safety requirements when developing products
  • Applying rigor to the development process
  • Defining a safety culture
  • Implementing safety features to minimize the impact of hardware failures
  • Assessing and analyzing the impact of safety features to ensure mitigation of hardware failures

Industry-accredited inspection companies, such as SGS-TUV Saar, are available to audit products and processes for compliance and certification of ISO 26262.
The ISO 26262 certification process includes multiple steps, policies, and reports and must start from the very beginning of product development. For example, the Failure Mode Effect and Diagnosis Analysis (FMEDA), a report that development teams generate, provides all the information regarding the adherence to ISO 26262 from a functional safety perspective.

Created by design and verification engineers, the report is a critical component of an ASIL assessment, not just for evidence of compliance but also for design targets and a rating assessment at the end of the development flow. Designated safety managers, separately from the development organization, who are fully trained to monitor the development process, milestones and product reviews, ensure all the documentation and traceability is completed throughout the SoC development flow as defined by the standard.

The FMEDA report also includes a summary of the safety features, their development, and verification. It clearly documents the safety features contained in the products and how these products react to the random faults that are injected into them. The FMEDA report is mandatory and is given to all parties involved in the product review process.

How ISO 26262 Certification is Implemented

A standard SoC or IP product development flow starts with register-transfer level (RTL) design, which is then implemented, verified, and validated in hardware and software in the final prototypes. An ISO 26262 compliant development adds additional steps over the standard design process including at the very start when defining a core architecture and specification. Designers define a safety plan that includes safety features and goals. The product team and safety manager review the safety plan and strategy to achieve the designated functional safety for the end application.

It is important to conduct a failure analysis by injecting faults to assess the safety level and the system’s reaction to those faults. The FMEDA shows a fault injection analysis for both permanent and transient faults to assess the impact. The analysis and assessments are clearly documented in the FMEDA report as part of the ISO 26262 certification process along with the safety manuals. This entire process is shown in Figure 2.

Figure 2: An example of a standard SoC or IP design with additional ISO 26262 certification steps and requirements

The safety manual in the ISO 26262 certification process defines the safety features in the product, which is critical to the operation of the product. The standard provides some guidelines as to the effectiveness of safety features that can detect possible failures. Safety features for IP product design fall into three categories: protection mechanisms, replication, and various.

  • Protection mechanisms, such as protecting the interface between the IP in the SoC architecture with the protection of elastic buffers, parity protection on the data path and configuration registers, and error correction code protection for both writes and reads.
  • Replication is a safety feature category that includes duplicating (or triplicating) key modules and using voting logic to ensure redundancy.
  • Various includes parity checks for all the state registers, single cycle pulse validity, various dedicated interrupts, and hot state machine protection for bad states.

The process to meet ISO 26262 functional safety certification is stringent from creating the FMEDA report, designating a safety plan that defines safety features for the target ASIL, to employing a safety manager and documenting and reviewing every milestone with all the stakeholders. In addition to meeting ISO 26262 functional safety requirements, integrated ADAS domain controller SoC development teams and the rest of the supply chain, including the design IP provider, must adhere to automotive reliability and quality requirements.

To meet the automotive reliability standard as defined by the automotive industry, automotive SoCs and IP must be designed and tested to meet very low defect densities which are measured by Defects Parts Per Million (DPPM). The automotive industry has a requirement for less than one DPPM, encouraging designers to set a goal of zero defects per million throughout the automotive product lifetime of 15 years.

Meeting temperature grade is another reliability requirement. For ADAS, the highest level of operating temperature is Grade 1 which requires up to 125 degrees Celsius ambient or 150 degrees Celsius junction temperatures. Each company within the automotive supply chain has a proprietary temperature mission profile to which they design and test their products. SoC and IP designers who are developing products for the different ADAS applications take the temperature mission profiles into account during the development process. Different requirements such as electromigration, transistor aging, and transistor self-heating must be considered against the temperature mission profile for the different devices.

Synopsys offers a portfolio of automotive-certified IP that is ASIL Ready ISO 26262 certified, designed, tested for grade 1 and 2 temperatures, and fully adhere to the automotive quality management process. For information view the DesignWare IP for automotive SoC web page.Based on a technical bulletin posting By Ron DiGiuseppe, Automotivie Segment Marketing Manager, Synopsys


Wi-Fi Standards Simplified

Wi-Fi Standards Simplified
by Bernard Murphy on 11-01-2018 at 7:00 am

In the world of communications, the industry fairly quickly got a handle on a naming convention for cellular technology generations that us non-communication geeks could understand – 2G, 3G, 4G and now 5G, (though some of us could never quite understand the difference between 4G and LTE, at least as those terms are widely and no doubt inexpertly used). This is a nice steady progression, easy for us uncultured masses to remember, with no confusing affixes.

Bluetooth has from the beginning followed a largely similar convention in generation naming from 1.0 up to current BT 5 which drops the ”.0” on the name, not coincidentally aligning numerically with the latest cellular standard. There is a sub-generation of 4.0 (at least initially), known widely as BLE (Bluetooth Low Energy) but we generally understand that this is a low-energy variant on the underlying standard.

Wi-Fi on the other hand soldiered on in geekdom, oblivious to the needs of the masses, by sticking a bewildering array of postfixes on the root 802.11 standard name. I guarantee that no-one in the general public would have any idea what you were talking about if you asked them about say 802.11n, whereas many would profess at least some familiarity with 5G, if with the name only. The Wi-Fi Alliance recognized this wasn’t great for marketing and has recently switched to a much easier naming convention, at least for the most recent generations.

For the newer generations of the standard, 802.11n becomes Wi-Fi 4, 802.11ac becomes Wi-Fi 5 and 802.11ax becomes Wi-Fi 6. For 802.11a/b/g I hear differing stories. Fortune magazine says that these names won’t change. Another contact says (plausibly) that these will be known as Wi-Fi generations 1, 2 and 3. Now isn’t that easier to understand? You may not know what they are but all of us can understand that as you progress from Wi-Fi 1 to Wi-Fi 6, you get better technology at each stage, for which you are prepared to shell out more money (see, that’s marketing).

We’re a bit more sophisticated than that, so what are 4-6? Wi-Fi 4 (aka the “n” version) supports transmission rates 5-6 times faster than the Wi-Fi 3 version, higher reliability and supports MIMO (multi-input, multi-output) where multiple antennae at receiver and/or transmitter further boost reliability and transmission rates.

Wi-Fi 5 (aka the “ac” version), introduced in 2014, increases rates by a factor 2 or more and more efficiently uses bandwidth so more users can be served at speed in one network. However, more efficient usage is only in the downlink. So you can all watch cat videos with minimal buffering but the network bogs down if you all want to share with friends. Wi-Fi 6 (aka the “ax” version), expected to be ratified by the end of 2019, goes one step further, packing multiple users more effectively in both downlink and uplink, serving an expected 4X improvement in throughput in high user density environments (think of a stadium). It also offers a 25% improvement in peak data rate.

So Wi-Fi generation naming is now understandable and mostly aligned with cellular and Bluetooth naming. There’s an apparent disconnect in that the latest Wi-Fi is Wi-Fi 6, whereas cellular is at 5G and Bluetooth is at BT5. But the official Wi-Fi 6 is a year away so perhaps numeric synchronization isn’t too far off. When we’re checking out phones or other devices, we should reasonably soon be able to look for all communications to be at level “N” (maybe 6?) no matter what the underlying technology. That will make life a lot simpler, certainly for me.

I have to thank Franz Dugand, Sales and Mktg Director for Connectivity at CEVA for these insights. Naturally CEVA has a wide range of Wi-Fi IPs across these standards, including Wi-Fi 6, ranging from low-power to high-performance to multi-gig rates. CEVA have been in the Wi-Fi core licensing business since 2002 so they’re very well known and established in the space. You can learn more about their RivieraWaves Wi-Fi platforms HERE.


Architecture for Machine Learning Applications at the Edge

Architecture for Machine Learning Applications at the Edge
by Tom Dillinger on 10-31-2018 at 2:01 pm

Machine learning applications in data centers (or “the cloud”) have pervasively changed our environment. Advances in speech recognition and natural language understanding have enabled personal assistants to augment our daily lifestyle. Image classification and object recognition techniques enrich our social media experience, and offer significant enhancements in medical diagnosis and treatment. These applications are typically based upon a deep neural network (DNN) architecture. DNN technology has been evolving since the origins of artificial intelligence as a field of computer science research, but has only taken off recently due to the improved computational throughput, optimized silicon hardware, and available software development kits (and significant financial investment, as well).

Although datacenter-based ML applications will no doubt continue to grow, an increasing focus is being applied to ML architectures optimized for “edge” devices. There are stringent requirements for ML at the edge – e.g., real-time throughput, power efficiency, and cost are critical constraints.

I recently spoke with Geoff Tate, CEO at Flex Logix Technologies, for his insights on ML opportunities at the edge, and specifically, a new product emphasis that FlexLogix is undertaking. First, a quick background on DNN’s.

Background

A “deep” neural network consists of multiple “layers” of nodes. At each node, a vector set of inputs is provided to a computational engine. The output of each node is further refined by a (potentially non-linear) activation function calculation, which is then forwarded to the nodes in the next layer. The final layer provides the DNN decision from the original input set – i.e., a “classification” result of an input image to a reference set of objects.

Figure 1. Illustration of a simple DNN, with 3 “hidden layers”. The computation at each layer is a matrix multiplication of the input vector and a matrix of weights.

Numerous DNN topologies are used in practice – the figure above depicts a simple, fully-connected multi-layer 2D design. (More complex “3D” topologies and implementations with feedback connections in the hidden layers are often used, which are optimal for specific types of inputs.)

Each node in the DNN above performs several computations, as shown in the figure below. At each node in the layer, a set of weights are multiplied against the input value, then summed – i.e., a “multiply-accumulate” (MAC) calculation. An (optional) bias value may be incorporated into the sum at each node. The MAC output is input to a normalizing “activation” function, which may also incorporate specific parameter values – activation function examples are illustrated below.

Figure 2. Expanded detail of the calculation at each node in a layer, and some examples of activation functions.

During the DNN training phase, a reference set of inputs is applied. The selection of the initial weights, biases, and activation parameters at each node is an active area of research, to optimize the training time. (The simplest method would be to pick values at random from a normalized distribution.) The input reference set proceeds throughforward evaluation and the DNN result compared to the expected output.

An error difference is calculated at the output layer. A backwards optimization phase is then performed, evaluating an error gradient dependence for the network parameters. Internal DNN values are then adjusted, and another forward evaluation pass performed. This training optimization iterates until the DNN classification results demonstrate acceptable accuracy on the input reference set.

The DNN values from training are subsequently used as part of the production inference engine, to which user data is now the input.

DNN in the Data Center

The initial DNN (training and inference) implementations in the datacenter utilized traditional von Neumann CPU and (DRAM) memory resources to perform the MAC and activation calculations. The DRAM bandwidth to the CPU core is typically the throughput bottleneck.

A transition to GPU-based cores for DNN calculation was then pursued, to leverage the SIMD dot product MAC calculations prevalent in GPU image processing. GPU’s have a drastically different architecture, with very wide internal vector datapaths – e.g., ~1024 bits wide. As a result, to improve core resource efficiency, a “batch” of inputs is evaluated concurrently – e.g., 32b floating-point DNN parameter values could be concatenated into a wide vector to evaluate a batch size of 32 inputs in parallel through the DNN layers. Yet, the local memory associated with each GPU core is relatively small (KB’s). Again, the (GDDR) memory bandwidth is a DNN performance-limiting factor.

New chip architectures are being aggressively pursued for ML applications – e.g., Google’s Tensor Processing Unit (TPU). And, due to the intense interest in the growing base of ML applications, numerous chip start-ups have recently received (initial round) VC funding – see the figure below.

Figure 3. Examples of ML hardware startups (from [1]).

Additionally, cloud service providers are deploying FPGA hardware to offer effective, easily-reconfigurable DNN capabilities. [2]

DNN’s using conventional CPU and GPU hardware architectures are throttled by the access bandwidth to retrieve the weights and biases for each layer evaluation. Training presents an additional data constraint as these parameter values are required to compute both the forward evaluation and the backward optimization error gradients, as well. As an example, the ResNet-50 DNN is commonly used as a reference benchmark for image classification, a complex (3D) 50-layer convolutional network topology. A forward pass evaluation utilizes ~26M weights. [3] Depending upon the data precision of these parameters, the memory bandwidth required to access these values for use in a layer computation is very high.

ML Applications at the Edge

My conversation with Geoff at Flex Logix was very enlightening. First, he shared some of the characteristics of edge applications.

“An ML application will typically pursue DNN training at the datacenter, and then transfer the DNN parameters to edge hardware for inference.”

“Often, a DNN hardware implementation quotes a peak throughput, in trillions of operations per second (TOPS), and a related power efficiency (TOPS/W). Yet, it is important to analyze what memory bandwidth and batch evaluation assumptions are used to calculate that throughput.”

“Edge customers will typically be utilizing (sensor) input data corresponding to ‘batch_size = 1’. Maybe a batch size of 2 or 4 is applicable, say if there are multiple cameras providing video frames per second input. The datacenter architectures that merge parallel input sets into large batch size DNN evaluations to optimize MAC efficiency just don’t apply at the inference edge.”

“High batch count increases overall classification latency, as well, as the parallel input set is being merged – that’s of little consequence for typical datacenter applications, but additional latency is not appropriate at the edge.”

I asked Geoff, “How is Flex Logix approaching this opportunity at the edge? What elements of the existing embedded FPGA technology are applicable?”

Geoff replied, “We have announced a new product initiative, NMAX. This architecture builds upon many of the existing features of our embedded FPGA, specifically:

  • a tile-based building block that is readily arrayed into an (m X n) configuration
  • a combination of logic LUT cell and MAC engines in a DSP-centric tile
  • a method for optimal embedding of SRAM macros of varying size between tiles (link)
  • a rich set of external connectivity options when embedded within an SoC design

A block diagram of a single “NMAX512” tile is illustrated below.

Figure 4. An architectural block diagram of the NMAX512 DNN tile, and an array of tiles depicting the L2-level SRAM between tiles.

Each tile contains 8 NMAX clusters. Each cluster contains 64 MAC’s using an 8b x 8b parameter data width (with options for 16b x 16b), with a 32b accumulate. There is a total of 512 MAC’s per tile. The programmable EFLX logic LUT’s perform the activation functions for the DNN layer. The weight and bias values for the layer are accessed from the local (L1) SRAM within the tile.

An embedded (L2) SRAM between tiles stores the intermediate DNN results and parameter values for successive layer calculations. New values are loaded into the L2 SRAM in the background during forward evaluation. The required data bandwidth for system DRAM memory is reduced significantly.

Geoff added, “The time to reconfigure the NMAX tile with new DNN layer with new DLL layer data (from L2) is very fast, on the order of 100’s of nsec.”

“How is the NMAX tile implementation for a DNN developed?”, I inquired.

Geoff answered,“ML developers utilize the familiar TensorFlow or Caffe languages to define their DNN topology. We will be releasing a new NMAX implementation flow. Users provide their TF or Caffe model, and the NMAX compiler fully maps the data and logic operations to the MAC clusters and reconfigurable EFLX LUT logic. All the sequencing of DNN layer evaluation is mapped automatically. The physical LUT placement and logic switch routing configuration is also automatic, as with a conventional embedded FPGA.”

Geoff continued, “Our preliminary performance models indicate we will be able to achieve ~1GHz clocking (TSMC 16FFC), or roughly ~1 TOPS throughput per tile(with LPDDR4 DRAM, L2 SRAM size optimized for the DNN). The distributed L2 SRAM helps maintain a very high MAC and activation function utilization.”

“Speaking of performance modeling, do you have NMAX benchmark data?”, I wondered.

Geoff replied, “We prepared the following data for NMAX compared to other architectures, such as the Nvidia Tesla T4, for the ResNet-50 DNN benchmark with Int8 parameters. Note that the NMAX architecture enables a wide span of tiled array sizes, with corresponding throughput scaling for the batch_size = (1, 2, 4) of greatest interest to edge customers. The initial MAC utilization and total power dissipation is much improved over other architectures, as well.”

Finally, I asked, “What is the NMAX deployment schedule?”

Geoff answered, “We are starting to engage partners now, in terms of potential NMAX sizes of interest. Our engineering team will be finalizing IP area/performance/power specs in 1H2019, as well as finalizing the NMAX compiler. A tapeout release with a specific tile and SRAM configuration will occur in 2H2019, to provide evaluation boards to customers.”

There is clearly a lot of activity (and VC investment) pursuing optimized DNN hardware architectures for datacenter applications. There is certainly also a large market for (embedded IP or discrete) hardware focused on the power/perf/cost constraints of the low batch-size ML applications at the edge. Flex Logix is leveraging their expertise in reconfigurable (DSP plus logic) functionality in pursuit of this opportunity.

It will be an interesting R&D area to follow, for sure.

-chipguy

References

[1]https://origin-blog.appliedmaterials.com/vc-opportunities-ai-developer-ecosystem

[2]Putnam, A., “The Configurable Cloud — Accelerating Hyperscale Datacenter Services with FPGA’s”,2017 IEEE 33rd International Conference on Data Engineering (ICDE),https://ieeexplore.ieee.org/document/7930129/ .

[3]https://www.graphcore.ai/posts/why-is-so-much-memory-needed-for-deep-neural-networks


Mentor’s Busy ITC and Major Test Product Updates

Mentor’s Busy ITC and Major Test Product Updates
by Tom Simon on 10-31-2018 at 1:00 pm

In conjunction with the 2018 International Test Conference, Mentor has several interesting test announcements. They also have a busy round of technical activities, including a number of technical papers, presentations, tutorials and a poster from a major customer about using Mentor. I’d like to touch on the two product related announcements, because they are pretty interesting.

There is probably no area other than automotive where reliability has received more focus lately. First off, zero defects are a prerequisite for ISO 26262. Also, the automotive semiconductor market is experiencing higher growth rates than almost any other sector. As such, automotive applications have become the reference for quality and reliability. It is pushing the development of improved methodologies and is setting the standard for the highest reliability.

Automotive is on a steep complexity growth curve, both historically and for the immediate future. It used to be that automotive semiconductors could leisurely stay behind the bleeding edge of technology. However, ADAS and new requirements for infotainment have brought automotive computing and data transfer requirements to the point where only the most advanced nodes will suffice. Going from Level 2 to Level 4/5 for ADAS will increase the number of sensor modules by around a factor of 5. Higher levels of automation will require more complex computational tasks. For example, AI algorithms may be used to anticipate pedestrian movements to help avoid auto versus pedestrian conflicts. All of this adds up to more complex chips and more of them, which will necessitate increased effort to ensure the highest reliability.

To address the needs of this market Mentor has added a new set of test patterns to deal with failure modes found in FinFET processes and 3D transistor structures. Mentor Tessent TestKompress already looks at each cell to determine areas that are vulnerable to defects. Now they are adding analysis that will look at cell to cell interactions looking for potential defect sources. Several Tessent customer are publicly reporting significant reductions in their DPPMs, in the ranger of 700 to 4300, by using the improved analyses offered in Mentor’s automotive-grade ATPG.

With the added automotive-grade test patterns it should be possible to replace many system level or functional test patterns. Mentor has added automated pattern generation that targets critical-area based interconnect bridges and opens, as well as cell internal and neighborhood defects.

Mentor’s other announcement concerns improving the efficiency of silicon bring-up. Mentor has created a networked connection between the DFT software and the testers themselves. They worked with Teradyne to interface UltraFLEX ATE to Tessent SiliconInsight. Because literally hundreds of IP blocks are being added into new SOC designs, IJTAG has seen strong adoption. With this comes the need for IJTAG debug tools. With the introduction of Mentor’s Interactive IJTAG, designers can get better insight into what is happening on the tester, right in their test software in real-time. Reduced iteration time can shorten bring-up from weeks to days. Interactive IJTAG speeds up the many complex mappings that are needed to generate the test program on the tester and then interpret the test results in a way that is meaningful to the designer.

The 2018 ITC features papers from Tessent customers discussing real world results with TestKompress Automotive-grade ATPG and SiliconInsight Interactive IJTAG. Mentor is vigorously involved in improving the state of the art in test. This shows in their ISO 26262 qualification for use on all ASIL ISO 26262 projects. System designers and end customers are the beneficiaries of their sustained efforts in test. There are more details on these new announcements and Mentor Test products on their website under TestKompress or SiliconInsight.