DAC2025 SemiWiki 800x100

Real-Time AI-driven Image Signal Processing with Reduced Memory Footprint and Processing Latency

Real-Time AI-driven Image Signal Processing with Reduced Memory Footprint and Processing Latency
by Kalar Rajendiran on 05-31-2023 at 10:00 am

AI ISP for Automotive, Low Light Image Enhancement

In our day to day lives, we all benefit from image signal processing (ISP), whether everyone realizes it or not. ISP is the technique of processing image data captured by an imaging device. It involves a series of algorithms that transform raw image data into a usable image by correcting for distortions, removing noise, adjusting brightness and contrast, and enhancing features. So, ISP by itself is not something new.

What is new though is leveraging artificial intelligence (AI) to enhance ISP to yield better results than traditional ISP can. Firstly, in the field of digital photography, AI can significantly enhance ISP capabilities. Traditional ISPs are effective at processing images, but AI can take this to the next level. For instance, AI can help in noise reduction, improving the image’s clarity, especially under low-light conditions. Moreover, AI algorithms can recognize various scenes or objects in the image, enabling automatic adjustments to different parameters like brightness, contrast, and saturation for optimal results.

In the world of autonomous vehicles, AI-enhanced ISPs can process real-time images to understand the vehicle’s surroundings better, aiding in decision-making. Traditional ISPs can struggle with different lighting conditions or object detection at high speeds, but AI can improve upon these aspects, enhancing the vehicle’s ability to react to potential hazards and improve overall road safety.

Lastly, in the realm of surveillance and security, AI-enhanced ISPs can process images from CCTV footage more effectively. AI can help detect suspicious activities, recognize faces, or identify objects left unattended, providing real-time alerts and enhancing overall security measures.

But how to implement AI-based ISP? It is easier said than done. There are a lot of challenges to overcome.  AI algorithms are fast advancing, requiring an AI-ISP solution to be programmable. AI models require a lot of computational power. Conventional  techniques combining separately developed ISP and NPU often require a lot of memory to store entire frames of images for processing. Accessing DDR-SDRAMs consume a lot of power. And this loosely coupled frame-based solution will introduce latencies measured in frames, which is unacceptable for many applications.

What is needed for today’s applications is real-time processing at low latency and low power consumption. Making the solution DDR-less is even more attractive as it will reduce the system power requirements significantly. Of course, NPUs are key to an AI-based ISP solution. But there is a lot more to arriving at a DDR-less, low latency, low power AI-ISP solution. This was the topic of Mankit Lo’s presentation at the recent Embedded Vision Summit conference. Mankit, who is the Chief Architect, NPU IP Development at VeriSilicon walked the audience through the various aspects that need to be addressed.

Solution Requirements

ISP Requirements

In traditional hardware for ISP, there are many modules in the pipeline stages to correct the potential artifacts of the imaging system. To perform AI-ISP, the chosen ISP should be flexible enough to allow the customer to pick and choose the modules to replace or enhance.

NPU Requirements

As ISP tasks are computationally huge and intense, the task is usually partitioned to be executed on many NPU cores. There is a lot of image overlap on the input side going into the NPU cores. Even for a 3×3 convolution layer inside the neural network, the overlapping requirement for just a few pixels could result in a huge overlap at the whole network level. The overlap needs to be minimized for reducing the memory, power, and computing demand on the system. The way to do it is through layer-level overlap sharing.

What is needed is an NPU that can handle raster lines and become part of the ISP processing pipeline, making the solution DDR-less, low latency, and low power. The NPU needs to be programmable to handle a changing AI network model landscape and deliver very good performance. It also needs to be able to support Per Layer Overlap Sharing which results in not requiring any overlap on the image input side.

VeriSilicon Offers

Specific to the topic of AI-ISP, VeriSilicon provides ISP, NPU, and the FLEXA-PSI interface to seamlessly connect these IPs. Refer to the block diagram of a VeriSilicon AI-ISP solution.

The customers need to just add their custom AI algorithms to the mix to complete their unique solution. VeriSilicon can also provide algorithms relating to the ISP such as noise reduction, demosaicing, and different types of detection such as facial detection and scene detection, etc.

Click here to learn more about VeriSilicon’s NPU IP offering.

Click here to learn more about VeriSilicon’s ISP IP offering.

VeriSilicon offers custom silicon services and a very broad portfolio of IP for many different markets. Learn more at VeriSilicon.com

Also Read:

VeriSilicon’s VeriHealth Chip Design Platform for Smart Healthcare Applications

VeriSilicon’s AI-ISP Breaks the Limits of Traditional Computer Vision Technologies


Investing in a sustainable semiconductor future: Materials Matter

Investing in a sustainable semiconductor future: Materials Matter
by Daniel Nenni on 05-31-2023 at 6:00 am

EMD LinkedIn Twitter Materials Matter

In 2020 TSMC established its Net Zero Project with a goal of net zero emissions by 2050. I remember wondering how could this possibly be done before 2050 or at all for that matter. After working with TSMC for 20+ years I have learned never to bet against them on any topic and green manufacturing is one of them, absolutely.

TSMC presented on green manufacturing at the recent symposium in Silicon Valley. Clearly energy and water resources are critical parts of any net zero project but carbon emissions are also a of great importance and that means materials. In regards to semiconductor manufacturing materials we have experts here on SemiWiki.

EMD Electronics has a presence in 66 countries and over 100 years of invaluable experience in the electronic materials space delivering a broad portfolio of semiconductor and display materials for cutting-edge electronics.

EMD Electronics recently published a paper: INVESTING IN A SUSTAINABLE SEMICONDUCTOR FUTURE – MATERIALS MATTER.

It is a very interesting comprehensive look at innovative sustainable semiconductor materials techniques that decrease carbon emissions, improve resource efficiency and productivity, and highly contribute to achieving net zero semiconductor.

“It is amazing where collaboration can take us. Sustainability is no longer the result of individuals; only by working together can we get closer to our goals for a sustainable future! Brita Grundke, Head of Sustainability, EMD Electronics”

Here is the introduction. This paper is freely available and well worth the read for semiconductor professionals at all levels:

Emissions from semiconductor manufacturing are a growing segment of global greenhouse gas (GHG) emissions. There are two reasons behind this trend. First, the demand for semiconductor chips is growing. Our technology appears in everything from mobile phones to automobiles, where the number of chips per vehicle increases every year. Data storage, which relies on the semiconductor industry, is exploding. Second, today’s manufacturing processes have more deposition and etch steps than ever. Each step consumes water and electricity and creates GHG emissions.

Semiconductor companies large and small talk about achieving climate neutrality by 2030. That sounds like a great goal. But merely achieving that goal won’t solve the emissions problem. How we get there matters.

Buying carbon offsets is an easy way out. It isn’t the best long-term answer, because many offsets are not as effective as they claim to be. Some may even make the problem worse, defeating the purpose [1]. And relying on offsets can make internal actions seem less pressing. It is best to see offsets as a temporary or last-choice option.

Semiconductor industry leaders are, of course, doing more than buying offsets. They are investing in renewable energy, improving the energy efficiency of their processes, and finding ways to reduce waste. These actions are helpful, and we must do more. Despite modest success in reducing emissions per wafer or per revenue, demand for semiconductor chips is growing faster than the improvements can handle. We need more drastic reductions, and that starts by examining the sources.

Bottom Line: Climate change is real and semiconductor manufacturing is under a microscope now that it is being regionalized due to the shortages and supply chain issues we suffered during the pandemic. If you really want to know why semiconductor manufacturing left the US it was due to the Environmental protection Agency crack down on water, ground, and air pollution. I grew up in Silicon Valley so I had front row seat to the environmental issues of semiconductor manufacturing. Now that semiconductor manufacturing is coming back to the US and other parts of the world sustainability is front and center once again.

Also Read:

Step into the Future with New Area-Selective Processing Solutions for FSAV

Integrating Materials Solutions with Alex Yoon of Intermolecular

Ferroelectric Hafnia-based Materials for Neuromorphic ICs


Coherent Optics: Synergistic for Telecom, DCI and Inter-Satellite Networks

Coherent Optics: Synergistic for Telecom, DCI and Inter-Satellite Networks
by Kalar Rajendiran on 05-30-2023 at 10:00 am

Data Connectivity to the Edge

The telecommunications industry has experienced significant growth in recent years, driven by the increasing demand for high-speed internet and data services. This growth has created a surge in traffic on optical networks, leading to the development of new telecom network architectures that can support the increasing demand for bandwidth.

Optical networking technologies, such as coherent optics, have traditionally been developed for telecom applications. However, with the growth of hyperscale data centers and the increasing demand for high-speed networking, these technologies are now also being adopted in data center applications. Traditionally, data centers have used copper or short-range optical cables to connect servers and storage devices within the same data center. However, as data volumes continue to grow and data center interconnect (DCI) requirements increase, coherent optical networking is becoming an attractive option for data centers. With coherent optical networking, data centers can achieve higher data transmission rates over longer distances, resulting in increased data capacity and lower latency. 400G was the first data rate where hyperscale data center applications outpaced telecom applications in the use of coherent optics.

Coherent optics enables the transmission of high-speed data over long distances by using advanced signal processing techniques to mitigate the effects of signal distortion and noise. This technology is essential for supporting the growing demand for high-speed internet and data services, particularly in areas where traditional copper-based networks are not feasible. This trend is likely to continue and proliferate further going forward, driven by the ongoing growth of cloud computing, big data, AI/ML workloads and other data-intensive applications.

Another driver of the shift towards optical interconnects has been the increasing complexity of satellite networks. As satellite networks become more complex, the need for high-speed, low-latency communication between satellites becomes more important. Optical interconnects are ideal for this type of communication, as they offer very low latency and can support high-speed data transfer between satellites.

Optical Telecom – Satellite Communications Synergies

Optical telecom synergies have played a significant role in the evolution of inter-satellite communication. Many of the technologies and techniques used in optical telecom networks have been adapted for use in inter-satellite communication. Innovations in optical digital signal processing (DSP) and system automation also offer several optimization opportunities with inter-satellite interconnects.

Improved Signal Quality: Optical DSP can be used to compensate for impairments in the optical signal, such as chromatic dispersion and polarization mode dispersion. This can improve the quality of the signal and reduce the bit error rate (BER), enabling high-quality communications over long distances.

Reduced Latency: System automation can also be used to optimize the routing of data between satellites, minimizing the number of hops and reducing latency. This can improve the responsiveness of the system and enhance the user experience.

Power-efficient Modulation Formats: Optical DSP can enable the use of power-efficient modulation formats, such as pulse-amplitude modulation (PAM), which can reduce the power consumption of the inter-satellite links while maintaining high data rates.

Energy-efficient Signal Processing: Optical DSP can also be optimized to perform signal processing operations more energy-efficiently. For example, parallel processing and low-power digital signal processing techniques can reduce the power consumption of the signal processing circuitry.

Interoperability Demonstration

At the recent Optical Fiber Communication (OFC) conference, Alphawave Semi showcased its ZeusCORE XLR test chip during the interoperability demonstration organized by the Optical Internetworking Forum (OIF). Alphawave Semi executives Loukas Paraschis, VP of Business Development and Tony Chan Carusone, CTO, presented on high-speed connectivity leadership. Their presentations touched on the growing synergies and optimization opportunities of inter-satellite interconnects and optical telecom through innovations in optical DSP and system automation.

Summary

As the volume of data traffic on optical networks continues to increase, it is essential to ensure that the cost of implementing and maintaining these networks remains affordable. This requires a delicate balance between increasing volume and decreasing costs, which can only be achieved through innovation and the development of highly-integrated co-designed solutions. These solutions combine multiple technologies and functions into a single device, reducing the complexity and cost of optical network infrastructure. This approach enables the development of more efficient, cost-effective optical networks that can meet the growing demand for bandwidth and high-speed data transmission.

To learn more about the ZeusCORE, visit the product page.

Also Read:

Alphawave Semi Showcases 3nm Connectivity Solutions and Chiplet-Enabled Platforms for High Performance Data Center Applications

Alphawave Semi at the Chiplet Summit

Alphawave IP is now Alphawave Semi for a very good reason!


Deep Learning for Fault Localization. Innovation in Verification

Deep Learning for Fault Localization. Innovation in Verification
by Bernard Murphy on 05-30-2023 at 6:00 am

Innovation New

A new look at fault localization and repair in debug using learning based on deep semantic features. Paul Cunningham (Senior VP/GM, Verification at Cadence), Raúl Camposano (Silicon Catalyst, entrepreneur, former Synopsys CTO and now Silvaco CTO) and I continue our series on research ideas. As always, feedback welcome.

The Innovation

This month’s pick is Improving Fault Localization and Program Repair with Deep Semantic Features and Transferred Knowledge. The authors presented the paper at the 2022 International Conference on Software Engineering and are from Beihang University in Beijing and Newcastle University in New South Wales.

The method goes beyond familiar spectrum-based (SBFL) and mutation-based (MBFL) localization techniques to use deep learning from pre-qualified datasets of bugs and committed fixes. The fix aspect is important here because it depends on very accurate localization of a fault (in fact localization and nature of the fix are closely linked). The paper uses SBFL and MBFL metrics as input to their deep learning method. The authors demonstrate their methods are more effective than selected SBFL and MBFL approaches and argue this is because other methods have either no semantic understanding or only a shallow understanding of semantic features of the design, whereas they intentionally have a deeper understanding.

Fully automatic repair might be a step too far for RTL debug, however suggested fixes are already familiar for spell and grammar checks, hinting that this feature might also be valuable in verification.

Paul’s view

One year on from our review of DeepFL, we take a look at another paper that tries to move the needle further. Automatic bug detection is now a regular topic with Cadence customers and expectations are high that deep neural networks or large language models can be used to dramatically improve time spent root causing bugs.

DeepFL used a RNN to rank code for bugs based on suspiciousness features (complexity-based, mutation-based, spectrum-based, text-based). This month’s paper adds an additional bug template matching step as a further input feature to improve accuracy. The bug template matcher is itself another RNN that matches individual code statements to one or more of 11 possible bug templates, e.g. a missed null pointer checker, incorrect type cast, incorrect statement position, or incorrect arithmetic operator.

The key contribution of the paper for me is the dataset the authors build to train their bug template matching network. They mine the entire github repository to find real bug fixes that match their bug templates. For each match they require that there be another matching statement elsewhere in the same source code file that is not part of the bug fix – i.e. so the dataset has both positive and false-positive matches. The final dataset has around 400,000 positive/false-positive bug fix pairs. Nice!

As with DeepFL, the authors benchmark their tool TRANSFER-FL using Defects4J. Results look decent – 171 of the 395 bugs in Defects4J are ranked Top-5 by TRANSFER-FL vs. 140 using DeepFL. However, from a commercial standpoint 171 is still less than half of the total 395 benchmark set. If you look at the average rank across all 395 it’s 80, a long way from Top-5, so a ways off commercial deployment. I’m looking forward to reviewing some large language model-based papers next year that move the needle yet further 😊

Raúl’s view

This month we move into the areas of fault localization and automatic program repair for SW, the reviewed paper explores these techniques for Java code. In May 2022 we reviewed DeepFL, which is similar to this paper in that it extends traditional spectrum- and mutation-based techniques for fault localization with deep learning models.

To state my conclusion upfront, perhaps automatic RTL or SystemC fault localization and code repair will become routine in the foreseeable future… The authors are optimistic regarding the applicability to other languages, “most of the fix templates can be generalized to other languages because of the generic representation of AST (Abstract Syntax Tree)” with the caveat that sufficient data needs to be available for training the different networks used in their approach. For the paper 2000 open-source Java projects from GitHub were collected to construct 11 fault localization datasets with a total of 392,567 samples (faulty statements for 11 bug types that have a bug fix commit); and a program repair dataset with 11 categories of bug fixes with a total of 408,091 samples, each sample consisting of a faulty statement with the contextual method and its corresponding bug type. An AST is used to do this matching.

The detailed approach, called TRANSFER, is rather complex and requires some time to digest, 67 references help to dive into the details. It leverages existing approaches for fault localization, 1) Spectrum-based features which take source code and relevant test cases as inputs and output a sorted list of code elements ordered by suspicion scores calculated from the execution of test cases, and 2) Mutation-based features which calculate suspicion scores by analyzing the changes of execution results between original code element and its mutants. It adds 3) Deep Semantic features obtained by using BiLSTM (Bidirectional Long Short-Term Memory) binary classifiers trained with the fault localization datasets. Program repair is done using a “fine-tuned” multi-classifier trained with the program repair dataset.

The bottom line is that TRANSFER outperforms existing approaches, successfully fixing 47 bugs (6 more than the best existing approaches) on the Defects4J benchmark.

Writing and debugging SW is already routinely assisted by AI such as GitHub Copilot; designing hardware, aka writing RTL or higher-level code, can’t be too far behind, perhaps the largest obstacle being the availability of the data required.

Also Read:

Opinions on Generative AI at CadenceLIVE

Takeaways from CadenceLIVE 2023

Anirudh Keynote at Cadence Live


WEBINAR: An Ideal Neural Processing Engine for Always-sensing Deployments

WEBINAR: An Ideal Neural Processing Engine for Always-sensing Deployments
by Daniel Nenni on 05-29-2023 at 10:00 am

Option 1

Always-sensing cameras are a relatively new method for users to interact with their smartphones, home appliances, and other consumer devices. Like always-listening audio-based Siri and Alexa, always-sensing cameras enable a seamless, more natural user experience. However, always-sensing camera subsystems require specialized processing due to the quantity and complexity of data generated.

But, how can always-sensing sub-systems be architected to meet the stringent power, latency, and privacy needs of the user? Despite ongoing improvements in energy storage density, next-generation devices always place increased demands on batteries. Even wall-powered devices face scrutiny, with consumers, businesses, and governments demanding lower power consumption. Latency is a huge factor as well; for the best user experience, devices must instantly react to user inputs, and always-sensing systems cannot compete with other processes which add unneeded latency and slow reasons. Privacy and data security are also significant concerns; always-sensing systems need to be architected to securely capture and process data from the camera without storing or exposing it.

So how can always-sensing be enabled in a power, latency, and privacy-friendly method? While many existing Application Processors (APs) have NPUs inside of them, those NPUs aren’t the ideal vehicle for always-sensing. A typical AP is a mix of heterogeneous computing cores, including CPUs, ISPs, GPU/DSPs, and NPUs. Each processor is designed for specific computing and potentially large processing loads. For example, a typical general-purpose NPU might provide 5-10 TOPS of performance, with a typical power consumption of around 4 TOPS/W and about 40% utilization. However, it is inefficient because it must be somewhat overdesigned to handle worst-case workloads.

Always-sensing neural networks are specifically created to require minimal processing, typically measured in GOPS — GOPS being one-thousandth of TOPS. While the NPU in an existing AP is capable of always-sensing AI processing, it’s not the right choice for various reasons. First, power consumption will significant, which is a non-starter for an always-on feature since it translates directly to reduced battery life. Second, since AP-based NPU is typically busy with other tasks, other processes can increase latency and negatively impact the user experience. Finally, privacy concerns essentially preclude using the application processor. This is because the always-sensing camera data needs to be isolated from the rest of the system and must not be stored within the device or transmitted off the device. This is necessary to limit the exposure of that data and reduce the chances of a nefarious party stealing the data.

The solution, then, is a dedicated NPU specifically designed and implemented to process always-sensing networks with an absolute minimum of area, power, and latency: the LittleNPU.

In this webinar, Expedera and SemiWiki explore how a dedicated always-sensing subsystem with a dedicated LittleNPU can address the power, latency, and privacy needs while providing an incredible user experience.

REGISTER HERE

Presented by
Sharad Chole, Expedera Chief Scientist and Co-founder

About this talk
Always-sensing cameras are emerging in smartphones, home appliances, and other consumer devices, much like the always-listening Siri or Google voice assistants. Always-on technologies enable a more natural and seamless user experience, allowing such features as automatic locking and unlocking of the device or display adjustment based on the user’s gaze. However, camera data has quality, richness, and privacy concerns which requires specialized Artificial Intelligence (AI) processing. However, existing system processors are ill-suited for always-sensing applications.

Without careful attention to Neural Processing Unit (NPU) design, an always-sensing sub-system will consume excessive power, suffer from excessive latency, or risk the privacy of the user, all leading to an unsatisfactory user experience. To process always-sensing data in a power, latency, and privacy-friendly manner, OEMs are turning to specialized “LittleNPU” AI processors. In this webinar, we’ll explore the architecture of always-sensing, discuss use cases, and provide tips for how OEMs, chipmakers, and system architects can successfully evaluate, specify, and deploy an NPU in an always-on camera sub-system.

About Expedera
Expedera provides scalable neural engine semiconductor IP that enables major improvements in performance, power, and latency while reducing cost and complexity in AI-inference applications. Third-party silicon-validated, Expedera’s solutions produce superior performance and are scalable to a wide range of applications from edge nodes and smartphones to automotive and data centers. Expedera’s Origin™ deep learning accelerator products are easily integrated, readily scalable, and can be customized to application requirements. The company is headquartered in Santa Clara, California. Visit expedera.com
Also Read:

Deep thinking on compute-in-memory in AI inference

Area-optimized AI inference for cost-sensitive applications

Ultra-efficient heterogeneous SoCs for Level 5 self-driving

CEO Interview: Da Chuang of Expedera


Why Secure Ethernet Connections?

Why Secure Ethernet Connections?
by Daniel Payne on 05-29-2023 at 6:00 am

Ethernet Security min

While web browsing I constantly glance for the padlock symbol to indicate that the site is encrypting any of my form data by using the https prefix, which means that an SSL (Secure Sockets Layer) certificate is being used by the web hosting company. I have peace of mind knowing that my credit card information cannot be easily stolen on secure web sites using an SSL certificate. For Ethernet networks there are also several potential security breaches:

  • Man-in-the-middle attacks
  • Eavesdropping
  • Denial of service
  • Privilege escalation

The downside risks of data theft are quite high to any organization, and being compliant with data security standards like Health Insurance Portability and Accountability Act of 1996 (HIPAA) in the US and the  General Data Protection Regulation (GDPR) are important.

To detect and prevent any unwanted data intrusions a system must have privacy measures like source validation and authentication.

Secure Ethernet Protocol

Thankfully, the IEEE created the Media Access Control Security (MACsec) protocol in 2006 to help secure Ethernet networks. The cryptography used to protect Ethernet traffic in MACsec is  Advanced Encryption Standard-Galois/Counter Mode Cryptography (AES-GCM), also used for CXL and PCI Express security. MACsec is applied at Layer 2 of the Open Systems Interconnection (OSI) networking model, enabling data integrity, confidentiality, replay protection and data origin authenticity.

MACsec uses AES-GCM cryptography

The encrypted connection with MACsec has several authentication steps:

  1. A Pre-Shared Key (PSK) for mutual peer authentication.
  2. A secure Connectivity association Key Name (CKN) is exchanged.
  3. The two endpoints decide which is the key server, and key client.
  4. The key server sends the Secure Association Key (SAK) to the key client.
  5. Encrypted data is ready for exchange.

You could take the time to become an expert on MACsec and design your own Ethernet security solution that conforms to the protocol, or consider using Ethernet IP from Synopsys that in addition to controllers and PHYs includes their MACsec Security Modules. Here’s what the Ethernet security IP from Synopsys looks like:

Synopsys Ethernet Solution with MACsec security

The benefits of using the Synopsys MACsec Security Modules include compliance with the IEEE 802.1AE standard, throughput scalability and configurability to tune solutions for specific applications and use cases with optimal latency, power, performance and area.

Synopsys not only has secure IP for Ethernet, it also extends into other interfaces: PCIe, CXL, USB, HDMI, DisplayPort, DDR, LPDDR, Die-to-Die, MIPI, UFS and eMMC.

Synopsys Secure Interfaces Solutions

Kalar Rajendiran wrote more about this topic in his January blog on SemiWiki.

Summary

Having a secure Ethernet makes a lot of sense, and with the IEEE protocol MACsec there’s a standard way to ensure that network data is not stolen or compromised. Malicious actors want to steal our web browsing activity and that extends into the realm of Ethernet traffic.

Instead of waiting to have your Ethernet data breached, why not be proactive by adding the MACsec protocol in your SoC implementation. Do the math on the cost of building your own MACsec compliant IP, then compare that to what Synopsys has designed and verified already. The semiconductor IP industry has grown steadily for many years now, and for good reason, IP blocks from a trusted vendor can offer a faster path to market while not requiring your engineering team to become domain experts in something new and unfamiliar.

Related Blogs


Podcast EP164: How Weebit Nano is Disrupting the Memory Market with Coby Hanoch

Podcast EP164: How Weebit Nano is Disrupting the Memory Market with Coby Hanoch
by Daniel Nenni on 05-26-2023 at 10:00 am

Dan is joined by Coby Hanoch, Coby joined Weebit Nano as CEO in 2017. He has 15 years of experience in engineering and engineering management roles, and 28 years of experience in sales management and executive roles.

Coby explains the unique features of Weebit Nano’s non-volatile ReRAM technology. He explores the technology’s extended compatibility across advanced process nodes and its speed and power advantages as compared to more traditional approaches such as flash. He discusses current and future deployment for both embedded and discrete applications.

The views, thoughts, and opinions expressed in these podcasts belong solely to the speaker, and not to the speaker’s employer, organization, committee or any other group or individual.


Micron Mandarin Memory Machinations- CHIPS Act semiconductor equipment hypocrisy

Micron Mandarin Memory Machinations- CHIPS Act semiconductor equipment hypocrisy
by Robert Maire on 05-26-2023 at 6:00 am

Xi Biden China CHIPs ACT

-We think China issue is overblown- Zero sum game theory
-US should respond Tit-for-Tat and then some- Chop Chinese Chips
-Chip equip makers duplicitous behavior- smacks of hypocrisy
-Exporting chip equip biz/jobs to Asia while asking for CHIPS Act$

Memory Market remains a zero sum game market much like petroleum

China’s “banning” of Micron memory products in China generated a large blowback, in our view overreaction, in Micron’s stock. The reality is that while it may cause some dislocation in the near term, in the longer run there will likely be less impact.

Much like the oil market, memory is a commodity product that is essentially interchangeable with one another. Yes, there may be differences between Texas sweet and Brent, they kind of all burn the same way at the end of the day.

If the US bans the import of Russian or Iranian oil, some other country will buy it instead. Such is the same with memory products. There may be near term dislocation but in the longer run market share between Samsung, SK & Micron will remain similar to today as just the shipping destination will vary.

Its certainly far from the negative impact the stock saw.

The US can speed up the rearrangement of customers

One very easy way to speed up the redistribution of market share would be for the US government to ban any and all Chinese memory products, including memory made by foreign companies in China, from being imported in any product into the US.

In this way Micron will pick up those empty slots while Chinese memory companies will replace Micron in China and we can all go home and call it a day…..

Seems a pretty simple Tit for Tat solution…and effectively nullifies China’s move

We see no reason not to…..

Perhaps up the ante by cutting off 300MM to China?

Right now China is by far the biggest buyer of semiconductor equipment in the world. The US continues to sell China fishing poles even though we have cut off their fish supply….

One problem we have seen is that there is a lot of dancing around which tools can and cannot be sent to China and where in China they can be sent to.

We have heard rumor of some tools being degraded in software in order to be exportable to China or perhaps having the label and name changed to get around export restrictions only to be able to be used for more advanced purposes later. China can even get around a lot of restrictions by multiple patterning and other difficult tricks.

Perhaps the simplest and easiest way is to just restrict China to 200MM (8 inch) and below wafer tools. It essentially would make any and all work arounds impossible, and highly limit how advanced the technology node could be.

It would be super easy to enforce, any idiot with a tape measure could do so. More importantly no dancing around or negotiating specifications and features. Essentially a truly impenetrable technology wall.

Perhaps a Tit for Tat response is not enough….just drop the hammer….

The not so hidden duplicity behind Applied’s “Epic” announcement

We were quite pleased and happy to hear Applied Material’s announced plan to spend billions of dollars on a new “Epic” R&D center in California. Kind of like the existing Maydan center on steroids. The spending and additional hiring over the next few years will be needed in any event to support the expected growth of the semiconductor industry and in order to keep up this is the type of effort that is needed. We view this not as above and beyond spending but spending in line with expectations if we look at current and expected growth and spending rates.

We would note that Applied added an interesting caveat in its press release that this spending is dependent upon CHIPS Act funding. Would they not fund increasing R&D without CHIPS Act money??

Perhaps much more interesting is the juxtaposition of the “Epic” announcement coming six months after another record expansion and spending effort by Applied Materials in Singapore in December. Obviously the celebration of spending all that money and moving many AMAT US jobs to Asia was not accompanied by US government officials in a big splashy announcement.

In December Applied announced “Singapore 2030”. An effort to double its operation in Singapore by building a $600M 700,000 sq foot facility for R&D and manufacturing in Singapore (about three times larger than the 180,000 sq foot “Epic” center in California). It is expected to create well over 1000 direct jobs and many indirect and construction jobs. Singapore is already Applied largest facility outside the US and this expansion will obviously make it larger than a single location in the US.

The spending in Singapore also includes academic partnerships, AI, ML robotics etc;. just like Epic.

If anything , Applied growth in Asia, and specifically Singapore, is at a much higher rate than the US and in fact is likely contributing and causing slowing manufacturing & R&D job growth that would have otherwise happened in Applied in California or Texas.

We find it somewhat amusing that Applied can spend all that money, by itself, in Singapore, creating & moving jobs there while it needs CHIPS Act money to expand in the US?

Is taxpayer CHIPS Act money being used for stock buy backs and dividends?

One other fact we found ironic and interesting juxtaposed timing about the need for CHIPS Act funding for Applied’s Epic center is their earnings announcement just a few days prior……

On the earnings call the CFO Brice said “The board of directors approved a 23% per share dividend increase which is the largest increase in 5 years and supplemented our share buyback program with a new $10 billion repurchase authorization. We believe our free cash flow can continue to grow and support increasing the dividend at an accelerated rate over the next several years which would double our previous dividend per share”

Somehow that does not sound like a company that desperately needs CHIPS Act corporate welfare to afford its baseline R&D spending and new “EPIC” center. Especially after the Singapore expansion.

The optics are even worse in that we are in an “epic” downturn in the industry yet we are accelerating exporting jobs, jacking up shareholder rewards while holding out a hand for CHIPS Act money.

We thought the CHIPS Act was not supposed to support, even indirectly, corporate welfare of shareholder rewards or exporting jobs.

Obviously not……

Lam Leaving Livermore

Applied Materials is far from alone in continuing to move jobs from the US to Asia. It seems part of Lam’s plan of accelerating and increasing manufacturing at its new location in Malaysia is to ramp down and eventually close its manufacturing operation in Livermore California. Essentially moving all those thousands of jobs as well as indirect jobs to Asia in the process.

Lam is also doing quite well despite the downturn and like Applied could make slightly less profit while keeping jobs in the US.

However we haven’t heard about Lam asking for CHIPS Act money……at least not yet.

We have seen this movie before…and it didn’t turn out well…CHIPS Act II

It is amazingly ironic that the semiconductor equipment industry is following in the exact same footsteps of the semiconductor industry that it serves.

Over the last 30 or so years the semiconductor industry exported chip manufacturing to Asia to save costs and increase profits while keeping chip design in the good ole USA.

Now the semiconductor equipment industry is doing the exact same thing despite the huge outcry of what happened to the semiconductor industry itself and now huge rescue and re-shoring efforts that are taking place.

Semi equipment companies are exporting manufacturing while keeping R&D in the US, hence the Epic R&D center, just like the semiconductor industry did and did not work out so well for the industry.

In our view the best way to re-shore the semiconductor industry is to not offshore it in the first place and certainly not support companies that do so for the sake of profitability over patriotism.

The alternative is to have the CHIPS Act II which will be aimed at re-shoring the equipment industry like semi industry before it a few years down the road. It would be much easier to fix right now before we export all equipment manufacturing…..

The very ultimate irony would be for all the new shiny semiconductor fabs being built in the US in Arizona, Idaho, Ohio & Texas etc; to have to import their equipment made in Asia by US companies who moved manufacturing there at the same time the fabs were being built in the US…… how stupid can you get?

Wake up and smell the coffee….

The stocks

While we have no desire to run out and buy Micron shares in the face of the disaster in the memory industry that will be long lived we do believe the sell of was unwarranted and overdone.

If the US does react with restrictions on Chinese memory, which we believe they should, Micron could easily jump back and more.

Nvidia is a shining light in an otherwise dark and gloomy industry and its earnings should help lend some emotional support. We would continue to own (and do so in our personal account) as one of the better and only plays in the semiconductor industry that is seeing a huge success.

With memory in the dumps and foundry/logic not much better we can still wait for better days to own the equipment stocks that are likely overbought in the recent run up.

About Semiconductor Advisors LLC

Semiconductor Advisors is an RIA (a Registered Investment Advisor), specializing in technology companies with particular emphasis on semiconductor and semiconductor equipment companies. We have been covering the space longer and been involved with more transactions than any other financial professional in the space. We provide research, consulting and advisory services on strategic and financial matters to both industry participants as well as investors. We offer expert, intelligent, balanced research and advice. Our opinions are very direct and honest and offer an unbiased view as compared to other sources.

Also Read:

AMAT- Trailing Edge & China Almost Offset Floundering Foundry & Missing Memory

LAM Not Yet at Bottom Memory Worsening Down 50%

ASML Wavering- Supports our Concern of Second Leg Down for Semis- False Bottom


Chiplet Interconnect Challenges and Standards

Chiplet Interconnect Challenges and Standards
by Daniel Payne on 05-25-2023 at 10:00 am

Multi die IP min

For decades now I’ve watched the incredible growth of SoCs in terms of die size, transistor count, frequency and complexity. Instead of placing all of the system complexity into a single, monolithic chip, there are now compelling reasons to use a multi-chip approach, like when the maximum die size limit is reached, or it’s more cost effective to use two or more smaller chiplets designed in a variety of technology nodes. Taking the multi-die system path introduces new chiplet interconnect challenges:

  • Reliable connectivity
  • High bandwidth
  • Low power
  • Low latency
  • Standards support

Fortunately for the industry there’s been a collective effort to develop standards, and the Universal Chiplet Interconnect Express™ (UCIe™) has gained traction by enabling package-level integration through a die-to-die interconnect along with a connectivity protocol, so that multiple vendors can grow an ecosystem through interoperability. UCIe covers three stack layers, and the PHY layer defines the electrical interface.

Synopsys has been delivering IP for many years now across many domains, like: Interface, Foundation, Processor, Security, Analog, Subsystems. They’ve also joined the UCIe Consortium, contributing to the specification of the standard. There’s a UCIe PHY IP from Synopsys, along with a UCIe Controller IP and verification IP.

Synopsys Multi-die IP

In March 2023 Synopsys announced that their UCIe PHY IP had a tape-out on the TSMC N3E process node.

For reliable connectivity the UCIe standard has up to 8 spare pins per direction, allowing repair of the functional links.

Link Repairs

Variations in the die-to-die interface signals are monitored by Signal Integrity Monitors (SIM), then the Monitoring, Test and Repair controller can determine the health of the multi-die system for predictive maintenance of the links. Synopsys has the Silicon Lifecycle Management tools to monitor the UCIe interface while its operating, detecting soft or hard errors.

Synopsys Monitoring, Test and Repair (MTR) controller

Bandwidth for UCIe using the Synopsys PHY IP is up to 5Tbps/mm efficiency. The Controller IP supports streaming protocols as well as PCI Express and CXL protocols, delivering secure, low-latency data.

Coming up to speed on the UCIe specification takes precious engineering time, so re-using protocol verification IP is going to save your team valuable time to market. Verification IP running on a software simulator provides a good start, then adding hardware emulation with Synopsys ZeBu and prototyping with Synopsys HAPS Platform offer more time savings to debug the whole system running software.

Routing the UCIe signals between dies is automated by the Synopsys 3DIC Compiler tool, and it works for 2.5D chiplets.

Summary

Systems engineers today still have to decide between two approaches for implementation, the traditional single-chip SoC, or the multi-die system. EDA vendors like Synopsys have long been automating the EDA tasks for a single-chip SoC, and they’ve also extended their automation into the realm of 2.5D by developing new EDA tools, verification and IP for multi-die systems.

The interconnect challenges of multi-die systems have been addressed through standardization efforts like UCIe, which lowers the risks for new projects considering chiplet-based systems. Synopsys is one of the few EDA and IP vendors with such broad support of multi-die systems.

Related Blogs


IP Lifecycle Management for Chiplet-Based SoCs

IP Lifecycle Management for Chiplet-Based SoCs
by Kalar Rajendiran on 05-24-2023 at 10:00 am

IP Object for IPLM

Chiplet-based System-on-Chips (SoCs) are becoming increasingly popular in the semiconductor industry due to their potential to improve design efficiency, increase performance, and reduce costs. While chiplets are seen as a way to reduce the cost of innovation, they introduce a lot of challenges too. Packaging, interconnect models, verification, provenance and traceability, IP/design ecosystem management and ensuring security are areas where challenges need to be addressed. Before mass migration to chiplet-based designs can happen, new infrastructure and tools are needed. More complex modeling is needed for interconnects, timing verification, thermal management. Trust concerns around how the chiplets are delivered and integrated into a SoC need to be addressed. This in turn involves making sure chiplets have not been compromised along their way from development through integration to delivery to the foundry/assembly house. Establishing provenance and a traceable integration process becomes a key requirement.

It is in the above context that Simon Butler gave a talk at the IP SoC 2023 conference in Santa Clara last month. He is the General Manager of the Methodics IPLM business unit at Perforce. A chiplet is essentially an IP in the form of a standalone die. Perforce and Simon are no strangers to IP Lifecycle management. Simon walked the audience through laying the foundation for establishing provenance and traceability, bill of materials (BoM) management, and enforcing BoM provenance. Provenance means the origin of something and the history of its ownership, used for establishing authenticity and quality.

Challenges to Establishing Provenance and Traceability

In chiplet-based SoCs, different chiplets can be used in various SoC designs. Therefore, the IP reuse becomes important in order to increase efficiency, reduce cost and time to market. But reusing IP from other designs may cause compatibility issues. In addition, different IP blocks may come from different sources, and knowing the ownership of IP is crucial.

Any system to be used for managing IP should be able to avoid incompatibility issues among all the chiplets and other IP blocks in a chiplet-based SoC. The system should also be able to help avoid IP infringement issues and confirm proper licensing exists. It should also be able to perform authenticity verification to ensure the chiplets were not compromised through backdoors during the manufacturing or delivery process.

Foundation for Establishing Provenance and Traceability

The system should be able to represent IP in a way for identifying the respective design files as well as corresponding metadata model defining the IP’s state and context. While there are many popular version control systems for data/file management, one that includes metadata management is key. Metadata model is where access permissions, quality grade, IP compatibility information, standards compliance details, etc. are maintained. As such, the metadata layer should also offer immutable fields such as timestamps, ownership and other user-defined items.

With the IP representation defined as above, the next critical piece is Bill of Materials (BoM) management.

Bill of Materials (BoM) Management

A BoM is essentially a hierarchically versioned tree of the building block components in a system. As an example, referring to the Figure below, SoC@4 refers to fourth version of the SoC and is guaranteed to contain the respective versions of the sub-components contained in that tree. This ensures traceability and the metadata overlays from the various sub-components provide context.  Thus, the BoM can be relied upon as the single source of truth for the configuration management of the SoC. BoM immutability should be enforced at the platform level so that the definition of a version cannot change after a release is made.

Enforcing BoM Provenance

By integrating IP fingerprinting techniques including blockchain-managed signatures for important metadata, immutability of IP contents can be locked down. Queries can be used on such an integrated platform to validate the trustworthiness of the components on the BoM. This helps to mitigate the risk of supply chain attacks, where malicious actors can introduce backdoors or other security vulnerabilities into the product.

Many industries have regulations that require the traceability of components used in electronic products. Enforcing the BoM provenance helps to ensure that the product complies with these regulations and can help with audits and other compliance requirements.

Summary

The best way to ensure BoM provenance is to use a traceable platform such as the one described above. IP release management, built-in traceability, IP discovery and reuse and workspace management across design managers are all included in Methodics IP Lifecycle Management solution.

You can listen toSimon’s talk on-demand here. You can download his presentation slides here.

Also Read:

Overcoming Semiconductor Supply Chain Bottlenecks

The CHIPS and Science Act, Cybersecurity, and Semiconductor Manufacturing

Solve Embedded Development Challenges With IP-Centric Planning