100X800 Banner (1)

Designing a Wafer-Scale Image Sensor for use in X-Rays

Designing a Wafer-Scale Image Sensor for use in X-Rays
by Daniel Payne on 06-22-2012 at 1:32 pm

At Intel we mused about designing wafer-scale integration (WSI) back in the 70’s however I just learned about how Dr.Renato Turchetta at the Science and Technology Facilities Council (STFC) designed a wafer-scale imaging sensor chip for X-Ray applications. I was also able to interview Dr. Turchetta to learn more about the challenges that they faced in creating this chip.

Continue reading “Designing a Wafer-Scale Image Sensor for use in X-Rays”


Will Microsoft Go Thermonuclear?

Will Microsoft Go Thermonuclear?
by Ed McKernan on 06-21-2012 at 8:20 pm

Microsoft is in trouble. Many of you already know that. Steve Ballmer has one last opportunity to set the company on a growth path or they will retreat into IBM legacy mode… ala the post 1990s Lou Gerstner era. And so they introduce a large tablet-convertible in direct competition with their PC partners Dell and HP. The End Game is coming sooner than anyone expected. Where is this headed?

A few weeks ago I considered writing a blog following the disastrous Dell quarterly earnings report where they admitted Apple’s iPAD was starting to inflict damage on their PC business. I was sure that this was going to be a forcing function on the Wintel Empire and tried to look at it from the Intel side. Would Intel get into the PC business as Dell and HP scaled back? I couldn’t complete the logic because an Intel entry is at conflict with Apple’s strong consumer push with their PCs.

No, the more logical conclusion is that Intel forges ahead with owning the high performance driven datacenter market (at some ever shrinking expense to HP and Dell) while Microsoft scrambles to replace HP and Dell in the client space where form factors are shrinking dramatically. DVD and HDDs are now being replaced by tiny, energy sipping NAND flash chips while application delivery over the cloud becomes a must.

Then there are the economics to consider but first a little history on tablets:

Microsoft has been dabbling in tablets for more than 20 years. Does anyone remember the flaming startup called Go? The stylus and hand writing recognition were considered the key to tablets taking off and Microsoft thought that they were closing in on a solution year after year. As an engineering friend of mine liked to say on a daily basis about his constantly delayed chip tapeout: “never been closer.”

The aim of hand writing recognition was to enhance Microsoft’s bread and butter Office Application environment with users casually writing in documents on a screen. It was a hard object (stylus) on a hard SURFACE (had to throw that in). It was clunky and unnatural. As is typical to all massively planned grand events, history chooses to take a detour and we ended up with smartphones and tablets controlled by the soft touch of a thumb and index finger.

Now to the Economics:

Apple’s operating margins are now greater than Microsoft (39% vs. 36.8%). Did you know that? If Microsoft stays on the Wintel course in the tablet and PC market working with traditional customers like Dell and HP they will face either shrinking margins or reduced volume to stay relevant. The alternative is to take the hardware into their own hands and deliver a complete solution to corporations and consumers while eliminating costs by going vertical. This is what the tablet announcement signifies and in a way it is leading down the IBM’s mainframe bundled solutions path. Eliminating Dell and HP off the top reduces the end prices of PC equipment by 15-25% depending on the configuration. It’s a good start towards being competitive but without a doubt a burning the ships moment.

Microsoft will need to go a step further and take ownership for some of the silicon that goes into the tablet and ultrabooks. Specifically they will need to acquire the likes of nVidia and AMD in order to develop low cost ARM and x86 processors and thereby effectively moving margins into Microsoft’s pocket. The new vertically integrated Microsoft then can replicate the Apple supply chain enabling them to hit low end market price points and to undercut Intel in the >$700 mobiles that use Ivy Bridge. Now this is a true Thermonuclear War. In the end, the 30 year old Wintel stasis can not coexist with Apple’s hyper growth.

Steve Ballmer has maybe two years (more likely one year) to put Microsoft in a leadership position that remains within spitting distance of Apple. If he is not successful, would it be even possible for Bill Gates to come out of retirement to forge a winning plan?

FULL DISCLOSURE: I am Long AAPL, INTC, QCOM, ALTR



EDA Tools to Optimize Memory Design

EDA Tools to Optimize Memory Design
by Daniel Payne on 06-21-2012 at 8:15 pm

I met with Amit Gupta, President and CEO of Solido at DAC on Tuesday to get an update on their EDA tools used in the design of memory, standard cells and low-power. In 2012 they’ve expanded to add three new software packages: Memory, Standard Cell, Low Power. They must be doing something right because at DAC this year I see more competitors jumping into the Fast Monte Carlo space. Continue reading “EDA Tools to Optimize Memory Design”


Mike Muller’s ARM Keynote at DAC 2012

Mike Muller’s ARM Keynote at DAC 2012
by Paul McLellan on 06-21-2012 at 7:30 pm

Mike Muller’s keynote focused on a lot of changes since the ARM1 was designed in 1983 when ARM the company did not exist and ARM was the next generation processor for Acorn Computer, which was really in the hobby market and had its first boost when they had a contract to design the BBC Microcomputer to go along with a computer literacy project. My minor claim to fame about the ARM1 was that I installed the VLSI Technology design software that was used to design it on two Apollo workstations (which hadn’t actually shipped so I had to go and install it at Apollo not Acorn).

Mike decided to dig up the layout of the ARM1, which turned out to be more of a challenge than expected, involving finding an Exabyte drive, converting VAX files to Linux and converting from CIF (which VLSI Technology used for layout) to GDSII. The original ARMs were full custom designs (this predated synthesis). ARM have recently announced a low end microcontroller, the Cortex-M0. It turns out that this is roughly the same number of transistors as the ARM1 so it makes an interesting basis for comparison.

The chips are about 26 years apart, or 13 process generations, so it should be about 1/2[SUP]13[/SUP] smaller (namely 1/8000) and it is actually 1/10,000 times smaller so about right. Performance should be about 6 performance scalings, so 64 times as fast but it is only 16 times as fast. This is because the 5V power supply that the ARM1 used should have scaled down to 8mV but in fact it has only scaled to 950mV and so the transistor threshold voltages haven’t scaled enough.

The big change in design is that the ARM1 took 6 months to layout. The Cortex-M0 took 32 minutes. Basically synthesis, place and route has automated the whole process whereas the ARM1 was custom. But that is really the only major improvement in design methodology.

Mike then looked at the design productivity. The ARM1 took 6 man-years (MY) and was 25,000 transistors. The dual Cortex-A15 took 150MY and is 150 million transistors. Luckily design productivity has increased 240 times. It is when you look at software that things are scary. The ARM1 graphics library was 0.25MY of work and was 150 lines of code (Loc). Assembly of course. The current ARM GPU, the Mali-T604 has OpenGl, open CL and other graphics support and is 190MY of work and 1M LoC. Just a 7 times increase in productivity.

Mike pointed out that the hardware people shouldn’t be complacent. We came up with synthesis and P&R so that we can essentially compile our chips. But we haven’t come up with anything comparable since. Software people have moved onto Python, cloud computing, development environment and lots of new goodies. So apart from the few people left having to write device drivers, the way software is being developed is changing fast.

Next Mike moved onto validation. The ARM1 was in 3um, had 24,000 transistors and took 6 MY to validate. The Cortex-M0 is in 20nm, 32,000 transistors (not much more) and took 11 MY to validate. But the big difference is the machine resources brought to bear. The ARM1 took 2,000 hours to validate but the Cortex-M0 took 1,439,000 hours to validate. Taking the speed of the machines into account this is 3,000,000 times less efficient. We waste a lot of computer cycles early, especially with constrained random verification, in order to avoid silicon respins later.

Mike feels we need to get more formal design techniques. Today formal verification is stuck on the side of the design process as an “optional extra” to be run by a specialist, rather than something embedded deeply in the design process. Without this we are stuck with constrained random and burning computer and verification engineer cycles, and formal verification not part of every designers job.

As an aside, Mike talked a bit about building ARM’s datacenter in the parking lot. It has 200TFLOPS and 93TB DRAM. It consumes 1.5MW (luckily there was a new housing tract going in across the street so they actually managed to get a line like that installed). The UPS consists of two parts, firstly 6 spinning flywheels which can each deliver 250KW for 20 seconds. Then two 785bhp quad turbo diesel generators that can get up to full power in 8 seconds. Looks pretty good too!

There is a video the keynote here. The video starts with the DAC award session and the keynote itself starts at 29 minutes in. Wait and let enough of the video load and then you can skip straight to the start (unless you really want to see a re-run of the award session).


Off topic: Matt

Off topic: Matt
by Paul McLellan on 06-20-2012 at 8:12 pm

This has pretty much nothing to do with EDA or semiconductor. OK, absolutely nothing. Years ago a friend pointed me at a video of a guy who used to be a video game designer (wait, a semiconductor connection) who decided to take his savings and travel the world. As he put it, “I wasn’t a very good video game designer. But I was good at travel.” He filmed himself dancing badly in front of famous places, put them together into a video and it went viral. A chewing gum company sponsored him to make two more, equally successful. Visa and hotels.com paid him to make ads…

I was at TED in Palm Springs three years ago (the TED for people who are too poor and not nearly famous enough to go to the real one) and he was there for one of the token events that was live in Palm Springs and everyone in Long Beach had to watch on video feed. I talked to him at the party that evening. He said he was trying to travel less since he had a girlfriend so couldn’t just live on the road for months on end. But she became the producer and the story carried on.

Anyway, after two years in the making, the 2012 video was released today. It’s a huge undertaking when you compare it to the first original one. I suspect this will be the last one for some time. You’ll see why when you see him dancing in Seattle (his home town) in the last scene.

I don’t understand YouTube statistics: 7000 people already “like” it, 2000 people have commented, but apparently only 301 have seen the video.

Enjoy.


Double Patterning Technology at DAC

Double Patterning Technology at DAC
by Daniel Payne on 06-20-2012 at 5:12 pm

David Abercrombie from Mentor Graphics met with me on Tuesday at DAC to provide an update on DPT – Double Patterning Technology, something new required for several layers starting at the 20nm node in order to get any IC yield. DPT is also part of Multiple-Patterning. Continue reading “Double Patterning Technology at DAC”


Custom IC Layout Automation at DAC

Custom IC Layout Automation at DAC
by Daniel Payne on 06-20-2012 at 3:56 pm

Three people from Ciranova met with me at DAC this year: Eric Filseth (CEO), David Millman (Marketing) and Lindor Hendrickson (CTO). They gave me an update on how the Helix tools are being used to automate the layout of custom IC designs at 28nm and smaller nodes. Continue reading “Custom IC Layout Automation at DAC”


IPL Alliance at DAC

IPL Alliance at DAC
by Daniel Payne on 06-20-2012 at 3:25 pm

Lunch on Tuesday at DAC was sponsored by the IPL Alliance and thankfully this year they skipped the attempt at humor and focused on interoperable PDKs. Presenting companies include: Synopsys, Dongbu HiTek, TowerJazz, X-FAB and Si2. Having both OpenPDK and iPDK on the same platform does sound like a peaceful co-existence to me, although I wasn’t too sure about that one year ago.

Continue reading “IPL Alliance at DAC”