Array
(
    [content] => 
    [params] => Array
        (
            [0] => /forum/threads/is-there-a-comparison-on-the-eda-tools-and-their-effectiveness.3201/page-2
        )

    [addOns] => Array
        (
            [DL6/MLTP] => 13
            [Hampel/TimeZoneDebug] => 1000070
            [SV/ChangePostDate] => 2010200
            [SemiWiki/Newsletter] => 1000010
            [SemiWiki/WPMenu] => 1000010
            [SemiWiki/XPressExtend] => 1000010
            [ThemeHouse/XLink] => 1000970
            [ThemeHouse/XPress] => 1010570
            [XF] => 2021770
            [XFI] => 1050270
        )

    [wordpress] => /var/www/html
)

Is there a comparison on the EDA tools and their effectiveness?

Joe Gianelli
As has already been stated above, EDA companies (and many other software companies) have you sign license agreements which prohibit disclosure of benchmark results. However, I would think a non-biased third party could set up a rigorous testing procedure, procure all the tools, and set up an evaluation service for design teams. This entity would need a slightly modified license agreement such that they could disclose results to their clients only. Who knows, maybe this service already exists.
 
Jean-François Agaësse

What you mentions looks like the approach that is familiar to SW engineers with the use of widely adopted Dhrystone or more recent benchmarks (such as antutu, ...). However, recent threads/disputes on www have pointed out that these benchmarks might be somewhat of twisted to highlight the benefits of a given architecture and do not provide the full vision. On the other end, they are so complex, that you can pick only ONE part of the outcomes and start creating the buzz disregarding results that do not go in your direction (see the recent "did Intel beat ARM" story)

Not going as far as this in duplicity, we all know that benchmarks only cover a part of the problems. So, applying them for what they are not designed for is pure nonsense. For instance Dhrystone might be used by HW designers to give a merit figure of their core wrt power dissipation, mainly because this is the only one they have. But we all know that the Dhrystone aims at measuring the efficient computation logic and doesn’t exercise realistic use cases – power wise - where the different levels of memory are exercised. Consequently, in the real life, the power figures and the silicon areas where power is burnt will we very different of what you get.

Assuming we’ve managed to solve the business case(s) of having an independent lab running the benchmark (meaning that Aart, Wally and few others agree on a free and fair comparison) we will still not be home free. Based on my experience in Physical Design I can tell that one bad tool will not give good results, but a tool proving OK in a benchmark is not granted to exhibit the same technical advantage on its competitors in the real life. This is because, unlike the well identified SW processing situations we find in SW benchmarks, - BE wise all designs situations are different (when it is about BE tools we speak about “heuristics” which means that a small variation in the initial conditions – design, technology, floorplan shape, tool settings, might give very different results). Moreover, company design flows / style / practices also differ. The long and fruitful partnerships between Semiconductor Companies and EDAs have resulted in some tools specially trimmed to a kind of dominant design style and flow, and that particular tool will not be at top for different methodologies.

So when a team is at par with the current solution, people are very unlikely to themselves look for benchmarking, or trust other’s results for swapping tool for a quite uncertain benefit. For this you need a revolution, like moving from channel routing to surface routing in the middle of 199’s.
But, in my perspective, but since half a decade in the conventional EDA area it is more about evolution than revolution. Doesn’t mean that what has been introduced is not good, but it looks more like variations on the ‘divide and conquer’ strategy a/o more efficient implementation of multithreading.
(But I would be happy to read why I am wrong, or a bit exaggerating). As a consequence, “Independent” official benchmarks are likely to never happen, because in the event they would show a revolution the EDA leaders will start explaining “this is not the right design for our tools” (aha! heard this many times as well?) and if only about limited benefit, No one will make a change for 20%, he has even not experimented himself.

At the end it looks like that based on such benchmark results, EDA vendors will assault the design directors advertising their tools for ‘solving some of the problems you may face. A kind of similar situation we all know when winter starts, with ads in the newspapers a/o on TV for pills fighting against some bacteria’s a/o viruses responsible for the flue. …. Usually you got the other ones, the pills you’ve been sold will not help to defeat.

JFA
 
Yu Wang

IC designers do have the data but I guess they could not share for various reasons. But I personally would like to see more if possible, considering we're in such a highly competitive industry. It would not be worse to know exactly how EDA vendors are doing and what customers are complaining about.
 
Shishir Gupta

Well, given today's highly competitive market, most eda tools from the big 3 are almost equal, give or take a little. There are 2 factors which really decide the effectiveness of any eda tool:

1. the cad flow its integrated in
2. the skill lievel or expertise of the engineer driving it

When it comes to #1, some of the larger companies have very efficient flows. They shield their engineers from the tool and its transparent to the engineer what tool they are using. They also have comprehensive checks to ensure data hand off from one tool to another tool or from one engineer to another engineer is complete <- this is usually where most design teams waste time. e.g. timing constraints hand-off from synthesis team to the backend team. This also means, these companies need the brightest minds in CAD, and can make do with average design engineers

On point #2, that has always been the case. Smart designers can make do with anything. They dont believe in reporting bugs or tool problems, they just find ways to work around it. But they are starting to become a rarer breed now....
 
John Beaudoin


Jeremy,
I've seen many benchmarks cover all that you mentioned. One company had 22 representative designs and benchmarked 3 SPICE simulators across the 22 designs. It was comprehensive and it outlined the sweet-spots of each tool. I was not allowed to know the results. As mentioned earlier, they were confidential to the company that did the benchmark (must admit I did take a peak when no one was looking). Most importantly, I assure you that the results were objective and comprehensive and done with the expert knowledge of each specific simulator. I will add that I have not seen any more professional benchmark than that one in 20 years in EDA.

EDA vendors do NOT want these kinds of benchmarks to get out. And even if one is faster in all categories, that does not mean that the results translate back to the human interface better than the competitor. And the set-up time and human interaction on the input is also a factor that sometimes is not considered. Ease-of-Use is always the asterisk in the evaluation.

And these reasons are why it takes a 3X or greater advantage to replace competitor based on a technological decision. Business decisions are usually the case for change. In most circumstances, if engineering says they want to change tools, they have to make a business case and not just say, "Well, it's faster."

In summary, this is why it is important to have a professional sales and marketing team who understand where to apply their time and resources based on business conditions at the customer. If you are 10% better in technology, then you are not going to replace the competition unless it is tied to other business reasons.

Back to the top, there are comparisons. I would trust any. Make your own based on your specific needs. Get the EDA vendors to do as much of the work for you as possible. Invite them in. Use their resources. And let them live or die on the performance of their tools.
 
Shishir Gupta
...
When it comes to #1, some of the larger companies have very efficient flows. They shield their engineers from the tool and its transparent to the engineer what tool they are using. They also have comprehensive checks to ensure data hand off from one tool to another tool or from one engineer to another engineer is complete .......

In your dreams (mine too). The current CAD flows are a godawful mess of tools that were acquired by the big EDA companies and never properly integrated. Most of the standard interfaces pre-date power management and are in no way ready for 3-D ICs.

Both Cadence and Synopsys have about 5 simulator products. As an expert in that area I can only think of about three simulation techniques, and you could certainly do them all in one tool. A friend at another company says their "secret sauce" is a list of hacks/workarounds that make the tools work in practice - of course they now have no incentive to actually get the tools fixed since that would benefit their competition.

The EDA vendors like it they way it is because they make money on licenses rather than results (working Silicon), and the more complicated the flow the harder it is to switch tools - e.g. nobody will switch simulators for something that is only 2x faster.
 
Hey Joe, yeah it is starting to shape up that it might just be me that acts as the platform to do just that. Who knows who I may talk to about this, I am quite sure that I can secure some funding from interested parties that see this as a value add to their firm or industry initiative.

However as others have stated, and just like Jeremy's recent comment, is that there are multiple vectors of how to evaluate the tools would need to be clearly defined and in what context their use. This is called a user based design requirements, I would likely propose doing a HOQ (House of Quality) analysis w/ inputs from users on what is goodness and what is not, that way we create a comprehensive set of benchmarking values for all to see, complete transparency, so that at the very least everyone knows what we are doing, and how we are doing it. That way everyone stays honest. What do you think?
 
Jeremy - you rock, awesome suggestions. However until the EDA vendors change tactics and embrace the needs of the customer's, a Market insurgency, for which I have been accused of having skills in, which may be necessary to get the vendors attention, and their willingness to participate in this little thing we call life, well shall see....it can be a very uncomfortable place to be when on the receiving end of a market insurgency, companies lose $$ and customers.

Please read what "Market Insurgents", and their kissing cousins, "Positive Deviants" are below, just to give you an idea about what it is that I am getting at how this can be achieved. To put it more succintly I am suggesting to take this issue a couple steps further than where the vendors can go. I have been accused of being both of these descriptive terms.

Market Insurgency:
How Markets Are Made and Broken by Social Activists | Stanford Graduate School of Business
Market rebels and radical innovation | McKinsey & Company
Rao, H.: Market Rebels: How Activists Make or Break Radical Innovations.

Positive Deviants:
Positive Deviant | Fast Company | Business + Innovation
A Q&A on Positive Deviance, Innovation and Complexity | Aid on the Edge of Chaos
http://opinionator.blogs.nytimes.com/2013/02/27/when-deviants-do-good/?_r=0

Background:
When I 1st saw this issue of the EDA tools a couple years ago, I assessed what was the most effective way to deal w/ the challenges, that have been highlighted by everyone here so far. When in discussion w/ one of the EDA vendors I knew that asking nice about getting what was needed was falling upon deaf ears. So I felt it necessary to propose what I am proposing to all of you here, now. So the "fluffy bunny" approach was not going to work w/ them.

Gents, (and ladies if you are reading this) - this isn't about me, this is about all of us and what we need to get done in our work, your work is designing good products and using the mfg envelope to deliver the products that customers want, and it is what gets you paid.

Me I design and specialize in innovation methodologies / frameworks / methods and tools for the high-tech industry, that is not in dispute.

I am seeking us to work together to make this happen for our own mutually inclusive benefit. All of our work is hard enough as it is w/out someone placing restrictions on us that prevents us from achieving what we need to.
 
The EDA vendors like it they way it is because they make money on licenses rather than results (working Silicon), and the more complicated the flow the harder it is to switch tools - e.g. nobody will switch simulators for something that is only 2x faster.

I was at Cadence when CEO Mike Fister attempted to get revenue for a tool (a new router) based on results (i.e some % of the money the customer saved due to area reduction). Every customer he approached refused. The license model seems pretty 'sticky.'
 
Hey Beth - what do you think of simguru's comment? It makes sense to me.

The other question that seems to be raised by simguru's comment is there a work around even if the Cloud based design flows are not in place, and get us a benchmark anyway?
 
Hey Beth - what do you think of simguru's comment? It makes sense to me.

The other question that seems to be raised by simguru's comment is there a work around even if the Cloud based design flows are not in place, and get us a benchmark anyway?

I work for Mentor now (although I'm not speaking for the company), but don't have much visibility into the growth of cloud-based EDA. That said, I have the impression that the big semiconductor companies are too concerned about security to use cloud-based EDA, and they don't need it anyway. Limited cloud-based EDA will be useful for small companies and we'll see more of it, but I predict it will always be a niche. As for public benchmarks, I agree with Jean-François Agaësse and some others that the variables are too great for there to be meaningful generic benchmarks. Plenty of semi companies *do* publish benchmarks on specific designs, you see them all the time in trade shows, but how a tool did on a Broadcom design will probably tell you little about how it will perform on yours.
 
Consider Open Source or Where There are Fewer Restrictions on Publication

Hey Simguru - I see what you've just written is coming from someone who actually knows how such a benchmark would actually need to be set up, and done so independently, so as not pervert the outcomes and favor one EDA provider over another. I don't think I need permission as I don't have an EDA license just yet, but I am very interested in the information and knowledge thereof.

I absolutely could provide such a location as well as publish the results so that they are available to the world, free of charge, your help in this would be much appreciated, if not outright you writing it. Someone we could do for the industry. I would just be happy enough to assist in the effort in any way that I can. I am not bound by such egregious, anti-customer centric licensing agreements. So how about this let's connect on LinkedIn and we can discuss specifics at our leisure? What do you think?

Thank you so much for your suggestion and proposal.

Best
Richard Platt - Managing Partner
Work: 503.421.9391
Email: richard.platt@sig-hq.com
http://www.linkedin.com/in/richardplatthttp://www.linkedin.com/in/richardplatt
Skype: richard.platt101

The real value may be in the benchmark designs themselves. You might start at the FPGA end of the market where you could afford the tools and publish results privately. Depending upon the tool set you are talking about there may not be enough distinct customers to justify a common benchmark effort (e.g. place and route for chip designs). Simulation tools for FPGAs might be one market where folks might pay something for an analysis. There is a long history of efforts in this area, including ork done by Tom Moxon that was published in EE Times and work done by Seva (Yatin Trivedi and Larry Saunders) in the 1990's, as well as the EDAC Interoperability Lab. The economics are tricky and the incentives are such that most large firms would prefer to run their own benchmarks as a barrier to competitors.

It's an interesting idea, I hope that you are able to find a way to make it work.

Sean Murphy SKMurphy, Inc. 408-252-9676
 
Jeremy Birch
The problem also varies by tool type. For simulators and other analysis tools there are standard inputs, commands, and a knowable golden result. For synthesis, place and route and other heuristic tools, the result may differ a large amount due to a small difference in the input data, the tool invocation varies widely, and there is generally not a knowable golden result. For these types of tool it is harder to know which tool is best for your type of design, and it is also not so easy to identify metrics by which to measure better or worse tools. For instance we might assume that all P&R results need to be DRC and LVS clean (although the benchmark data may not allow this to be achieved), but what do you measure after that? A result with more wire and more vias might be bad (longer delays) or might be good (higher yield, lower coupling etc). To determine the best result then depends on the use of analysis tools which might vary again in their analysis.

Tools which excel at large square designs with lots of metals may be truly awful at long thin designs with 2 or 3 metals, and vice versa. Producing meaningful benchmarks that help a wide variety of customers would be pretty tricky.
 
Timo D. Hämäläinen
The business is tool centric (bottom-up), but should be process centric (top-down) and thats why it is difficult to measure the productivity or even model the process. Most often developers say it is simply impossible because the same tools can be used in so many different ways.
 
Gabe Moretti

Because when a company purchases a license for a competitive tool from any vendor but a struggling start-up that company signs a non-disclosure agreement that prohibits the kind of comparison you want. The "hype" is defeated by learned evaluation, not by some report whose value is lost after less that a quarter. Moreover a tool "effectiveness" depends in a significant way from the ease of integration into a specific flow, that varies from company to company. Your meter, "Degree of Effectiveness" is relative to the using company, the skill of the engineers that are working with the tool, and the nature of the specific design. One can prove practically anything with a benchmark, you "just' have to use the right design. Let's not be lazy, let's do the work of evaluating within the environment and the type of designs the tool will be used in. And, by the way, at 20 nm and below the results also depend on the foundry used as well.
 
Frank W. Bennett

Whenever I hear this topic come up I'm reminded of the contest John Cooley ran in 1978 at an EDA conference "Verilog verses VHDL" or at that time Cadence verses Synopsys. The results of that can still be found at and other places like:

http://www.bawankule.com/verilogcenter/contest.html

This was a good measurement of productivity and if we are considering buying a $70,000 license for software to run on a $1500 PC, I think it matters. The GPL Cver is a good example how open source code (free) can replace Verilog ($17k). Cver covers 110% of the original Cadence functionality in an amasingly few lines of code. It doesn't handle encryped HDL source code and a faster compiled HDL code with support was available as CVC from Tachyon Design Automation. Hope these guys are still around but a target for aquisition from the likes of Mentor who generally buys up the little fish, i.e. Code Sourcer.

I'm a big proponent of open source tools like Linux, Kicad, GnuCash. In addition to rooting for the little guy, open source provides the source code with the guarantee of better support of the executable and the output. Another example, is Kicad, schematic and PCB layout application has pretty much caught up with the $10k PCB packages. The outputs, your design source, are ASCII which can be tweeked with your favorite text editing tools. or join the developer's group. The user group provides a good measure of ease of use when the requests run along the lines of footprint or netlist format requests.

EDA support has failed under the Windows operating system and PCB design is
soon to follow.
 
Frank what an awesome comment, if you have other insights please share them. This was extremely insightful and at least helpful to me, much appreciate your inputs
 
I understand what you are saying Beth, but I disagree w/ respectfully I see it differently, I do not subscribe to the model presented by the EDA/DFM vendors. I believe our community deserves better. So perhaps because you work for Mentor as you self admit, that we will have to agree to disagree, I have seen and heard from others who also share my take on things.
 
Back
Top