( ESNUG 547 Item 5 ) -------------------------------------------- [02/27/15]

Subject: Bernard Murphy on DVcon, UVM, Lib-bu, SW vs. HW engineers, specs

    Hi, John,

    I thought I'd follow in this pattern and give your readers our
    own internal DVcon'14 Trip Report in anticipation of the
    DVcon'15 coming up next week.
       
        - Bernard Murphy
          Atrenta, Inc.                            San Jose, CA

           ----    ----    ----    ----    ----    ----   ----

DVCON ATTENDANCE NUMBERS

          2006 :################################ 650
          2007 :################################### 707
          2008 :######################################## 802
          2009 :################################ 665
          2010 :############################## 625
          2011 :##################################### 749
          2012 :########################################## 834
          2013 :############################################ 883
          2014 :############################################ 879

Attendance was flat over last year, though the number of exhibitors grew
(40 this year, including 15 first-timers and 10 international).

My main take-aways were:

    - There is a common theme around software-driven SoC verification.
      This was generally agreed in multiple forums.  But this is not
      using the production software, not even a branch of that software.
      This software is purpose-built for verification.

    - UVM and constrained random are great for IP and maybe some sub-
      systems, but not at all so for SoCs.  Speakers openly discouraged
      developing UVM or any standard in this direction.

    - UVM nicely formalizes what it can do.  We now need breakthrough
      innovation to get us to what it can't do (well).

    - Standards stifle, rather than accelerate progress on verification
      for SoCs.

There was emphasis this year on adding chip design content in addition to
verification.  This seemed to come across mainly through tutorials and
papers on UPF, especially UPF 2.1 - not a bad place to start.

           ----    ----    ----    ----    ----    ----   ----
           ----    ----    ----    ----    ----    ----   ----
           ----    ----    ----    ----    ----    ----   ----

LIP-BU'S KEYNOTE ADDRESS

For those of you who haven't seen the stats, Lip-Bu divided the chip market
into Mobility, Cloud and Internet of Things

    - Mobility (1B phones, 400M tablets today, 50%+ compound growth)
    - Cloud (less easy to divide into units but also 50%+ compound growth)
    - Internet of Things (perhaps 50B units, possible growth to >14T units)

He also pointed out that systems companies are becoming much more significant
as drivers in at least the first two categories.

     

Amazon, Apple and Google are obvious examples.  Lip-Bu talks frequently to
his friends in these and other companies (like Facebook and Twitter) to
encourage them to see hardware as a differentiator -- that ultimately they
cannot win on software alone.

     

Conversely for the HW developers in the audience, Lip-Bu cited that SW,
verification, and validation accounted for 80% of a chip's total development
costs.

     

He also cited how design reuse and 3rd party IP cores were exponentially
growing in current SoC design.  (Perhaps as an oblique reference to how CDNS
was expanding in IP vs. SNPS?)

     

He made a pitch for the importance of ESL design, pointing to the recent
Cadence acquisition of  Forte. He also stressed that system-level analysis
of characteristics like power and signal integrity are becoming increasingly
important.

Lip-Bu continues to see opportunity for innovation in silicon design, but
this needs to be more aware of the eco-system (IP, tools, software, foundry,
OEM).  Interactions between these are becoming more strongly coupled; so
opportunities for innovation will be prominent in the overlaps.


BERNARD'S REACTION TO LIP-BU'S KEYNOTE

This speech was a bit of a let-down.  Lip-Bu gave a polished but not very
informative address.  The standard complaint, that he doesn't have technical
depth in EDA, is given.  But we know he is a strong business leader, an
expert in finance, and an active investor.  I was looking forward to a talk
more along these lines -- a nice contrast to all the UVM, UPF, formal and
other discussions that filled the week.   But his DVcon keynote was really
a replay of common knowledge, even among us gear-heads.

I'm sure Lip-Bu has plenty of keen business insights.  Perhaps he didn't
want to tip his hand.  Perhaps his insights are competitively significant.
Who knows?

           ----    ----    ----    ----    ----    ----   ----
           ----    ----    ----    ----    ----    ----   ----
           ----    ----    ----    ----    ----    ----   ----

    NOTE: To shorten this report, I have provided a digest of comments
    from the panels rather than blow-by-blow quoting.  Apologies in
    advance to anyone whose brilliant insights I failed to include.

        - Bernard Murphy, CTO of Atrenta


      Panel: "Is Software The Missing Piece In Verification?"

  Moderator: Ed Sperling - SemiEngineering.com

  Panelists:

           - Tom Anderson - Breker Systems
           - Kenneth Knowlson - Intel
           - Steve Chappell - Synopsys
           - Sandeep Pendharkar - Vyavya Labs
           - Frank Schirrmeister - Cadence
           - Mark Olen - Mentor Graphics

Great discussion, not always obviously on-topic, but lively and adversarial;
the best kind of panel.  Frank dominated the responses, though others also
made worthy contributions.  No bad thing, Frank is an entertaining and
opinionated guy.  He introduced us to the concept of "Elvis" and "non-Elvis"
software.  (Elvis software is the software that leaves the building.)


Q: "What is software-driven verification, and is it different from system-
    level verification or just marketing hype?"

    - Software-driven verification is using the CPUs on an SoC design
      along with purpose-built software as the primary mechanism to
      drive full-chip verification, in contrast to a testbench.  In
      deployment, the device uses the processors, so it is unreasonable
      to assume you can adequately test it without involving the
      processors.  But this is not the same as testing the production
      software.  Here you are using specialized software as part of
      the hardware test.

    - What if the design doesn't contain a CPU?

    - We have seen cases where people add a simple processor, either
      open-source, or a small ARM core to create an embedded testbench.
      This makes the software easier to handle than an SV testbench.

    - To answer why is this happening now: there are so many engines
      involved in verification, from TLM techniques to RTL techniques:
      simulation, formal, emulation, FPGA-based.  A key advantage of
      using software is that you can reuse verification across many
      of these techniques.  You can even use it in post-silicon debug.

    - Most test software is typically not a branch of the production
      software.  Of course you need to test the Elvis software at
      some point, but the non-Elvis software is designed to test
      bring-up, all the peripherals start in the right order, that
      sort of thing.

    - We are seeing a lot more IPs these days are highly programmable.
      So it makes sense that when you are verifying the design, an
      optimum way to drive that is through software.

    - But validating with the product software is also a part of the
      process.  You just don't want to confuse that with more detailed
      verification.

    - You need to be careful to distinguish what software aspect we
      are talking about - production software, test software or models?

    - This is not a new problem.  Customers since 1995 have been using
      software to drive hardware verification.  It hasn't been easy but
      they have been finding a way.  About 10 years ago, hot-swap
      technology made it simpler to switch between speed and accuracy,
      and now we're getting to more clever approaches.

    - As we get more processors and more code, we need to be able to
      reuse it more.  The more we can reuse between the Elvis and
      non-Elvis software, the better.


Q: "You've got a room-full of people who turned up at 8:30 in the morning
    to hear about something that is almost 20 years old.  Why now?"

    - We're in the hockey stick is what has changed.  Diagnostic software
      is piece that is changing.  And the need to run verification across
      engines is now paramount, which is a big change.

    - Software becomes the testbench for driving verification - and we
      need to be able to reproduce bugs found post-silicon using same
      testbench I ran on the netlist.

    - We see real tension between validation people and the embedded
      software people.  This is partly a capacity/complexity problem;
      no one person understands all details about the hardware and
      software.  Making communication between hardware and software
      more efficient is key. Is a common environment possible?


Q: "How does this change verification? What has to change?"

    - You need to have a processor in the system.  Almost like a next-
      level BIST with an embedded processor for test.  Plus you
      need an education change - the verification engineer needs to
      understand software - either needs to generate the software
      themselves or at least understand what is happening in the
      software - both Elvis and non-Elvis.

    - We are finding that the validation software exercises the hardware
      differently than production software.  But we struggle to define
      comprehensive use-cases to drive directed testing.

    - Graph-based verification is one approach that can work to
      communicate between different teams - it can also define a common
      level between different engines.

    - There was a lot of debate on graph-based approaches and some
      confusion on graphical tools (undesirable) versus supporting the
      concept.  There seemed to be general agreement that the approach
      is a good start, but less agreement about scalability to SoC.
      There is a proposal to Accellera to create a standard for
      graph-based design.

    - Perhaps model-based approaches are the ultimate direction.


Q: "What does software-driven verification do for mixed-signal, RF, etc.?"

    - Cadence is trying to understand how much of the system the customer
      needs to model - not just the chip, but out to the air interface,
      loss due to thunderstorms, etc.  What we and our customers are still
      trying to figure out is what bugs you find when simulate everything
      (digital and analog) together.  Where does total system simulation
      add value?

    - The ability to do this is still nascent.  We need seamless modeling
      across the flow.  Need a framework to run software on, then be able
      to take that to any of the platforms to run verification.  The SW
      people need to start on FPGA platform, which gives them ability to
      get quick turns.  The SW people also need to understand the HW, and
      HW teams need to provide SW with a reasonably validated platform.

    - When SW people meet HW people, the first thing they do is to exchange
      business cards.  EDA vendors are often enabling these meetings.  Do
      we need a new kind of engineer or a designer that can straddle
      between teams - i.e. SW engineers working in the HW team?

    - We are developing systems/glue engineers; building specific firmware
      working with design people.

    - Engineers building non-Elvis software are embedded people on loan
      because hardware people can't write software.

    - Many engineers use our Labview to test post silicon, and they are
      thinking about the total system.  Do the people who do this scale
      to the pre-silicon task?


Q: "Can we get good coverage of an SoC?  Or is it spinning out of control?
    Do we need to just test to 'good enough'?"

    - You can get good coverage at different levels of design, but you
      are covering different things.

    - If there is a good level of automation to translate between different
      levels, then there can be more investment at higher levels, but
      otherwise it is difficult to get engineers to invest in high-level
      (abstract) testing if they cannot leverage to other testing - that
      just looks like more work with uncertain return.

    - We need to figure out how to do more with less, and only with the
      relevant pieces.

    - We talk about language, but that doesn't have to be syntax - it is
      more about communicating between teams and between engines.

    - Verification reuse is key.


Q: "Verification costs are now 80-85% of the NRE.  Is there a way we can
    knock down the cost?"

    - Automation is always a part of the solution, but we should be
      realistic.  More likely we will hold steady through higher
      productivity.  Complexity is still growing faster than
      productivity.  Verification will always be an unbounded problem.

    - We need to look beyond what is available today.  Standards are
      good, but they cement only what is already well known.  We need
      to look for breakthrough approaches.

    - A truly seamless flow would reduce learning curve, and increase
      productivity.

    - Hardware verification won't go away with production software
      validation.  Production software testing needs weeks to run on
      a real platform - that's a non-starter for hardware verification.
      Non-Elvis software enablement is the real growth area.

IMHO, by far this was the best panel of DVcon'14.

           ----    ----    ----    ----    ----    ----   ----

   Editor's Note: This is continued in ESNUG 547 #6 -- Bernard on UVM.

           ----    ----    ----    ----    ----    ----   ----

Related Articles:

    Bernard Murphy on DVcon, Lib-bu, SW vs. HW engineers, specs, IP
    Bernard on UVM, SW-driven verification, C-to-RTL, and Jim Hogan

Join    Index    Next->Item






   
 Sign up for the DeepChip newsletter.
Email
 Read what EDA tool users really think.














Feedback About Wiretaps ESNUGs SIGN UP! Downloads Trip Reports Advertise

"Relax. This is a discussion. Anything said here is just one engineer's opinion. Email in your dissenting letter and it'll be published, too."
This Web Site Is Modified Every 2-3 Days
Copyright 1991-2024 John Cooley.  All Rights Reserved.
| Contact John Cooley | Webmaster | Legal | Feedback Form |

   !!!     "It's not a BUG,
  /o o\  /  it's a FEATURE!"
 (  >  )
  \ - / 
  _] [_     (jcooley 1991)