Big Data: Design Progress Analytics & IP Theft Protection

Shiv Sikand, IC Manage (DAC 2015 edited transcript)

There’s a lot of talk about analytics and big data. There is a fundamental difference – analytics and analytical thinking requires a cultural shift in the way you track and plan your activities. Those changing thought processes are based on where we are today. Most schedules and organization are very, very reactive.

Transition from Reactive to Proactive

We have standard reports. People fill in data. They report that information at status meetings. There’s usually a leader or a project leader who tries to marshal and corral that information and present it in a succinct fashion during weekly meetings or bi-weekly meetings, and try to draw through and find out what data is relevant to the overall scope of the project, versus very individualized technical problems that a group maybe facing.

Then of course there’s this whole notion of semi-automated data gathering where you’re really requiring or relying on the engineers to themselves populate this data and present it to you. As you know people will tend to color that data because they don’t necessarily understand the full significance of it in the overall project context, and they only want to present the data that they feel is relevant. But the truth is much deeper than that.

Now just because you gather all the big data, doesn’t automatically mean you are no longer reactive at all:  I’ve built a big data system, I grab all this stuff, and now suddenly I’m still reactive. The transition from reactive to proactive is difficult. You don’t get it out of the box. Forward looking decisions, and I think Simon [Burke, Xilinx] pointed that out, really need new types of analysis. That analysis typically is continually changing. You have different constraints as you go through your project lifecycle and then different constraints between projects, because they could have different flavors.

Big Data Requires Powerful Framework

In a typical semiconductor company you’ve got different flavors of devices and IP is being delivered through different mechanisms. The key is to have the ability to keep writing analysis, identify the key indicators, and then drill down deeper to find out exactly what’s going on.  To do that you really need a powerful framework.

What are those framework challenges? The first one is that we’re talking about very, very large amounts of unstructured data. This is machine data on all the jobs that ran. But unless you can correlate the actual events that drove the creation of that data it’s very difficult to write a good model. That’s where a very well structured SCM repository can help drive that correlation.

It’s also very important to remember that project leads are not database gurus, they’re not data scientists. What they want is something that’s usable. They don’t want to have to train a group of people to run those analytics for them. They want to be able to do it in a fairly scalable fashion.

In terms of framework challenges making data useful requires a certain level of optimization. Yes, you do need to grab a lot of your data and put it into the system, but you also want to make sure that it’s optimized, and that you don’t put a lot of wasteful stuff in there, because it just becomes harder to find that needle in the haystack.

Predictive Modeling for Engineers and Managers

You need optimization, you need some notion of predictive modeling, and there’s always a fair amount of statistical analysis. Usually that statistical analysis is based on historical trends or historical precedents that allow you to build your base model. Once you have your base model, then you can extrapolate it based on your complexity and your scaling factors.

The key here is that vendors need to provide that value. Most of that value comes through providing library functions so that you can compute the metrics that you want, and provide easy programmatic interfaces so that not everyone has to be a database expert or a data scientist. Then provide presentation layers so that visualization, which is a very big part of analytics, comes up very, very quickly.

As Simon has mentioned, the tape-out prediction comes after at least one design and you get actionable information and accuracy improves over time. It’s not some magic thing that you’re going to press a button and get a result from a tool.


Shiv Sikand is Vice President of Engineering at IC Manage