Advertisement
News
EEtimes
News the global electronics community can trust
eetimes.com
power electronics news
The trusted news source for power-conscious design engineers
powerelectronicsnews.com
EPSNews
News for Electronics Purchasing and the Supply Chain
epsnews.com
elektroda
The can't-miss forum engineers and hobbyists
elektroda.pl
eetimes eu
News, technologies, and trends in the electronics industry
eetimes.eu
Products
Electronics Products
Product news that empowers design decisions
electronicproducts.com
Datasheets.com
Design engineer' search engine for electronic components
datasheets.com
eem
The electronic components resource for engineers and purchasers
eem.com
Design
embedded.com
The design site for hardware software, and firmware engineers
embedded.com
Elector Schematics
Where makers and hobbyists share projects
electroschematics.com
edn Network
The design site for electronics engineers and engineering managers
edn.com
electronic tutorials
The learning center for future and novice engineers
electronics-tutorials.ws
TechOnline
The educational resource for the global engineering community
techonline.com
Tools
eeweb.com
Where electronics engineers discover the latest toolsThe design site for hardware software, and firmware engineers
eeweb.com
Part Sim
Circuit simulation made easy
partsim.com
schematics.com
Brings you all the tools to tackle projects big and small - combining real-world components with online collaboration
schematics.com
PCB Web
Hardware design made easy
pcbweb.com
schematics.io
A free online environment where users can create, edit, and share electrical schematics, or convert between popular file formats like Eagle, Altium, and OrCAD.
schematics.io
Product Advisor
Find the IoT board you’ve been searching for using this interactive solution space to help you visualize the product selection process and showcase important trade-off decisions.
transim.com/iot
Transim Engage
Transform your product pages with embeddable schematic, simulation, and 3D content modules while providing interactive user experiences for your customers.
transim.com/Products/Engage
About
AspenCore
A worldwide innovation hub servicing component manufacturers and distributors with unique marketing solutions
aspencore.com
Silicon Expert
SiliconExpert provides engineers with the data and insight they need to remove risk from the supply chain.
siliconexpert.com
Transim
Transim powers many of the tools engineers use every day on manufacturers' websites and can develop solutions for any company.
transim.com

Baidu Releases AI Benchmark

By   09.26.2016 0

SAN JOSE, Calif. – Calling for 100x faster processors, China Web giant Baidu released DeepBench, an open source benchmark for how fast processors train neural networks for machine learning.

DeepBench is available online along with first results from Intel and Nvidia processors running it. The benchmark tests low-level operations such as matrix multiplication, convolutions, handing recurrent layers and the time it takes for data to be shared with all processors in a cluster.

Machine learning has emerged as a critical workload for Web giants such as Baidu, Google, Facebook and others. The workloads come in many flavors serving applications such as speech, object and video recognition and automatic language translation.

Today the job of training machine learning models “is limited by compute, if we had faster processors we’d run bigger models…in practice we train on a reasonable subset of data that can finish in a matter of months,” said Greg Diamos, a senior researcher at Baidu’s Silicon Valley AI Lab.

The lab has found, for example, it can reduce by 40% errors in automatic language translation for every order-of-magnitude performance improvement in computing. “We could use improvements of several orders of magnitude–100x or greater,” said Diamos.

No striking results emerged from initial tests on NVidia TitanX, TitanX Pascal and M40 GPUs and Intel Xeon Phi processors, according to the Baidu researchers.

“Even for the same type of operations like matrix multiplies, depending on the sizes of the models and the ways they are used, performance varies even on the same processor,” said Diamos. “We aren’t as concerned about minor differences between these processors–we want both to be 10 times faster,” he said.

Nevertheless, one executive took occasion to claim bragging rights.

“Baidu’s DeepBench results clearly highlight Pascal as the performance leader across all deep learning workloads,” said Ian Buck, vice president of accelerated computing at Nvidia. “When full applications are benchmarked, such as Caffe AlexNet or Caffe VGG-19, our internal testing shows Pascal is up to 6x faster than [Intel’s] Knights Landing,” he added.

The Baidu researchers also hope other processor vendors and data center operators contribute to expanding and running chips on the benchmark.

“I’d be personally interested in results on AMD GPUs and ASICs from startups with custom hardware for whom it may be difficult to run full models — this might be an easier way to for them to get their capabilities out there,” Diamos said, noting the lab currently uses systems with eight Nvidia TitanX processors to run its speech recognition models.

DeepBench tests low-level hardware libraries, not higher level AI frameworks the data centers create such as Baidu’s PaddlePaddle and Google’s TensorFlow.

“At the framework level, there’s a huge amount of difference in the models, and different models for different apps, but below the frameworks they use a few common operations,” said Sharan Narang a software engineer in Baidu’s AI lab. “The hope is we can find the core common operations that are more actionable for hardware makers,” he said.

The researchers are considering whether they need a benchmark for inferencing, the separate job of using the models to find patterns in data. Today Baidu and Microsoft use FPGAs on servers to accelerate that less compute-intensive work.

Whether competing data center operators collaborate on DeepBench remains to be seen.

To date, Google has made its TensorFlow framework open source and created an ASIC to accelerate inferencing jobs on its servers. For its part, Facebook released an open source server using multiple Nvidia GPUs for training neural networks.

A handful of startups are working on processors optimized for training neural networks including Wave Computing which presents its work at an event this week and Nervana which Intel recently acquired.

— Rick Merritt, Silicon Valley Bureau Chief, EE Times Circle me on Google+

Related posts:

0 comments
Post Comment
joe.raffa   2016-09-27 11:42:38

http://www.meetup.com/ACM-SIGAI-Bay-Area-Chapter/events/233406926/

 

WaveJin   2016-09-27 17:04:54

Great article Rick. At Wave we welcome the goals of DeepBench. Although in its current form, it is useful mainly for comparing GPUs and CPUs, we are looking forward to contributing to this effort to enable fair comparison of non-GPU architectures. More fundamental to Baidu's goal of achieving 100x improvement in compute power, we believe this can only be met through discontinuous versus continuous innovation. If we only focus on incremental improvements of low-level operations such as matrix multiplications or convolutions, we will miss out on opportunities for key innovations.

Wave's CTO, Dr. Chris Nicol, recently wrote an interesting blog about the need for discontinuous innovation in computing architectures to achieve real and game-changing improvements in computing systems for deep learning rather than being constrained by certain types of benchmarks and associated designs.

http://wavecomp.com/blog/benchmarks-can-either-encourage-or-constrain-innovation

While the initial work on DeepBench is very useful, it needs to be complemented with the willingness to think out-of-the-box and innovate at all levels to meet the 100x challenge posed by Baidu. Thank you again for a really insightful article and we are looking forward to continuing discussions on this topic.

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.