Monday, October 15, 2018
Search
  
Submit your own News for
inclusion in our Site.
Click here...
Breaking News
New Cyber-shot HX99 Is Possibly the World's Smallest Travel High Zoom Camera
New Palm Phone Is About the Size of a Credit Card
HP Arms ZBooks with Intel Core i9 Processors
Dell Expands UltraSharp Monitor Family With New UltraSharp 49-inch Dual QHD Monitor
ARM and Intel to Secure Internet of Things
Adobe Announces New Creative Cloud at MAX 2018
8K TV Shipments to Reach More Than 400,000 Units in 2019, IHS Markit Says
Sharp Releases New AQUOS 8K LCD TVs With 8K Satellite Tuner
Active Discussions
Which of these DVD media are the best, most durable?
How to back up a PS2 DL game
Copy a protected DVD?
roxio issues with xp pro
Help make DVDInfoPro better with dvdinfomantis!!!
menu making
Optiarc AD-7260S review
cdrw trouble
 Home > News > General Computing > Microso...
Last 7 Days News : SU MO TU WE TH FR SA All News

Wednesday, August 23, 2017
Microsoft Unveils Project Brainwave for Real-time AI


Today at Hot Chips 2017, a cross-Microsoft team unveiled a new deep learning acceleration platform, codenamed Project Brainwave.

Microsoft has selected Intel Stratix 10 FPGAs as a key hardware accelerator in its new accelerated deep learning platform. This FPGA-based accelerated deep learning platform is capable of delivering "real-time AI," which will allow cloud infrastructure to process and transmit data as fast as it comes in, with ultralow latency. In the cloud, delivering real-time AI is becoming more important as systems are required to process live data streams, including video, sensors or search queries, and rapidly deliver the data back to users.

The Project Brainwave system is built with three main layers:

  • A high-performance, distributed system architecture;
  • A hardware DNN engine synthesized onto FPGAs; and
  • A compiler and runtime for low-friction deployment of trained models.

First, Project Brainwave leverages the massive FPGA infrastructure that Microsoft has been deploying over the past few years. By attaching high-performance FPGAs directly to Microsoft's datacenter network, the company can serve DNNs as hardware microservices, where a DNN can be mapped to a pool of remote FPGAs and called by a server with no software in the loop. This system architecture both reduces latency, since the CPU does not need to process incoming requests, and allows very high throughput, with the FPGA processing requests as fast as the network can stream them.

Second, Project Brainwave uses a powerful "soft" DNN processing unit (or DPU), synthesized onto commercially available FPGAs. A number of companies are building hardened DPUs. Although some of these chips have high peak performance, they must choose their operators and data types at design time, which limits their flexibility. Project Brainwave takes a different approach, providing a design that scales across a range of data types, with the desired data type being a synthesis-time decision. The design combines both the ASIC digital signal processing blocks on the FPGAs and the synthesizable logic to provide a greater and more optimized number of functional units. This approach exploits the FPGA's flexibility in two ways. First, Microsoft has defined highly customized, narrow-precision data types that increase performance without real losses in model accuracy. Second, the company can incorporate research innovations into the hardware platform quickly (typically a few weeks), which is essential in this fast-moving space. As a result, Microsoft says it has achieved performance comparable to - or greater than - many of these hard-coded DPU chips but are delivering the promised performance today.

Third, Project Brainwave incorporates a software stack designed to support the wide range of popular deep learning frameworks. The system already supports Microsoft Cognitive Toolkit and Google's Tensorflow, and Microsoft's plan is to support many others. The company has defined a graph-based intermediate representation, to which they convert models trained in the popular frameworks, and then compile down to our high-performance infrastructure.

The system is architected to show high actual performance across a wide range of complex models, with batch-free execution. Companies and researchers building DNN accelerators often show performance demos using convolutional neural networks (CNNs). Since CNNs are so compute intensive, it is comparatively simple to achieve high performance numbers. Those results are often not representative of performance on more complex models from other domains, such as LSTMs or GRUs for natural language processing. Another technique that DNN processors often use to boost performance is running deep neural networks with high degrees of batching. While this technique is effective for throughput-based architectures-as well as off-line scenarios such as training-it is less effective for real-time AI. With large batches, the first query in a batch must wait for all of the many queries in the batch to complete. Microsoft's system, designed for real-time AI, can handle complex, memory-intensive models such as LSTMs, without using batching to juice throughput.

Even on early Stratix 10 silicon, the ported Project Brainwave system ran a large GRU model-five times larger than Resnet-50 - with no batching, and achieved record-setting performance. The demo used Microsoft's custom 8-bit floating point format ("ms-fp8"), which does not suffer accuracy losses (on average) across a range of models. Microsoft showed Stratix 10 sustaining 39.5 Teraflops on this large GRU, running each request in under one millisecond. At that level of performance, the Brainwave architecture sustains execution of over 130,000 compute operations per cycle, driven by one macro-instruction being issued each 10 cycles.

Microsoft is working to bring this real-time AI system to users in Azure, so its customers can benefit from Project Brainwave directly, complementing the indirect access through our services such as Bing.



Previous
Next
LG to Establish Factory for Electric Vehicle Components in the U.S.        All News        Panasonic Touts 1TB Optical Disc For Storing Cold Data
LG to Establish Factory for Electric Vehicle Components in the U.S.     General Computing News      Google to Detail Titan Security Chip for Cloud Services

Get RSS feed Easy Print E-Mail this Message

Related News
Micron Announces New $100 Million Venture Investment in AI
New Intel Vision Accelerator Solutions Speed Up Deep Learning and Artificial Intelligence on Edge Devices
Huawei Unveils AI Strategy and New Chips
Microsoft's Mobile Phone Keyboard SwiftKey Translates As You Text
Microsoft Focuses on Security and Brings AI to the Masses at Ignite 2018
Samsung AI Forum Underlines the Importance of Unsupervised Learning
Facebook Confirms Internal Silicon Team
Cisco Unveils Server for Artificial Intelligence and Machine Learning
Samsung Opens a New AI Center in New York City
New Google Tool Spots Child Abuse in Photos
Broadcom to Design 7-nm AI processor For Wave: report
Microsoft Calls for Public Regulation of AI Face Recognition Software

Most Popular News
 
Home | News | All News | Reviews | Articles | Guides | Download | Expert Area | Forum | Site Info
Site best viewed at 1024x768+ - CDRINFO.COM 1998-2018 - All rights reserved -
Privacy policy - Contact Us .