Intel Delivers Real Time AI in Microsoft's New Accelerated Deep Learning Platform
August 24, 2017 -- Artificial intelligence (AI) and deep learning (DL) promise to transform how people engage with the world and how businesses operate in order to make smarter decisions, and Intel® technology is enabling this transformation. Today, Intel is highlighting how its AI technology is being leveraged by one of the world’s largest cloud service providers to run complex deep learning models, which enable applications ranging from facial recognition and voice recognition on smartphones to autonomous driving.
Microsoft has selected Intel® Stratix® 10 FPGAs as a key hardware accelerator in its new accelerated deep learning platform – code-named Project Brainwave. This FPGA-based accelerated deep learning platform is capable of delivering “real-time AI,” which will allow cloud infrastructure to process and transmit data as fast as it comes in, with ultralow latency. In the cloud, delivering real-time AI is becoming more important as systems are required to process live data streams, including video, sensors or search queries, and rapidly deliver data back to users.
Through a blog issued today, Microsoft detailed how Project Brainwave, leveraging Intel FPGAs, is capable of handling extremely challenging deep learning models at “unprecedented” levels of performance and flexibility. Microsoft demonstrated its FPGA-based deep learning platform at Hot Chips 2017, a symposium that showcases the latest advancements in semiconductor technology.
Microsoft was the first major cloud service provider to deploy FPGAs in its public cloud infrastructure and the technology advancements it is demonstrating today with Intel Stratix 10 FGPAs enable the acceleration of deep neural networks (DNNs) that replicate “thinking” in a manner that is conceptually similar to that of the human brain.
AI is a rapidly evolving field that requires multiple technologies to efficiently manage various workload requirements. Intel offers a broad set of technologies to enable the market’s evolution; including Intel Xeon® processors, Intel FPGAs and Intel Nervana™ ASIC technology. Compared to dedicated deep learning hardware accelerators that are optimized to run a single workload, the flexibility of Intel’s FPGAs enables users to customize the hardware to meet specific workload requirements, and reconfigure the hardware rapidly as deep learning workloads and use models change. Intel Stratix 10 FPGAs combine hardened processor blocks that deliver high levels of sustained performance and efficiency, with a programmable fabric for user customization.
Many silicon AI accelerators today require grouping multiple requests together (called “batching”) to achieve high performance. Project Brainwave, leveraging the Intel Stratix 10 technology, demonstrated over 39 Teraflops of achieved performance on a single request, setting a new standard in the cloud for real-time AI computation. Stratix 10 FPGAs set a new level of cloud performance for real-time AI computation, with record low latency, record performance and batch-free execution of AI requests.
“We exploit the flexibility of Intel FPGAs to incorporate new innovations rapidly, while offering performance comparable to, or greater than, many ASIC-based deep learning process units,” said Doug Burger, distinguished engineer at Microsoft Research NExT.
Microsoft is currently working to deploy Project Brainwave in the Azure cloud so that customers eventually can run complex deep learning models at record-setting performance.
Machine learning platforms powered by Intel technology help transform data into actionable business intelligence through accelerated model training, fast scoring highly scalable infrastructure. In conjunction with the latest Intel Xeon Scalable processors, Intel FPGAs are customizable and programmable to deliver low latency and flexible precision, with higher performance per watt for deep learning inference compared with Intel Xeon processors alone.
To learn more about Intel’s comprehensive portfolio of AI solutions, visit www.intel.com/ai
More:
- Revving up Microsoft Project Brainwave with Intel FPGAs (Dan McNamara Blog)
- Artificial Intelligence (Press Kit)
|
Intel FPGA Hot IP
Related News
- NEUCHIPS Secures $20 Million in Series B2 Funding to Deliver AI Inference Platform for Deep Learning Recommendation
- Mentor's Catapult HLS enables Chips&Media to deliver deep learning hardware accelerator IP in half the time
- Microsoft Outlines Hardware Architecture for Deep Learning on Intel FPGAs
- Jetstream Media Technologies' New Real-Time Video Effect IP Core Delivers Easy to Use, Fun and Personalized Memories
- VideoLocus Introduces World's First Real-Time H.264/MPEG-4 AVC Standard Definition Video Encoder and Video Evaluation Platform
Breaking News
- TSMC drives A16, 3D process technology
- Frontgrade Gaisler Unveils GR716B, a New Standard in Space-Grade Microcontrollers
- Blueshift Memory launches BlueFive processor, accelerating computation by up to 50 times and saving up to 65% energy
- Eliyan Ports Industry's Highest Performing PHY to Samsung Foundry SF4X Process Node, Achieving up to 40 Gbps Bandwidth at Unprecedented Power Levels with UCIe-Compliant Chiplet Interconnect Technology
- CXL Fabless Startup Panmnesia Secures Over $60M in Series A Funding, Aiming to Lead the CXL Switch Silicon Chip and CXL IP
Most Popular
- Cadence Unveils Arm-Based System Chiplet
- CXL Fabless Startup Panmnesia Secures Over $60M in Series A Funding, Aiming to Lead the CXL Switch Silicon Chip and CXL IP
- Esperanto Technologies and NEC Cooperate on Initiative to Advance Next Generation RISC-V Chips and Software Solutions for HPC
- Eliyan Ports Industry's Highest Performing PHY to Samsung Foundry SF4X Process Node, Achieving up to 40 Gbps Bandwidth at Unprecedented Power Levels with UCIe-Compliant Chiplet Interconnect Technology
- Arteris Selected by GigaDevice for Development in Next-Generation Automotive SoC With Enhanced FuSa Standards
E-mail This Article | Printer-Friendly Page |