Optimizing AI and Machine Learning with eFPGAs
By Cheng Wang, Flex Logix, Inc.
August 6th, 2018, eecatalog.com
Why the performance and flexibility offered by eFPGA is turning out to be a game changer for anyone designing AI and machine learning and struggling to meet the compute demands.
The market for artificial intelligence (AI) and machine learning applications has been growing substantially over the last several years. Designers have a tough row to hoe when it comes to satisfying these applications’ seemingly insatiable compute hunger. They are finding that traditional Von Neumann processor architectures are not optimal solutions for the neural networks fundamental to AI and machine learning.
When GPUs are used to train neural networks, they require floating pointing math that is very compute intensive. However, using integer math for inference, designers can speed computation by turning to FPGAs for neural network processing. Many companies are starting to recognize this, with Microsoft’s Project Brainwave, which uses FPGA chips to accelerate AI, as a perfect example.
E-mail This Article | Printer-Friendly Page |
|
Related Articles
- Accelerating SoC Evolution With NoC Innovations Using NoC Tiling for AI and Machine Learning
- Performance Evaluation of machine learning algorithms for cyber threat analysis SDN dataset
- Optimizing Electronics Design With AI Co-Pilots
- Exploring Machine Learning testing and its tools and frameworks
- Artificial Intelligence (AI) utilizing deep learning techniques to enhance ADAS
New Articles
Most Popular
- Streamlining SoC Design with IDS-Integrate™
- System Verilog Assertions Simplified
- System Verilog Macro: A Powerful Feature for Design Verification Projects
- Enhancing VLSI Design Efficiency: Tackling Congestion and Shorts with Practical Approaches and PnR Tool (ICC2)
- PCIe error logging and handling on a typical SoC