LeapMind Announces Efficiera v2 Ultra-Low Power AI Inference Accelerator IP
Enhance product for wider-range application based on initial market introduction, results and evaluation
November 30, 2021 – Tokyo Japan, LeapMind Co., Ltd., a leading creator of the standard in edge artificial intelligence (AI), today announced version 2 (hereinafter: v2) of the ultra-low power AI inference accelerator IP Efficiera, scheduled to be available in December 2021. Efficiera is highly valued for its power saving, high performance, space saving and performance scalability features. Efficiera v2’s improved features and benefits expand the range of applications by providing wider coverage of performance with broader product availability while maintaining the circuit scale of the minimum configurations. These enhancements are based on learnings from the initial product introduction and market evaluation.
"Last year, we officially launched the commercial version of v1, and many companies evaluated Efficiera, and by the end of September 2021, we signed license agreements with eight domestic companies. Our corporate motto is concretely feeling that the "spreading to the world new devices that use machine learning " that we have set as our corporate philosophy is steadily progressing through the provision of v1. We will continue to strive to popularize AI through further technological innovation and product lineup expansion." said Soichi Matsuda, CEO of LeapMind.
Efficiera is an ultra-low power AI inference accelerator IP specialized for convolutional neural network (CNN) inference processing that runs as a circuit on FPGA or application-specific integrated circuit (ASIC) devices. The ultra-small quantization technology minimizes the number of quantization bits to 1 - 2 bits, maximizing the power and area efficiency of convolution that accounts for most of inference processing without the need for advanced semiconductor manufacturing processes or special cell libraries. By using this product, deep learning functions can be incorporated into a variety of edge devices, including consumer electronics such as home appliances, industrial equipment such as construction machinery, surveillance cameras, broadcasting equipment, as well as small machines and robots that are constrained by power, cost, and thermals — all of which have been technically difficult in the past.
"Since the official release of v1, we have aimed to develop the world's most power efficient DNN accelerator. We have strengthened the design and verification method and development process. We developed v2 so that it can respond to adoption in ASICs and application-specific standard products (ASSPs). We are also developing an inference learning model on the deep learning side to maximize the benefits of ultra-small quantization technology. The greatest strength of LeapMind is that we can provide a technique to master these two wheels,” said Dr. Hiroyuki Tokunaga, Director and CTO of LeapMind.
By improving design and verification methodology and reviewing the development process, product quality can be applied not only to FPGA but also to ASIC and ASSP. LeapMind starts to provide a model development environment, Network Development Kit (NDK) which enables users to develop deep learning models for Efficiera, which has not been done previously.
Other Notable Specifications and Features of Efficiera V2
Concept of Efficiera v2
Hardware features
・Performance scalability up to 48 times by multiplexing MAC array + multi-core.
・v2 allows you to triple the number of MAC arrays in the Convolution pipeline to v1 (x1, x4 selectable), and expands performance scalability by allowing you to select up to 4 cores.
・Hardware execution of Skip connection and Pixel Embedding in addition to convolution and quantization.
・Resources equivalent to Efficiera v1 were used with the same configuration and analyzed to understand execution time and assess if additional hardware functions are required.
Integration into SoC
・AMBA AXI interface
・AMBA AXI continued to be used as the interface with the external and the interface when viewed as a black box is the same as before however it is now easier to migrate from the existing design.
・Single clock domain
Target frequency in FPGA
・The operating frequency of the FPGA is the same as before and although it depends on the device, about 150 to 250 MHz can be expected.
・256 GOP/s @ 125MHz (1 core)
・Up to 12 TOP/s @ 250MHz (2 cores)
・Provided as encrypted RTL
Network Development Kit (NDK)
・Package of code and information which is required to create ultra-small quantization DL model for Efficiera.
・Immediate use is possible for developers working with deep learning model for GPU
・DL framework supports PyTorch and TensorFlow 2
・The learning environment is a GPU-equipped Linux server
・The inference environment is a device equipped with Efficiera
・Provide support from LeapMind
LeapMind is welcoming trials and feedback from all interested parties, including system on chip (SoC) vendors and end-user product designers. To obtain Efficiera v2, please contact us at business@leapmind.io. For more product information, please visit: https://leapmind.io/en/business/ip/.
About LeapMind
LeapMind Inc. was founded in 2012 with the corporate mission "To create innovative devices with machine learning and make them available everywhere." Total investment in LeapMind to date has reached 4.99 billion yen (as of May 2021). The company's strength is in extremely low bit quantization for compact deep learning solutions. It has a proven track record of achievement with over 150 companies, many of which are centered in manufacturing, including the automobile industry. It is also developing its Efficiera semiconductor IP, based on its experience in the development of both software and hardware
|
Related News
- LeapMind's "Efficiera" Ultra-low Power AI Inference Accelerator IP Was Verified RTL Design for ASIC/ASSP Conversion
- Official Commercial Launch of Efficiera Ultra-Low Power AI Inference Accelerator IP Core
- LeapMind Announces the Beta Release of their Ultra-low Power Consumption AI Inference Accelerator IP
- LeapMind Unveils "Efficiera", the New Ultra Low Power AI Inference Accelerator IP
- LeapMind's Ultra Low-Power AI accelerator IP "Efficiera" Achieved industry-leading power efficiency of 107.8 TOPS/W
Breaking News
- Jury is out in the Arm vs Qualcomm trial
- Ceva Seeks To Exploit Synergies in Portfolio with Nano NPU
- Synopsys Responds to U.K. Competition and Markets Authority's Phase 1 Announcement Regarding Ansys Acquisition
- Alphawave Semi Scales UCIe™ to 64 Gbps Enabling >20 Tbps/mm Bandwidth Density for Die-to-Die Chiplet Connectivity
- RaiderChip Hardware NPU adds Falcon-3 LLM to its supported AI models
Most Popular
E-mail This Article | Printer-Friendly Page |