NVM OTP NeoBit in Maxchip (180nm, 160nm, 150nm, 110nm, 90nm, 80nm)
LeapMind's New AI Chip Paves the Way for Unprecedented Cost-Effective AI Computing
Tokyo, Japan: October 16, 2023 - LeapMind Inc, has announced the commencement of the new AI chip development that accelerates the training and inference of large-scale AI models.
The cost of training advanced AI models, including large-scale language models (LLMs), has increased significantly over the past decade due to larger model sizes and greater computational complexity. This rising cost has become a major bottleneck in AI progress.
To create high-quality AI models, a substantial number of processors are needed for parallel computing, which requires a sizable budget. However, with the availability of cost-effective processors, it's now possible to develop improved AI models within the same expense. The shift in demand for AI training processors is moving from sheer performance to cost-effectiveness.
In response to this situation, we've initiated the development of a new AI training and inference processor, referred to as "AI chip." Leveraging our expertise from AI accelerator development for edge devices, this new AI chip targets a computing performance of 2 PFLOPS (petaflops) while aiming for a cost performance 10 times higher than that of an equivalent GPU. We anticipate that this product will be ready for shipment by the end of 2025 at the latest.
Matsuda, Soichi Chief Executive Officer of LeapMind, says “We have achieved a high level of success and a proven track record in the development of edge AI inference accelerators. On the server side, we will accelerate the evolution of next-generation AI by developing new AI chips that leverage our accumulated technological expertise to accelerate the computing process of AI models."
New AI Chip overview
The new AI chip has three major characteristics; 1) Designed for AI model training and inference, 2)Low-bit representation, and 3) Open-source drivers and compilers
Designed for AI model training and inference
Considering AI model training and inference as computational tasks, the following characteristics emerge:
- Matrix multiplication stands out as a computational bottleneck.
- These tasks can be easily executed in parallel.
- There are very few conditional branches involved.
Our design approach leverages the features above, creating AI chips specialized for AI model training and inference rather than pursuing performance improvements as a general-purpose computing machine. For instance, we minimize transistors by eliminating the branch prediction unit due to the scarcity of conditional branches in the program.
Low-Bit Representation
The primary bottleneck in AI model calculations is matrix multiplication, which involves an extensive number of multiplications and additions. Multipliers typically require large circuits, but we reduce the number of necessary transistors by adopting lower bit-width data types, such as fp8. Additionally, downsizing the data processed effectively utilizes DRAM bandwidth, which has been a bottleneck in recent years.
Open-source drivers and Compilers
Developing advanced software stacks is essential for AI model development, and no single company can provide all the required components. An open-source software ecosystem involving multiple companies already exists, and to be a part of this ecosystem, it's crucial to engage as an open-source software community member.
We will release comprehensive hardware specifications and software, including drivers and compilers, under OSI-compliant licenses to contribute to the open-source community.
Application
This new AI chip supports training and inference for a wide range of neural networks, including generative AI models such as diffusion models, as well as training for large-scale language models.
Specific performance values, benchmarks, and other details about the new AI chip will be updated on our tech blog.
[LeapMind Tech Blog](https://leapmind.io/tech-blog/)
About LeapMind Inc.
Founded in 2012, LeapMind is a Japan-based technology company that develops and provides artificial intelligence (AI) accelerators and its software, tasked with revolutionizing business and society by contributing to the advancement of AI technology. With its expertise and ability to provide AI solutions, LeapMind is committed to developing both the software and hardware necessary to meet customer needs and drive innovation toward a sustainable future. (as of October 2023).
|
Related News
- intoPIX JPEG XS Powers Village Island's VICO-2L-XS for Cost-effective SDI/IP Conversion
- Xiphera's Customisable nQrux™ Confidential Computing Engine Protects Cloud, Edge, and AI Environments
- LeapMind's Ultra Low-Power AI accelerator IP "Efficiera" Achieved industry-leading power efficiency of 107.8 TOPS/W
- LeapMind's "Efficiera" Ultra-low Power AI Inference Accelerator IP Was Verified RTL Design for ASIC/ASSP Conversion
- Arm's solution to the future needs of AI, security and specialized computing is v9
Breaking News
- Arm loses out in Qualcomm court case, wants a re-trial
- Jury is out in the Arm vs Qualcomm trial
- Ceva Seeks To Exploit Synergies in Portfolio with Nano NPU
- Synopsys Responds to U.K. Competition and Markets Authority's Phase 1 Announcement Regarding Ansys Acquisition
- Alphawave Semi Scales UCIe™ to 64 Gbps Enabling >20 Tbps/mm Bandwidth Density for Die-to-Die Chiplet Connectivity
Most Popular
E-mail This Article | Printer-Friendly Page |