RaiderChip unveils its fully Hardware-Based Generative AI Accelerator: The GenAI NPU
The new embedded accelerator boosts inference speed by 2.4x, combining complete privacy and autonomy with a groundbreaking innovation: it eliminates the need for CPUs.
Spain, January 27, 2024 -- RaiderChip has officially launched the GenAI NPU, a fully hardware-based accelerator that sets new standards for efficiency and scalability in Generative AI. The GenAI NPU retains the key features of its predecessor, the GenAI v1: offline operation and autonomous functionality.
Additionally, it becomes fully stand-alone by embedding all Large Language Models (LLMs) operations directly into its hardware, thereby eliminating the need for CPUs.
RaiderChip GenAI NPU running the Llama 3.2 1B LLM model and streaming its output to a terminal
Thanks to its fully hardware-based design, the GenAI NPU achieves unprecedented levels of efficiency, unattainable by hybrid designs. According to RaiderChip CTO Victor Lopez: “By eliminating latency caused by hardware-software communication, we achieve superior performance while removing external dependencies, such as CPUs. The performance that you see is what you will get, regardless of the target electronic system where the accelerator is integrated. This improves energy efficiency and ensures fully predictable performance—advantages which make the GenAI NPU the ideal solution for embedded systems.”
Furthermore, the new design optimizes token generation speed per available memory bandwidth, multiplying it by 2.4x, while enabling the use of more cost-efficient memories like DDR or LPDDR without relying on expensive options such as HBM to achieve excellent performance. It also delivers equivalent results with fewer components, reducing size, cost, and energy consumption. These features allow for the development of more affordable and sustainable generative AI solutions, with faster return on investment and seamless integration into a variety of products tailored to different needs.
With this innovation, RaiderChip strengthens its strategy of offering optimized solutions based on affordable hardware, designed to bring generative AI to the Edge. These solutions ensure complete privacy and security for applications thanks to their ability to operate entirely offline and on-premises, while eliminating dependence on the cloud and recurring monthly subscriptions.
WANT TO KNOW MORE?
|
Related News
- RaiderChip raises 1 Million Euros in seed capital to market its innovative generative AI accelerator: the GenAI v1.
- RaiderChip launches its Generative AI hardware accelerator for LLM models on low-cost FPGAs
- RaiderChip Hardware NPU adds Falcon-3 LLM to its supported AI models
- GenAI v1-Q launched with 4 bits Quantization support to accelerate larger LLMs at the Edge
- Vybium, develops European AI/ML accelerators based on the Stream Computing NPU IP
Breaking News
- Arteris Wins Two Gold and One Silver Stevie® Awards in the 2025 American Business Awards®
- Faraday Adds QuickLogic eFPGA to FlashKit‑22RRAM SoC for IoT Edge
- Xylon Introduces Xylon ISP Studio
- Crypto Quantique announces QRoot Lite - a lightweight and configurable root-of-trust IP for resource-constrained IoT devices
- BOS Semiconductors to Partner with Intel to Accelerate Automotive AI Innovation
Most Popular
- Andes Technology and Imagination Technologies Showcase Android 15 on High-Performance RISC-V Based Platform
- TSMC Unveils Next-Generation A14 Process at North America Technology Symposium
- Synopsys and TSMC Usher In Angstrom-Scale Designs with Certified EDA Flows on Advanced TSMC A16 and N2P Processes
- Certus Semiconductor Joins TSMC IP Alliance Program to Enhance Custom I/O and ESD Solutions
- M31 Collaborates with TSMC to Advance 2nm eUSB2 IP Innovation
![]() |
E-mail This Article | ![]() |
![]() |
Printer-Friendly Page |