Ultra-low-power AI/ML processor and accelerator
NPU Clusters:
• Optimized for spatial neural networks (e.g. CNNs, ResNets, MobileNets)
• Sparsity exploitation
• Peak MAC performance: 160 GOPS
FETA Cluster:
• Optimized for temporal neural networks (e.g. RNNs like LSTM or GRU)
• Smart temporal feature extraction engine
View Ultra-low-power AI/ML processor and accelerator full description to...
- see the entire Ultra-low-power AI/ML processor and accelerator datasheet
- get in contact with Ultra-low-power AI/ML processor and accelerator Supplier
Block Diagram of the Ultra-low-power AI/ML processor and accelerator
Video Demo of the Ultra-low-power AI/ML processor and accelerator
Emotion detection running from a coin cell battery
AI IP
- RT-630 Hardware Root of Trust Security Processor for Cloud/AI/ML SoC FIPS-140
- RT-630-FPGA Hardware Root of Trust Security Processor for Cloud/AI/ML SoC FIPS-140
- NPU IP for Embedded AI
- RISC-V-based AI IP development for enhanced training and inference
- Tessent AI IC debug and optimization
- NPU IP family for generative and classic AI with highest power efficiency, scalable and future proof