Cervell™ NPU

C8
Higher‑Throughput NPU IP Core

Scaled for heavier inference and higher throughput  

Key Benefits

Boost AI throughput via a unified fabric that saturates bandwidth for maximum compute density

Same code, more performance

Same code, more performance

Move from edge to rack under one architecture

Transformer‑ready

Transformer‑ready

Gazillion™ helps absorb KV‑cache bandwidth so utilization stays high

Architecture Highlights

AI Accelerator

64-bit RISC-V AI Accelerator

64‑bit RISC‑V CPU + RVV 1.0 Vector + programmable Tensor unit

Sustained DRAM bandwidth

Sustained DRAM bandwidth

Gazillion™ memory streaming for sustained DRAM bandwidth

Simple Linux Programming

Simple Linux Programming

Linux‑ready bring‑up; unified programming model with C1

Software Path

Single RISC‑V ISA
Single RISC‑V ISA from
C1 C8 Cluster Hardware

Typical Deployments

Vision at scale

Efficient, low-latency AI inference, enabling massive deployment of vision processing

Compact gateways

Recommendation systems

Enables quick, personalized user recommendations by accelerating core ML algorithms

Compact gateways

LLM/transformer inference nodes

Dedicated acceleration of matrix multiplication for high-throughput, low-latency LLM/Transformer inference

Compact gateways

Find out more

Our IP is silicon-ready, and in silicon implementations. Speak to us about reference designs

Learn more