SNAP-V: A RISC-V SoC with Configurable Neuromorphic Acceleration for Small-Scale Spiking Neural Networks

TL;DR

SNAP-V: A RISC-V SoC optimized for small-scale SNN inference, average synaptic energy 1.05 pJ.

cs.AR 🔴 Advanced 2026-03-12 9 views
Kanishka Gunawardana Sanka Peeris Kavishka Rambukwella Thamish Wanduragala Saadia Jameel Roshan Ragel Isuru Nawinne
Edge Computing Neuromorphic Computing Network-on-Chip Spiking Neural Networks Accelerator

Key Findings

Methodology

This paper presents SNAP-V, a RISC-V-based neuromorphic SoC specifically optimized for small-scale Spiking Neural Networks (SNNs). The system integrates two accelerator variants: Cerebra-S (bus-based) and Cerebra-H (Network-on-Chip-based), to achieve efficient SNN inference. A RISC-V core manages tasks, with both accelerators featuring parallel processing nodes and distributed memory. Experimental results show an average accuracy deviation of 2.62% between software and hardware inference, and an average synaptic energy of 1.05 pJ per synaptic operation in 45 nm CMOS technology.

Key Results

  • Result 1: An average accuracy deviation of 2.62% between software and hardware inference across multiple network configurations, demonstrating the precision of the hardware implementation.
  • Result 2: An average synaptic energy of 1.05 pJ per synaptic operation in 45 nm CMOS technology, indicating high energy efficiency.
  • Result 3: Cerebra-H achieves a maximum clock frequency of 96.24 MHz, a 9.46-fold improvement over Cerebra-S, significantly enhancing real-time responsiveness.

Significance

This research bridges the gap between large-scale neuromorphic hardware and small-scale embedded systems by designing the SNAP-V SoC. The system excels in energy efficiency and real-time performance, making it suitable for applications in robotics, edge intelligence, and low-power adaptive control. By enabling flexible interfacing and general-purpose computation within a unified architecture, SNAP-V provides a resource-efficient solution for practical deployment of small-scale SNNs, promoting the adoption of neuromorphic computing in real-world applications.

Technical Contribution

The technical contribution of SNAP-V lies in its innovative architectural design, combining the openness and modularity of RISC-V to enable general-purpose processing and event-driven computation on a single platform. The designs of Cerebra-S and Cerebra-H offer different performance baselines and optimization insights, particularly the hierarchical Network-on-Chip architecture of Cerebra-H, which effectively addresses communication bottlenecks and enhances system efficiency and performance.

Novelty

SNAP-V is the first SoC to combine RISC-V with small-scale SNN accelerators, providing flexible interfacing and efficient event-driven computation. Unlike existing large-scale neuromorphic platforms, SNAP-V is specifically designed for small-scale applications, avoiding resource wastage and significantly improving energy efficiency.

Limitations

  • Limitation 1: While SNAP-V performs excellently in small-scale SNN applications, its architecture may not be suitable for large-scale neural networks, especially in scenarios requiring higher computational power.
  • Limitation 2: Although Cerebra-S is simple in design, it suffers from communication and memory access bottlenecks, affecting overall system performance.
  • Limitation 3: The current implementation is primarily based on 45 nm CMOS technology, which may require re-optimization in more advanced processes.

Future Work

Future research directions include exploring the applicability of SNAP-V in broader application scenarios, particularly performance optimization in more advanced process technologies. Additionally, further integration of more sensor interfaces and control logic could enhance the system's versatility and adaptability.

AI Executive Summary

Spiking Neural Networks (SNNs) have garnered significant attention in edge computing due to their low power consumption and computational efficiency. However, existing implementations either use conventional System on Chip (SoC) architectures that suffer from memory-processor bottlenecks, or large-scale neuromorphic hardware that is inefficient and wasteful for small-scale SNN applications. This paper presents SNAP-V, a RISC-V-based neuromorphic SoC with two accelerator variants: Cerebra-S (bus-based) and Cerebra-H (Network-on-Chip-based), optimized for small-scale SNN inference, integrating a RISC-V core for management tasks, with both accelerators featuring parallel processing nodes and distributed memory.

Experimental results show close agreement between software and hardware inference, with an average accuracy deviation of 2.62% across multiple network configurations, and an average synaptic energy of 1.05 pJ per synaptic operation in 45 nm CMOS technology. These results demonstrate that the proposed solution enables accurate, energy-efficient SNN inference suitable for real-time edge applications.

The design goal of SNAP-V is to bridge the gap between large-scale neuromorphic hardware and practical, small-scale embedded systems. By leveraging the openness and modularity of RISC-V, the system facilitates efficient management of SNN workloads while supporting additional embedded tasks such as sensor processing, control logic, and signal encoding within a unified platform, eliminating the need for external microcontrollers or companion chips.

The designs of Cerebra-S and Cerebra-H offer different performance baselines and optimization insights, particularly the hierarchical Network-on-Chip architecture of Cerebra-H, which effectively addresses communication bottlenecks and enhances system efficiency and performance. By enabling flexible interfacing and general-purpose computation within a unified architecture, SNAP-V provides a resource-efficient solution for practical deployment of small-scale SNNs, promoting the adoption of neuromorphic computing in real-world applications.

However, while SNAP-V performs excellently in small-scale SNN applications, its architecture may not be suitable for large-scale neural networks, especially in scenarios requiring higher computational power. Additionally, the current implementation is primarily based on 45 nm CMOS technology, which may require re-optimization in more advanced processes. Future research directions include exploring the applicability of SNAP-V in broader application scenarios, particularly performance optimization in more advanced process technologies.

Deep Analysis

Background

Spiking Neural Networks (SNNs) represent the third generation of neural network models, extending conventional artificial networks by incorporating temporal dynamics and event-driven computation. Unlike traditional Artificial Neural Networks (ANNs), where neurons communicate through continuous activations, SNNs transmit discrete spike events, enabling sparse and asynchronous processing. This characteristic leads to substantially lower energy consumption and improved temporal precision, which are particularly advantageous for edge and embedded systems. At the computational level, a range of neuron models has been proposed to emulate biological spiking behavior. Highly detailed models such as Hodgkin–Huxley provide strong biological realism but are computationally intensive and impractical for hardware implementation. Simplified models, most notably the Leaky Integrate-and-Fire (LIF) neuron, abstract spiking dynamics using lightweight equations that significantly reduce computational and hardware overhead. Information in SNNs is encoded as sequences of spikes using neural coding schemes that directly influence computational efficiency. Common strategies include rate coding, where firing frequency represents signal strength, and temporal codes such as time-to-first-spike (TTFS) and phase coding, which exploit precise spike timing for low-latency computation. The choice of neuron model and coding scheme strongly impacts implementation complexity, power consumption, and latency in neuromorphic hardware. In this work, the LIF model is selected due to its deterministic timing behavior and hardware simplicity, making it well suited for compact neuromorphic architectures targeting real-time embedded operation.

Core Problem

While large-scale SNNs aim to replicate complex cortical networks, many embedded applications operate efficiently on compact, task-specific architectures. In this context, small-scale SNNs can be defined as networks with approximately 10 to 2000 neurons. Networks of this size are commonly used in robotics, signal processing, and sensor-driven applications, where energy efficiency, deterministic behavior, and low latency are critical. Despite the availability of large-scale neuromorphic platforms such as Intel’s Loihi (131,072 neurons) and SpiNNaker (up to 1 billion neurons in multi-chip configurations), many real-world applications require significantly smaller networks, leading to substantial hardware underutilization. In robotics, Gridbot employs only 1,321 neurons for spatial navigation, while NeuroPod uses merely 22 neurons on the massive SpiNNaker platform for insect-inspired locomotion. Control applications demonstrate even more extreme mismatches: autonomous racing systems operate with 39 neurons, lane-keeping controllers require 34 neurons, and event-based PID control for quadrotors utilizes approximately 35-40 neurons. Beyond control tasks, robot localization systems typically employ 700–800 neurons, while micro-scale engine optimization operates with as few as 8 neurons. These examples show that small-scale SNNs are not simplified versions of large networks but domain-optimized architectures tailored for specific spatiotemporal tasks with predictable latency, efficient memory access, and minimal communication overhead, characteristics that would be better served by dedicated small-scale neuromorphic platforms rather than running on massively over-provisioned hardware.

Innovation

The core innovations of SNAP-V lie in its architectural design, combining the openness and modularity of RISC-V to enable general-purpose processing and event-driven computation on a single platform. Specific innovations include:

1) The design and implementation of a RISC-V integrated neuromorphic SoC, optimized for inference on small-scale SNNs.

2) A neuromorphic accelerator with two progressive designs: Cerebra-S (Small) establishing a baseline for performance analysis and optimization insights, and Cerebra-H (High-performance), a clustered architecture with hierarchical NoC, refined based on Cerebra-S findings.

3) RTL-level functional validation demonstrating close agreement between software and hardware inference, with an average accuracy deviation below 3%.

4) A detailed power and energy characterization revealing memory-dominated system behavior and achieving low synaptic energy per operation.

Methodology

The design and implementation methodology of SNAP-V includes the following steps:

  • �� Leveraging RISC-V's open and modular ISA to achieve design flexibility and compatibility with open-source SoC frameworks.
  • �� Tight integration of the neuromorphic subsystem with RISC-V cores using the Rocket Custom Coprocessor (RoCC) interface.
  • �� Issuing custom SNN instructions from SpikeCore via the RoCC interface for efficient configuration, spike-data transfer, and synchronization with the accelerator's event-driven pipeline.
  • �� Implementing an Accelerator Controller for instruction decoding, dual NoC management, and time synchronization between the CPU and the accelerator.
  • �� Using a hardware-implemented rate coding scheme to convert sensor data into spike trains and route them to neuron clusters through the NoC.
  • �� Integrating spike encoding and decoding at the hardware level to achieve tighter coupling between sensing, computation, and actuation.

Experiments

The experimental design includes comparing software and hardware inference across multiple network configurations to evaluate system accuracy and energy efficiency. Experiments are conducted in 45 nm CMOS technology, measuring average synaptic energy per operation. RTL-level functional validation demonstrates close agreement between software and hardware inference. Experiments also compare the performance and energy efficiency of Cerebra-S and Cerebra-H, particularly the effectiveness of Cerebra-H's hierarchical Network-on-Chip architecture in addressing communication bottlenecks.

Results

Experimental results show an average accuracy deviation of 2.62% between software and hardware inference across multiple network configurations, and an average synaptic energy of 1.05 pJ per synaptic operation in 45 nm CMOS technology. Cerebra-H achieves a maximum clock frequency of 96.24 MHz, a 9.46-fold improvement over Cerebra-S, significantly enhancing real-time responsiveness. The hierarchical Network-on-Chip architecture of Cerebra-H effectively addresses communication bottlenecks, enhancing system efficiency and performance.

Applications

SNAP-V is suitable for applications in robotics, edge intelligence, and low-power adaptive control. Its flexible interfacing and general-purpose computation capabilities enable efficient management of SNN workloads on a single platform, while supporting additional embedded tasks such as sensor processing, control logic, and signal encoding. SNAP-V provides a resource-efficient solution for practical deployment of small-scale SNNs, promoting the adoption of neuromorphic computing in real-world applications.

Limitations & Outlook

While SNAP-V performs excellently in small-scale SNN applications, its architecture may not be suitable for large-scale neural networks, especially in scenarios requiring higher computational power. Additionally, although Cerebra-S is simple in design, it suffers from communication and memory access bottlenecks, affecting overall system performance. The current implementation is primarily based on 45 nm CMOS technology, which may require re-optimization in more advanced processes. Future research directions include exploring the applicability of SNAP-V in broader application scenarios, particularly performance optimization in more advanced process technologies.

Plain Language Accessible to non-experts

Imagine you have a smart kitchen with various appliances and tools. SNAP-V is like the intelligent brain of this kitchen, efficiently managing and coordinating all kitchen activities. The RISC-V core acts like the head chef, overseeing planning and management, while Cerebra-S and Cerebra-H are like specialized cooks, each handling different cooking tasks. Cerebra-S is like a chef skilled in simple dishes, quickly preparing basic meals, while Cerebra-H is like an experienced master chef, capable of handling complex dishes and multitasking. The entire system communicates through an intelligent network (like the kitchen's internal communication system) to coordinate cooperation among the chefs, ensuring every dish is served at the optimal time. This smart kitchen not only efficiently handles daily cooking tasks but can also adjust and optimize according to different cooking needs and conditions.

ELI14 Explained like you're 14

Hey there, buddy! Imagine you're playing a super cool game with lots of characters, each with their own tasks. SNAP-V is like the game's super commander, coordinating and managing all the characters' actions. The RISC-V core is like the game's main console, responsible for overall planning, while Cerebra-S and Cerebra-H are like different game characters, each handling different tasks. Cerebra-S is like a speedy runner character, quickly completing simple tasks, while Cerebra-H is like a powerful fighter character, capable of handling complex tasks and multitasking. The whole system communicates through an intelligent network (like the game's chat system) to coordinate cooperation among the characters, ensuring every task is completed smoothly. This super commander not only efficiently manages all tasks in the game but can also adjust and optimize according to different game scenarios and challenges.

Glossary

RISC-V

RISC-V is an open instruction set architecture with modularity and extensibility, allowing designers to customize the instruction set as needed.

In this paper, RISC-V serves as the foundational architecture for the SNAP-V SoC, supporting the integration of neuromorphic accelerators.

SoC (System on Chip)

An SoC is a chip that integrates multiple computing components, including processors, memory, and interfaces, to implement a complete computing system.

The SNAP-V designed in this paper is a RISC-V SoC optimized for small-scale SNNs.

SNN (Spiking Neural Network)

An SNN is a neural network model that communicates through discrete spike events, simulating the behavior of biological neurons, characterized by low power consumption and high temporal precision.

The design goal of SNAP-V is to optimize the inference performance of small-scale SNNs.

Cerebra-S

Cerebra-S is a neuromorphic accelerator in SNAP-V, employing a bus-based architecture to provide a basic performance baseline.

Cerebra-S is used to evaluate the basic performance of SNAP-V in small-scale SNN applications.

Cerebra-H

Cerebra-H is a neuromorphic accelerator in SNAP-V, employing a hierarchical Network-on-Chip architecture to optimize communication and energy efficiency.

Cerebra-H enhances the overall performance of SNAP-V through improved architectural design.

Network-on-Chip (NoC)

NoC is an architecture for on-chip communication, connecting different computing units through a network topology to achieve efficient data transmission.

Cerebra-H uses a Network-on-Chip architecture to optimize communication between neurons.

Synaptic Energy

Synaptic energy refers to the energy consumed per synaptic operation in a neural network, typically measured in picojoules (pJ).

In SNAP-V, the average synaptic energy per operation is 1.05 pJ.

Event-Driven Computation

Event-driven computation is a computing paradigm where the system processes based on the occurrence of events rather than continuous computation.

SNAP-V achieves efficient SNN inference through event-driven computation.

Leaky Integrate-and-Fire (LIF) Neuron

The LIF neuron is a simplified neuron model that simulates spiking dynamics using lightweight equations, reducing computational and hardware overhead.

SNAP-V selects the LIF model for the design of its neuromorphic accelerators.

Rate Coding

Rate coding is a neural encoding strategy where firing frequency represents signal strength, used to convey information in neural networks.

The coding hardware unit in SNAP-V employs a rate coding scheme to convert sensor data into spike trains.

Open Questions Unanswered questions from this research

  • 1 How can SNAP-V's performance and energy efficiency be optimized in more advanced process technologies? The current implementation is primarily based on 45 nm CMOS technology, and future iterations may need redesign and optimization at smaller process nodes to enhance system performance and efficiency.
  • 2 How can more sensor interfaces and control logic be integrated into SNAP-V to enhance the system's versatility and adaptability? The current design focuses primarily on the inference performance of small-scale SNNs, and future expansions may be needed to support more embedded tasks.
  • 3 How can more efficient memory management and data transmission be achieved in SNAP-V to further reduce system energy consumption? The existing design has optimized communication and memory access, but further improvements in memory management strategies could further reduce energy consumption.
  • 4 How can more flexible neuron models and coding schemes be implemented in SNAP-V to support a broader range of applications? The current design is primarily based on the LIF model and rate coding scheme, and future exploration of other neuron models and coding schemes may be needed to support more complex tasks.
  • 5 How can more efficient hardware-software co-design be achieved in SNAP-V to improve system development efficiency and performance? The current design has achieved tight integration between hardware and software, but further improvements in co-design strategies could enhance system performance and development efficiency.

Applications

Immediate Applications

Robotic Navigation

SNAP-V can be used in robotic navigation systems, enabling real-time path planning and environmental awareness through efficient SNN inference, suitable for resource-constrained embedded platforms.

Edge Intelligence Devices

SNAP-V is suitable for edge intelligence devices, such as smart cameras and sensors, enabling real-time data processing and analysis through low-power event-driven computation.

Low-Power Control Systems

SNAP-V can be used in low-power control systems, such as drones and autonomous vehicles, enabling real-time control and decision-making through efficient SNN inference.

Long-term Vision

Smart Home Systems

SNAP-V can be integrated into smart home systems, enabling automated control and management of smart devices through efficient SNN inference, improving quality of life.

Medical Monitoring Devices

SNAP-V can be used in medical monitoring devices, enabling real-time health data analysis and anomaly detection through low-power event-driven computation, enhancing healthcare services.

Abstract

Spiking Neural Networks (SNNs) have gained significant attention in edge computing due to their low power consumption and computational efficiency. However, existing implementations either use conventional System on Chip (SoC) architectures that suffer from memory-processor bottlenecks, or large-scale neuromorphic hardware that is inefficient and wasteful for small-scale SNN applications. This work presents SNAP-V, a RISC-V-based neuromorphic SoC with two accelerator variants: Cerebra-S (bus-based) and Cerebra-H (Network-on-Chip (NoC)-based) which are optimized for small-scale SNN inference, integrating a RISC-V core for management tasks, with both accelerators featuring parallel processing nodes and distributed memory. Experimental results show close agreement between software and hardware inference, with an average accuracy deviation of 2.62% across multiple network configurations, and an average synaptic energy of 1.05 pJ per synaptic operation (SOP) in 45 nm CMOS technology. These results show that the proposed solution enables accurate, energy-efficient SNN inference suitable for real-time edge applications.

cs.AR cs.NE

References (20)

A 0.086-mm$^2$ 12.7-pJ/SOP 64k-Synapse 256-Neuron Online-Learning Digital Spiking Neuromorphic Processor in 28-nm CMOS

C. Frenkel, M. Lefebvre, J. Legat et al.

2018 291 citations ⭐ Influential View Analysis →

A 4096-Neuron 1M-Synapse 3.8-pJ/SOP Spiking Neural Network With On-Chip STDP Learning and Sparse Weights in 10-nm FinFET CMOS

Gregory K. Chen, Raghavan Kumar, H. Sumbul et al.

2019 179 citations ⭐ Influential

Darwin3: A large-scale neuromorphic chip with a Novel ISA and On-Chip Learning

De Ma, Xiaofei Jin, Shichun Sun et al.

2023 65 citations ⭐ Influential View Analysis →

Spiking Neural Networks and Their Applications: A Review

Kashu Yamazaki, Viet-Khoa Vo-Ho, D. Bulsara et al.

2022 501 citations

Lapicque's introduction of the integrate-and-fire model neuron (1907).

L. F. Abbott, F. Laurence, Abbott

1999 721 citations

Efficient Video and Audio Processing with Loihi 2

S. Shrestha, Jonathan Timcheck, Paxon Frady et al.

2023 63 citations View Analysis →

End to End Learning of Spiking Neural Network Based on R-STDP for a Lane Keeping Vehicle

Zhenshan Bing, Claus Meschede, Kai Huang et al.

2018 79 citations

Loihi: A Neuromorphic Manycore Processor with On-Chip Learning

Mike Davies, N. Srinivasa, Tsung-Han Lin et al.

2018 3287 citations

Scalable NoC-based Neuromorphic Hardware Learning and Inference

Haowen Fang, Amar Shrestha, De Ma et al.

2018 26 citations View Analysis →

A Small, Low Cost Event-Driven Architecture for Spiking Neural Networks on FPGAs

J. P. Mitchell, Catherine D. Schuman, T. Potok

2020 29 citations

Scalable Hierarchical Network-on-Chip Architecture for Spiking Neural Network Hardware Implementations

Snaider Carrillo, J. Harkin, L. McDaid et al.

2013 113 citations

GAP-8: A RISC-V SoC for AI at the Edge of the IoT

E. Flamand, D. Rossi, Francesco Conti et al.

2018 210 citations

Caspian: A Neuromorphic Development Platform

J. P. Mitchell, Catherine D. Schuman, R. Patton et al.

2020 43 citations

A Scalable Multicore Architecture With Heterogeneous Memory Structures for Dynamic Neuromorphic Asynchronous Processors (DYNAPs)

S. Moradi, Ning Qiao, F. Stefanini et al.

2017 547 citations View Analysis →

Neural Coding in Spiking Neural Networks: A Comparative Study for Robust Neuromorphic Systems

Wenzhe Guo, M. Fouda, A. Eltawil et al.

2021 242 citations

Efficient Neuromorphic Signal Processing with Loihi 2

G. Orchard, E. P. Frady, D. B. Rubin et al.

2021 258 citations View Analysis →

SYNtzulu: A Tiny RISC-V-Controlled SNN Processor for Real-Time Sensor Data Analysis on Low-Power FPGAs

Gianluca Leone, Matteo Antonio Scrugli, Lorenzo Badas et al.

2025 12 citations

Gridbot: An autonomous robot controlled by a Spiking Neural Network mimicking the brain's navigational system

Guangzhi Tang, K. Michmizos

2018 47 citations View Analysis →

Real-Time Evolution and Deployment of Neuromorphic Computing at The Edge

Catherine D. Schuman, Steven R. Young, Bryan P. Maldonado et al.

2021 10 citations

Event-based PID controller fully realized in neuromorphic hardware: a one DoF study

R. Stagsted, A. Vitale, Alpha Renner et al.

2020 37 citations