spikingjelly VS norse

Compare spikingjelly vs norse and see what are their differences.

spikingjelly

SpikingJelly is an open-source deep learning framework for Spiking Neural Network (SNN) based on PyTorch. (by fangwei123456)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
spikingjelly norse
1 6
1,153 615
- 2.3%
8.7 6.4
15 days ago 17 days ago
Python Python
GNU General Public License v3.0 or later GNU Lesser General Public License v3.0 only
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

spikingjelly

Posts with mentions or reviews of spikingjelly. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2022-04-04.
  • Has anyone used Spiking Neural Networks (SNNs) for image processing?
    2 projects | /r/computervision | 4 Apr 2022
    Surrogate gradient learning w/ backpropagation: for short, you can use backpropagation with SNNs (by a little trick during the backward pass). Super easy to implement, super efficient. You have a deep SNN trained via backprop with any type of input you want. Personally, that is completely my jam. Maybe you can use such paradigm to easily train an SNN in your biomed image dataset. Good repos: SnnTorch comes with the best tutorials to explain SNNs and surrogate gradient learning. This is the fastest way to understand the field and begin to implement you solution. Nevertheless, spikingjelly remains a better option when it comes to implement your ideas (better memory efficiency, etc). Good mention to lava-dl, with which you can train a neural network and directly transfer it into neuromorphic hardware (Intel Loihi) if you have access to this kind of chip.

norse

Posts with mentions or reviews of norse. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-04-24.
  • Neuromorphic learning, working memory, and metaplasticity in nanowire networks
    2 projects | news.ycombinator.com | 24 Apr 2023
    This gives you a ludicrous advantage over current neural net accelerators. Specifically 3-5 orders is magnitude in energy and time, as demonstrated in the BranScaleS system https://www.humanbrainproject.eu/en/science-development/focu...

    Unfortunately, that doesn't solve the problem of learning. Just because you can build efficient neuromorphic systems doesn't mean that we know how to train them. Briefly put, the problem is that a physical system has physical constraints. You can't just read the global state in NWN and use gradient descent as we would in deep learning. Rather, we have to somehow use local signals to approximate local behaviour that's helpful on a global scale. That's why they use Hebbian learning in the paper (what fires together, wires together), but it's tricky to get right and I haven't personally seen examples that scale to systems/problems of "interesting" sizes. This is basically the frontier of the field: we need local, but generalizable, learning rules that are stable across time and compose freely into higher-order systems.

    Regarding educational material, I'm afraid I haven't seen great entries for learning about SNNs in full generality. I co-author a simulator (https://github.com/norse/norse/) based on PyTorch with a few notebook tutorials (https://github.com/norse/notebooks) that may be helpful.

    I'm actually working on some open resources/course material for neuromorphic computing. So if you have any wishes/ideas, please do reach out. Like, what would a newcomer be looking for specifically?

  • [D] The Complete Guide to Spiking Neural Networks
    3 projects | /r/MachineLearning | 10 Apr 2023
    Surrogate gradients and BPTT, this is what is implemented in Norse https://github.com/Norse/Norse. It is also possible to compute exact gradients using the Eventprop algorithm.
  • [P] Norse - Deep learning with spiking neural networks (SNNs) in PyTorch
    1 project | /r/MachineLearning | 16 Jun 2022
    1 project | /r/MachineLearning | 16 Jun 2021
  • Show HN: Deep learning with spiking neural networks (SNNs) in PyTorch
    1 project | news.ycombinator.com | 16 Jun 2021
  • Don't Mess with Backprop: Doubts about Biologically Plausible Deep Learning
    4 projects | news.ycombinator.com | 15 Feb 2021
    That repo is slightly outdated, development now continues at https://github.com/norse/norse.

What are some alternatives?

When comparing spikingjelly and norse you can also consider the following projects:

snntorch - Deep and online learning with spiking neural networks in Python

lava-dl - Deep Learning library for Lava

Spiking-Neural-Network - Pure python implementation of SNN

bindsnet - Simulation of spiking neural networks (SNNs) using PyTorch.

Pytorch - Tensors and Dynamic neural networks in Python with strong GPU acceleration

CUDA-Guide - CUDA Guide

norse - Deep learning for spiking neural networks

ocaml-torch - OCaml bindings for PyTorch

Neuromorphic-Computing-Guide - Learn about the Neumorphic engineering process of creating large-scale integration (VLSI) systems containing electronic analog circuits to mimic neuro-biological architectures.

lava - A Software Framework for Neuromorphic Computing