notebooks VS NYU-DLSP20

Compare notebooks vs NYU-DLSP20 and see what are their differences.

notebooks

Notebooks illustrating the use of Norse, a library for deep-learning with spiking neural networks. (by norse)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
notebooks NYU-DLSP20
2 2
24 6,625
- -
0.0 6.1
over 1 year ago 4 months ago
Jupyter Notebook Jupyter Notebook
- GNU General Public License v3.0 or later
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

notebooks

Posts with mentions or reviews of notebooks. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-04-24.
  • Neuromorphic learning, working memory, and metaplasticity in nanowire networks
    2 projects | news.ycombinator.com | 24 Apr 2023
    This gives you a ludicrous advantage over current neural net accelerators. Specifically 3-5 orders is magnitude in energy and time, as demonstrated in the BranScaleS system https://www.humanbrainproject.eu/en/science-development/focu...

    Unfortunately, that doesn't solve the problem of learning. Just because you can build efficient neuromorphic systems doesn't mean that we know how to train them. Briefly put, the problem is that a physical system has physical constraints. You can't just read the global state in NWN and use gradient descent as we would in deep learning. Rather, we have to somehow use local signals to approximate local behaviour that's helpful on a global scale. That's why they use Hebbian learning in the paper (what fires together, wires together), but it's tricky to get right and I haven't personally seen examples that scale to systems/problems of "interesting" sizes. This is basically the frontier of the field: we need local, but generalizable, learning rules that are stable across time and compose freely into higher-order systems.

    Regarding educational material, I'm afraid I haven't seen great entries for learning about SNNs in full generality. I co-author a simulator (https://github.com/norse/norse/) based on PyTorch with a few notebook tutorials (https://github.com/norse/notebooks) that may be helpful.

    I'm actually working on some open resources/course material for neuromorphic computing. So if you have any wishes/ideas, please do reach out. Like, what would a newcomer be looking for specifically?

  • Event-Based Backpropagation for Exact Gradients in Spiking Neural Networks
    1 project | news.ycombinator.com | 2 Jun 2021
    We've written some documentation around our neuron equations in Python that explains this: https://norse.github.io/norse/auto_api/norse.torch.functiona...

    See also our tutorial on neuron parameter optimization to understand how it's useful for machine learning: https://github.com/norse/notebooks#level-intermediate

    Disclaimer: I'm a co-author of the library Norse

    Regarding the target audience, it's actually not entirely clear to me. This lies in the intersection between computational neuroscience and deep learning. Which isn't a huge set of people. Meaning, you're questions are valid and we (as researchers) have a lot of communication to do to explain why this is interesting and important.

NYU-DLSP20

Posts with mentions or reviews of NYU-DLSP20. We have used some of these posts to build our list of alternatives and similar projects.

What are some alternatives?

When comparing notebooks and NYU-DLSP20 you can also consider the following projects:

DeepLearningExamples - State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure.

webdataset - A high-performance Python-based I/O system for large (and small) deep learning problems, with strong support for PyTorch.

fastai - The fastai deep learning library

Real-time-Object-Detection-for-Autonomous-Driving-using-Deep-Learning - My Computer Vision project from my Computer Vision Course (Fall 2020) at Goethe University Frankfurt, Germany. Performance comparison between state-of-the-art Object Detection algorithms YOLO and Faster R-CNN based on the Berkeley DeepDrive (BDD100K) Dataset.

Made-With-ML - Learn how to design, develop, deploy and iterate on production-grade ML applications.

nlp-class - A Natural Language Processing course taught by Professor Ghassemi

bitcoin_price_prediction - This project tries to prediction the bitcoin price with machine and deep learning.

ydata-profiling - 1 Line of code data quality profiling & exploratory data analysis for Pandas and Spark DataFrames.

dl-colab-notebooks - Try out deep learning models online on Google Colab

ML-Workspace - 🛠 All-in-one web-based IDE specialized for machine learning and data science.

Siren-fastai2 - Unofficial implementation of 'Implicit Neural Representations with Periodic Activation Functions'

data-science-learning - Repository of code and resources related to different data science and machine learning topics. For learning, practice and teaching purposes.