accelerate
Embedded language for high-performance array computations (by AccelerateHS)
grenade
Deep Learning in Haskell (by HuwCampbell)
Our great sponsors
accelerate | grenade | |
---|---|---|
9 | 5 | |
886 | 1,438 | |
0.5% | - | |
5.3 | 5.6 | |
14 days ago | 5 months ago | |
Haskell | Haskell | |
BSD 3-clause "New" or "Revised" License | BSD 2-clause "Simplified" License |
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
accelerate
Posts with mentions or reviews of accelerate.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2023-02-23.
-
Should I use newer ghc?
Someone has opened a PR for accelerate here https://github.com/AccelerateHS/accelerate/pull/525 (sadly seems not actively maintained at the moment, but that can always change if people care enough). I agree for an executable you should freeze your dependencies and compiler version, and using 8.10 is fine. Although there are tons of improvements in 9.2+
-
Haskell deep learning tutorials [Blog]
Backprop is a neat library. However, I guess its use case is if you actually don't want to go for anything standard like Torch or TF (perhaps for research?) For instance, if I were to use something like Accelerate for GPU acceleration, or some other computation-oriented library, then I would mix it with Backprop. Previously, I have benefited from Backprop in a ConvNet tutorial and I liked it.
-
I made a petition to get the accelerate project for Haskell some funding.
Wait, really? Here's a conversation I had with him: https://github.com/AccelerateHS/accelerate/discussions/528
-
Who is researching array languages these days?
I know Accelerate is being developed at Utrecht University in the Netherlands. You can look at publications by Trevor McDonell to get a taste of what they are doing.
-
Next Decade in Languages: User Code on the GPU
I’m personally a big fan of http://www.acceleratehs.org / https://github.com/AccelerateHS/accelerate-llvm
-
Introduction to Doctests in Haskell
Looking for a few projects that make use of it, I found accelerate, hawk, polysemy and pretty-simple, so I'll be interested to poke around in their code and see how they have things set up.
-
Monthly Hask Anything (March 2022)
There's accelerate for GPU computing and hmatrix for bindings to BLAS and LAPACK.
-
Idris2+WebGL, part #12: Linear algebra with linear types... not great
I'm toying with the idea of replacing vector values with vector generators, where e.g. v1 + v2 is not evaluated to a new vector, but to a vector program. This is similar to the approaches of Accelerate and TensorFlow. On the flip side, I don't think I could get rid of the overhead, and I expect much smaller computation loads than aforementioned libraries, so overheads could be very significant. The added benefit of using vector generators is that the generator could not only be evaluated, but also be turned into a Latex formula.
grenade
Posts with mentions or reviews of grenade.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2023-01-23.
-
Haskell deep learning tutorials [Blog]
Grenade is fun, but it does not support CUDA, so it will limit you. I would say that this was a great experiment that has influenced the Hasktorch library in different ways (let me know if I am wrong).
- Dhall: A Gateway Drug to Haskell
-
Haskell for Artificial Intelligence?
FWIW there's an interesting library called grenade which offers nice types for constructing neural nets. I haven't used it, and this is not my areas of expertise, but it looks cool!
-
Rank 3 Stencils for "Efficient Parallel Stencil Convolution in Haskell" (Repa)
When I wrote grenade I used the im2col trick to turn convolutions into a single matrix multiplication, which could then be done in hmatrix.
- What are some ways I could tickle my (beginner) haskell-brain with something *useful*?
What are some alternatives?
When comparing accelerate and grenade you can also consider the following projects:
dhall - Maintainable configuration files
hasktorch - Tensors and neural networks in Haskell
accelerate-bignum - Fixed-length large integer arithmetic for Accelerate
liblinear-enumerator - Haskell bindings to liblinear
accelerate-cuda - DEPRECATED: Accelerate backend for NVIDIA GPUs
simple-neural-networks - Simple parallel neural networks implementation in pure Haskell
hyper-haskell-server - The strongly hyped Haskell interpreter.
CV - Haskell wrappers and utilities for OpenCV machine vision library
accelerate-fft - FFT library for Haskell based on the embedded array language Accelerate
nn - A tiny neural network 🧠
feldspar-compiler - This is the compiler for the Feldspar Language.
hnn - haskell neural network library
accelerate vs dhall
grenade vs hasktorch
accelerate vs accelerate-bignum
grenade vs liblinear-enumerator
accelerate vs accelerate-cuda
grenade vs simple-neural-networks
accelerate vs hyper-haskell-server
grenade vs CV
accelerate vs accelerate-fft
grenade vs nn
accelerate vs feldspar-compiler
grenade vs hnn