ArrayFire
flashlight
Our great sponsors
ArrayFire | flashlight | |
---|---|---|
6 | 16 | |
4,404 | 5,152 | |
1.2% | 1.2% | |
7.8 | 7.4 | |
24 days ago | 25 days ago | |
C++ | C++ | |
BSD 3-clause "New" or "Revised" License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
ArrayFire
-
Learn WebGPU
Loads of people have stated why easy GPU interfaces are difficult to create, but we solve many difficult things all the time.
Ultimately I think CPUs are just satisfactory for the vast vast majority of workloads. Servers rarely come with any GPUs to speak of. The ecosystem around GPUs is unattractive. CPUs have SIMD instructions that can help. There are so many reasons not to use GPUs. By the time anyone seriously considers using GPUs they're, in my imagination, typically seriously starved for performance, and looking to control as much of the execution details as possible. GPU programmers don't want an automagic solution.
So I think the demand for easy GPU interfaces is just very weak, and therefore no effort has taken off. The amount of work needed to make it as easy to use as CPUs is massive, and the only reason anyone would even attempt to take this on is to lock you in to expensive hardware (see CUDA).
For a practical suggestion, have you taken a look at https://arrayfire.com/ ? It can run on both CUDA and OpenCL, and it has C++, Rust and Python bindings.
-
seeking C++ library for neural net inference, with cross platform GPU support
What about Arrayfire. https://github.com/arrayfire/arrayfire
-
[D] Deep Learning Framework for C++.
Low-overhead — not our goal, but Flashlight is on par with or outperforming most other ML/DL frameworks with its ArrayFire reference tensor implementation, especially on nonstandard setups where framework overhead matters
-
[D] Neural Networks using a generic GPU framework
Looking for frameworks with Julia + OpenCL I found array fire. It seems quite good, bonus points for rust bindings. I will keep looking for more, Julia completely fell off my radar.
- Windows 11 va bloquer les bidouilles qui facilitent l'emploi d'un navigateur alternatif à Edge
-
Arrayfire progressive performance decline?
Your Problem may be the lazy evaluation, see this issue: https://github.com/arrayfire/arrayfire/issues/1709
flashlight
-
MatX: Efficient C++17 GPU numerical computing library with Python-like syntax
I think a comparison to PyTorch, TensorFlow and/or JAX is more relevant than a comparison to CuPy/NumPy.
And then maybe also a comparison to Flashlight (https://github.com/flashlight/flashlight) or other C/C++ based ML/computing libraries?
Also, there is no mention of it, so I suppose this does not support automatic differentiation?
-
Project Resources
This Facebook ai project seems reasonably structured after looking at its CMakeLists.txt. CMake is a build generator for c++, it's how you make binaries to run your project: https://github.com/flashlight/flashlight
-
Meta AI Open Sources Flashlight: Fast and Flexible Machine Learning Toolkit in C++
Continue reading | Check out the paper and github link
- Flashlight: A C++ standalone library for machine learning
-
[D] Deep Learning Framework for C++.
I built and maintain Flashlight, a C++-first library for ML/DL. We built Flashlight to be:
- [R] C++ for Machine Learning
-
What is the most used library for AI in C++ ?
I’ve never used it, but Facebook’s flashlight looks interesting
-
Python.
Flashlight bro, not flash. Read again
-
Mozilla Common Voice Adds 16 New Languages and 4,600 New Hours of Speech
I've had good results with https://github.com/flashlight/flashlight/blob/master/flashli.... Seems to work well with spoken english in a variety of accents. Biggest limitation is that the architecture they have pretrained models for doesn't really work well with clips longer than ~15 seconds, so you have to segment your input files.
- [D] C++ in Machine Learning.
What are some alternatives?
Thrust - [ARCHIVED] The C++ parallel algorithms library. See https://github.com/NVIDIA/cccl
TTS - 🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production
Boost.Compute - A C++ GPU Computing Library for OpenCL
DeepSpeech - DeepSpeech is an open source embedded (offline, on-device) speech-to-text engine which can run in real time on devices ranging from a Raspberry Pi 4 to high power GPU servers.
VexCL - VexCL is a C++ vector expression template library for OpenCL/CUDA/OpenMP
PaddleSpeech - Easy-to-use Speech Toolkit including Self-Supervised Learning model, SOTA/Streaming ASR with punctuation, Streaming TTS with text frontend, Speaker Verification System, End-to-End Speech Translation and Keyword Spotting. Won NAACL2022 Best Demo Award.
Pytorch - Tensors and Dynamic neural networks in Python with strong GPU acceleration
STT - 🐸STT - The deep learning toolkit for Speech-to-Text. Training and deploying STT models has never been so easy.
CUB - THIS REPOSITORY HAS MOVED TO github.com/nvidia/cub, WHICH IS AUTOMATICALLY MIRRORED HERE.
NeMo - A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
Taskflow - A General-purpose Parallel and Heterogeneous Task Programming System
DNS-Challenge - This repo contains the scripts, models, and required files for the Deep Noise Suppression (DNS) Challenge.