openvino
TornadoVM
openvino | TornadoVM | |
---|---|---|
17 | 22 | |
5,962 | 1,108 | |
3.8% | 1.4% | |
10.0 | 9.9 | |
about 14 hours ago | 4 days ago | |
C++ | Java | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
openvino
- FLaNK Stack 05 Feb 2024
- QUIK is a method for quantizing LLM post-training weights to 4 bit precision
- Intel OpenVINO 2023.1.0 released
- Intel OpenVINO 2023.1.0 released, open-source toolkit for optimizing and deploying AI inference
- OpenVINO 2023.1.0 released
- [N] Intel OpenVINO 2023.1.0 released, open-source toolkit for optimizing and deploying AI inference
-
Powering Anomaly Detection for Industry 4.0
Anomalib is an open-source deep learning library developed by Intel that makes it easy to benchmark different anomaly detection algorithms on both public and custom datasets, all by simply modifying a config file. As the largest public collection of anomaly detection algorithms and datasets, it has a strong focus on image-based anomaly detection. It’s a comprehensive, end-to-end solution that includes cutting-edge algorithms, relevant evaluation methods, prediction visualizations, hyperparameter optimization, and inference deployment code with Intel’s OpenVINO Toolkit.
TornadoVM
-
Intel Gaudi 3 AI Accelerator
You don't need to use C++ to interface with CUDA or even write it.
A while ago NVIDIA and the GraalVM team demoed grCUDA which makes it easy to share memory with CUDA kernels and invoke them from any managed language that runs on GraalVM (which includes JIT compiled Python). Because it's integrated with the compiler the invocation overhead is low:
https://developer.nvidia.com/blog/grcuda-a-polyglot-language...
And TornadoVM lets you write kernels in JVM langs that are compiled through to CUDA:
https://www.tornadovm.org
There are similar technologies for other languages/runtimes too. So I don't think that will cause NVIDIA to lose ground.
- Java VectorAPI compatiblity with TornadoVM GPU programming framework
- Java GPU pre/post processing with ONNX RT and TornadoVM
- FLaNK Stack 05 Feb 2024
- FLaNK 25 December 2023
- GPU Acceleration for Python, JavaScript, Ruby from Java with Truffle
- TornadoVM v1.0 Released
- TornadoVM 1.0
-
From CPU to GPU and FPGAs: Supercharging Java Applications with TornadoVM [video]
Presented by Juan Fumero, PhD & Research Fellow (The University of Manchester, UK) during the JVM Language Summit 2023 (Santa Clara CA).
More information on TornadoVM can be found at https://www.tornadovm.org/
Tags: #Java #JVMLS #GPU #FPGA #OpenJDK #GraalVM #AI
What are some alternatives?
TensorRT - NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
Aparapi - The New Official Aparapi: a framework for executing native Java and Scala code on the GPU.
deepsparse - Sparsity-aware deep learning inference runtime for CPUs
openapi4j - OpenAPI 3 parser, JSON schema and request validator.
mediapipe - Cross-platform, customizable ML solutions for live and streaming media.
GraalVMREPL - REPL (read–eval–print loop) shell built on top of JavaFX and GraalVM stack, incorporating GraalJS, GraalPython, TruffleRuby and FastR
stable-diffusion - Go to lstein/stable-diffusion for all the best stuff and a stable release. This repository is my testing ground and it's very likely that I've done something that will break it.
kattlo-cli - Kattlo CLI Project
neural-compressor - SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
junodb - JunoDB is PayPal's home-grown secure, consistent and highly available key-value store providing low, single digit millisecond, latency at any scale.
nebuly - The user analytics platform for LLMs
jr - JR: streaming quality random data from the command line