IRTools.jl
Dagger.jl
IRTools.jl | Dagger.jl | |
---|---|---|
2 | 4 | |
107 | 578 | |
0.9% | 1.2% | |
5.4 | 8.9 | |
6 days ago | 8 days ago | |
Julia | Julia | |
MIT License | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
IRTools.jl
- From Julia to Rust
-
Ask HN: Show me your Half Baked project
Which is the concept behind Cassette.jl (https://github.com/jrevels/Cassette.jl) and IRTools.jl (https://github.com/MikeInnes/IRTools.jl).
Dagger.jl
- Dagger: a new way to build CI/CD pipelines
-
DTable a new distributed table implementation in Julia using Dagger.jl
Firstly, I'll say that we already have work started to implement out-of-core directly in Dagger: https://github.com/JuliaParallel/Dagger.jl/pull/289.
With that PR in place, it should be possible to define a "storage device" which is backed by a database. I haven't had a chance to actually try this, since the PR still needs quite some work and testing, but it's definitely something on my radar!
- From Julia to Rust
-
Cerebras’ New Monster AI Chip Adds 1.4T Transistors
I'm not sure that's necessarily the domain of a low-level package like CUDA.jl though (which I assume you're referring to). That kind of interface is more the domain of higher-level packages like https://github.com/JuliaParallel/Dagger.jl/ and to a lesser extent https://juliagpu.github.io/KernelAbstractions.jl/stable/. Moreover, the jury is still out on whether the built-in Distributed module is an ideal abstraction for every use-case (clusters, heterogeneous compute, etc.)
WRT Nx, my biggest question is how they'll crack the problem of still needing big balls of C++ and the shims everywhere to get acceleration. Creating a compiler that generates efficient GPU or other accelerator code is a massive research project with no clear winners, never mind the challenge of reconciling the very mutation-heavy needs of GPU compute with a mostly immutable language model.
What are some alternatives?
JET.jl - An experimental code analyzer for Julia. No need for additional type annotations.
earthly - Super simple build framework with fast, repeatable builds and an instantly familiar syntax – like Dockerfile and Makefile had a baby.
Juleps - Julia Enhancement Proposals
julia - The Julia Programming Language
pyodide - Pyodide is a Python distribution for the browser and Node.js based on WebAssembly
DuckDB.jl
StaticArrays.jl - Statically sized arrays for Julia
determined - Determined is an open-source machine learning platform that simplifies distributed training, hyperparameter tuning, experiment tracking, and resource management. Works with PyTorch and TensorFlow.
glow - Compiler for Neural Network hardware accelerators
Metatheory.jl - General purpose algebraic metaprogramming and symbolic computation library for the Julia programming language: E-Graphs & equality saturation, term rewriting and more.
Symbolics.jl - Symbolic programming for the next generation of numerical software