jax
Pytorch
jax | Pytorch | |
---|---|---|
89 | 387 | |
31,945 | 89,253 | |
1.6% | 2.3% | |
10.0 | 10.0 | |
6 days ago | about 15 hours ago | |
Python | Python | |
Apache License 2.0 | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
jax
- I want a good parallel computer
-
Show HN: Localscope–Limit scope of Python functions for reproducible execution
localscope is a small Python package that disassembles functions to check if they access global variables they shouldn't. I wrote this a few years ago to detect scope bugs which are common in Jupyter notebooks. It's recently come in handy writing jax code (https://github.com/jax-ml/jax) because it requires pure functions. Thought I'd share.
- Zest
-
KlongPy: High-Performance Array Programming in Python
If you like high-performance array programming a la "numpy with JIT" I suggest looking at JAX. It's very suitable for general numeric computing (not just ML) and a very mature ecosystem.
https://github.com/jax-ml/jax
-
PyTorch is dead. Long live Jax
Nope, changing graph shape requires recompilation: https://github.com/google/jax/discussions/17191
- cuDF – GPU DataFrame Library
-
Rebuilding TensorFlow 2.8.4 on Ubuntu 22.04 to patch vulnerabilities
I found a GitHub issue that seemed similar (missing ptxas) and saw a suggestion to install nvidia-cuda-toolkit. Alright: but that exploded the container size from 6.5 GB to 12.13 GB … unacceptable 😤 (Incidentally, this is too large for Cloud Shell to build on its limited persistent disk.)
-
The Elements of Differentiable Programming
The dual numbers exist just as surely as the real numbers and have been used well over 100 years
https://en.m.wikipedia.org/wiki/Dual_number
Pytorch has had them for many years.
https://pytorch.org/docs/stable/generated/torch.autograd.for...
JAX implements them and uses them exactly as stated in this thread.
https://github.com/google/jax/discussions/10157#discussionco...
As you so eloquently stated, "you shouldn't be proclaiming things you don't actually know on a public forum," and doubly so when your claimed "corrections" are so demonstrably and totally incorrect.
-
Julia GPU-based ODE solver 20x-100x faster than those in Jax and PyTorch
On your last point, as long as you jit the topmost level, it doesn't matter whether or not you have inner jitted functions. The end result should be the same.
Source: https://github.com/google/jax/discussions/5199#discussioncom...
-
Apple releases MLX for Apple Silicon
The design of MLX is inspired by frameworks like NumPy, PyTorch, Jax, and ArrayFire.
Pytorch
-
Fine-tuning LLMs locally: A step-by-step guide
Next, we define a training loop that uses our prepared data and optimizes the weights of the model. Here's an example using PyTorch:
-
Ask HN: Why hasn't AMD made a viable CUDA alternative?
> But that does not seem to be the strategy, which implies it is not so simple?
That is exactly what has been happening [1], and not just in pytorch. Geohot has been very dedicated in working with AMD to upgrade their station in this space. If you hang out in the tinygrad discord, you can see this happening in real time.
> those I have talked to say they depend on a lot more than just one or two key libraries.
Theres a ton of libraries out there yes, but if we're talking about python and the libraries in question are talking to GPUs its going to be exceedingly rare that theyre not using one of these under the hood: pytorch, tensorflow, jax, keras, et al.
There are of course exceptions to this, particular if you're not using python for your ML work (which is actually common for many companies running inference at scale and want better runtime performance, training is a different story). But ultimately the core ecosystem does work just fine with AMD GPUs, provided you're not doing any exotic custom kernel work.
[1] https://github.com/pytorch/pytorch/pulls?q=is%3Aopen+is%3Apr...
-
10 Useful Tools and Libraries for Python Developers
4. PyTorch - A Deep Learning Powerhouse
-
torch.export()
GraphModule compiles every instruction into low-level ATen operations.
-
10 Must-Have AI Tools to Supercharge Your Software Development
8. TensorFlow and PyTorch: These frameworks support AI and machine learning integrations, allowing developers to build and deploy intelligent models and workflows. TensorFlow is widely used for deep learning applications, offering pre-trained models and extensive documentation. PyTorch provides flexibility and ease of use, making it ideal for research and experimentation. Both frameworks support neural network design, training, and deployment in production environments. Download TensorFlow here and Download PyTorch here.
-
Automating Enhanced Due Diligence in Regulated Applications
Frameworks like TensorFlow and PyTorch can help you build and train models for various tasks, such as risk scoring, anomaly detection, and pattern recognition.
-
Must-Know 2025 Developer’s Roadmap and Key Programming Trends
Python’s Growth in Data Work and AI: Python continues to lead because of its easy-to-read style and the huge number of libraries available for tasks from data work to artificial intelligence. Tools like TensorFlow and PyTorch make it a must-have. Whether you’re experienced or just starting, Python’s clear style makes it a good choice for diving into machine learning. Actionable Tip: If you’re new to Python, try projects that combine data with everyday problems. For example, build a simple recommendation system using Pandas and scikit-learn.
-
Decorator JITs: Python as a DSL
Basically this style of code - https://github.com/pytorch-labs/attention-gym/pull/84/files - has issues like this - https://github.com/pytorch/pytorch/pull/137452 https://github.com/pytorch/pytorch/issues/144511 https://github.com/pytorch/pytorch/issues/145869
For some higher level context, see https://pytorch.org/blog/flexattention/
-
Building an AI-powered Financial Data Analyzer with NodeJS, Python, SvelteKit, and TailwindCSS - Part 0
The AI Service will be built using aiohttp (asynchronous Python web server) and integrates PyTorch, Hugging Face Transformers, numpy, pandas, and scikit-learn for financial data analysis.
- PyTorch 2.6.0 Release
What are some alternatives?
Numba - NumPy aware dynamic Python compiler using LLVM
Pandas - Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more
dex-lang - Research language for array processing in the Haskell/ML family
tensorflow - An Open Source Machine Learning Framework for Everyone
julia - The Julia Programming Language
tinygrad - You like pytorch? You like micrograd? You love tinygrad! ❤️