xla VS determined

Compare xla vs determined and see what are their differences.

xla

Enabling PyTorch on XLA Devices (e.g. Google TPU) (by pytorch)

determined

Determined is an open-source machine learning platform that simplifies distributed training, hyperparameter tuning, experiment tracking, and resource management. Works with PyTorch and TensorFlow. (by determined-ai)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
xla determined
8 10
2,296 2,868
1.7% 2.5%
9.9 9.9
5 days ago 3 days ago
C++ Go
GNU General Public License v3.0 or later Apache License 2.0
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

xla

Posts with mentions or reviews of xla. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-01-26.
  • Who uses Google TPUs for inference in production?
    1 project | news.ycombinator.com | 11 Mar 2024
    > The PyTorch/XLA Team at Google

    Meanwhile you have an issue from 5 years ago with 0 support

    https://github.com/pytorch/xla/issues/202

  • Google TPU v5p beats Nvidia H100
    2 projects | news.ycombinator.com | 26 Jan 2024
    PyTorch has had an XLA backend for years. I don't know how performant it is though. https://pytorch.org/xla
  • Why Did Google Brain Exist?
    2 projects | news.ycombinator.com | 26 Apr 2023
    It's curtains for XLA, to be precise. And PyTorch officially supports XLA backend nowadays too ([1]), which kind of makes JAX and PyTorch standing on the same foundation.

    1. https://github.com/pytorch/xla

  • Accelerating AI inference?
    4 projects | /r/tensorflow | 2 Mar 2023
    Pytorch supports other kinds of accelerators (e.g. FPGA, and https://github.com/pytorch/glow), but unless you want to become a ML systems engineer and have money and time to throw away, or a business case to fund it, it is not worth it. In general, both pytorch and tensorflow have hardware abstractions that will compile down to device code. (XLA, https://github.com/pytorch/xla, https://github.com/pytorch/glow). TPUs and GPUs have very different strengths; so getting top performance requires a lot of manual optimizations. Considering the the cost of training LLM, it is time well spent.
  • [D] Colab TPU low performance
    2 projects | /r/MachineLearning | 18 Nov 2021
    While apparently TPUs can theoretically achieve great speedups, getting to the point where they beat a single GPU requires a lot of fiddling around and debugging. A specific setup is required to make it work properly. E.g., here it says that to exploit TPUs you might need a better CPU to keep the TPU busy, than the one in colab. The tutorials I looked at oversimplified the whole matter, the same goes for pytorch-lightning which implies switching to TPU is as easy as changing a single parameter. Furthermore, none of the tutorials I saw (even after specifically searching for that) went into detail about why and how to set up a GCS bucket for data loading.
  • How to train large deep learning models as a startup
    5 projects | news.ycombinator.com | 7 Oct 2021
  • Distributed Training Made Easy with PyTorch-Ignite
    7 projects | dev.to | 10 Aug 2021
    XLA on TPUs via pytorch/xla.
  • [P] PyTorch for TensorFlow Users - A Minimal Diff
    1 project | /r/MachineLearning | 9 Mar 2021
    I don't know of any such trick except for using TensorFlow. In fact, I benchmarked PyTorch XLA vs TensorFlow and found that the former's performance was quite abysmal: PyTorch XLA is very slow on Google Colab. The developers' explanation, as I understood it, was that TF was using features not available to the PyTorch XLA developers and that they therefore could not compete on performance. The situation may be different today, I don't know really.

determined

Posts with mentions or reviews of determined. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-01-05.
  • Open Source Advent Fun Wraps Up!
    10 projects | dev.to | 5 Jan 2024
    17. Determined AI | Github | tutorial
  • ML Experiments Management with Git
    4 projects | news.ycombinator.com | 2 Nov 2023
    Use Determined if you want a nice UI https://github.com/determined-ai/determined#readme
  • Determined: Deep Learning Training Platform
    1 project | news.ycombinator.com | 24 Mar 2023
  • Queueing/Resource Management Solutions for Self Hosted Workstation?
    1 project | /r/mlops | 23 Jan 2023
    I looked up and found [Determined Platform](determined.ai), tho it looks a very young project that I don't know if it's reliable enough.
  • Ask HN: Who is hiring? (June 2022)
    22 projects | news.ycombinator.com | 1 Jun 2022
    - Developer Support Engineer (~1/3 client facing, triaging feature requests and bug reports, etc; 2/3 debugging/troubleshooting)

    We are developing enterprise grade artificial intelligence products/services for AI engineering teams and fortune 500 companies and need more software devs to fill the increasing demand.

    Find out more at https://determined.ai/. If AI piques your curiosity or you want to interface with highly skilled engineers in the community, apply within (search "determined ai" at careers.hpe.com and drop me a message at asnell AT hpe PERIOD com).

  • How to train large deep learning models as a startup
    5 projects | news.ycombinator.com | 7 Oct 2021
    Check out Determined https://github.com/determined-ai/determined to help manage this kind of work at scale: Determined leverages Horovod under the hood, automatically manages cloud resources and can get you up on spot instances, T4's, etc. and will work on your local cluster as well. Gives you additional features like experiment management, scheduling, profiling, model registry, advanced hyperparameter tuning, etc.

    Full disclosure: I'm a founder of the project.

  • [D] managing compute for long running ML training jobs
    2 projects | /r/MachineLearning | 21 Jun 2021
    These are some of the problems we are trying to solve with the Determined training platform. Determined can be run with or without k8s - the k8s version inherits some of the scheduling problems of k8s, but the non-k8s version uses a custom gang scheduler designed for large scale ML training. Determined offers a priority scheduler that allows smaller jobs to run while being able to schedule a large distributed job whenever you need, by setting a higher priority.
  • Cerebras’ New Monster AI Chip Adds 1.4T Transistors
    4 projects | news.ycombinator.com | 22 Apr 2021
    Ah I see - I think we're pretty much on the same page in terms of timetables. Although if you include TPU, I think it's fair to say that custom accelerators are already a moderate success.

    Updated my profile. I've been working on DL training platforms and distributed training benchmarking for a bit so I've gotten a nice view into the GPU/TPU battle.

    Shameless plug: you should check out the open-source training platform we are building, Determined[1]. One of the goals is to take our hard-earned expertise on training infrastructure and build a tool where people don't need to have that infrastructure expertise. We don't support TPUs, partially because a lack of demand/TPU availability, and partially because our PyTorch TPU experiments were so unimpressive.

    [1] GH: https://github.com/determined-ai/determined, Slack: https://join.slack.com/t/determined-community/shared_invite/...

  • [D] Software stack to replicate Azure ML / Google Auto ML on premise
    2 projects | /r/MachineLearning | 3 Feb 2021
    Take a look at Determined https://github.com/determined-ai/determined
  • AWS open source news and updates No.41
    13 projects | dev.to | 25 Oct 2020
    determined is an open-source deep learning training platform that makes building models fast and easy. This project provides a CloudFormation template to bootstrap you into AWS and then has a number of tutorials covering how to manage your data, train and then deploy inference endpoints. If you are looking to explore more open source machine learning projects, then check this one out.

What are some alternatives?

When comparing xla and determined you can also consider the following projects:

NCCL - Optimized primitives for collective multi-GPU communication

ColossalAI - Making large AI models cheaper, faster and more accessible

pytorch-lightning - Build high-performance AI models with PyTorch Lightning (organized PyTorch). Deploy models with Lightning Apps (organized Python to build end-to-end ML systems). [Moved to: https://github.com/Lightning-AI/lightning]

Dagger.jl - A framework for out-of-core and parallel execution

why-ignite - Why should we use PyTorch-Ignite ?

aws-virtual-gpu-device-plugin - AWS virtual gpu device plugin provides capability to use smaller virtual gpus for your machine learning inference workloads

pocketsphinx - A small speech recognizer

cfn-diagram - CLI tool to visualise CloudFormation/SAM/CDK stacks as visjs networks, draw.io or ascii-art diagrams.

ignite - High-level library to help with training and evaluating neural networks in PyTorch flexibly and transparently.

goofys - a high-performance, POSIX-ish Amazon S3 file system written in Go

ompi - Open MPI main development repository

alpa - Training and serving large-scale neural networks with auto parallelization.