xla

Enabling PyTorch on XLA Devices (e.g. Google TPU) (by pytorch)

Xla Alternatives

Similar projects and alternatives to xla

  • Pytorch

    336 xla VS Pytorch

    Tensors and Dynamic neural networks in Python with strong GPU acceleration

  • InvokeAI

    239 xla VS InvokeAI

    InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.

  • WorkOS

    The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.

    WorkOS logo
  • pytorch-lightning

    Discontinued Build high-performance AI models with PyTorch Lightning (organized PyTorch). Deploy models with Lightning Apps (organized Python to build end-to-end ML systems). [Moved to: https://github.com/Lightning-AI/lightning] (by PyTorchLightning)

  • Megatron-LM

    18 xla VS Megatron-LM

    Ongoing research training transformer models at scale

  • ompi

    Open MPI main development repository

  • pocketsphinx

    A small speech recognizer

  • NCCL

    3 xla VS NCCL

    Optimized primitives for collective multi-GPU communication

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • determined

    Determined is an open-source machine learning platform that simplifies distributed training, hyperparameter tuning, experiment tracking, and resource management. Works with PyTorch and TensorFlow.

  • ignite

    3 xla VS ignite

    High-level library to help with training and evaluating neural networks in PyTorch flexibly and transparently. (by pytorch)

  • why-ignite

    Why should we use PyTorch-Ignite ?

  • glow

    6 xla VS glow

    Compiler for Neural Network hardware accelerators (by pytorch)

  • snowboy

    6 xla VS snowboy

    Future versions with model training module will be maintained through a forked version here: https://github.com/seasalt-ai/snowboy

  • gloo

    2 xla VS gloo

    Collective communications library with various primitives for multi-machine training. (by facebookincubator)

  • mojo.vim

    3 xla VS mojo.vim

    Vim Syntax Highlighting for the Mojo programming language

  • pytorch-lightning

    Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.

  • Spoken-Keyword-Spotting

    In this repository, we explore using a hybrid system consisting of a Convolutional Neural Network and a Support Vector Machine for Keyword Spotting task.

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a better xla alternative or higher similarity.

xla reviews and mentions

Posts with mentions or reviews of xla. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-01-26.
  • Who uses Google TPUs for inference in production?
    1 project | news.ycombinator.com | 11 Mar 2024
    > The PyTorch/XLA Team at Google

    Meanwhile you have an issue from 5 years ago with 0 support

    https://github.com/pytorch/xla/issues/202

  • Google TPU v5p beats Nvidia H100
    2 projects | news.ycombinator.com | 26 Jan 2024
    PyTorch has had an XLA backend for years. I don't know how performant it is though. https://pytorch.org/xla
  • Why Did Google Brain Exist?
    2 projects | news.ycombinator.com | 26 Apr 2023
    It's curtains for XLA, to be precise. And PyTorch officially supports XLA backend nowadays too ([1]), which kind of makes JAX and PyTorch standing on the same foundation.

    1. https://github.com/pytorch/xla

  • Accelerating AI inference?
    4 projects | /r/tensorflow | 2 Mar 2023
    Pytorch supports other kinds of accelerators (e.g. FPGA, and https://github.com/pytorch/glow), but unless you want to become a ML systems engineer and have money and time to throw away, or a business case to fund it, it is not worth it. In general, both pytorch and tensorflow have hardware abstractions that will compile down to device code. (XLA, https://github.com/pytorch/xla, https://github.com/pytorch/glow). TPUs and GPUs have very different strengths; so getting top performance requires a lot of manual optimizations. Considering the the cost of training LLM, it is time well spent.
  • [D] Colab TPU low performance
    2 projects | /r/MachineLearning | 18 Nov 2021
    While apparently TPUs can theoretically achieve great speedups, getting to the point where they beat a single GPU requires a lot of fiddling around and debugging. A specific setup is required to make it work properly. E.g., here it says that to exploit TPUs you might need a better CPU to keep the TPU busy, than the one in colab. The tutorials I looked at oversimplified the whole matter, the same goes for pytorch-lightning which implies switching to TPU is as easy as changing a single parameter. Furthermore, none of the tutorials I saw (even after specifically searching for that) went into detail about why and how to set up a GCS bucket for data loading.
  • How to train large deep learning models as a startup
    5 projects | news.ycombinator.com | 7 Oct 2021
  • Distributed Training Made Easy with PyTorch-Ignite
    7 projects | dev.to | 10 Aug 2021
    XLA on TPUs via pytorch/xla.
  • [P] PyTorch for TensorFlow Users - A Minimal Diff
    1 project | /r/MachineLearning | 9 Mar 2021
    I don't know of any such trick except for using TensorFlow. In fact, I benchmarked PyTorch XLA vs TensorFlow and found that the former's performance was quite abysmal: PyTorch XLA is very slow on Google Colab. The developers' explanation, as I understood it, was that TF was using features not available to the PyTorch XLA developers and that they therefore could not compete on performance. The situation may be different today, I don't know really.
  • A note from our sponsor - SaaSHub
    www.saashub.com | 26 Apr 2024
    SaaSHub helps you find the best software and product alternatives Learn more →

Stats

Basic xla repo stats
8
2,291
9.9
1 day ago

Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com