Megatron-DeepSpeed

Ongoing research training transformer language models at scale, including: BERT & GPT-2 (by bigscience-workshop)

Megatron-DeepSpeed Alternatives

Similar projects and alternatives to Megatron-DeepSpeed

  • gpt-neox

    An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.

  • fairscale

    PyTorch extensions for high performance and large scale training.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a better Megatron-DeepSpeed alternative or higher similarity.

Megatron-DeepSpeed reviews and mentions

Posts with mentions or reviews of Megatron-DeepSpeed. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2022-02-17.
  • [D] DeepSpeed vs PyTorch native API
    3 projects | /r/MachineLearning | 17 Feb 2022
    Both EleutherAI's gpt-neox and the BigScience project use DeepSpeed under the hood, probably because DeepSpeed still remains the best component for training large models. So really dependent on your scale if DeepSpeed is still your answer, or if you can get away with these native PyTorch alternatives.

Stats

Basic Megatron-DeepSpeed repo stats
1
1,242
2.4
about 1 month ago

bigscience-workshop/Megatron-DeepSpeed is an open source project licensed under GNU General Public License v3.0 or later which is an OSI approved license.

The primary programming language of Megatron-DeepSpeed is Python.


Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com