Python Transformer

Open-source Python projects categorized as Transformer

Top 23 Python Transformer Projects

  • transformers

    🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

  • Project mention: AI enthusiasm #6 - Finetune any LLM you want💡 | dev.to | 2024-04-16

    Most of this tutorial is based on Hugging Face course about Transformers and on Niels Rogge's Transformers tutorials: make sure to check their work and give them a star on GitHub, if you please ❤️

  • mmdetection

    OpenMMLab Detection Toolbox and Benchmark

  • WorkOS

    The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.

    WorkOS logo
  • vllm

    A high-throughput and memory-efficient inference and serving engine for LLMs

  • Project mention: Mistral AI Launches New 8x22B Moe Model | news.ycombinator.com | 2024-04-09

    The easiest is to use vllm (https://github.com/vllm-project/vllm) to run it on a Couple of A100's, and you can benchmark this using this library (https://github.com/EleutherAI/lm-evaluation-harness)

  • best-of-ml-python

    🏆 A ranked list of awesome machine learning Python libraries. Updated weekly.

  • RWKV-LM

    RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.

  • Project mention: Do LLMs need a context window? | news.ycombinator.com | 2023-12-25

    https://github.com/BlinkDL/RWKV-LM#rwkv-discord-httpsdiscord... lists a number of implementations of various versions of RWKV.

    https://github.com/BlinkDL/RWKV-LM#rwkv-parallelizable-rnn-w... :

    > RWKV: Parallelizable RNN with Transformer-level LLM Performance (pronounced as "RwaKuv", from 4 major params: R W K V)

    > RWKV is an RNN with Transformer-level LLM performance, which can also be directly trained like a GPT transformer (parallelizable). And it's 100% attention-free. You only need the hidden state at position t to compute the state at position t+1. You can use the "GPT" mode to quickly compute the hidden state for the "RNN" mode.

    > So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding (using the final hidden state).

    > "Our latest version is RWKV-6,*

  • LaTeX-OCR

    pix2tex: Using a ViT to convert images of equations into LaTeX code.

  • Project mention: Detexify LaTeX Handwriting Symbol Recognition | news.ycombinator.com | 2023-11-14
  • PaddleSpeech

    Easy-to-use Speech Toolkit including Self-Supervised Learning model, SOTA/Streaming ASR with punctuation, Streaming TTS with text frontend, Speaker Verification System, End-to-End Speech Translation and Keyword Spotting. Won NAACL2022 Best Demo Award.

  • Project mention: Open Source Libraries | /r/AudioAI | 2023-10-02

    PaddlePaddle/PaddleSpeech

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • petals

    🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading

  • Project mention: Mistral Large | news.ycombinator.com | 2024-02-26

    So how long until we can do an open source Mistral Large?

    We could make a start on Petals or some other open source distributed training network cluster possibly?

    [0] https://petals.dev/

  • faster-whisper

    Faster Whisper transcription with CTranslate2

  • Project mention: Using Groq to Build a Real-Time Language Translation App | dev.to | 2024-04-05

    For our real-time STT needs, we'll employ a fantastic library called faster-whisper.

  • PaddleSeg

    Easy-to-use image segmentation library with awesome pre-trained model zoo, supporting wide-range of practical tasks in Semantic Segmentation, Interactive Segmentation, Panoptic Segmentation, Image Matting, 3D Segmentation, etc.

  • LMFlow

    An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.

  • Project mention: Your weekly machine learning digest | /r/learnmachinelearning | 2023-07-03
  • trax

    Trax — Deep Learning with Clear Code and Speed

  • Project mention: Replit's new Code LLM was trained in 1 week | news.ycombinator.com | 2023-05-03

    and the implementation https://github.com/google/trax/blob/master/trax/models/resea... if you are interested.

    Hope you get to look into this!

  • text-generation-inference

    Large Language Model Text Generation Inference

  • Project mention: Zephyr 141B, a Mixtral 8x22B fine-tune, is now available in Hugging Chat | news.ycombinator.com | 2024-04-12

    I wanted to write that TGI inference engine is not Open Source anymore, but they have reverted the license back to Apache 2.0 for the new version TGI v2.0: https://github.com/huggingface/text-generation-inference/rel...

    Good news!

  • jukebox

    Code for the paper "Jukebox: A Generative Model for Music"

  • Project mention: Open Source Libraries | /r/AudioAI | 2023-10-02

    openai/jukebox: Music Generation

  • mmsegmentation

    OpenMMLab Semantic Segmentation Toolbox and Benchmark.

  • GPT2-Chinese

    Chinese version of GPT2 training code, using BERT tokenizer.

  • bertviz

    BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)

  • Project mention: StreamingLLM: tiny tweak to KV LRU improves long conversations | news.ycombinator.com | 2024-02-13

    This seems only to work cause large GPTs have redundant, undercomplex attentions. See this issue in BertViz about attention in Llama: https://github.com/jessevig/bertviz/issues/128

  • BERT-pytorch

    Google AI 2018 BERT pytorch implementation

  • Informer2020

    The GitHub repository for the paper "Informer" accepted by AAAI 2021.

  • lm-evaluation-harness

    A framework for few-shot evaluation of language models.

  • Project mention: Mistral AI Launches New 8x22B Moe Model | news.ycombinator.com | 2024-04-09

    The easiest is to use vllm (https://github.com/vllm-project/vllm) to run it on a Couple of A100's, and you can benchmark this using this library (https://github.com/EleutherAI/lm-evaluation-harness)

  • OpenPrompt

    An Open-Source Framework for Prompt-Learning.

  • manga-image-translator

    Translate manga/image 一键翻译各类图片内文字 https://cotrans.touhou.ai/

  • Project mention: [DISC] - The angel who came to pick me up is a Gal (Oneshot by Shiraishi Kouhei) | /r/manga | 2023-09-06

    OCR works pretty good. ocr.space, ocr.best and cotrans.touhou.ai/ are all pretty nice.

  • SwinIR

    SwinIR: Image Restoration Using Swin Transformer (official repository)

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
NOTE: The open source projects on this list are ordered by number of github stars. The number of mentions indicates repo mentiontions in the last 12 Months or since we started tracking (Dec 2020). The latest post mention was on 2024-04-16.

Python Transformer related posts

Index

What are some of the best open-source Transformer projects in Python? This list will help you:

Project Stars
1 transformers 124,557
2 mmdetection 27,658
3 vllm 17,656
4 best-of-ml-python 15,302
5 RWKV-LM 11,579
6 LaTeX-OCR 10,711
7 PaddleSpeech 10,069
8 petals 8,631
9 faster-whisper 8,578
10 PaddleSeg 8,227
11 LMFlow 7,975
12 trax 7,948
13 text-generation-inference 7,722
14 jukebox 7,554
15 mmsegmentation 7,342
16 GPT2-Chinese 7,342
17 bertviz 6,356
18 BERT-pytorch 5,979
19 Informer2020 4,890
20 lm-evaluation-harness 4,848
21 OpenPrompt 4,141
22 manga-image-translator 4,127
23 SwinIR 4,060
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com