SaaSHub helps you find the best software and product alternatives Learn more โ
Top 23 Python Transformer Projects
-
[{'id': 'AI Music Generation', 'text': 'Love this new Space built by @enzostvs + @Xenova for Transformers.js: Generate your own AI music (In-browser generation) with AI Jukebox \n\nhttps://huggingface.co/spaces/enzostvs/ai-jukebox', 'score': 0.8460421562194824}, {'id': 'Kolmogorov Arnold Networks', 'text': 'Transformers are not all we need, that is being proven repeatedly now as more alternative frameworks emerge. Another such framework is Kolmogorov Arnold Network based Transformers. I break down exactly how these differ from Perceptron based Transformers and give you the link to my Colab where I create a model based on the research paper that absolutely destroys a standard Transformers based model. Check out the video here: https://www.youtube.com/watch?v=Sw0euxNZCc4', 'score': 0.8424240350723267}, {'id': 'GitHub Issue 8771', 'text': 'This issue is just a treasure ! A bit deprecated i guess, but things are in their historical context. (personally, still need more to understand better)\nhttps://github.com/huggingface/transformers/issues/8771\n\U0001fae1 to the man @stas ', 'score': 0.8417709469795227}]
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
-
nn
๐งโ๐ซ 60+ Implementations/tutorials of deep learning papers with side-by-side notes ๐; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), ๐ฎ reinforcement learning (ppo, dqn), capsnet, distillation, ... ๐ง
-
vLLM stands for virtual large language models. It is one of the open source fast inferencing and serving libraries. As the name suggests, โvirtualโ encapsulates the concept of virtual memory and paging from operating systems, which allows addressing the problem of maximum utilization of resources and providing faster token generation by utilizing PagedAttention. Traditional LLM serving involves storing large attention keys and value tensors in GPU memory, leading to inefficient memory usage.
-
-
Best of ml python
-
Project mention: Leveraging Large Language Models for Advanced Multilingual Text-to-Speech | news.ycombinator.com | 2024-11-05
-
Project mention: Show HN: Synthesize TikZ Graphics Programs for Scientific Figures and Sketches | news.ycombinator.com | 2024-06-06
already claim to (at least partially) support this.
[1] https://github.com/lukas-blecher/LaTeX-OCR
-
RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
https://github.com/BlinkDL/RWKV-LM#rwkv-discord-httpsdiscord... lists a number of implementations of various versions of RWKV.
https://github.com/BlinkDL/RWKV-LM#rwkv-parallelizable-rnn-w... :
> RWKV: Parallelizable RNN with Transformer-level LLM Performance (pronounced as "RwaKuv", from 4 major params: R W K V)
> RWKV is an RNN with Transformer-level LLM performance, which can also be directly trained like a GPT transformer (parallelizable). And it's 100% attention-free. You only need the hidden state at position t to compute the state at position t+1. You can use the "GPT" mode to quickly compute the hidden state for the "RNN" mode.
> So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding (using the final hidden state).
> "Our latest version is RWKV-6,*
-
Project mention: Play 3.0 mini โ A lightweight, reliable, cost-efficient Multilingual TTS model | news.ycombinator.com | 2024-10-14
Hi, I don't know what's SOTA, but I got good results with these (open source, on-device) :
https://github.com/SYSTRAN/faster-whisper (speech-to-text)
-
PaddleSpeech
Easy-to-use Speech Toolkit including Self-Supervised Learning model, SOTA/Streaming ASR with punctuation, Streaming TTS with text frontend, Speaker Verification System, End-to-End Speech Translation and Keyword Spotting. Won NAACL2022 Best Demo Award.
-
petals
๐ธ Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
Project mention: Serving AI from the Basement โ 192GB of VRAM Setup | news.ycombinator.com | 2024-09-08 -
Project mention: Best LLM Inference Engines and Servers to Deploy LLMs in Production | dev.to | 2024-06-05
GitHub repository: https://github.com/huggingface/text-generation-inference
-
PaddleSeg
Easy-to-use image segmentation library with awesome pre-trained model zoo, supporting wide-range of practical tasks in Semantic Segmentation, Interactive Segmentation, Panoptic Segmentation, Image Matting, 3D Segmentation, etc.
-
-
LMFlow
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
-
Project mention: Maxtext: A simple, performant and scalable Jax LLM | news.ycombinator.com | 2024-04-23
Is t5x an encoder/decoder architecture?
Some more general options.
The Flax ecosystem
https://github.com/google/flax?tab=readme-ov-file
or dm-haiku
https://github.com/google-deepmind/dm-haiku
were some of the best developed communities in the Jax AI field
Perhaps the โtraxโ repo? https://github.com/google/trax
Some HF examples https://github.com/huggingface/transformers/tree/main/exampl...
Sadly it seems much of the work is proprietary these days, but one example could be Grok-1, if you customize the details. https://github.com/xai-org/grok-1/blob/main/run.py
-
-
-
The easiest is to use vllm (https://github.com/vllm-project/vllm) to run it on a Couple of A100's, and you can benchmark this using this library (https://github.com/EleutherAI/lm-evaluation-harness)
-
Project mention: StreamingLLM: tiny tweak to KV LRU improves long conversations | news.ycombinator.com | 2024-02-13
This seems only to work cause large GPTs have redundant, undercomplex attentions. See this issue in BertViz about attention in Llama: https://github.com/jessevig/bertviz/issues/128
-
-
-
manga-image-translator
Translate manga/image ไธ้ฎ็ฟป่ฏๅ็ฑปๅพ็ๅ ๆๅญ https://cotrans.touhou.ai/
Python Transformer discussion
Python Transformer related posts
-
Leveraging Large Language Models for Advanced Multilingual Text-to-Speech
-
Quantized Llama models with increased speed and a reduced memory footprint
-
How to Run Llama 3 405B on Home Devices? Build AI Cluster
-
Fish Speech 1.3 is released โ Clone Any Voice in 15s
-
Codestral Mamba
-
Show HN: Synthesize TikZ Graphics Programs for Scientific Figures and Sketches
-
Best LLM Inference Engines and Servers to Deploy LLMs in Production
-
A note from our sponsor - SaaSHub
www.saashub.com | 1 Dec 2024
Index
What are some of the best open-source Transformer projects in Python? This list will help you:
Project | Stars | |
---|---|---|
1 | transformers | 135,627 |
2 | nn | 56,623 |
3 | vllm | 30,713 |
4 | mmdetection | 29,672 |
5 | best-of-ml-python | 18,182 |
6 | fish-speech | 14,662 |
7 | LaTeX-OCR | 12,836 |
8 | RWKV-LM | 12,707 |
9 | faster-whisper | 12,676 |
10 | PaddleSpeech | 11,191 |
11 | petals | 9,260 |
12 | text-generation-inference | 9,162 |
13 | PaddleSeg | 8,726 |
14 | mmsegmentation | 8,321 |
15 | LMFlow | 8,291 |
16 | trax | 8,102 |
17 | jukebox | 7,833 |
18 | GPT2-Chinese | 7,475 |
19 | lm-evaluation-harness | 7,059 |
20 | bertviz | 6,962 |
21 | BERT-pytorch | 6,160 |
22 | Informer2020 | 5,476 |
23 | manga-image-translator | 5,419 |