SaaSHub helps you find the best software and product alternatives Learn more →
Top 23 Python Transformer Projects
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.Project mention: Fine-Tuned Llama2 Inserting Unnecessary Delimiters | /r/LocalLLaMA | 2023-11-04
While its tough to say something specifc since we dont know how exactly you trained it or the prompt format of your training input or how you are performing inference, one thing I found when I faced similar types of issues is that the model does not know when to stop. Some of it is because the fast llama tokenizer does not add the token when encoding your inputs. So you can either add that token explicitly in your input text for each sample or use the slow llama tokenizer. Check llama_recipes github repo for the exact issue https://github.com/huggingface/transformers/issues/22794. The other most probable thing you might want to check is if the model.generate output contains the exact input tokens too. That is the expected behavior of some models (like llama2 or mpt) for example when you use vanilla transformers for inference.
OpenMMLab Detection Toolbox and BenchmarkProject mention: Semantic segementation | /r/computervision | 2023-04-12
When I look for benchmarks I always start here https://paperswithcode.com/task/instance-segmentation/codeless it has the lists of datasets to measure models accross lots o papers. Many are very specific models with low support or community but it gives you a good idea of the state of the art. It also lists repositories related to good community. https://github.com/open-mmlab/mmdetection seems very active and the one that is being used the most, you could use the models that it has integrated in its model zoo, within the same repository. It has the benchmarks to compare those same models and some of them are from 2022
Collect and Analyze Billions of Data Points in Real Time. Manage all types of time series data in a single, purpose-built database. Run at any scale in any environment in the cloud, on-premises, or at the edge.
🏆 A ranked list of awesome machine learning Python libraries. Updated weekly.Project mention: Ask HN: How to get back into AI? | news.ycombinator.com | 2022-12-10
For Python, here's a nice compilation: https://github.com/ml-tooling/best-of-ml-python/blob/main/RE...
A high-throughput and memory-efficient inference and serving engine for LLMsProject mention: OpenAI Delays | news.ycombinator.com | 2023-12-01
Heh, so it totally depends on the use case.
I use GPT4 constantly to chat through issues I am working on and get different perspectives. I cannot do that with local models.
On the other hand, I have been processing a ton of text transcripts with a fine tuned llama2 13b model i've been working on, and for the tasks I have fine-tuned on, my local model is producing better results than GPT4, often taking a task that I had to do in multiple steps with GPT4, and being able to complete it in a single shot.
I can run my local model through vLLM on my workstation at around the same tokens/sec as I can spend maxing out my API limits with GPT3.5-turbo (~$20/hr) while running on 2x 3090's. I'm hitting the vLLM (OpenAI clone) chat/completions endpoint. My model implements the HF chat_templates feature, and I worked on adding support for that to vLLM: https://github.com/vllm-project/vllm/pull/1756 (llama.cpp is talking about adding support for it too) so I could easily swap out my model in my data pipeline in place of GPT3.5/GPT4, and I wouldn't have to keep maintaining that code on my side.
So, with these transcripts I've been:
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
Easy-to-use Speech Toolkit including Self-Supervised Learning model, SOTA/Streaming ASR with punctuation, Streaming TTS with text frontend, Speaker Verification System, End-to-End Speech Translation and Keyword Spotting. Won NAACL2022 Best Demo Award.Project mention: Open Source Libraries | /r/AudioAI | 2023-10-02
Trax — Deep Learning with Clear Code and SpeedProject mention: Replit's new Code LLM was trained in 1 week | news.ycombinator.com | 2023-05-03
and the implementation https://github.com/google/trax/blob/master/trax/models/resea... if you are interested.
Hope you get to look into this!
Learn any GitHub repo in 59 seconds. Onboard AI learns any GitHub repo in minutes and lets you chat with it to locate functionality, understand different parts, and generate new code. Use it for free at www.getonboard.dev.
Easy-to-use image segmentation library with awesome pre-trained model zoo, supporting wide-range of practical tasks in Semantic Segmentation, Interactive Segmentation, Panoptic Segmentation, Image Matting, 3D Segmentation, etc.Project mention: [Medical Segmentation] The all-in-one 3D medical image segmentation toolkit. From data annotation to model deployment, you are welcome to try it all！ | /r/ArtificialInteligence | 2022-12-19
pix2tex: Using a ViT to convert images of equations into LaTeX code.Project mention: Detexify LaTeX Handwriting Symbol Recognition | news.ycombinator.com | 2023-11-14
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.Project mention: Your weekly machine learning digest | /r/learnmachinelearning | 2023-07-03
Code for the paper "Jukebox: A Generative Model for Music"Project mention: Open Source Libraries | /r/AudioAI | 2023-10-02
openai/jukebox: Music Generation
Chinese version of GPT2 training code, using BERT tokenizer.
OpenMMLab Semantic Segmentation Toolbox and Benchmark.Project mention: [D] The MMSegmentation library from OpenMMLab appears to return the wrong results when computing basic image segmentation metrics such as the Jaccard index (IoU - intersection-over-union). It appears to compute recall (sensitivity) instead of IoU, which artificially inflates the performance metrics. | /r/MachineLearning | 2023-03-06
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
Faster Whisper transcription with CTranslate2Project mention: Distil-Whisper: distilled version of Whisper that is 6 times faster, 49% smaller | news.ycombinator.com | 2023-10-31
That's the implication. If the distil models are same format as original openai models then the Distil models can be converted for faster-whisper use as per the conversion instructions on https://github.com/guillaumekln/faster-whisper/
So then we'll see whether we get the 6x model speedup on top of the stated 4x faster-whisper code speedup.
Google AI 2018 BERT pytorch implementation
The GitHub repository for the paper "Informer" accepted by AAAI 2021.
An Open-Source Framework for Prompt-Learning.
SwinIR: Image Restoration Using Swin Transformer (official repository)Project mention: Certain directories (e.g. SwinIR) are empty (version: Empire Media Science A1111 Web UI Installer) | /r/StableDiffusion | 2023-03-17
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
Translate manga/image 一键翻译各类图片内文字 https://cotrans.touhou.ai/Project mention: [DISC] - The angel who came to pick me up is a Gal (Oneshot by Shiraishi Kouhei) | /r/manga | 2023-09-06
OCR works pretty good. ocr.space, ocr.best and cotrans.touhou.ai/ are all pretty nice.
The OCR approach is rephrased as Segmentation Transformer: https://arxiv.org/abs/1909.11065. This is an official implementation of semantic segmentation for HRNet. https://arxiv.org/abs/1908.07919
Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast.Project mention: FLaNK Stack Weekly for 14 Aug 2023 | dev.to | 2023-08-14
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
Python Transformer related posts
Detexify LaTeX Handwriting Symbol Recognition
5 projects | news.ycombinator.com | 14 Nov 2023
Pix2tex: Using a ViT to convert images of equations into LaTeX code
1 project | /r/hypeurls | 7 Nov 2023
Pix2tex: Using a ViT to convert images of equations into LaTeX code
5 projects | news.ycombinator.com | 3 Nov 2023
Open Source Libraries
25 projects | /r/AudioAI | 2 Oct 2023
A look at Apple’s new Transformer-powered predictive text model
4 projects | news.ycombinator.com | 16 Sep 2023
OpenAI compatible API for LLMs (and embeddings) with ready-to-use Docker images
1 project | news.ycombinator.com | 14 Aug 2023
[D] What is the most efficient version of OpenAI Whisper?
7 projects | /r/MachineLearning | 12 Jul 2023
A note from our sponsor - #<SponsorshipServiceOld:0x00007f0f9b43be88>
www.saashub.com | 3 Dec 2023
What are some of the best open-source Transformer projects in Python? This list will help you: