Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →
Top 23 Python pretrained-model Projects
-
pytorch-image-models
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more
-
Scout Monitoring
Free Django app performance insights with Scout Monitoring. Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.
-
PaddleNLP
👑 Easy-to-use and powerful NLP and LLM library with 🤗 Awesome model zoo, supporting wide-range of NLP tasks from research to industrial applications, including 🗂Text Classification, 🔍 Neural Search, ❓ Question Answering, ℹ️ Information Extraction, 📄 Document Intelligence, 💌 Sentiment Analysis etc.
-
Qwen
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
-
petals
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
LMFlow
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
-
awesome-pretrained-chinese-nlp-models
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
-
facenet-pytorch
Pretrained Pytorch face detection (MTCNN) and facial recognition (InceptionResnet) models
-
Efficient-AI-Backbones
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
-
Chinese-CLIP
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
-
PINTO_model_zoo
A repository for storing models that have been inter-converted between various frameworks. Supported frameworks are TensorFlow, PyTorch, ONNX, OpenVINO, TFJS, TFTRT, TensorFlowLite (Float32/16/INT8), EdgeTPU, CoreML.
-
Pretrained-Language-Model
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
-
OFA
Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
Project mention: How to count tokens in frontend for Popular LLM Models: GPT, Claude, and Llama | dev.to | 2024-05-21Thanks to transformers.js, we can run the tokenizer and model locally in the browser. Transformers.js is designed to be functionally equivalent to Hugging Face's transformers python library, meaning you can run the same pretrained models using a very similar API.
virtual dj and others stem separator is shrinked model of this https://github.com/deezer/spleeter you will get better results downloading original + their large model.
Qwen: https://github.com/QwenLM/Qwen
Project mention: Instance segmentation of small objects in grainy drone imagery | /r/computervision | 2023-12-09Also, I’d suggest considering switching to the segmentation-models library - it provides U-Net models with a variety of pretrained backbones of as encoders. The author also put out a PyTorch version. https://github.com/qubvel/segmentation_models.pytorch https://github.com/qubvel/segmentation_models
Things like [petals](https://github.com/bigscience-workshop/petals) exist, distributed computing over willing participants. Right now corporate cash is being rammed into the space so why not snap it up while you can, but the moment it dries up projects like petals will see more of the love they deserve.
I envision a future where crypto-style booms happen over tokens useful for purchasing priority computational time, which is earned by providing said computational time. This way researchers can daisy-chain their independent smaller rigs together into something with gargantuan capabilities.
CogVLM is very good in my (brief) testing: https://github.com/THUDM/CogVLM
The model weights seem to be under a non-commercial license, not true open source, but it is "open access" as you requested.
Project mention: Fast Llama 2 on CPUs with Sparse Fine-Tuning and DeepSparse | news.ycombinator.com | 2023-11-23Interesting company. Yannic Kilcher interviewed Nir Shavit last year and they went into some depth: https://www.youtube.com/watch?v=0PAiQ1jTN5k DeepSparse is on GitHub: https://github.com/neuralmagic/deepsparse
Python pretrained-models related posts
-
CogAgent-18B – visual-based GUI Agent capabilities
-
Are stems a good way of making mashups
-
What do you think. When should we expect the next SDXL version?
-
Big News!
-
Anybody here know what AI model does Steinberg's Spectralayers use to do stem separation?
-
Gemini: Google's most capable AI model yet
-
Open-source LLMs with Image Interpretation
-
A note from our sponsor - InfluxDB
www.influxdata.com | 1 Jun 2024
Index
What are some of the best open-source pretrained-model projects in Python? This list will help you:
Project | Stars | |
---|---|---|
1 | transformers | 126,915 |
2 | pytorch-image-models | 30,220 |
3 | spleeter | 25,067 |
4 | PaddleNLP | 11,583 |
5 | Qwen | 11,682 |
6 | segmentation_models.pytorch | 8,951 |
7 | petals | 8,763 |
8 | LMFlow | 8,062 |
9 | CodeGeeX | 7,840 |
10 | EfficientNet-PyTorch | 7,715 |
11 | mmf | 5,425 |
12 | PaddleClas | 5,290 |
13 | CogVLM | 5,355 |
14 | awesome-pretrained-chinese-nlp-models | 4,332 |
15 | facenet-pytorch | 4,232 |
16 | Efficient-AI-Backbones | 3,856 |
17 | Chinese-CLIP | 3,771 |
18 | PINTO_model_zoo | 3,351 |
19 | mmpretrain | 3,225 |
20 | Pretrained-Language-Model | 2,971 |
21 | deepsparse | 2,902 |
22 | OFA | 2,343 |
23 | asteroid | 2,135 |
Sponsored