Python pretrained-models

Open-source Python projects categorized as pretrained-models

Top 23 Python pretrained-model Projects

pretrained-models
  1. transformers

    🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

    Project mention: QwQ-32B: Embracing the Power of Reinforcement Learning | news.ycombinator.com | 2025-03-05

    Huggingface's transformers library supports something similar to this. You set a minimum length, and until that length is reached, the end of sequence token has no chance of being output.

    https://github.com/huggingface/transformers/blob/51ed61e2f05...

    S1 does something similar to put a lower limit on its reasoning output. End of thinking is represented with the <|im_start|> token, followed by the word 'answer'. IIRC the code dynamically adds/removes <|im_start|> to the list of suppressed tokens.

    Both of these approaches set the probability to zero, not something small like you were suggesting.

  2. CodeRabbit

    CodeRabbit: AI Code Reviews for Developers. Revolutionize your code reviews with AI. CodeRabbit offers PR summaries, code walkthroughs, 1-click suggestions, and AST-based analysis. Boost productivity and code quality across all major languages with each PR.

    CodeRabbit logo
  3. pytorch-image-models

    The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNetV4, MobileNet-V3 & V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more

    Project mention: This PR content was generated automatically using cover-agent | news.ycombinator.com | 2024-11-19

    Those are some pointless tests.

    E.g. test_activation_stats_functions [1] that just checks that the returned value is a float, and that it can take random numbers as input.

    test_get_state_dict_custom_unwrap [2] is probably supposed to check that custom_unwrap is invoked, but since it doesn't either record being called, or transform its input, the assertions can't actually check that it was called.

    [1] https://github.com/huggingface/pytorch-image-models/pull/233...

    [2] https://github.com/huggingface/pytorch-image-models/pull/233...

  4. spleeter

    Deezer source separation library including pretrained models.

    Project mention: Synchronizing pong to music with constrained optimization | news.ycombinator.com | 2024-09-03

    Absolutely wonderful!

    > "We obtain these times from MIDI files, though in the future I’d like to explore more automated ways of extracting them from audio."

    Same here. In case it helps: I suspect a suitable option is (python libs) Spleeter (https://github.com/deezer/spleeter) to split stems and Librosa (https://github.com/librosa/librosa) for beat times. I haven't ventured into this yet though so I may be off. My ultimate goal is to be able to do it 'on the fly', i.e. in a live music setting being able to generate visualisations a couple of seconds ahead being played along with the track.

    Not sure if this is unsavory self promotion, but I am in the middle of documenting something similar at the moment.

    Experiments #1 - A Mutating Maurer Rose | Syncing Scripted Geometric Patterns to Music: https://www.youtube.com/watch?v=bfU58rBInpw

    It generates a mutating Maurer Rose using react-native-svg on my RN stack, synced to a music track I created in Suno AI.

  5. Qwen

    The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.

    Project mention: Running Qwen, Nearly as Powerful as DeepSeek, on a MacBook Pro | dev.to | 2025-02-05

    Qwen (Qwen GitHub Repository) has been gaining attention recently as a powerful open-source large language model (LLM). I decided to give it a spin on my MacBook Pro using Ollama, a platform designed for running local LLMs. While Qwen2.5-Max boasts the highest performance, my setup could only handle the smaller Qwen2.5 (32B) model. Here's what I found!

  6. PaddleNLP

    👑 Easy-to-use and powerful NLP and LLM library with 🤗 Awesome model zoo, supporting wide-range of NLP tasks from research to industrial applications, including 🗂Text Classification, 🔍 Neural Search, ❓ Question Answering, ℹ️ Information Extraction, 📄 Document Intelligence, 💌 Sentiment Analysis etc.

  7. open_clip

    An open source implementation of CLIP.

    Project mention: Xkcd 1425 (Tasks) turns ten years old today | news.ycombinator.com | 2024-09-25

    Here’s some code: https://github.com/mlfoundations/open_clip?tab=readme-ov-fil...

  8. segmentation_models.pytorch

    Semantic segmentation models with 500+ pretrained convolutional and transformer-based backbones.

  9. SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
  10. petals

    🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading

    Project mention: Serving AI from the Basement – 192GB of VRAM Setup | news.ycombinator.com | 2024-09-08
  11. CodeGeeX

    CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)

  12. LMFlow

    An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.

  13. EfficientNet-PyTorch

    A PyTorch implementation of EfficientNet

  14. InternLM

    Official release of InternLM series (InternLM, InternLM2, InternLM2.5, InternLM3).

    Project mention: InterLM: Open-source LLM with 1M context window | news.ycombinator.com | 2024-10-28
  15. CogVLM

    a state-of-the-art-level open visual language model | 多模态预训练模型

  16. PaddleClas

    A treasure chest for visual classification and recognition powered by PaddlePaddle

  17. mmf

    A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)

  18. awesome-pretrained-chinese-nlp-models

    Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合

  19. Chinese-CLIP

    Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.

  20. facenet-pytorch

    Pretrained Pytorch face detection (MTCNN) and facial recognition (InceptionResnet) models

  21. Efficient-AI-Backbones

    Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.

  22. PINTO_model_zoo

    A repository for storing models that have been inter-converted between various frameworks. Supported frameworks are TensorFlow, PyTorch, ONNX, OpenVINO, TFJS, TFTRT, TensorFlowLite (Float32/16/INT8), EdgeTPU, CoreML.

  23. mmpretrain

    OpenMMLab Pre-training Toolbox and Benchmark

  24. deepsparse

    Sparsity-aware deep learning inference runtime for CPUs

  25. Pretrained-Language-Model

    Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.

  26. SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
NOTE: The open source projects on this list are ordered by number of github stars. The number of mentions indicates repo mentiontions in the last 12 Months or since we started tracking (Dec 2020).

Python pretrained-models discussion

Log in or Post with

Python pretrained-models related posts

  • Analyzing Hugging Face Posts with Graphs and Agents

    7 projects | dev.to | 21 Nov 2024
  • This PR content was generated automatically using cover-agent

    2 projects | news.ycombinator.com | 19 Nov 2024
  • How to Run Llama 3 405B on Home Devices? Build AI Cluster

    2 projects | news.ycombinator.com | 29 Jul 2024
  • CogAgent-18B – visual-based GUI Agent capabilities

    2 projects | news.ycombinator.com | 16 Dec 2023
  • Are stems a good way of making mashups

    1 project | /r/Beatmatch | 10 Dec 2023
  • What do you think. When should we expect the next SDXL version?

    1 project | /r/StableDiffusion | 10 Dec 2023
  • Big News!

    1 project | /r/OnePieceMangaCut | 9 Dec 2023
  • A note from our sponsor - CodeRabbit
    coderabbit.ai | 16 Mar 2025
    Revolutionize your code reviews with AI. CodeRabbit offers PR summaries, code walkthroughs, 1-click suggestions, and AST-based analysis. Boost productivity and code quality across all major languages with each PR. Learn more →

Index

What are some of the best open-source pretrained-model projects in Python? This list will help you:

# Project Stars
1 transformers 141,182
2 pytorch-image-models 33,472
3 spleeter 26,487
4 Qwen 17,389
5 PaddleNLP 12,421
6 open_clip 11,222
7 segmentation_models.pytorch 10,177
8 petals 9,492
9 CodeGeeX 8,431
10 LMFlow 8,375
11 EfficientNet-PyTorch 7,991
12 InternLM 6,815
13 CogVLM 6,408
14 PaddleClas 5,588
15 mmf 5,546
16 awesome-pretrained-chinese-nlp-models 5,146
17 Chinese-CLIP 4,952
18 facenet-pytorch 4,750
19 Efficient-AI-Backbones 4,156
20 PINTO_model_zoo 3,737
21 mmpretrain 3,576
22 deepsparse 3,114
23 Pretrained-Language-Model 3,047

Sponsored
CodeRabbit: AI Code Reviews for Developers
Revolutionize your code reviews with AI. CodeRabbit offers PR summaries, code walkthroughs, 1-click suggestions, and AST-based analysis. Boost productivity and code quality across all major languages with each PR.
coderabbit.ai