pruning

Open-source projects categorized as pruning

Top 20 pruning Open-Source Projects

  • deepsparse

    Sparsity-aware deep learning inference runtime for CPUs

  • Project mention: Fast Llama 2 on CPUs with Sparse Fine-Tuning and DeepSparse | news.ycombinator.com | 2023-11-23

    Interesting company. Yannic Kilcher interviewed Nir Shavit last year and they went into some depth: https://www.youtube.com/watch?v=0PAiQ1jTN5k DeepSparse is on GitHub: https://github.com/neuralmagic/deepsparse

  • Torch-Pruning

    [CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs

  • Project mention: [P] Help: I want to compress EfficientnetV2 using pruning. | /r/MachineLearning | 2023-06-28

    I also tried structured pruning from https://github.com/VainF/Torch-Pruning, as they report EfficientNetV2 to be "prunable", but got much worse results. However, the advantage of this approach is that it keeps the model dense, and you can get a real speed-up with common GPUs, while unstructured pruning sparsifies the model and you need hardware that can exploit such sparsity.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • sparseml

    Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models

  • neural-compressor

    SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime

  • aimet

    AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.

  • model-optimization

    A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.

  • mmrazor

    OpenMMLab Model Compression Toolbox and Benchmark.

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
  • autorestic

    Config driven, easy backup cli for restic.

  • Project mention: Duplicity | news.ycombinator.com | 2024-01-24

    I really like restic, and am personally happy to use it via the command line. It's very fast and efficient! However, I do wish there was better tooling / wrappers around it. For example, Pika Backup is a popular UI for Borg of which no equivalent exists for Restic. I'd love to be able to set something simple up on my partner's Macbook.

    For my own purposes, I've been using a script I found on Github[0] for a while, but it only really supports Backblaze B2 AFAIK.[1]

    I've been meaning to try autorestic[2] and resticprofile[3] as they are potentially more flexible than the script I'm currently using, and prestic[4] looks intriguing for my partner's use, but seems to have very few users. And the fact that there are so many competing tools makes it difficult to land on one.

    [0] https://github.com/erikw/restic-automatic-backup-scheduler

    [1] https://github.com/erikw/restic-automatic-backup-scheduler/i...

    [2] https://github.com/cupcakearmy/autorestic

    [3] https://github.com/creativeprojects/resticprofile

    [4] https://github.com/ducalex/prestic

  • nncf

    Neural Network Compression Framework for enhanced OpenVINOโ„ข inference

  • Project mention: FLaNK Stack Weekly 06 Nov 2023 | dev.to | 2023-11-06
  • optimum-intel

    ๐Ÿค— Optimum Intel: Accelerate inference with Intel optimization tools

  • Sparsebit

    A model compression and acceleration toolbox based on pytorch.

  • sparsify

    ML model optimization product to accelerate inference.

  • only_train_once

    OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM

  • wyng-backup

    Fast Time Machine-like backups for logical volumes & disk images

  • Project mention: My SSD suddenly died. I only lost 10 minutes of data, thanks to ZFS | news.ycombinator.com | 2023-08-22

    For people who don't want to use ZFS but are okay with LVM: wyng-backup (formerly sparsebak)

    https://github.com/tasket/wyng-backup

  • UPop

    [ICML 2023] UPop: Unified and Progressive Pruning for Compressing Vision-Language Transformers.

  • Project mention: Show HN: Compress vision-language and unimodal AI models by structured pruning | news.ycombinator.com | 2023-07-31
  • delve

    PyTorch model training and layer saturation monitor (by delve-team)

  • Generalizing-Lottery-Tickets

    This repository contains code to replicate the experiments given in NeurIPS 2019 paper "One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizers"

  • OWL

    Official Pytorch Implementation of "Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity" (by luuyin)

  • Project mention: Outlier Weighed Layerwise Sparsity: A Missing Secret Sauce for Pruning LLMs | news.ycombinator.com | 2023-10-10

    Paper abstract: Large Language Models (LLMs), renowned for their remarkable performance across diverse domains, present a challenge due to their colossal model size when it comes to practical deployment. In response to this challenge, efforts have been directed toward the application of traditional network pruning techniques to LLMs, uncovering a massive number of parameters can be pruned in one-shot without hurting performance. Building upon insights gained from pre-LLM models, particularly BERT-level language models, prevailing LLM pruning strategies have consistently adhered to the practice of uniformly pruning all layers at equivalent sparsity levels, resulting in robust performance. However, this observation stands in contrast to the prevailing trends observed in the field of vision models, where non-uniform layerwise sparsity typically yields substantially improved results. To elucidate the underlying reasons for this disparity, we conduct a comprehensive analysis of the distribution of token features within LLMs. In doing so, we discover a strong correlation with the emergence of outliers, defined as features exhibiting significantly greater magnitudes compared to their counterparts in feature dimen- sions. Inspired by this finding, we introduce a novel LLM pruning methodology that incorporates a tailored set of non-uniform layerwise sparsity ratios specif- ically designed for LLM pruning, termed as Outlier Weighed Layerwise sparsity (OWL). The sparsity ratio of OWL is directly proportional to the outlier ratio observed within each layer, facilitating a more effective alignment between layer- wise weight sparsity and outlier ratios. Our empirical evaluation, conducted across the LLaMA-V1 family and OPT, spanning various benchmarks, demonstrates the distinct advantages offered by OWL over previous methods. For instance, our approach exhibits a remarkable performance gain, surpassing the state-of-the-art Wanda and SparseGPT by 61.22 and 6.80 perplexity at a high sparsity level of 70%, respectively. Codes are available at https://github.com/luuyin/OWL.

  • thesis

    Master's thesis, Uni Passau (by harshildarji)

  • Pi-SqueezeDet

    Pruning SqueezeDet for inference on Raspberry PI

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
NOTE: The open source projects on this list are ordered by number of github stars. The number of mentions indicates repo mentiontions in the last 12 Months or since we started tracking (Dec 2020).

pruning related posts

  • [P] Help: I want to compress EfficientnetV2 using pruning.

    1 project | /r/MachineLearning | 28 Jun 2023
  • [R] ๐Ÿค–๐ŸŒŸ Unlock the Power of Personal AI: Introducing ChatLLaMA, Your Custom Personal Assistant! ๐Ÿš€๐Ÿ’ฌ

    9 projects | /r/MachineLearning | 19 Mar 2023
  • What backup plan do I use?

    1 project | /r/linuxquestions | 10 Mar 2023
  • Intel Textual Inversion Training on Hugging Face

    1 project | /r/StableDiffusion | 22 Dec 2022
  • [R] New sparsity research (oBERT) enabled 175X increase in CPU performance for MLPerf submission

    2 projects | /r/MachineLearning | 10 Sep 2022
  • [R] BERT-Large: Prune Once for DistilBERT Inference Performance

    2 projects | /r/MachineLearning | 16 Jul 2022
  • [R] How well do sparse ImageNet models transfer? Prune once and deploy anywhere for inference performance speedups! (arxiv link in comments)

    2 projects | /r/MachineLearning | 26 Jun 2022
  • A note from our sponsor - InfluxDB
    www.influxdata.com | 2 May 2024
    Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more โ†’

Index

What are some of the best open-source pruning projects? This list will help you:

Project Stars
1 deepsparse 2,873
2 Torch-Pruning 2,307
3 sparseml 1,976
4 neural-compressor 1,964
5 aimet 1,908
6 model-optimization 1,470
7 mmrazor 1,365
8 autorestic 1,099
9 nncf 819
10 optimum-intel 324
11 Sparsebit 320
12 sparsify 315
13 only_train_once 261
14 wyng-backup 237
15 UPop 83
16 delve 77
17 Generalizing-Lottery-Tickets 50
18 OWL 39
19 thesis 15
20 Pi-SqueezeDet 2

Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com