nncf
nncf | gpt-llm-trainer | |
---|---|---|
2 | 4 | |
825 | 3,814 | |
7.9% | - | |
9.7 | 5.4 | |
5 days ago | about 2 months ago | |
Python | Jupyter Notebook | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
nncf
- FLaNK Stack Weekly 06 Nov 2023
-
NN mixed-precision quantization framework that supports TF?
I am aware of NNCF (https://github.com/openvinotoolkit/nncf), but it doesn't support mixed precision quantization for TF. What other frameworks support that for TF? (implement HAWQ or AutoQ algorithms for example)
gpt-llm-trainer
- FLaNK Stack Weekly 06 Nov 2023
-
Show HN: Fine-tune your own Llama 2 to replace GPT-3.5/4
Very nice, thanks!
Check out what Matt Shumer put together as well: https://github.com/mshumer/gpt-llm-trainer.
I have used his trainer for auto distillation of GPT-4 into GPT3.5 fine tunes, but plan to do the same for Llama as well.
Cheers!
-
[D] Anyone tried gpt-llm-trainer?
Hey guys, so I stumbled upon this Linkedin post, this guy was showing a jupyter notebook on google colab and was explaining step by step how to train your own model to accomplish very specific tasks, and I believe the base model he was using Llama 2 7B Fine tuning version. This is the github link: https://github.com/mshumer/gpt-llm-trainer
- GPT-LLM-Trainer
What are some alternatives?
model-optimization - A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
axolotl - Go ahead and axolotl questions
json-masker - High-performance JSON masker library in Java with no runtime dependencies
OpenPipe - Turn expensive prompts into cheap fine-tuned models
Open3D-ML - An extension of Open3D to address 3D Machine Learning tasks
Llama-2-Onnx
deepsparse - Sparsity-aware deep learning inference runtime for CPUs
trieve - All-in-one infrastructure for building search, recommendations, and RAG. Trieve combines search language models with tools for tuning ranking and relevance.
open_model_zoo - Pre-trained Deep Learning models and demos (high quality and extremely fast)
siamese-nn-semantic-text-similarity - A repository containing comprehensive Neural Networks based PyTorch implementations for the semantic text similarity task, including architectures such as: Siamese LSTM Siamese BiLSTM with Attention Siamese Transformer Siamese BERT.
vllm - A high-throughput and memory-efficient inference and serving engine for LLMs