nncf
siamese-nn-semantic-text-similarity
nncf | siamese-nn-semantic-text-similarity | |
---|---|---|
2 | 2 | |
825 | 45 | |
7.9% | - | |
9.7 | 0.0 | |
4 days ago | about 2 years ago | |
Python | Python | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
nncf
- FLaNK Stack Weekly 06 Nov 2023
-
NN mixed-precision quantization framework that supports TF?
I am aware of NNCF (https://github.com/openvinotoolkit/nncf), but it doesn't support mixed precision quantization for TF. What other frameworks support that for TF? (implement HAWQ or AutoQ algorithms for example)
siamese-nn-semantic-text-similarity
What are some alternatives?
gpt-llm-trainer
adapters - A Unified Library for Parameter-Efficient and Modular Transfer Learning
model-optimization - A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
TrAVis - TrAVis: Visualise BERT attention in your browser
json-masker - High-performance JSON masker library in Java with no runtime dependencies
haystack - :mag: LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
Open3D-ML - An extension of Open3D to address 3D Machine Learning tasks
llmware - Providing enterprise-grade LLM-based development framework, tools, and fine-tuned models.
deepsparse - Sparsity-aware deep learning inference runtime for CPUs
RWKV-LM - RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
open_model_zoo - Pre-trained Deep Learning models and demos (high quality and extremely fast)
bertviz - BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)