st-moe-pytorch
Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch (by lucidrains)
mixtral-offloading
Run Mixtral-8x7B models in Colab or consumer desktops (by dvmazur)
st-moe-pytorch | mixtral-offloading | |
---|---|---|
1 | 3 | |
231 | 2,242 | |
- | - | |
7.8 | 8.6 | |
3 months ago | about 1 month ago | |
Python | Python | |
MIT License | MIT License |
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
st-moe-pytorch
Posts with mentions or reviews of st-moe-pytorch.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2023-12-05.
mixtral-offloading
Posts with mentions or reviews of mixtral-offloading.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2024-03-27.
-
DBRX: A New Open LLM
Waiting for Mixed Quantization with MQQ and MoE Offloading [1]. With that I was able to run Mistral 8x7B on my 10 GB VRAM rtx3080... This should work for DBRX and should shave off a ton of VRAM requirement.
1. https://github.com/dvmazur/mixtral-offloading?tab=readme-ov-...
- Mixtral in Colab
- Run Mixtral-8x7B models in Colab or consumer desktops
What are some alternatives?
When comparing st-moe-pytorch and mixtral-offloading you can also consider the following projects:
OpenMoE - A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
lightning-mlflow-hf - Use QLoRA to tune LLM in PyTorch-Lightning w/ Huggingface + MLflow
spaCy - 💫 Industrial-strength Natural Language Processing (NLP) in Python
dbrx - Code examples and resources for DBRX, a large language model developed by Databricks