dbrx
OLMo
dbrx | OLMo | |
---|---|---|
4 | 3 | |
2,407 | 4,007 | |
94.6% | 10.0% | |
5.9 | 9.9 | |
9 days ago | 4 days ago | |
Python | Python | |
GNU General Public License v3.0 or later | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
dbrx
-
Hello OLMo: A Open LLM
One thing I wanted to add and call attention to is the importance of licensing in open models. This is often overlooked when we blindly accept the vague branding of models as “open”, but I am noticing that many open weight models are actually using encumbered proprietary licenses rather than standard open source licenses that are OSI approved (https://opensource.org/licenses). As an example, Databricks’s DBRX model has a proprietary license that forces adherence to their highly restrictive Acceptable Use Policy by referencing a live website hosting their AUP (https://github.com/databricks/dbrx/blob/main/LICENSE), which means as they change their AUP, you may be further restricted in the future. Meta’s Llama is similar (https://github.com/meta-llama/llama/blob/main/LICENSE ). I’m not sure who can depend on these models given this flaw.
-
DBRX: A New Open LLM
Sorry, I forgot to link the repository and missed the edit window by the time I realized.
[1] https://github.com/databricks/dbrx
OLMo
-
Meta Llama 3
Olmo from AI2. They released the model weights plus training data and training code.
link: https://allenai.org/olmo
-
Hello OLMo: A Open LLM
It looks like the weights [0] and code [1] are Apache licensed, but the training data [2] is using the license that OP is quoting from.
[0] https://huggingface.co/allenai/OLMo-7B
[1] https://github.com/allenai/OLMo
[2] https://huggingface.co/datasets/allenai/dolma
- FLaNK Stack Weekly 12 February 2024
What are some alternatives?
llama - Inference code for Llama models
electric - Local-first sync layer for web and mobile apps. Build reactive, realtime, local-first apps directly on Postgres.
xTuring - Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6
RustPython - A Python Interpreter written in Rust
mixtral-offloading - Run Mixtral-8x7B models in Colab or consumer desktops
Real-Time-Voice-Cloning - Clone a voice in 5 seconds to generate arbitrary speech in real-time
makeMoE - From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)
flink-cdc - Flink CDC is a streaming data integration tool
rich-cli - Rich-cli is a command line toolbox for fancy output in the terminal
tortoise-tts - A multi-voice TTS system trained with an emphasis on quality
FLiPStackWeekly - FLaNK AI Weekly covering Apache NiFi, Apache Flink, Apache Kafka, Apache Spark, Apache Iceberg, Apache Ozone, Apache Pulsar, and more...
openai-java - OpenAI Api Client in Java