gemma
gemma.cpp
gemma | gemma.cpp | |
---|---|---|
3 | 8 | |
2,080 | 5,560 | |
10.1% | 8.4% | |
5.6 | 9.3 | |
4 days ago | 3 days ago | |
Jupyter Notebook | C++ | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
gemma
-
What are LLMs? An intro into AI, models, tokens, parameters, weights, quantization and more
Medium models: Roughly between 1B to 10B parameters. This is where Mistral 7B, Phi-3, Gemma from Google DeepMind, and wizardlm2 sit. Fun fact: GPT 2 was a medium sized model, much smaller than its latest versions.
- Gemma – a family of lightweight, state-of-the art open models from Google
-
Gemma: New Open Models
We've documented the architecture (including key differences) in our technical report here (https://goo.gle/GemmaReport), and you can see the architecture implementation in our Git Repo (https://github.com/google-deepmind/gemma).
gemma.cpp
-
LLaMA Now Goes Faster on CPUs
For C++, also check out our https://github.com/google/gemma.cpp/blob/main/gemma.cc, which has direct calls to MatVec.
- FLaNK Stack 26 February 2024
-
Gemma.cpp: lightweight, standalone C++ inference engine for Gemma models
Looks like they're working on it: https://github.com/google/gemma.cpp/issues/16
- Source code of Google Gemma model in C++
-
Gemma: New Open Models
They have implemented the model also on their own C++ inference engine: https://github.com/google/gemma.cpp
What are some alternatives?
gemma_pytorch - The official PyTorch implementation of Google's Gemma models
llamafile - Distribute and run LLMs with a single file.
ollama - Get up and running with Llama 3, Mistral, Gemma, and other large language models.
mud-pi - A simple MUD server in Python, for teaching purposes, which could be run on a Raspberry Pi
xpk - xpk (Accelerated Processing Kit, pronounced x-p-k,) is a software tool to help Cloud developers to orchestrate training jobs on accelerators such as TPUs and GPUs on GKE.
htmx - </> htmx - high power tools for HTML
plantuml - Generate diagrams from textual description
prql - PRQL is a modern language for transforming data — a simple, powerful, pipelined SQL replacement
lnav - Log file navigator
ibis - the portable Python dataframe library
lotion - An open-source Notion UI built with Vue 3