fastLLaMa
llama.py
fastLLaMa | llama.py | |
---|---|---|
6 | 1 | |
402 | 28 | |
- | - | |
7.1 | 7.5 | |
12 months ago | about 1 year ago | |
C | C | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
fastLLaMa
-
[N] OpenLLaMA: An Open Reproduction of LLaMA
If your GPU isn't good enough, you could use llama.cpp, which runs on CPU, or one of its forks like fastLLaMa.
-
Serge... Just works
possible through fastllama in python or gpt-llama.cpp an API wrapper around llama.cpp
-
llama-cpp-python VS fastLLaMa - a user suggested alternative
2 projects | 25 Apr 2023
It is better, Lots of low level cpp optimisations that are better
-
[P] LoRA adapter switching at runtime to enable Base model to inherit multiple personalities
u/_Arsenie_Boca_ you can have a look at this discussion for more info https://github.com/PotatoSpudowski/fastLLaMa/discussions/48
-
[P] fastLLaMa, A python wrapper to run llama.cpp
Repo Link
llama.py
What are some alternatives?
llama-cpp-python - Python bindings for llama.cpp
Local-LLM-Langchain - Load local LLMs effortlessly in a Jupyter notebook for testing purposes alongside Langchain or other agents. Contains Oobagooga and KoboldAI versions of the langchain notebooks with examples.
llama - Inference code for Llama models
llamazoo - Large Model Collider - The Platform for serving LLM models [Moved to: https://github.com/gotzmann/collider]
llama.cpp - LLM inference in C/C++
llama.go - llama.go is like llama.cpp in pure Golang!
serge - A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API.
kruk - Ukrainian instruction-tuned language models and datasets
gpt-llama.cpp - A llama.cpp drop-in replacement for OpenAI's GPT endpoints, allowing GPT-powered apps to run off local llama.cpp models instead of OpenAI.
alpaca-lora - Instruct-tune LLaMA on consumer hardware