MemGPT
llama.cpp
MemGPT | llama.cpp | |
---|---|---|
15 | 777 | |
9,252 | 57,984 | |
- | - | |
9.9 | 10.0 | |
7 days ago | about 1 hour ago | |
Python | C++ | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
MemGPT
-
AI21 Labs Unveils Jamba: The First Production-Grade Mamba-Based AI Model
On a side note: working over longer contexts also reminds me of MemGPT(https://github.com/cpacker/MemGPT)
- FLaNK Weekly 18 Dec 2023
-
At this point we don’t necessarily need higher context windows. We need better truncation.
The MemGPT project is taking on this challenge.
-
Putting Together the Pieces of Transformative AI
Long Term Memory - Voyager, MemGPTand LongMem
-
[R] MemGPT: Towards LLMs as Operating Systems - UC Berkeley 2023 - Is able to create unbounded/infinite LLM context!
Blog: https://memgpt.ai/
-
MemGPT: Towards LLMs as Operating Systems - UC Berkeley 2023 - Is able to create unbounded/infinite LLM context!
Github: https://github.com/cpacker/MemGPT
-
MemGPT – LLMs with self-editing memory for unbounded context
Hey all, MemGPT authors here! Happy to answer any questions about the implementation.
If you want to try it out yourself, we have a Discord bot up-and-running on the MemGPT server (https://discord.gg/9GEQrxmVyE) where you can see the memory editing in action - as you chat you'll see MemGPT update its profile about you (and itself).
Everything's open source, so can also try running MemGPT locally using the code here: https://github.com/cpacker/MemGPT. In the repo we also have a document-focused example where you can chat with MemGPT about the LlamaIndex API docs.
- MemGPT — a combination of OS and GPT
llama.cpp
-
IBM Granite: A Family of Open Foundation Models for Code Intelligence
if you can compile stuff, then looking at llama.cpp (what ollama uses) is also interesting: https://github.com/ggerganov/llama.cpp
the server is here: https://github.com/ggerganov/llama.cpp/tree/master/examples/...
And you can search for any GGUF on huggingface
-
Ask HN: Affordable hardware for running local large language models?
Yes, Metal seems to allow a maximum of 1/2 of the RAM for one process, and 3/4 of the RAM allocated to the GPU overall. There’s a kernel hack to fix it, but that comes with the usual system integrity caveats. https://github.com/ggerganov/llama.cpp/discussions/2182
- Xmake: A modern C/C++ build tool
-
Better and Faster Large Language Models via Multi-Token Prediction
For anyone interested in exploring this, llama.cpp has an example implementation here:
https://github.com/ggerganov/llama.cpp/tree/master/examples/...
- Llama.cpp Bfloat16 Support
-
Fine-tune your first large language model (LLM) with LoRA, llama.cpp, and KitOps in 5 easy steps
Getting started with LLMs can be intimidating. In this tutorial we will show you how to fine-tune a large language model using LoRA, facilitated by tools like llama.cpp and KitOps.
- GGML Flash Attention support merged into llama.cpp
-
Phi-3 Weights Released
well https://github.com/ggerganov/llama.cpp/issues/6849
- Lossless Acceleration of LLM via Adaptive N-Gram Parallel Decoding
- Llama.cpp Working on Support for Llama3
What are some alternatives?
tidybot - TidyBot: Personalized Robot Assistance with Large Language Models
ollama - Get up and running with Llama 3, Mistral, Gemma, and other large language models.
LongMem - Official implementation of our NeurIPS 2023 paper "Augmenting Language Models with Long-Term Memory".
gpt4all - gpt4all: run open-source LLMs anywhere
mlx-examples - Examples in the MLX framework
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
Efficient-LLMs-Survey - Efficient Large Language Models: A Survey
GPTQ-for-LLaMa - 4 bits quantization of LLaMA using GPTQ
FLiPStackWeekly - FLaNK AI Weekly covering Apache NiFi, Apache Flink, Apache Kafka, Apache Spark, Apache Iceberg, Apache Ozone, Apache Pulsar, and more...
ggml - Tensor library for machine learning
LLMCompiler - [ICML 2024] LLMCompiler: An LLM Compiler for Parallel Function Calling
alpaca.cpp - Locally run an Instruction-Tuned Chat-Style LLM