gpt-llama.cpp
long_term_memory
gpt-llama.cpp | long_term_memory | |
---|---|---|
12 | 12 | |
587 | 300 | |
- | - | |
8.2 | 9.3 | |
11 months ago | 9 months ago | |
JavaScript | Python | |
MIT License | GNU Affero General Public License v3.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
gpt-llama.cpp
-
Attempt to run Llama on a remote server with chatbot-ui
hi! I really like the solution https://github.com/keldenl/gpt-llama.cpp which helps to deploy https://github.com/mckaywrigley/chatbot-ui on the local model. I am running this together with Wizard7b or 13b locally and it works fine, but when I tried to upload to a remote server I met an error.
-
Introducing Basaran: self-hosted open-source alternative to the OpenAI text completion API
sounds like you’re asking for exactly this? https://github.com/keldenl/gpt-llama.cpp
- LLaMA and AutoAPI?
-
New big update to GPTNicheFinder: better trends analysis and scoring system, cleaned up UI and verbose in the terminal for people who want to see what is going on and to verify the results
I salut you good sir. This is an amazing idea. I don't have time but it will be interesting idea to use this wrapper https://github.com/keldenl/gpt-llama.cpp which simulates GPT endpoint for local lama, so basically we can have amazing tool for completely free use. If somebody test it please let me know underneath my comment!
-
I build an AI powered writing tools, an AI co-author
I would gladly buy your product to run with a local model, like Vicuna ggml , also see https://github.com/keldenl/gpt-llama.cpp/
-
Serge... Just works
possible through fastllama in python or gpt-llama.cpp an API wrapper around llama.cpp
-
Embeddings?
https://github.com/keldenl/gpt-llama.cpp supports embeddings, and it even takes in openai type requests and returns openai compatible responses!
-
I built a completely Local AutoGPT with the help of GPT-llama running Vicuna-13B
https://github.com/keldenl/gpt-llama.cpp
- I build a completely Local and portable AutoGPT with the help of gpt-llama, running on Vicuna-13b
-
Adding Long-Term Memory to Custom LLMs: Let's Tame Vicuna Together!
There's a (kind of) working Auto-GPT solution that uses Vicuna https://github.com/keldenl/gpt-llama.cpp/blob/master/docs/Auto-GPT-setup-guide.md
long_term_memory
-
Looking for the long-term memory extension.
what you're probably thinking of is this: https://github.com/wawawario2/long_term_memory
-
Instruct Models not remembering previous responses?
I do believe there are some proposed solutions in order to retain memories such as LTM or Long Term Memory. I think there are some extensions for ooba's webui that implements that. see: https://github.com/wawawario2/long_term_memory
- Long-term memory (LTM) extension for oobabooga's Text Generation Web UI
-
I just made an easy GUI for changing start Parameters
If I understand correctly, one of the extensions you have enabled as an option for the gui to be able to select is the long_term_memory extension. That extension enables people to store in their conversations with the model into a long-term memory upon the next chat. I didn't know if you had a method that interfaced with that so that you could enable the long-term memory read or not. Here is where it explains how long_term_memory works: https://github.com/wawawario2/long_term_memory
-
Is there a way to feed in documents similar to the Llama Index?
I saw [this](https://github.com/wawawario2/long_term_memory) as a possible option, but now that I know about SuperBooga I am now not sure which would be better for my purposes.
- Suggestions for long term memories
- CarterAI's StableVicuna 13B with RHLF training. Now available quantised in GGML and GPTQ.
-
Working on long term memories for the AI
I thought about Langchain, but it already does the functionality that we have built in which is.. just feed the context back to the model. What I am aiming to do.. is inject relevant memories with the combined text and mix it with the prompt. So the responses will be very fine tuned to the prompt itself based on the context/and memory! so when you say prompt something and it has a memory.. it will take the memory + prompt and the context of the conversation and give a very fine tuned response. https://github.com/wawawario2/long_term_memory
-
Adding Long-Term Memory to Custom LLMs: Let's Tame Vicuna Together!
Is this going to function at all similarly to https://github.com/wawawario2/long_term_memory ?
-
Advanced character documentation?
There is the long term memory extension that might assist with this. I found out about it though and haven't gotten it to load successfully yet.. https://github.com/wawawario2/long_term_memory
What are some alternatives?
llama_index - LlamaIndex is a data framework for your LLM applications
StartUI-oobabooga-webui - WebUI StartGUI is a Python graphical user interface (GUI) written with PyQT5, that allows users to configure settings and start the oobabooga web user interface (WebUI). It provides a convenient way to adjust various parameters and launch the WebUI with the desired settings.
Auto-LLM-Local - Created my own python script similar to AutoGPT where you supply a local llm model like alpaca13b (The main one I use), and the script can access the supplied tools to achieve your objective. Code fully works as far as I can tell. Takes me 5 minutes per chain on my slow laptop.
SillyTavern-Extras - Extensions API for SillyTavern.
langchain - ⚡ Building applications with LLMs through composability ⚡ [Moved to: https://github.com/langchain-ai/langchain]
semantic-kernel - Integrate cutting-edge LLM technology quickly and easily into your apps
text-generation-webui-extensions
langchain - 🦜🔗 Build context-aware reasoning applications
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
AGiXT - AGiXT is a dynamic AI Agent Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.
annoy_ltm - annoy long term memory experiment for oobabooga/text-generation-webui