gpt-llama.cpp
langchain
gpt-llama.cpp | langchain | |
---|---|---|
12 | 34 | |
587 | 84,427 | |
- | 3.9% | |
8.2 | 10.0 | |
11 months ago | about 6 hours ago | |
JavaScript | Python | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
gpt-llama.cpp
-
Attempt to run Llama on a remote server with chatbot-ui
hi! I really like the solution https://github.com/keldenl/gpt-llama.cpp which helps to deploy https://github.com/mckaywrigley/chatbot-ui on the local model. I am running this together with Wizard7b or 13b locally and it works fine, but when I tried to upload to a remote server I met an error.
-
Introducing Basaran: self-hosted open-source alternative to the OpenAI text completion API
sounds like youβre asking for exactly this? https://github.com/keldenl/gpt-llama.cpp
- LLaMA and AutoAPI?
-
New big update to GPTNicheFinder: better trends analysis and scoring system, cleaned up UI and verbose in the terminal for people who want to see what is going on and to verify the results
I salut you good sir. This is an amazing idea. I don't have time but it will be interesting idea to use this wrapper https://github.com/keldenl/gpt-llama.cpp which simulates GPT endpoint for local lama, so basically we can have amazing tool for completely free use. If somebody test it please let me know underneath my comment!
-
I build an AI powered writing tools, an AI co-author
I would gladly buy your product to run with a local model, like Vicuna ggml , also see https://github.com/keldenl/gpt-llama.cpp/
-
Serge... Just works
possible through fastllama in python or gpt-llama.cpp an API wrapper around llama.cpp
-
Embeddings?
https://github.com/keldenl/gpt-llama.cpp supports embeddings, and it even takes in openai type requests and returns openai compatible responses!
-
I built a completely Local AutoGPT with the help of GPT-llama running Vicuna-13B
https://github.com/keldenl/gpt-llama.cpp
- I build a completely Local and portable AutoGPT with the help of gpt-llama, running on Vicuna-13b
-
Adding Long-Term Memory to Custom LLMs: Let's Tame Vicuna Together!
There's a (kind of) working Auto-GPT solution that uses Vicuna https://github.com/keldenl/gpt-llama.cpp/blob/master/docs/Auto-GPT-setup-guide.md
langchain
-
Deploy LangServe Application to AWS
Limited by the current packaging method of Pluto, it does not yet support LangChain's Template Ecosystem. Coming soon
- Construyendo un asistente genAI de WhatsApp con Amazon Bedrock
- Show HN: SpRAG β Open-source RAG implementation for challenging real-world tasks
-
Aider: AI pair programming in your terminal
Big fan of Aider.
We are interesting in integrating Aider as a tool for Dosu https://dosu.dev/ to help it navigate and modify a codebase on issues like this https://github.com/langchain-ai/langchain/issues/8263#issuec...
-
π¦ Llama-2-GGML-CSV-Chatbot π€
Developed using Langchain and Streamlit technologies for enhanced performance.
-
Building a WhatsApp generative AI assistant with Amazon Bedrock and Python
Tip: Kenton Blacutt, an AWS Associate Cloud App Developer, collaborated with Langchain, creating the Amazon Dynamodb based memory class that allows us to store the history of a langchain agent in an Amazon DynamoDB.
-
π Top Open Source Projects of 2023 π
LangChain was first released in October 2022 as an open-source side project, a framework that makes developing AI applications more flexible. It got so popular that it was promptly turned into a startup.
-
Fuck You, Show Me the Prompt
> Furthermore, the prompt has a spelling error (Let'w) and also overly focuses on the negative about identifying errors - which makes me skeptical that this prompt has been optimized or tested.
Fixed in https://github.com/langchain-ai/langchain/commit/7c6009b76f0...
- LangChain Repository Disappeared
-
π Local & Open Source AI: a kind ollama & LlamaIndex intro
Being able to plug third party frameworks (Langchain, LlamaIndex) so you can build complex projects
What are some alternatives?
llama_index - LlamaIndex is a data framework for your LLM applications
Auto-LLM-Local - Created my own python script similar to AutoGPT where you supply a local llm model like alpaca13b (The main one I use), and the script can access the supplied tools to achieve your objective. Code fully works as far as I can tell. Takes me 5 minutes per chain on my slow laptop.
semantic-kernel - Integrate cutting-edge LLM technology quickly and easily into your apps
long_term_memory - A gradio web UI for running Large Language Models like GPT-J 6B, OPT, GALACTICA, LLaMA, and Pygmalion.
haystack - :mag: LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
langchain - β‘ Building applications with LLMs through composability β‘ [Moved to: https://github.com/langchain-ai/langchain]
griptape - Modular Python framework for AI agents and workflows with chain-of-thought reasoning, tools, and memory.
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
AGiXT - AGiXT is a dynamic AI Agent Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.
private-gpt - Interact with your documents using the power of GPT, 100% privately, no data leaks