NeMo-Guardrails
simpleAI
NeMo-Guardrails | simpleAI | |
---|---|---|
13 | 11 | |
3,373 | 319 | |
4.7% | - | |
9.9 | 7.3 | |
5 days ago | 11 months ago | |
Python | Python | |
GNU General Public License v3.0 or later | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
NeMo-Guardrails
- NeMO Guardrails from Nvidia
-
Run and create custom ChatGPT-like bots with OpenChat
- https://github.com/NVIDIA/NeMo-Guardrails/
- LangChain: The Missing Manual
-
The Dual LLM pattern for building AI assistants that can resist prompt injection
Here's "jailbreak detection", in the NeMo-Guardrails project from Nvidia:
https://github.com/NVIDIA/NeMo-Guardrails/blob/327da8a42d5f8...
I.e. they ask the llm if the prompt will break the llm. (I believe that more data /some evaluation on how well this performs is intended to be released. Probably fair to call this stuff "not battle tested".)
-
How To Setup a Model With Guardrails?
I have been playing around with some models locally and creating a discord bot as a fun side project, and I wanted to setup some guardrails on inputs / outputs of the bot to make sure that it isn't violating any ethical boundaries. I was going to use Nvidia's Nemo guardrails, but they only support openai currently. Are there any other good ways to control inputs?
-
RasaGPT: First headless LLM chatbot built on top of Rasa, Langchain and FastAPI
Thanks, I hadn't seen those. I did find https://github.com/NVIDIA/NeMo-Guardrails earlier but haven't looked into it yet.
I'm not sure it solves the problem of restricting the information it uses though. For example, as a proof of concept for a customer, I tried providing information from a vector database as context, but GPT would still answer questions that were not provided in that context. It would base its answers on information that was already crawled from the customer website and in the model. That is concerning because the website might get updated but you can't update the model yourself (among other reasons).
- How do we prevent prompt injection in a GPT API app?
- Nvidia NeMo Guardrails – open-source guardrails to conversational systems
-
Should LangChain be used in Prod?
you can use guard rails with langchain - https://github.com/NVIDIA/NeMo-Guardrails
simpleAI
-
[P] I got fed up with LangChain, so I made a simple open-source alternative for building Python AI apps as easy and intuitive as possible.
Not related to my own project SimpleAI despite the name, but looks like we can easily make the two work together, to keep it « simple ». Nice work!
-
Run and create custom ChatGPT-like bots with OpenChat
Using this as an opportunity to mention my own related project, perhaps it can end up on your nice list one day. :)
https://github.com/lhenault/SimpleAI
- [D] OpenAI API vs. Open Source Self hosted for AI Startups
-
StableLM released
You could have a look at a project I’ve been working on, SimpleAI, doing exactly this by replicating the OpenAI endpoints (you can then use their JS client for integration). Adding StableLM should be straightforward, I plan to add it to the examples in the upcoming days once I have a bit of time.
-
[P] LoopGPT: A Modular Auto-GPT Framework
I’ve built SimpleAI with exactly these kinds of use cases in mind. That should allow supporting any model with minimal / no change to your project. Good job and good luck with LoopGPT, that looks nice!
-
Using the API in Node
You could give this a shot: https://github.com/lhenault/simpleAI
-
[D] Would a Tesla M40 provide cheap inference acceleration for self-hosted LLMs?
I don't know if this applies to your use case but this would probably work if you are looking for an llm to help with programming. Haven't really played around with it but this may work for general llm tasks, it doesn't have a web UI though.
-
Alpaca, LLaMa, Vicuna [D]
As per llama.cpp specifically, you can indeed add any model, it's just a matter of doing a bit of glue code and declaring it in your models.toml config. It's quite straightforward thanks to some provided tools for Python (see here for instance). For any other language it's a matter of integrating it through the gRPC interface (which shouldn't be too hard for Llama.cpp if you're comfortable in C++). I'm planning to also add support for REST for model in the backend at some point too.
-
[D] Is there currently anything comparable to the OpenAI API?
Shameless plug but I’ve been recently working on SimpleAI, a project replicating the main endpoints from OpenAI API, allowing you to seamlessly switch from their API to your own one, as it’s compatible with OpenAI client.
-
[P] SimpleAI : A self-hosted alternative to OpenAI API
I wanted to share with you SimpleAI, a self-hosted alternative to OpenAI API.
What are some alternatives?
guidance - A guidance language for controlling large language models. [Moved to: https://github.com/guidance-ai/guidance]
OpenChat - LLMs custom-chatbots console ⚡
langchainrb - Build LLM-powered applications in Ruby
dalai - The simplest way to run LLaMA on your local machine
guidance - A guidance language for controlling large language models.
AlpacaDataCleaned - Alpaca dataset from Stanford, cleaned and curated
lmql - A language for constraint-guided and efficient LLM programming.
gptcli - ChatGPT in command line with OpenAI API (gpt-3.5-turbo/gpt-4/gpt-4-32k)
basaran - Basaran is an open-source alternative to the OpenAI text completion API. It provides a compatible streaming API for your Hugging Face Transformers-based text generation models.
StableLM - StableLM: Stability AI Language Models
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
loopgpt - Modular Auto-GPT Framework