RasaGPT
promptfoo
RasaGPT | promptfoo | |
---|---|---|
8 | 20 | |
2,172 | 2,757 | |
- | 19.2% | |
5.6 | 9.9 | |
6 months ago | 3 days ago | |
Python | TypeScript | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
RasaGPT
-
(1/2) May 2023
RasaGPT: headless LLM chatbot platform built on top of Rasa and Langchain (https://github.com/paulpierre/RasaGPT)
-
AI Weekly rundown (May 7- May 13, 2023): OpenAI's Shap·E, Multimodal GPT, Anthropic's constitutional AI and a lot more.
(Source)
- RasaGPT: First headless LLM chatbot built on top of Rasa, Langchain and FastAPI
promptfoo
- Google CodeGemma: Open Code Models Based on Gemma [pdf]
- AI Infrastructure Landscape
- Promptfoo – Testing and Evaluation for LLMs
-
Show HN: Prompt-Engineering Tool: AI-to-AI Testing for LLM
Super interesting. We've been experimenting with [promptfoo](https://github.com/promptfoo/promptfoo) at my work, and this looks very similar.
- GitHub – promptfoo/promptfoo: Test your prompts
-
I asked 60 LLMs a set of 20 questions
In case anyone's interested in running their own benchmark across many LLMs, I've built a generic harness for this at https://github.com/promptfoo/promptfoo.
I encourage people considering LLM applications to test the models on their _own data and examples_ rather than extrapolating general benchmarks.
This library supports OpenAI, Anthropic, Google, Llama and Codellama, any model on Replicate, and any model on Ollama, etc. out of the box. As an example, I wrote up an example benchmark comparing GPT model censorship with Llama models here: https://promptfoo.dev/docs/guides/llama2-uncensored-benchmar.... Hope this helps someone.
- Ask HN: Prompt Manager for Developers
- DeepEval – Unit Testing for LLMs
- Show HN: Knit – A Better LLM Playground
- Show HN: CLI for testing and evaluating LLM outputs
What are some alternatives?
pandas-ai - Chat with your database (SQL, CSV, pandas, polars, mongodb, noSQL, etc). PandasAI makes data analysis conversational using LLMs (GPT 3.5 / 4, Anthropic, VertexAI) and RAG.
shap-e - Generate 3D objects conditioned on text or images
LLMChat - A Discord chatbot that supports popular LLMs for text generation and ultra-realistic voices for voice chat.
prompt-engineering - Tips and tricks for working with Large Language Models like OpenAI's GPT-4.
langchain-chatbot - AI Chatbot for analyzing/extracting information from data in conversational format.
WizardLM - Family of instruction-following LLMs powered by Evol-Instruct: WizardLM, WizardCoder and WizardMath
codeinterpreter-api - 👾 Open source implementation of the ChatGPT Code Interpreter
chat-ui - Open source codebase powering the HuggingChat app
langchain-llm-katas - This is a an open-source project designed to help you improve your skills with AI engineering using LLMs and the langchain library
litellm - Call all LLM APIs using the OpenAI format. Use Bedrock, Azure, OpenAI, Cohere, Anthropic, Ollama, Sagemaker, HuggingFace, Replicate (100+ LLMs)
sketch - AI code-writing assistant that understands data content
ChainForge - An open-source visual programming environment for battle-testing prompts to LLMs.