gpt-engineer
promptfoo
gpt-engineer | promptfoo | |
---|---|---|
44 | 5 | |
50,698 | 328 | |
1.5% | - | |
9.9 | 10.0 | |
6 days ago | 10 months ago | |
Python | TypeScript | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
gpt-engineer
-
๐ 7 AI Tools to Improve your productivity: A Deep Dive ๐ชโจ
2๏ธโฃ GPT-engineer ๐งช
-
Ask HN: Will AI take no-code to the next level?
you should check out this:
https://github.com/AntonOsika/gpt-engineer
-
Ask HN: How can ChatGPT be effectively utilized in the work
3. https://github.com/AntonOsika/gpt-engineer
- GPT Engineer
-
[P] Looking for a new problem to solve with LLMs and AI.
Framework to Generate HQ code (I'm a mod at gpt-engineer)
-
GPT-Prompt-Engineer
BTW, GPT-Engineer is openly collecting all of your data: user prompts and other metadata. And they were even defending it until they received some strong responses from the community: https://github.com/AntonOsika/gpt-engineer/issues/415 They now explicitly ask for consent regarding user data, but can we really trust their motives?
- Ai create entire project
- GPT-Engineer โ Specify what you want it to build, the AI builds it
-
I'm creating a GPT3.5-based javascript Game Engine
You might be interested in gpt-engineer https://github.com/AntonOsika/gpt-engineer
- GPT-Engineer
promptfoo
-
Ollama v0.1.33 with Llama 3, Phi 3, and Qwen 110B
Jumping in because I'm a big believer in (1) local LLMs, and (2) evals specific to individual use cases.
[0] https://github.com/typpo/promptfoo
- Meta Llama 3
-
Launch HN: Talc AI (YC S23) โ Test Sets for AI
Congrats on the launch!
I've been interested in automatic testset generation because I find that the chore of writing tests is one of the reasons people shy away from evals. Recently landed eval testset generation for promptfoo (https://github.com/typpo/promptfoo), but it is non-RAG so more simplistic than your implementation.
Was also eyeballing this paper https://arxiv.org/abs/2401.03038, which outlines a method for generating asserts from prompt version history that may also be useful for these eval tools.
-
GPT-Prompt-Engineer
Thanks for the promptfoo mention. For anyone else who might prefer deterministic, programmatic evaluation of LLM outputs, I've been building promptfoo: https://github.com/typpo/promptfoo
Example asserts include basic string checks, regex, is-json, cosine similarity, etc.
What are some alternatives?
gpt-pilot - The first real AI developer
rebuff - LLM Prompt Injection Detector
aider - aider is AI pair programming in your terminal
ChainForge - An open-source visual programming environment for battle-testing prompts to LLMs.
MetaGPT - ๐ The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
gateway - A Blazing Fast AI Gateway. Route to 100+ LLMs with 1 fast & friendly API.
developer - the first library to let you embed a developer agent in your own app!
shap-e - Generate 3D objects conditioned on text or images
hive-metastore - Apache Hive Metastore as a Standalone server in Docker
sugarcane-ai - npm like package ecosystem for Prompts ๐ค
StanfordQuadruped
plandex - An AI coding engine for building complex, real-world software with LLMs