lmql
marvin
lmql | marvin | |
---|---|---|
30 | 16 | |
3,342 | 4,773 | |
2.9% | 3.6% | |
9.5 | 9.9 | |
6 days ago | 3 days ago | |
Python | Python | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
lmql
- Show HN: Fructose, LLM calls as strongly typed functions
-
Prompting LLMs to constrain output
have been experimenting with guidance and lmql. a bit too early to give any well formed opinions but really do like the idea of constraining llm output.
-
[D] Prompt Engineering Seems Like Guesswork - How To Evaluate LLM Application Properly?
the only time i've ever felt like it was anything other than guesswork was using LMQL . not coincidentally, LMQL works with LLMs as autocomplete engines rather than q&a ones.
-
Guidance for selecting a function-calling library?
lqml
-
Show HN: Magentic – Use LLMs as simple Python functions
This is also similar in spirit to LMQL
https://github.com/eth-sri/lmql
- Show HN: LLMs can generate valid JSON 100% of the time
- LangChain Agent Simulation – Multi-Player Dungeons and Dragons
-
The Problem with LangChain
LLM calls are just function calls, so most functional composition is already afforded by any general-purpose language out there. If you need fancy stuff, use something like Python‘s functools.
Working on https://github.com/eth-sri/lmql (shameless plug, sorry), we have always found that compositional abstractions on top of LMQL are mostly there already, once you internalize prompts being functions.
- Is there a UI that can limit LLM tokens to a preset list?
-
Local LLMs: After Novelty Wanes
LMQL is another.
marvin
-
Show HN: Marvin 2.0 – a lightweight, multi-modal AI toolkit
Hey HN! We just released Marvin 2.0.
Marvin is an AI toolkit for developers who want to use LLMs with traditional software. We still see significant challenges integrating LLMs because of how difficult it is to get them to reliably accept and return structured data. Marvin consists of independent, functional tools that address this problem in a variety of ways.
Marvin has always been focused on using LLMs to work with native Python datatypes and Pydantic models. In 2.0 we've expanded this significantly with dedicated APIs for the most common use cases we've seen over the last year: classification, entity extraction, transforming data to types, and generating synthetic data. Marvin 2.0 is also fully multi-modal and supports images as inputs for classification, extraction, and transformation tasks (as well as simple image and speech generation). We've also introduces a Pythonic interface to OpenAI's assistants API, which now powers all of Marvin's interactive components.
We've tried to make an LLM framework that "sparks joy" and captures that same feeling you had the first time you saw an LLM in action. Try it out and let us know what you think!
(Repo: https://github.com/PrefectHQ/marvin)
-
Show HN: Magentic – Use LLMs as simple Python functions
Seems a lot like https://github.com/PrefectHQ/marvin?
The prompting you do seems an awfully like:
https://www.askmarvin.ai/prompting/prompt_function/
-
Amazon CodeWhisperer, Free for Individual Use, Is Now Generally Available
You can try the decorator ai_fn in marvin https://github.com/PrefectHQ/marvin
-
4-Apr-2023
Marvin: a batteries-included library for building AI-powered software. Marvin's job is to integrate AI directly into your codebase by making it look and feel like any other function (https://github.com/PrefectHQ/marvin)
-
Magic - AI functions for Typescript
Sure! I was inspired by this Python library: https://github.com/PrefectHQ/marvin
-
Show HN: A ChatGPT TUI with custom bots
I see Langchain has support for Azure chat models, and Marvin is built on Langchain so it may not be so difficult! Tracking issue here: https://github.com/PrefectHQ/marvin/issues/189
- FLaNK Stack Weekly 3 April 2023
- Meet Marvin: A batteries-included library for building AI-powered software, aka “woah-code”
-
Show HN: Marvin – build AI functions that use an LLM as a runtime
We have a related issue open (https://github.com/PrefectHQ/marvin/issues/64) but haven't designed anything yet.
What are some alternatives?
guidance - A guidance language for controlling large language models.
bpytop - Linux/OSX/FreeBSD resource monitor
guidance - A guidance language for controlling large language models. [Moved to: https://github.com/guidance-ai/guidance]
LocalAI - :robot: The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
simpleaichat - Python package for easily interfacing with chat apps, with robust features and minimal code complexity.
magentic - Seamlessly integrate LLMs as Python functions
NeMo-Guardrails - NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.
aide - LLM shell and document interogator
guardrails - Adding guardrails to large language models.
use_gpt_as_programming_lang - use gpt as programming language
basaran - Basaran is an open-source alternative to the OpenAI text completion API. It provides a compatible streaming API for your Hugging Face Transformers-based text generation models.
the-algorithm