flashtext
outlines
flashtext | outlines | |
---|---|---|
8 | 31 | |
5,535 | 5,649 | |
- | 8.6% | |
0.0 | 9.7 | |
6 months ago | 8 days ago | |
Python | Python | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
flashtext
-
Show HN: LLMs can generate valid JSON 100% of the time
I have some other comment on this thread where I point out why I don’t think it’s superficial. Would love to get your feedback on that if you feel like spending more time on this thread.
But it’s not obscure? FlashText was a somewhat popular paper at the time (2017) with a popular repo (https://github.com/vi3k6i5/flashtext). Their paper was pretty derivative of Aho-Corasick, which they cited. If you think they genuinely fucked up, leave an issue on their repo (I’m, maybe to your surprise lol, not the author).
Anyway, I’m not a fan of the whatabboutery here. I don’t think OG’s paper is up to snuff on its lit review - do you?
-
[P] what is the most efficient way to pattern matching word-to-word?
The library flashtext basically creates these tries based on keywords you give it.
-
What is the most efficient way to find substrings in strings?
Seems like https://github.com/vi3k6i5/flashtext would be better suited here.
-
[P] Library for end-to-end neural search pipelines
I started developing this tool after using haystack. Pipelines are easier to build with cherche because of the operators. Also, cherche offers FlashText, Lunr.py retrievers that are not available in Haystack and that I needed for the project I wanted to solve. Haystack is clearly more complete but I think also more complex to use.
-
How can I speed up thousands of re.subs()?
For the text part not requiring regex, https://github.com/vi3k6i5/flashtext might help
-
My first NLP pipeline using SpaCy: detect news headlines with company acquisitions
Spacy for parsing the Headlines, remove stop words etc. might be ok but I think the problem is quite narrow so a set of fixed regex searches might work quite well. If regex is too slow, try: https://github.com/vi3k6i5/flashtext
-
What tech do I need to learn to programmatically parse ingredients from a recipe?
I would probably use something like [flashtext](https://github.com/vi3k6i5/flashtext) which should not be too hard to port to kotlin.
- Quickest way to check that 14000 strings arent in An original string.
outlines
-
Show HN: LLM-powered NPCs running on your hardware
[4] https://github.com/outlines-dev/outlines/tree/main
-
Advanced RAG with guided generation
The next step is defining how to guide generation. For this step, we'll use the Outlines library. Outlines is a library for controlling how tokens are generated. It applies logic to enforce schemas, regular expressions and/or specific output formats such as JSON.
-
Anthropic's Haiku Beats GPT-4 Turbo in Tool Use
No benchmarks, just my anecdotal experience trying to get local LLM's to respond with JSON. The method above works for my use case nearly 100% of the time. Other things I've tried (e.g. `outlines`[0]) are really slow or don't work at all. Would love to hear what others have tried!
0 - https://github.com/outlines-dev/outlines
-
Show HN: Chess-LLM, using constrained-generation to force LLMs to battle it out
As I was playing with the Outlines library (https://outlines-dev.github.io/outlines/), I discussed with my friend Maxime how funny it would be if we set up a way to pair LLMs in chess matches till one wins. The first time I tried it, it required substantial prompt engineering to get some of those LLMs to propose valid moves. Large language models can mostly stay focused and even play rather well; see https://news.ycombinator.com/item?id=37616170 for example. However small language models aren't as easy to convince.
Some of those LLMs have seen very little chess notation and so after the first few opening moves there aren't any valid tactics, let alone strategy, so they would end up either repeating the same move, or hallucinate moves that are not valid (Kxe5, but there would be a queen on e5!)
Then Outlines came along and we could force them to pick valid moves with little cost! Maxime worked super fast and got a first version of this idea as a gradio space.
I think it is pretty fun to see the (mostly terrible, but otherwise valid) chess that those LLMs play. Maybe it will even be instructive to how we can create small LLMs that can play much better than the ones on the leaderboard.
Anyway, you can check it out here:
https://huggingface.co/spaces/mlabonne/chessllm
What is interactive about it: you can pick the LLMs from available models on HuggingFace (within reason, small LLMs are preferable so that the space does not crash) or push one of your own small models to HF and have it fight with others. At the end of the game the leaderboard is updated.
Hope you find it fun!
-
Show HN: Prompts as (WASM) Programs
> The most obvious usage of this is forcing a model to output valid JSON
Isn't this something that Outlines [0], Guidance [1] and others [2] already solve much more elegantly?
0. https://github.com/outlines-dev/outlines
1. https://github.com/guidance-ai/guidance
2. https://github.com/sgl-project/sglang
- Show HN: Fructose, LLM calls as strongly typed functions
-
Unlocking the frontend – a call for standardizing component APIs pt.2
And I think “just” Markdown doesn’t quite cut it for safe guidance. For example: directly generating content for your components. But I’m really excited about tooling like outlines appearing, with a greater focus on guided generation for structured data. Because this is often what we actually need!
-
Ask HN: What are some actual use cases of AI Agents?
It's pretty easy to force a locally running model to always output valid JSON: when it gives you probabilities for the next tokens, discard all tokens that would result in invalid JSON at that point (basically reverse parsing), and then apply the usual techniques to pick the completion only from the remaining tokens. You can even validate against a JSON schema that way, so long as it is simple enough.
There are a bunch of libraries for this already, e.g.: https://github.com/outlines-dev/outlines
-
Launch HN: AgentHub (YC W24) – A no-code automation platform
https://github.com/outlines-dev/outlines/blob/7fae436345e621... squares with my experience using LLMs for anything real
sequence = generator("Alice had 4 apples and Bob ate 2. Write an expression for Alice's apples:")
-
Ollama Python and JavaScript Libraries
There are "smaller" models, for example tinyllama 1.1B (tiny seems like an exaggeration). PHI2 is 2.7B parameters. I can't name a 500M parameter model but there is probably one.
The problem is they are all still broadly trained and so they end up being Jack of all trades master of none. You'd have to fine tune them if you want them good at some narrow task and other than code completion I don't know that anyone has done that.
If you want to generate json or other structured output, there is Outlines https://github.com/outlines-dev/outlines that constrains the output to match a regex so it guarantees e.g. the model will generate a valid API call, although it could still be nonsense if the model doesn't understand, it will just match the regex. There are other similar tools around.
What are some alternatives?
KeyBERT - Minimal keyword extraction with BERT
guidance - A guidance language for controlling large language models.
rake-nltk - Python implementation of the Rapid Automatic Keyword Extraction algorithm using NLTK.
jsonformer - A Bulletproof Way to Generate Structured JSON from Language Models
magnitude - A fast, efficient universal vector embedding utility package.
json-schema-spec - The JSON Schema specification
Optimus - :truck: Agile Data Preparation Workflows made easy with Pandas, Dask, cuDF, Dask-cuDF, Vaex and PySpark
torch-grammar
yake - Single-document unsupervised keyword extraction
Constrained-Text-Genera
gensim - Topic Modelling for Humans
langroid - Harness LLMs with Multi-Agent Programming