haystack
kiri
Our great sponsors
haystack | kiri | |
---|---|---|
54 | 12 | |
13,633 | 240 | |
5.8% | 0.0% | |
9.9 | 3.2 | |
2 days ago | almost 3 years ago | |
Python | Python | |
Apache License 2.0 | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
haystack
-
Release Radar • March 2024 Edition
View on GitHub
-
First 15 Open Source Advent projects
4. Haystack by Deepset | Github | tutorial
-
Generative AI Frameworks and Tools Every Developer Should Know!
Haystack can be classified as an end-to-end framework for building applications powered by various NLP technologies, including but not limited to generative AI. While it doesn't directly focus on building generative models from scratch, it provides a robust platform for:
-
Best way to programmatically extract data from a set of .pdf files?
But if you want an API that you can use to develop your own flow, Haystack from Deepset could be worth a look.
-
Which LLM framework(s) do you use in production and why?
Haystack for production. We cannot afford breaking changes in our production apps. Its stable, documentation is excellent and did I mention its' STABLE!??
- Overview: AI Assembly Architectures
-
Llama2 and Haystack on Colab
I recently conducted some experiments with Llama2 and Haystack (https://github.com/deepset-ai/haystack), the NLP/LLM framework.
The notebook can be helpful for those trying to load Llama2 on Colab.
1) Installed Transformers from the main branch (and other libraries)
- Build with LLMs for production with Haystack – has 10k stars on GitHub
- Show HN: Haystack – Production-Ready LLM Framework
-
Langchain Is Pointless
there is an alternative that is production-grade - deepset haystack https://haystack.deepset.ai/
p.s. i am contributor so there could be bias
kiri
-
[P][D] NLP question - Question Answering AI
I'm one of the authors of Backprop, a library built for transfer learning.
-
Backprop: Use and finetune models in a single line of code
I'd like to share Backprop, an open source library I've been co-authoring for the last few months.
-
[P] Backprop Model Hub: a curated list of state-of-the-art models
We've also got an open-source library that makes using + finetuning these models possible in a few lines of code.
- Show HN: Backprop – a simple library to use and finetune state-of-the-art models
- Show HN: Backprop – a library to easily finetune and use state-of-the-art models
-
[P] Backprop: a library to easily finetune and use state-of-the-art models
I'd like to share Backprop, a Python library I've been co-authoring for the last few months. Our goal is to make finetuning and using models as easy as possible, even without extensive ML experience.
-
GPT Neo: open-source GPT-3-like model with pretrained weights available
You might get some really promising results with finetuning.
If anything, you could build writing assistance that almost automates responses.
I've been co-authoring a library that lets you finetune such models in a single line of code.
https://github.com/backprop-ai/backprop
In specific the text generation finetuning example should be what you are looking for: https://github.com/backprop-ai/backprop/blob/main/examples/F...
Hope this helps, happy to chat more about it. Pretty curious about the results.
-
NLP Model for extracting specific text from raw text
Here's an example Jupyter Notebook for finetuning T5. Full disclosure, I work on this library myself -- but it could be helpful.
-
[D] Need help with document classifier and later prediction of text
I'm working on a library that hopefully makes working with some of these a bit easier -- here's an example notebook for running text classification with the BART checkpoint, if you're interested. If you need more task-specific finetuning for text classification, that's going to be rolled out in the near future.
-
Generating notes from text
I'm working on a library that includes a few different ML tasks, including summarisation. It uses a pretrained version of Google's T5 transformer model, which we host on Hugging Face with some details on how it was trained.
What are some alternatives?
langchain - 🦜🔗 Build context-aware reasoning applications
gpt-neox - An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
langchain - ⚡ Building applications with LLMs through composability ⚡ [Moved to: https://github.com/langchain-ai/langchain]
simpletransformers - Transformers for Information Retrieval, Text Classification, NER, QA, Language Modelling, Language Generation, T5, Multi-Modal, and Conversational AI
gpt-neo - An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
qagnn - [NAACL 2021] QAGNN: Question Answering using Language Models and Knowledge Graphs 🤖
BentoML - The most flexible way to serve AI/ML models in production - Build Model Inference Service, LLM APIs, Inference Graph/Pipelines, Compound AI systems, Multi-Modal, RAG as a Service, and more!
CLIP - CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
label-studio - Label Studio is a multi-type data labeling and annotation tool with standardized output format
Questgen.ai - Question generation using state-of-the-art Natural Language Processing algorithms
jina - ☁️ Build multimodal AI applications with cloud-native stack
BERTweet - BERTweet: A pre-trained language model for English Tweets (EMNLP-2020)