vault-ai
marqo
vault-ai | marqo | |
---|---|---|
80 | 118 | |
3,355 | 4,890 | |
0.8% | 1.1% | |
5.9 | 9.7 | |
5 months ago | 3 days ago | |
JavaScript | Python | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
vault-ai
- I built an open source website that lets you upload large files, such as in-depth novels/ebooks or academic papers, and ask GPT4 questions based on your specific knowledge base. So far, I've tested it with long books like the Odyssey and random research PDFs, and I'm shocked at how incisive it is.
-
Any better alternatives to fine-tuning GPT-3 yet to create a custom chatbot persona based on provided knowledge for others to use?
There's this GitHub repo for Pinecone Vector with custom knowledge base: VaultAI. But I'm sure the costs would be exorbitant at scale. Basically trains it on specific files, but the API is expensive as expected. Edit: I didn't read and thought you were talking about training your own, sorry. But I'll leave the second paragraph up anyways lol. Someone mentioned LLaMA and another Falcon, the latter of which I hadn't heard of but which looks good too.
-
I built an open source website that lets you upload large files such as academic PDFs or books and ask ChatGPT questions based on your custom knowledge base. So far, I've tried it with long ebooks like Plato's Republic, old letters, and random academic PDFs, and it works shockingly well.
Check out the instructions readme here! You may need a little bit of command line know-how but chatgpt can help guide you if you provide it the contents of the readme
- Are there any good free GPT-powered AI summarizer for very long text?
-
I built an open source website that lets you upload large files, such as long ebooks or academic papers, and ask ChatGPT questions about your specific knowledge base. So far, I've tested it with long e-books like the Odyssey and random research PDFs, and I'm shocked at how incisive it is
Yes, this use-case is a perfect fit actually – This deals very well with any type of manual with lots of human readable text (as opposed to charts or code). It is also better at answering more specific questions, so the example you gave regarding diagnosing engine issues is a really good match for what this is capable of. If you want to try it out you can check out the deployed version of the code here: https://vault.pash.city
-
Any help condensing academic journal articles using ChatGPT?
Have you tried Vault AI? Saw it pop up on a couple of other Reddits!
- OP Vault ChatGPT: Give ChatGPT long-term memory using the OP Stack (OpenAI + Pinecone Vector Database). Upload your own custom knowledge base files (PDF, txt, epub, etc) using a simple React frontend.
-
April 2023
OP Vault ChatGPT: Give ChatGPT long-term memory using the OP Stack (https://github.com/pashpashpash/vault-ai)
- Using ChatGPT to read multiple PDFs and create writing using them as sources
marqo
-
Why You Shouldn’t Invest In Vector Databases?
In cases where a company possesses a strong technological foundation and faces a substantial workload demanding advanced vector search capabilities, its ideal solution lies in adopting a specialized vector database. Prominent options in this domain include Chroma (having raised $20 million), Zilliz (having raised $113 million), Pinecone (having raised $138 million), Qdrant (having raised $9.8 million), Weaviate (having raised $67.7 million), LanceDB (YC W22), Vespa, Marqo, and others. Many of these players have secured significant funding in recent years and are well-positioned to capture notable market share. These vector databases offer efficient storage, indexing, and similarity search functionalities for vectors. They often incorporate specific optimizations tailored for vector data, such as similarity search based on inverted indexes and efficient vector computations. As a result, they cater to the requirements of companies operating in areas like recommendation systems, image search, and natural language processing.
-
Ask HN: What's your serverless stack for AI/LLM apps in production?
I have a hosted code-first agent builder platform in production, so I respond these question a lot from our customers.
1. Probably the best is fly.io IMHO. It has a nice balance between running ephemeral containers that can support long running tasks, and quickly booting up to respond to a tool call. [1]
2. If your task is truly long running, (I'm thinking several minutes), probably wise to put trigger [2] or temporal [3] under it.
3. A mix of prompt caching, context shedding, progressive context enrichment [4].
4. I'm building a platform that can be self-hosted to do a few of the above, so I can't speak to this. But most of my customers do not.
5. To start with, a simple postgres table and pgvector is all you need. But I've recently been delighted with the DX of Upstash vector [5]. They handle the embeddings for you and give you a text-in, text-out experience. If you want more control, and savings on a higher scale, have heard good things about marqo.ai [6].
Happy to talk more about this at length. (E-mail in the profile)
[1] https://fly.io/docs/reference/architecture/
[2] trigger.dev
[3] temporal.io
[4] https://www.inferable.ai/blog/posts/llm-progressive-context-...
[5] https://upstash.com/docs/vector/overall/getstarted
[6] https://www.marqo.ai/
- Pinecone integrates AI inferencing with vector database
- AI Search That Understands the Way Your Customer's Think
-
Are we at peak vector database?
We (Marqo) are doing a lot on 1 and 2. There is a huge amount to be done on the ML side of vector search and we are investing heavily in it. I think it has not quite sunk in that vector search systems are ML systems and everything that comes with that. I would love to chat about 1 and 2 so feel free to email me (email is in my profile). What we have done so far is here -> https://github.com/marqo-ai/marqo
-
Qdrant, the Vector Search Database, raised $28M in a Series A round
Marqo.ai (https://github.com/marqo-ai/marqo) is doing some interesting stuff and is oss. We handle embedding generation as well as retrieval (full disclosure, I work for Marqo.ai)
-
Ask HN: Is there any good semantic search GUI for images or documents?
Take a look here https://github.com/marqo-ai/local-image-search-demo. It is based on https://github.com/marqo-ai/marqo. We do a lot of image search applications. Feel free to reach out if you have other questions (email in profile).
-
90x Faster Than Pgvector – Lantern's HNSW Index Creation Time
That sounds much longer than it should. I am not sure on your exact use-case but I would encourage you to check out Marqo (https://github.com/marqo-ai/marqo - disclaimer, I am a co-founder). All inference and orchestration is included (no api calls) and many open-source or fine-tuned models can be used.
-
Embeddings: What they are and why they matter
Try this https://github.com/marqo-ai/marqo which handles all the chunking for you (and is configurable). Also handles chunking of images in an analogous way. This enables highlighting in longer docs and also for images in a single retrieval step.
-
Choosing vector database: a side-by-side comparison
As others have correctly pointed out, to make a vector search or recommendation application requires a lot more than similarity alone. We have seen the HNSW become commoditised and the real value lies elsewhere. Just because a database has vector functionality doesn’t mean it will actually service anything beyond “hello world” type semantic search applications. IMHO these have questionable value, much like the simple Q and A RAG applications that have proliferated. The elephant in the room with these systems is that if you are relying on machine learning models to produce the vectors you are going to need to invest heavily in the ML components of the system. Domain specific models are a must if you want to be a serious contender to an existing search system and all the usual considerations still apply regarding frequent retraining and monitoring of the models. Currently this is left as an exercise to the reader - and a very large one at that. We (https://github.com/marqo-ai/marqo, I am a co-founder) are investing heavily into making the ML production worthy and continuous learning from feedback of the models as part of the system. Lots of other things to think about in how you represent documents with multiple vectors, multimodality, late interactions, the interplay between embedding quality and HNSW graph quality (i.e. recall) and much more.
What are some alternatives?
paper-qa - High accuracy RAG for answering questions from scientific documents with citations
ai-pdf-chatbot-langchain - AI PDF chatbot agent built with LangChain & LangGraph
Weaviate - Weaviate is an open-source vector database that stores both objects and vectors, allowing for the combination of vector search with structured filtering with the fault tolerance and scalability of a cloud-native database.
chatgpt-memory - Allows to scale the ChatGPT API to multiple simultaneous sessions with infinite contextual and adaptive memory powered by GPT and Redis datastore.
Milvus - Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search