Show HN: Danswer – open-source question answering across all your docs

This page summarizes the projects mentioned and recommended in the original post on news.ycombinator.com

Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
  • danswer

    Gen-AI Chat for Teams - Think ChatGPT if it had access to your team's unique knowledge.

  • Otherwise, things were easy to start up and play around with, even for an AI noob like me. Both their web and text-upload source connectors worked without issue in my testing.

    [1]: https://github.com/danswer-ai/danswer/pull/139

  • qdrant

    Qdrant - High-performance, massive-scale Vector Database for the next generation of AI. Also available in the cloud https://cloud.qdrant.io/

  • WorkOS

    The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.

    WorkOS logo
  • Milvus

    A cloud-native vector database, storage for next generation AI applications

  • Check this out. Built on a vector database (https://github.com/milvus-io/milvus) and a semantic cache (https://github.com/zilliztech/GPTCache)

    https://osschat.io/

  • GPTCache

    Semantic cache for LLMs. Fully integrated with LangChain and llama_index.

  • Check this out. Built on a vector database (https://github.com/milvus-io/milvus) and a semantic cache (https://github.com/zilliztech/GPTCache)

    https://osschat.io/

  • llama.cpp

    LLM inference in C/C++

  • Nowadays falcon-40b is probably more accurate than gpt4j, here's to hoping we get llama.cpp support for falcon builds soon [0]!

    [0]: https://github.com/ggerganov/llama.cpp/issues/1602

  • ggllm.cpp

    Falcon LLM ggml framework with CPU and GPU support

  • The GGLLM fork seems to be the leading falcon winner for now [1]

    It comes with its own variant of the GGML sub format "ggcv1" but there's quants available on HF [2]

    Although if you have a GPU I'd go with the newly released AWQ quantization instead [3] the performance is better.

    (I may or may not have a mild local LLM addiction - and video cards cost more then drugs)

    [1] https://github.com/cmp-nct/ggllm.cpp

    [2] https://huggingface.co/TheBloke/falcon-7b-instruct-GGML

    [3] https://huggingface.co/abhinavkulkarni/tiiuae-falcon-7b-inst...

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts