SaaSHub helps you find the best software and product alternatives Learn more →
Top 15 sentence-transformer Open-Source Projects
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
beir
A Heterogeneous Benchmark for Information Retrieval. Easy to use, evaluate your models across 15+ diverse IR datasets.
-
StoryToolkitAI
An editing tool that uses AI to transcribe, understand content and search for anything in your footage, integrated with ChatGPT and other AI models
-
DiffCSE
Code for the NAACL 2022 long paper "DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings"
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
-
Python-Schema-Matching
A python tool using XGboost and sentence-transformers to perform schema matching task on tables.
-
Llama-2-GGML-CSV-Chatbot
The Llama-2-GGML-CSV-Chatbot is a conversational tool leveraging the powerful Llama-2 7B language model. It facilitates multi-turn interactions based on uploaded CSV data, allowing users to engage in seamless conversations.
-
balena
BALanced Execution through Natural Activation : a human-computer interaction methodology for code running.
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
Project mention: [D] Is it better to create a different set of Doc2Vec embeddings for each group in my dataset, rather than generating embeddings for the entire dataset? | /r/MachineLearning | 2023-10-28I'm using Top2Vec with Doc2Vec embeddings to find topics in a dataset of ~4000 social media posts. This dataset has three groups:
RAG is very difficult to do right. I am experimenting with various RAG projects from [1]. The main problems are:
- Chunking can interfer with context boundaries
- Content vectors can differ vastly from question vectors, for this you have to use hypothetical embeddings (they generate artificial questions and store them)
- Instead of saving just one embedding per text-chuck you should store various (text chunk, hypothetical embedding questions, meta data)
- RAG will miserably fail with requests like "summarize the whole document"
- to my knowledge, openAI embeddings aren't performing well, use a embedding that is optimized for question answering or information retrieval and supports multi language. Also look into instructor embeddings: https://github.com/embeddings-benchmark/mteb
1 https://github.com/underlines/awesome-marketing-datascience/...
The BEIR project might be what you're looking for: https://github.com/beir-cellar/beir/wiki/Leaderboard
There's StoryToolkitAI, it's free (But requires Davinci Resolve Studio). It can transcribe and generate subtitles quite accurately. It also has a feature to translate subtitles to English. I haven't tried the translate feature yet, but I've been using this tool for my work a lot. It also supports more languages than resolve's built in transcription and auto subtitle tool.
huggingface.co/Llama-2-GGML-CSV-Chatbot
pip install git+https://github.com/jacobmarks/emoji_search.git
By “this”, I mean an open-source semantic emoji search engine, with both UI-centric and CLI versions. The Python CLI library can be found here, and the UI-centric version can be found here. You can also play around with a hosted (also free) version of the UI emoji search engine online here.
Project mention: BALanced Execution Through Natural Activation: A HCI Methodology | news.ycombinator.com | 2023-12-31
Project mention: Tranformer-based Denoising AutoEncoder for ST Unsupervised pre-training | news.ycombinator.com | 2024-02-04A new PyPI package for training sentence embedding models in just 2 lines.
The acquisition of sentence embeddings often necessitates a substantial volume of labeled data. However, in many cases and fields, labeled data is rarely accessible, and the procurement of such data is costly. In this project, we employ an unsupervised process grounded in pre-trained Transformers-based Sequential Denoising Auto-Encoder (TSDAE), introduced by the Ubiquitous Knowledge Processing Lab of Darmstadt, which can realize a performance level reaching 93.1% of in-domain supervised methodologies.
The TSDAE schema comprises two components: an encoder and a decoder. Throughout the training process, TSDAE translates tainted sentences into uniform-sized vectors, necessitating the decoder to reconstruct the original sentences utilizing this sentence embedding. For good reconstruction quality, the semantics must be captured well in the sentence embeddings from the encoder. Subsequently, during inference, the encoder is solely utilized to form sentence embeddings.
GitHub : https://github.com/louisbrulenaudet/tsdae
Installation :
sentence-transformers related posts
-
Do you Know! Llama ?
-
BEIR: A Heterogeneous Benchmark for Information Retrieval
-
Smarter Summaries with Finetuning GPT-3.5 and Chain of Density
-
Benefits of hybrid search
-
[Discussion] Convince me that this training set contamination is fine (or not)
-
Ask HN: What's the best framework for text classification (few-shot learning)?
-
Is it worth using LLMs like GPT-3 for text classification?
-
A note from our sponsor - SaaSHub
www.saashub.com | 9 May 2024
Index
What are some of the best open-source sentence-transformer projects? This list will help you:
Project | Stars | |
---|---|---|
1 | Top2Vec | 2,847 |
2 | setfit | 2,001 |
3 | mteb | 1,421 |
4 | beir | 1,393 |
5 | StoryToolkitAI | 584 |
6 | DiffCSE | 286 |
7 | elastic_transformers | 159 |
8 | nalcos | 53 |
9 | Python-Schema-Matching | 24 |
10 | Llama-2-GGML-CSV-Chatbot | 8 |
11 | emoji_search | 8 |
12 | emoji-search-plugin | 5 |
13 | balena | 5 |
14 | tsdae | 3 |
15 | search-engine | 1 |
Sponsored