Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality. Learn more →
Top 6 Python pre-trained-language-model Projects
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
HugNLP
HugNLP is a unified and comprehensive NLP library based on HuggingFace Transformer. Please hugging for NLP now!😊 HugNLP will released to @HugAILab
I'd like to share with you today the Chinese-Alpaca-Plus-13B-GPTQ model, which is the GPTQ format quantised 4bit models of Yiming Cui's Chinese-LLaMA-Alpaca 13B for GPU reference.
Here’s another one - it’s older but has some interesting charts and graphs.
https://arxiv.org/abs/2303.18223
Project mention: [D] Is it better to create a different set of Doc2Vec embeddings for each group in my dataset, rather than generating embeddings for the entire dataset? | /r/MachineLearning | 2023-10-28I'm using Top2Vec with Doc2Vec embeddings to find topics in a dataset of ~4000 social media posts. This dataset has three groups:
Project mention: Knowledge Rumination for Pre-trained Language Models | /r/BotNewsPreprints | 2023-05-16Previous studies have revealed that vanilla pre-trained language models (PLMs) lack the capacity to handle knowledge-intensive NLP tasks alone; thus, several works have attempted to integrate external knowledge into PLMs. However, despite the promising outcome, we empirically observe that PLMs may have already encoded rich knowledge in their pre-trained parameters but fails to fully utilize them when applying to knowledge-intensive tasks. In this paper, we propose a new paradigm dubbed Knowledge Rumination to help the pre-trained language model utilize those related latent knowledge without retrieving them from the external corpus. By simply adding a prompt like ``As far as I know'' to the PLMs, we try to review related latent knowledge and inject them back to the model for knowledge consolidation. We apply the proposed knowledge rumination to various language models, including RoBERTa, DeBERTa, GPT-3 and OPT. Experimental results on six commonsense reasoning tasks and GLUE benchmarks demonstrate the effectiveness of our proposed approach, which further proves that the knowledge stored in PLMs can be better exploited to enhance the downstream performance. Code will be available in https://github.com/zjunlp/knowledge-rumination.
Python pre-trained-language-models related posts
- Chinese-Alpaca-Plus-13B-GPTQ
- How to train a new language that is not in base model?
- Open Source Chinese LLMs
- HugNLP: A Unified and Comprehensive Library for Natural Language Processing
- OpenPrompt: An Open-Source Toolkit for Prompt-Learning
-
A note from our sponsor - InfluxDB
www.influxdata.com | 26 Apr 2024
Index
What are some of the best open-source pre-trained-language-model projects in Python? This list will help you:
Project | Stars | |
---|---|---|
1 | Chinese-LLaMA-Alpaca | 17,251 |
2 | LLMSurvey | 8,716 |
3 | OpenPrompt | 4,146 |
4 | Top2Vec | 2,839 |
5 | HugNLP | 247 |
6 | knowledge-rumination | 16 |
Sponsored