Extracting-Training-Data-from-Large-Langauge-Models
aitg
Extracting-Training-Data-from-Large-Langauge-Models | aitg | |
---|---|---|
1 | 1 | |
26 | 4 | |
- | - | |
5.1 | 0.0 | |
almost 2 years ago | over 1 year ago | |
Python | Python | |
MIT License | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Extracting-Training-Data-from-Large-Langauge-Models
-
Paper Implementation - Extracting Training Data from Large Language Models
Code - https://github.com/shreyansh26/Extracting-Training-Data-from-Large-Langauge-Models
aitg
What are some alternatives?
Neural-Scam-Artist - Web Scraping, Document Deduplication & GPT-2 Fine-tuning with a newly created scam dataset.
nvc-gpt3-chat - This is the code I used to create a small private SMS text chat system that employs GPT3 from OpenAI and "nonviolent communication", an algorithmically based method of conflict resolution. Hopefully the chat helps users process conflict.
funcbygpt - tell the function what to do and gpt will write the code
ALAE - [CVPR2020] Adversarial Latent Autoencoders
finetune-gpt2xl - Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed
bertviz - BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
jukebox - Code for the paper "Jukebox: A Generative Model for Music"
kogpt - KakaoBrain KoGPT (Korean Generative Pre-trained Transformer)