finetune-gpt2xl
chatgpt-comparison-detection
finetune-gpt2xl | chatgpt-comparison-detection | |
---|---|---|
9 | 1 | |
421 | 1,190 | |
- | 3.3% | |
0.0 | 4.8 | |
11 months ago | 5 months ago | |
Python | Python | |
MIT License | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
finetune-gpt2xl
-
Fine-tuning?
git clone the finetuning repo https://github.com/Xirider/finetune-gpt2xl go into the finetuning repo, install the rest of the requirements, pip install -r requirements.txt
- Training text-generating models locally
-
Dataset For GPT Fine-Tuning
I would like to understand a little better how to organize texts for Fine-Tuning, especially for GPT Neo. I plan to use this repo procedure, where is the following notice,
-
How to share the finetuned model
In the code suggested in the video (and in the repo) the flag --fp16 is used. But reading the "DeepSpeed Integration" article it is said that,
- [D] I made a script that does all the work to deploy GPT-NEO on Windows 10. (Please Test)
-
[Project] Estimating fine-tuning cost
Finetuning GPT-NEO 2.7B on Wikitext (180mb) took me about 45 minutes on one preemptible V100 instance on google cloud. It cost 1.30$ per hour and therefore around 1 $. Here are the steps: https://github.com/Xirider/finetune-gpt2xl
-
[P] Guide: Finetune GPT2-XL (1.5 Billion Parameters, the biggest model) on a single 16 GB VRAM V100 Google Cloud instance with Huggingface Transformers using DeepSpeed
Here i explain the setup and commands to get it running: https://github.com/Xirider/finetune-gpt2xl
- Guide: Finetune GPT2-XL (1.5 Billion Parameters, the biggest model) on a single 16 GB VRAM V100 Google Cloud instance with Huggingface Transformers using DeepSpeed
chatgpt-comparison-detection
-
Hi friends, we bring you the first bilingual ChatGPT detection toolset and would love your feedback~
Project GitHub page: ChatGPT Comparison Corpus (C3), Detectors, and more! 🔥
What are some alternatives?
detoxify - Trained models & code to predict toxic comments on all 3 Jigsaw Toxic Comment Challenges. Built using ⚡ Pytorch Lightning and 🤗 Transformers. For access to our API, please email us at [email protected].
spacy-llm - 🦙 Integrating LLMs into structured NLP pipelines
Extracting-Training-Data-from-Large-Langauge-Models - A re-implementation of the "Extracting Training Data from Large Language Models" paper by Carlini et al., 2020
ScribeAI - ChatGPT extension for VSCode.
Caption-Anything - Caption-Anything is a versatile tool combining image segmentation, visual captioning, and ChatGPT, generating tailored captions with diverse controls for user preferences. https://huggingface.co/spaces/TencentARC/Caption-Anything https://huggingface.co/spaces/VIPLab/Caption-Anything
SuiSense - Using Artificial Intelligence to distinguish between suicidal and depressive messages (4th Place Congressional App Challenge)
chatgpt-extractive-shortener - Shortens a paragraph of text with ChatGPT, using successive rounds of word-level extractive summarization.
openai-function-calling - Helper functions to generate JSON schema dicts for OpenAI ChatGPT function calling requests.
sllm - A command line interface for the OpenAI LLMs.
gpt-chatbot-cli - it's basically chatgpt-cli, uses api.
qxresearch-event-1 - Python hands on tutorial with 50+ Python Application (10 lines of code) @xiaowuc2
bza - turn books, articles, plaintext or webpages into interactive read eval print loops, manage bookmarks in your own local database