Style-Transfer-in-Text
gpt-2-simple
Style-Transfer-in-Text | gpt-2-simple | |
---|---|---|
3 | 13 | |
1,620 | 3,403 | |
0.2% | 0.1% | |
1.2 | 0.0 | |
about 2 years ago | over 2 years ago | |
Python | ||
- | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Style-Transfer-in-Text
-
Suggestions on Cool NLP Projects!
I've been working on a way to train GPT-2 with HuggingFace Transformers (using aitextgen) to write short stories trained on public domain authors! I've been planning to look through these style transfer resources if you're interested.
-
Training GPT-2 with HuggingFace Transformers to sound like a certain author
Check out the "Stylistic Related Papers" and especially the "Controlled Text Generation" sections of this Github papers list -- sounds like a similar problem to what you're thinking of.
-
Something like style transfer for language processing?
There's been quite a bit of work on extending the same concept to text, with "style" often being used to refer to aspects like sentiment/aspect. https://github.com/fuzhenxin/Style-Transfer-in-Text is a good collection of recent work in NLP in this area. Specifically for authorial style, I'm aware of this recent one. Also relevant might be earlier work on author obfuscation and imitation.
gpt-2-simple
-
Show HN: WhatsApp-Llama: A clone of yourself from your WhatsApp conversations
Tap the contact's name in WhatsApp (I think it only works on a phone) and at the bottom of that screen there's Export Chat.
For finetuning GPT-2 I think I used this thing on Google Colab. (My friend ran it on his GPU, it should be doable on most modern-ish GPUs.)
https://github.com/minimaxir/gpt-2-simple
I tried doing something with this a few months ago though and it was a bit of a hassle to get running (needed to use a specific python version for some dependencies...), I forget the details sorry!
-
indistinguishable
I mentioned in a different reply that I used https://github.com/minimaxir/gpt-2-simple
-
training gpt on your own sources - how does it work? gpt2 v gpt3? and how much does it cost?
You will need a few hundred bucks, python experience, and a simple implementation such as this repo https://github.com/minimaxir/gpt-2-simple
-
I (re)trained an AI using the 36 lessons of Vivec, the entirety of C0DA, the communist manifesto and the top posts of /r/copypasta and asked it the most important/unanswered lore questions. What are the lore implications of these insights?
I just used the gpt-2-simple python package and ran it overnight in an jupyter notebook, but you could copy the code to any python compiler and it should also work.
-
How do I start a personal project?
I'll note that if you're just doing text generation it is a simple project as far as ML goes, there are some nice libraries you can use that require minimal ML knowlege -eg https://github.com/minimaxir/gpt-2-simple
-
I created a twitter account that posts AI generated Canucks related tweets. I call it "Canucks Artificial Insider".
Then, I use the GPT-2 AI libraries, wrapped in a python library GPT-2 Simple to generate the content. My actual code is basically just their code sample, so basically 6 lines of python. With GPT-2, you train the existing AI to your specific dataset, which in my case is this text file of tweets.
-
Training GPT-2 with HuggingFace Transformers to sound like a certain author
gpt_2_simple is your best bet! Its super easy to use, you just need to downgrade TensorFlow and some other packages in your environment.
-
These Magic cards don't exist - Generating names for new cards using machine learning and GPT-2.
I used the GPT-2 Simple program by minimaxir to train the algorithm on every card in Magic's history that was released in a main expansion. Then I generated about 2,000 (it was actually more, but the algorithm really liked giving me cards that already exist) new names which I searched through to find the best ones.
-
No rush, mostly curious (training/finetuned models)
Might I suggest starting Starting here, to learn on Simple GPT2. They have a Google Colab Notebook if your CPU GPU is shit, and what helped me learn best is dissect the code, and basically make my own Colab notebook piece by piece, learning what each function does.
-
Selecting good hyper-parameters for fine-tuning a GPT-2 model?
The last couple of months, I've been running a Twitter bot that posts GPT-2-generated content, trained off of Tweets from existing accounts using gpt-2-simple. In my more recent training sessions, it seems like the quality of the output has been decreasing; it often gives outputs that are just barely modified from the original training data, if not verbatim.
What are some alternatives?
awesome-refreshing-llms - EMNLP'23 survey: a curation of awesome papers and resources on refreshing large language models (LLMs) without expensive retraining.
rex-gym - OpenAI Gym environments for an open-source quadruped robot (SpotMicro)
arpa
textgenrnn - Easily train your own text-generating neural network of any size and complexity on any text dataset with a few lines of code.
Speech-Separation-Paper-Tutorial - A must-read paper for speech separation based on neural networks
ctrl-sum - Resources for the "CTRLsum: Towards Generic Controllable Text Summarization" paper