Basic-UI-for-GPT-J-6B-with-low-vram
adaptnlp
Our great sponsors
Basic-UI-for-GPT-J-6B-with-low-vram | adaptnlp | |
---|---|---|
4 | 2 | |
113 | 414 | |
- | 0.0% | |
0.0 | 0.0 | |
over 2 years ago | over 2 years ago | |
Jupyter Notebook | Jupyter Notebook | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Basic-UI-for-GPT-J-6B-with-low-vram
-
How to run this service with a local GPU?
You need a lot of VRAM to run the AI models, scaling somewhat with the amount of parameters a model uses. The most advanced model Pygmalion has is 6 billion parameters, which requires a minimum of 16GB of VRAM to run locally at decent speeds. There are methods of running 6b locally on low VRAM machines as listed here: https://github.com/arrmansa/Basic-UI-for-GPT-J-6B-with-low-vram but even then, the generations would be excruciatingly slow, and the lowest VRAM card used with this method has 6GB of VRAM.
-
Tesla M40 and GPT-J-6B
While waiting however I came across https://github.com/arrmansa/Basic-UI-for-GPT-J-6B-with-low-vram which allows you to use some of system memory to run the model. I was able to get a version working with 2.7B on my 2060 6GB with KoboldAI. The github above has an error that prevents it from working (https://github.com/arrmansa/Basic-UI-for-GPT-J-6B-with-low-vram/issues/1), but other than that it works.
-
How is any of this even possible?
Just to add to this, there is a low VRAM version of GPT-J here (suggest 16GB RAM + 8GB GPU).
-
GPT-J 6B locally on my computer
I found this yesterday, is it somehow possible to use this with KoboldAI to run GPT-J on weaker graphics cards?
adaptnlp
-
Tools to use for Semantic-searching Question Answering System
Check out adaptnlp
-
Case Sensitivity using HuggingFace & Google's T5 model (base)
Yes, there are capitals in the tokenizer vocabulary of t5-base and t5-small, so both support capitalization. A few days ago I was using t5-small through adaptnlp for extractive summarization and capitalization was working fine (https://github.com/Novetta/adaptnlp). AdaptNLP is basically just a transformers wrapper, so if you can't figure out a solution, you could just dissect their source code.
What are some alternatives?
gpt-neo_dungeon - Colab notebooks to run a basic AI Dungeon clone using gpt-neo-2.7B
keytotext - Keywords to Sentences
Behavior-Sequence-Transformer-Pytorch - This is a pytorch implementation for the BST model from Alibaba https://arxiv.org/pdf/1905.06874.pdf
fastai - The fastai deep learning library
clip-italian - CLIP (Contrastive Language–Image Pre-training) for Italian
gector - Official implementation of the papers "GECToR – Grammatical Error Correction: Tag, Not Rewrite" (BEA-20) and "Text Simplification by Tagging" (BEA-21)
pytorch-sentiment-analysis - Tutorials on getting started with PyTorch and TorchText for sentiment analysis.
browser-ml-inference - Edge Inference in Browser with Transformer NLP model
nn - 🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
Transformers-Tutorials - This repository contains demos I made with the Transformers library by HuggingFace.
pytorch-generative - Easy generative modeling in PyTorch.
ML-Workspace - 🛠 All-in-one web-based IDE specialized for machine learning and data science.