community-events
gpt-2
community-events | gpt-2 | |
---|---|---|
8 | 64 | |
379 | 21,146 | |
2.1% | 1.1% | |
7.2 | 2.5 | |
5 months ago | 27 days ago | |
Jupyter Notebook | Python | |
- | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
community-events
-
Controlling Stable Diffusion with JAX & Diffusers using TPU v4
Best applications that will come out of this sprint will receive prizes. You can find more information here. If you want to get started, simply join huggingface.co/discord, take the role 🧨 Diffusers and head to #jax-diffusers-ideas to share your idea or join one of the teams, and fill this form: https://forms.gle/t3M7aNPuLL9V1sfa9
-
JAX & Diffusers to Control Stable Diffusion (with TPUs ⚡️ )
It will start on 17th of April. To join us, you can join huggingface.co/join/discord and take the role Diffusers from #role-assignment. After this, simply fill the form provided in this guide to later get access to TPUs. https://github.com/huggingface/community-events/tree/main/jax-controlnet-sprint
- “Control Stable Diffusion” Sprint kicks off with free TPU-v4 from Google
-
Free compute to train custom ControlNet by Hugging Face
Details and sign-up: https://github.com/huggingface/community-events/tree/main/jax-controlnet-sprint
-
How can I create a dataset to refine Whisper AI from old videos with subtitles?
For the training, I extremely recommend checking out the Whisper Fine-Tuning Event. It has a python script to train in one command, tons of tips, even a walkthrough video.
- I am using OpenAi's whisper transcription/translation model. I am wondering if I can improve it's performance by optimizing the audio files somehow. What features of audio files should I look into to make the whisper model perform better?
-
[N] Gradio Blocks + Hugging Face event is starting this week. A hackathon type event from May 17th to May 31st with prizes in which we will create interactive web demos for state-of-the-art machine learning models
We are happy to invite you to the Gradio Blocks Party - a community event in which we will create interactive demos for state-of-the-art machine learning models. Demos are powerful because they allow anyone — not just ML engineers — to try out models in the browser, give feedback on predictions, identify trustworthy models. The event will take place from May 17th to 31st. We will be organizing this event on Github and the Hugging Face discord channel. Prizes will be given at the end of the event, see the Prizes section
-
Dall-E 2
If you're interested in generative models, Hugging Face is putting on an event around generative models right now called the HugGAN sprint, where they're giving away free access to compute to train models like this.
You can join it by following the steps in the guide here: https://github.com/huggingface/community-events/tree/main/hu...
There will also be talks from awesome folks at EleutherAI, Google, and Deepmind
gpt-2
-
What are LLMs? An intro into AI, models, tokens, parameters, weights, quantization and more
Medium models: Roughly between 1B to 10B parameters. This is where Mistral 7B, Phi-3, Gemma from Google DeepMind, and wizardlm2 sit. Fun fact: GPT 2 was a medium sized model, much smaller than its latest versions.
- Sam Altman is still trying to return as OpenAI CEO
- Build Personal ChatGPT Using Your Data
-
Are the recent advancements in AI technology primarily driven by recent discoveries or the progress in hardware capabilities and the abundance of available data?
"Our model, called GPT-2 (a successor to GPT), was trained simply to predict the next word in 40GB of Internet text. Due to our concerns about malicious applications of the technology, we are not releasing the trained model. As an experiment in responsible disclosure, we are instead releasing a much smaller model for researchers to experiment with, as well as a technical paper. "
-
BING IS NOW THE DEFAULT SEARCH FOR CHATGPT
They did release GPT-2 under the MIT License.
-
Don Knuth Plays with ChatGPT
Did you arrive at this certainty through reading something other than what OpenAI has published? The document [0] that describes the training data for GPT-2 makes this assertion hilarious to me.
[0]: https://github.com/openai/gpt-2/blob/master/model_card.md#da...
- Was frustriert euch an der Nutzung oder der Diskussion um KI?
- The AI
-
Help with pet project to learn - Running ChatGPT-2 at home
I made a clone of https://github.com/openai/gpt-2 on my local laptop
- По поводу опасности ИИ и предложений остановить разработки на 6 месяцев.
What are some alternatives?
dalle-2-preview
dalle-mini - DALL·E Mini - Generate images from a text prompt
minGPT - A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
bevy_retro - Plugin pack for making 2D games with Bevy
Real-Time-Voice-Cloning - Clone a voice in 5 seconds to generate arbitrary speech in real-time
lm-human-preferences - Code for the paper Fine-Tuning Language Models from Human Preferences
gpt-neo - An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
gpt-3 - GPT-3: Language Models are Few-Shot Learners
sentencepiece - Unsupervised text tokenizer for Neural Network-based text generation.
glide-text2im - GLIDE: a diffusion-based text-conditional image synthesis model
jukebox - Code for the paper "Jukebox: A Generative Model for Music"