tensorrtx
community-events
tensorrtx | community-events | |
---|---|---|
3 | 8 | |
6,584 | 377 | |
- | 1.6% | |
8.4 | 7.2 | |
6 days ago | 5 months ago | |
C++ | Jupyter Notebook | |
MIT License | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
tensorrtx
-
A Three-pronged Approach to Bringing ML Models Into Production
In terms of the latter, this is quite common when employing non-standard SOTA models. You may discover a variety of TensorRT implementations on the web if you want to use popular models—for example, in the project where we needed to train an object-detection algorithm on Rutorch and deploy it on Triton, we used many cases of PyTorch -> TensorRT -> Triton. The implementation of the model on TensoRT was taken from here. You may also be interested in this repository, as it contains many current implementations supported by developers.
-
Dall-E 2
I'll try them out. I have an RTX 2070, which apparently supports fp16. But it only has 8GB RAM.
I used the instructions here to check: https://github.com/wang-xinyu/tensorrtx/blob/master/tutorial...
-
Increasing usb cam FPS with Yolov5 on a Jetson Xavier NX?
Optimize your model using TensorRT. There is a good implementation here: https://github.com/wang-xinyu/tensorrtx/tree/master/yolov5
community-events
-
Controlling Stable Diffusion with JAX & Diffusers using TPU v4
Best applications that will come out of this sprint will receive prizes. You can find more information here. If you want to get started, simply join huggingface.co/discord, take the role 🧨 Diffusers and head to #jax-diffusers-ideas to share your idea or join one of the teams, and fill this form: https://forms.gle/t3M7aNPuLL9V1sfa9
-
JAX & Diffusers to Control Stable Diffusion (with TPUs ⚡️ )
It will start on 17th of April. To join us, you can join huggingface.co/join/discord and take the role Diffusers from #role-assignment. After this, simply fill the form provided in this guide to later get access to TPUs. https://github.com/huggingface/community-events/tree/main/jax-controlnet-sprint
- “Control Stable Diffusion” Sprint kicks off with free TPU-v4 from Google
-
Free compute to train custom ControlNet by Hugging Face
Details and sign-up: https://github.com/huggingface/community-events/tree/main/jax-controlnet-sprint
-
How can I create a dataset to refine Whisper AI from old videos with subtitles?
For the training, I extremely recommend checking out the Whisper Fine-Tuning Event. It has a python script to train in one command, tons of tips, even a walkthrough video.
- I am using OpenAi's whisper transcription/translation model. I am wondering if I can improve it's performance by optimizing the audio files somehow. What features of audio files should I look into to make the whisper model perform better?
-
[N] Gradio Blocks + Hugging Face event is starting this week. A hackathon type event from May 17th to May 31st with prizes in which we will create interactive web demos for state-of-the-art machine learning models
We are happy to invite you to the Gradio Blocks Party - a community event in which we will create interactive demos for state-of-the-art machine learning models. Demos are powerful because they allow anyone — not just ML engineers — to try out models in the browser, give feedback on predictions, identify trustworthy models. The event will take place from May 17th to 31st. We will be organizing this event on Github and the Hugging Face discord channel. Prizes will be given at the end of the event, see the Prizes section
-
Dall-E 2
If you're interested in generative models, Hugging Face is putting on an event around generative models right now called the HugGAN sprint, where they're giving away free access to compute to train models like this.
You can join it by following the steps in the guide here: https://github.com/huggingface/community-events/tree/main/hu...
There will also be talks from awesome folks at EleutherAI, Google, and Deepmind
What are some alternatives?
TensorRT - NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
dalle-2-preview
tensorflow-yolov4-tflite - YOLOv4, YOLOv4-tiny, YOLOv3, YOLOv3-tiny Implemented in Tensorflow 2.0, Android. Convert YOLO v4 .weights tensorflow, tensorrt and tflite
dalle-mini - DALL·E Mini - Generate images from a text prompt
v-diffusion-pytorch - v objective diffusion inference code for PyTorch.
bevy_retro - Plugin pack for making 2D games with Bevy
lm-human-preferences - Code for the paper Fine-Tuning Language Models from Human Preferences
gpt-3 - GPT-3: Language Models are Few-Shot Learners
SegmentationCpp - A c++ trainable semantic segmentation library based on libtorch (pytorch c++). Backbone: VGG, ResNet, ResNext. Architecture: FPN, U-Net, PAN, LinkNet, PSPNet, DeepLab-V3, DeepLab-V3+ by now.
glide-text2im - GLIDE: a diffusion-based text-conditional image synthesis model