stable-diffusion-nvidia-docker
nvidia-gpu-scheduler
Our great sponsors
stable-diffusion-nvidia-docker | nvidia-gpu-scheduler | |
---|---|---|
6 | 1 | |
346 | 7 | |
- | - | |
7.0 | 0.0 | |
6 months ago | over 1 year ago | |
Python | Python | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
stable-diffusion-nvidia-docker
- Does Stable Diffusion support NVLink?
- The guy behind the viral fake photo of the Pope in a puffy coat says using AI to make images of celebrities 'might be the line' — and calls for greater regulation
- Utilizing Multiple GPUs - Repurposing Mining Rig
- Can we start a list of Stable Diffusion 2.0 compatible UI's?
- Using several computers (GPUs) to speed up Stable Diffusion computation times
-
Clustering GPUs for use with SD
Any update? I just searched through the discord for dataparallel and found someone mentioned this link to a docker image which appears to support multi GPU but I don't have any experience with docker and haven't seen where dataparallel is used in any of the files. I'm also searching through all the mentions of K80.
nvidia-gpu-scheduler
-
[D] How to be more productive while doing Deep Learning experiments?
Sure. No, a simple bash script is not enough. In my case, we have several machines shared in the department, some with GPUs, some without. What I have is a python script that gets a list of jobs and then it schedule them in the first available machine (according to memory/CPU/GPU availability). Unfortunately, what I have is really entangled with our computing platform (Docker-based with a shared filesystem) and not really easy to have it as standalone project (that's why I said "know you infrastructure"). The most similar thing that I could find online is this project. I believe there are then some HPC tools that could be useful (e.g. Slurm), but that's way too much for what we need.
What are some alternatives?
deforum-stable-diffusion
detectron2 - Detectron2 is a platform for object detection, segmentation and other visual recognition tasks.
dream-textures - Stable Diffusion built-in to Blender
fastapi-cloud-tasks - GCP's Cloud Tasks + Cloud Scheduler + FastAPI = Partial replacement for celery.
image-super-resolution - 🔎 Super-scale your images and run experiments with Residual Dense and Adversarial Networks.
pytorch-lightning - Build high-performance AI models with PyTorch Lightning (organized PyTorch). Deploy models with Lightning Apps (organized Python to build end-to-end ML systems). [Moved to: https://github.com/Lightning-AI/lightning]
Stable-Diffusion-2.0-CPU-or-GPU-Colab-Gradio - Config files for my GitHub profile.
tmux - tmux source code
Mask_RCNN_Pytorch - Mask R-CNN for object detection and instance segmentation on Pytorch
Sacred - Sacred is a tool to help you configure, organize, log and reproduce experiments developed at IDSIA.
dream-factory - Multi-threaded GUI manager for mass creation of AI-generated art with support for multiple GPUs.
aim - Aim 💫 — An easy-to-use & supercharged open-source experiment tracker.