Our great sponsors
- InfluxDB - Access the most powerful time series database as a service
- ONLYOFFICE ONLYOFFICE Docs — document collaboration in your environment
- Sonar - Write Clean Python Code. Always.
- CodiumAI - TestGPT | Generating meaningful tests for busy devs
-
-
python: 3.10.6 • torch: 1.13.1+cu117 • xformers: 0.0.16+814314d.d20230119 • commit: 54674674 • checkpoint: 61a37adf76 i get 18.79it/s .. with all shebangs installed ... triton, deepspeed, tensorrt .. did not tested with torch 2.0
-
InfluxDB
Access the most powerful time series database as a service. Ingest, store, & analyze all types of time series data in a fully-managed, purpose-built database. Keep data forever with low-cost storage and superior data compression.
-
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
I tried installing PyTorch 2.0.0, with triton from here microsoft/DeepSpeed#2694, compiling my own xformers and it made my inference even slower. From 17-18it/s 512x512, Batch size: 1, any sampling method to around 16-17it/s but especially with Batch size: 8, from 5.65it/s to 4.66it/s.
Related posts
- PyTorch Primitives in WebGPU for the Browser
- Cleared AWS Machine Learning - Specialty exam.. Happy to help!!!
- Cerebras Open Sources Seven GPT models and Introduces New Scaling Law
- OpenAI’s policies hinder reproducible research on language models
- Seeking Advice on Optimizing Stable Diffusion with AMD Graphics Card