min-dalle
mesh-transformer-jax
min-dalle | mesh-transformer-jax | |
---|---|---|
31 | 52 | |
3,474 | 6,213 | |
- | - | |
0.0 | 0.0 | |
over 1 year ago | over 1 year ago | |
Python | Python | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
min-dalle
- Open source Python libraries for AI image generation that you can install on an Amazon GPU instance, like min(DALL-E) and Pixray?
- List of open source machine learning AI image generation/text-to-image libraries that can be installed on an Amazon GPU instance? e.g. MinDall-E, Disco Diffusion, Pixray
-
Free/open-source AI Text-To-Image Models that can be run on AWS?
min(DALL·E).
- I'm building a timeline for generative image ML models. What's missing?
-
DALL·E Now Available in Beta
Additionally, it's also open-sourced on GitHub and can be self-hosted, with easy instructions to do so: https://github.com/kuprel/min-dalle
-
dalle update
For CPU, even highly-optimized models like mindalle are prohibitively slow.
-
Hii everyone ,Can I build the dalle mini from scratch or not?? Please help!!
Maybe you would be interested in this GitHub repo.
-
World of Warcraft Character Beanie Babies
These were generated with DALL-E Mega via min-dalle, which is a more advanced version of DALL-E Mini with better visual fidelity (less blurry) but otherwise similar results.
- Show HN: Generate webpage summary images with DALL-E mini
-
"min(DALL·E)" is "a minimal implementation of Boris Dayma's DALL·E Mini in PyTorch. It has been stripped to the bare essentials necessary for doing inference." This uses the DALL-E Mega model. The Google Colab notebook using a Tesla T4 GPU takes 35 seconds to generate 4 images, and 17 seconds for 1.
GitHub repo (contains links to Colab notebook and web app at site Replicate[dot]com). The times mentioned in the title don't included setup time.
mesh-transformer-jax
-
Large Language Models: Compairing Gen2/Gen3 Models (GPT-3, GPT-J, MT5 and More)
GPT-J is a LLM case study with two goals: Training a LLM with a data source containing unique material, and using the training frameworkMesh Transformer JAX to achieve a high training efficiency through parallelization. There is no research paper about GPT-J, but on its GitHub pages, the model, different checkpoints, and the complete source code for training is given.
-
[R] Parallel Attention and Feed-Forward Net Design for Pre-training and Inference on Transformers
This idea has already been proposed in ViT-22B and GPT-J-6B.
- Show HN: Finetune LLaMA-7B on commodity GPUs using your own text
-
[D] An Instruct Version Of GPT-J Using Stanford Alpaca's Dataset
Sure. Here's the repo I used for the fine-tuning: https://github.com/kingoflolz/mesh-transformer-jax. I used 5 epochs, and appart from that I kept the default parameters in the repo.
- Boss wants me to use ChatGPT for work, but I refuse to input my personal phone number. Any advice?
-
Let's build GPT: from scratch, in code, spelled out by Andrej Karpathy
You can skip to step 4 using something like GPT-J as far as I understand: https://github.com/kingoflolz/mesh-transformer-jax#links
The pretrained model is already available.
-
Best coding model?
The Github repo suggests it's possible you can change the number of checkpoints to make it run on a GPU.
- Ask HN: What language models can I fine-tune at home?
-
selfhosted/ open-source ChatGPT alternative?
GPT-J, which uses mesh-transformer-jax: https://github.com/kingoflolz/mesh-transformer-jax
-
GPT-J, an open-source alternative to GPT-3
They hinted at it in the screenshot, but the goods are linked from the https://6b.eleuther.ai page: https://github.com/kingoflolz/mesh-transformer-jax#gpt-j-6b (Apache 2)
What are some alternatives?
dalle-mini - DALL·E Mini - Generate images from a text prompt
DeepSpeed - DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
dalle-playground - A playground to generate images from any text prompt using Stable Diffusion (past: using DALL-E Mini)
tensorflow - An Open Source Machine Learning Framework for Everyone
CogVideo - Text-to-video generation. The repo for ICLR2023 paper "CogVideo: Large-scale Pretraining for Text-to-Video Generation via Transformers"
gpt-neo - An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
imagen-pytorch - Implementation of Imagen, Google's Text-to-Image Neural Network, in Pytorch
jax - Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
KoboldAI-Client
DALLE2-pytorch - Implementation of DALL-E 2, OpenAI's updated text-to-image synthesis neural network, in Pytorch
alpaca-lora - Instruct-tune LLaMA on consumer hardware