Skip to content

Dockerfile + notes on getting DALL-E Mega running locally

Notifications You must be signed in to change notification settings

manimino/dalle-docker

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Dockerfile for DALL-E Mega

Follow instructions in Boris Dayma's DALL-E mini repo to download the latest DALL-E Mega model.

This Dockerfile will run the notebooks from that repo and generate you some images. Just like the Huggingface space does. Haven't tried training yet.

On an RTX 3090, it takes about 5.6 seconds per image (so 50 seconds for a 9-pack of images).

Image generation also worked on CPU, but took several minutes per image.

The images look great:

Spiderman generated image

Instructions

Build this Dockerfile:

docker build -t dallemega .

Run it:

Replace YOUR_PATH_HERE with wherever wandb downloaded your dall-e checkpoint to:

docker run --gpus all -it --rm -v $(realpath ~/notebooks):/notebooks -v /YOUR_PATH_HERE/artifacts/:/artifacts -p 8888:8888 dallemega

That will launch a Jupyter notebook instance, open a browser to localhost:8888/?token=... and copy the token from stdout.

If your docker run command fails, you may need to install nvidia-container-toolkit. That's an OS-specific process so... Google around. Good luck.

Run the inference notebook

Make a notebook in /notebooks and copy/paste stuff into it from the dalle-mini inference notebook.

Replace the DALLE_MODEL path with yours. So if you have checkpoint 15, it looks like:

DALLE_MODEL = "/artifacts/mega-1:v15"

About

Dockerfile + notes on getting DALL-E Mega running locally

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages