Skip to content

Latest commit

 

History

History
229 lines (147 loc) · 8.04 KB

INSTALL.md

File metadata and controls

229 lines (147 loc) · 8.04 KB

dream-creator Installation

This guide will walk you through multiple ways to setup dream-creator on Ubuntu and Windows. If you wish to install PyTorch and dream-creator on a different operating system like MacOS, installation guides can be found here.

Note that in order to reduce their size, the pre-packaged binary releases (pip, Conda, etc...) have removed support for some older GPUs, and thus you will have to install from source in order to use these GPUs.

Ubuntu:

With A Package Manager:

The pip and Conda packages ship with CUDA and cuDNN already built in, so after you have installed PyTorch with pip or Conda, you can skip to installing dream-creator.

Github and pip:

Following the pip installation instructions here, you can install PyTorch with the following commands:

# in a terminal, run the commands
cd ~/
pip install torch torchvision

Or:

cd ~/
pip3 install torch torchvision

Now continue on to installing dream-creator to install dream-creator.

Conda:

Following the Conda installation instructions here, you can install PyTorch with the following command:

conda install pytorch torchvision -c pytorch

Now continue on to installing dream-creator to install dream-creator.

From Source:

(Optional) Step 1: Install CUDA

If you have a CUDA-capable GPU from NVIDIA then you can speed up dream-creator with CUDA.

Instructions for downloading and installing the latest CUDA version on all supported operating systems, can be found here.

(Optional) Step 2: Install cuDNN

cuDNN is a library from NVIDIA that efficiently implements many of the operations (like convolutions and pooling) that are commonly used in deep learning.

After registering as a developer with NVIDIA, you can download cuDNN here. Make sure that you use the appropriate version of cuDNN for your version of CUDA.

Follow the download instructions on Nvidia's site to install cuDNN correctly.

Note that the cuDNN backend can only be used for GPU mode.

(Optional) Steps 1-3: Install PyTorch with support for AMD GPUs using Radeon Open Compute Stack (ROCm)

It is recommended that if you wish to use PyTorch with an AMD GPU, you install it via the official ROCm dockerfile: https://rocm.github.io/pytorch.html

  • Supported AMD GPUs for the dockerfile are: Vega10 / gfx900 generation discrete graphics cards (Vega56, Vega64, or MI25).

PyTorch does not officially provide support for compilation on the host with AMD GPUs, but a user guide posted here apparently works well.

ROCm utilizes a CUDA porting tool called HIP, which automatically converts CUDA code into HIP code. HIP code can run on both AMD and Nvidia GPUs.

Step 3: Install PyTorch

To install PyTorch from source on Ubuntu (Instructions may be different if you are using a different OS):

cd ~/
git clone --recursive https://github.com/pytorch/pytorch
cd pytorch
python setup.py install

cd ~/
git clone --recursive https://github.com/pytorch/vision
cd vision
python setup.py install

To check that your torch installation is working, run the command python or python3 to enter the Python interpreter. Then type import torch and hit enter.

You can then type print(torch.version.cuda) and print(torch.backends.cudnn.version()) to confirm that you are using the desired versions of CUDA and cuDNN.

To quit just type exit() or use Ctrl-D.

Now continue on to installing dream-creator to install dream-creator.

Windows Installation

If you wish to install PyTorch on Windows From Source or via Conda, you can find instructions on the PyTorch website: https://pytorch.org/

Github and pip

First, you will need to download Python 3 and install it: https://www.python.org/downloads/windows/. I recommend using the executable installer for the latest version of Python 3.

Then using https://pytorch.org/, get the correct pip command, paste it into the Command Prompt (CMD) and hit enter:

pip install torch===1.6.0 torchvision===0.7.0 -f https://download.pytorch.org/whl/torch_stable.html

After installing PyTorch, download the dream-creator Github repository and extract/unzip it to the desired location.

Then copy the file path to your dream-creator folder, and paste it into the Command Prompt, with cd in front of it and then hit enter.

In the example below, the dream-creator folder was placed on the desktop:

cd C:\Users\<User_Name>\Desktop\dream-creator-master

You can now continue on to installing dream-creator, skipping the git clone step.

Install dream-creator

First we clone dream-creator from GitHub:

cd ~/
git clone https://github.com/ProGamerGov/dream-creator.git
cd dream-creator

To test that dream-creator works, we'll first need to create a test dataset:

python data_tools/create_test_data.py -output_dir test_data -num_images 1200 -shape_size 200,200

Now we calculate the mean and standard deviation of our newly created test dataset:

python data_tools/calc_ms.py -data_path test_data

The calc_ms.py script will then output something that looks like this:

-data_mean 126.7094,126.0554,126.9935 -data_sd 73.655,73.3885,73.7058

Then using the mean and standard deviation parameter inputs we generated above, we can now train a GoogleNet model using the test dataset:

python train_googlenet.py -data_path test_data -balance_classes -batch_size 96 -num_epochs 10 -data_mean 126.7094,126.0554,126.9935 -data_sd 73.655,73.3885,73.7058

If everything is working properly with the training script, then you should see output like this:

Total 2400 images, split into 2 classes
Classes:
 {'ellipse': 0, 'rectangle': 1}
Model has 10,021,318 learnable parameters

Epoch 1/10
------------
train Loss: 0.9526 Acc: 0.8250
  Time Elapsed 0m 23s
val Loss: 0.0319 Acc: 0.9896
  Time Elapsed 0m 27s

Epoch 2/10
------------
train Loss: 0.3052 Acc: 0.9260
  Time Elapsed 0m 49s
val Loss: 0.0699 Acc: 0.9646
  Time Elapsed 0m 52s

Epoch 3/10
------------
train Loss: 0.2157 Acc: 0.9417
  Time Elapsed 1m 15s
val Loss: 0.0142 Acc: 0.9938
  Time Elapsed 1m 19s

Epoch 4/10
------------
train Loss: 1.6389 Acc: 0.5719
  Time Elapsed 1m 42s
val Loss: 0.6930 Acc: 0.5125
  Time Elapsed 1m 45s

Epoch 5/10
------------
train Loss: 1.3215 Acc: 0.5089
  Time Elapsed 2m 8s
val Loss: 0.6986 Acc: 0.5125
  Time Elapsed 2m 12s

After training has finished, we can now visualize the newly created DeepDream model's FC layers using the following command:

python vis_multi.py -model_file bvlc_out010.pth -num_iterations 200

The vis_multi.py script should end up creating two output images, where one image has more circlelike features and the other has more squarelike features. Using more complex datasets that have more classes and images will yield far better looking results. You can find a list of image collection tools, possible sources of images, and duplicate image detection tools on the dream-creator wiki.

Finally, to visualize a single layer and channel or to DeepDream your own image, we can use the following command:

# Random Noise content image
python vis.py -model_file bvlc_out010.pth -layer mixed5a -channel 7 -num_iterations 200

# With Content image
python vis.py -model_file bvlc_out010.pth -layer mixed5a -channel 7 -content_image examples/small/fc_flowers.jpg -image_size 512,512 -num_iterations 200