Skip to content

Implementation of Nalbach et al. 2017 paper.

Notifications You must be signed in to change notification settings

sjeongab/DeepShading

 
 

Repository files navigation

Deep Shading

Convolutional Neural Networks for Screen-Space Shading

Our project is based on Nalbach et al. 2017 paper. In this project, a set of buffers are provided to a Deep Convolutional Network in order to synthetize differend shading effects (such as Ambient Occlusion, Depth of Field, Global Illumination and Sub-surface Scattering). The set of buffers depends os the shading effect we want to synthetize.

Input Bufers:

Result (Ambient Occlusion):

Requirements for the project

  1. Python 3.x
  2. Tensorflow 1.10
  3. Keras
  4. OpenCV 3.4(for loading,resizing images)
  5. h5py(for saving trained model)
  6. pyexr

Steps to run the repo

  1. Clone the repo
  2. Download the dataset (https://deep-shading-datasets.mpi-inf.mpg.de/)
  3. Install the requirements
  4. Generate the .tfrecord for trainning and validation (Use the DataReader.py)
  5. Run "Shading.py"

References

[1] Oliver Nalbach, Elena Arabadzhiyska, Dushyant Mehta, Hans-Peter Seidel, Tobias Ritschel Deep Shading: Convolutional Neural Networks for Screen-Space Shading to appear in Proc. EGSR 2017

About

Implementation of Nalbach et al. 2017 paper.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 95.1%
  • Python 4.9%