Git Product home page Git Product logo

unprocessing's Introduction

Unprocessing Images for Learned Raw Denoising

Reference code for the paper Unprocessing Images for Learned Raw Denoising. Tim Brooks, Ben Mildenhall, Tianfan Xue, Jiawen Chen, Dillon Sharlet, Jonathan T. Barron CVPR 2019

If you use this code, please cite our paper:

@inproceedings{brooks2019unprocessing,
  title={Unprocessing Images for Learned Raw Denoising},
  author={Brooks, Tim and Mildenhall, Ben and Xue, Tianfan and Chen, Jiawen and Sharlet, Dillon and Barron, Jonathan T},
  booktitle={IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
  year={2019},
}

The code is implemented in Tensorflow and the required packages are listed in requirements.txt.

Evaluation on Darmstadt Noise Dataset

In our paper, we evaluate on the Darmstadt Noise Dataset. Here are our Darmstadt results. We highly recommend this dataset for measuring denoise performance on real photographs, as the dataset contains real noisy images, which after denoising and upon submission to the Darmstadt website will be compared against real clean ground truth. Here are instructions to download this dataset. You'll also need to download our trained models and unzip them into ./models/. Once downloaded, replace the provided dnd_denoise.py file with the version in this repository and follow the instructions below to run an unprocessing denoiser on this data.

/google-research$ python -m unprocessing.dnd_denoise \
    --model_ckpt=/path/to/models/unprocessing_srgb_loss/model.ckpt-3516383 \
    --data_dir=/path/to/darmstadt/data \
    --output_dir=/path/to/darmstadt/ouputs

Then follow instructions in the Darmstadt README file, including running bundle_submissions.py to prepare for submission to the Darmstadt Noise Dataset online benchmark.

Training on MIRFlickr Images

In our paper, we train on source images from MIRFlickr. We used the full MIRFLICKR-1M dataset, which includes 1 million source images, although the smaller MIRFLICKR-25000, which contains 25 thousand source images, can be used as well and is easier to download and store. Both versions are freely avaiable—here are instructions to download this dataset. Once downloaded, break images into mirflickr/train and mirflickr/test subdirectories. Then run the command below to train on MIRFlickr. Note that you may wish to downsample the source images prior to training to help remove JPEG compression artifacts, which we found slightly helpful in our paper.

/google-research$ python -m unprocessing.train \
    --model_dir=/path/to/models/unprocessing_mirflickr \
    --train_pattern=/path/to/mirflickr/train \
    --test_pattern=/path/to/mirflickr/test

Training for Different Cameras

Our models are trained to work best on the Darmstadt Noise Dataset, which contains 4 cameras of various types, and low to moderate amounts of noise. They generalize well to other images, such as those from the HDR+ Burst Photography Dataset. However, if you would like to train a denoiser to work best on images from different cameras, you may wish to modify random_ccm(), random_gains() and random_noise_levels() in py/unprocess.py to best match the distribution of image metadata from your cameras. See our paper for details of how we modeled such metadata for Darmstadt. If your cameras have a special Bayer pattern outside of those supported, you will also need to modify mosaic() and demosaic() to match.

Evaluation on Different Real Data

To run evaluation on real raw photographs outside of Darmstadt and HDR+ datasets will require loading and running the trained model similar to dnd_denoise.py, packing input raw images into a 4-channel image of Bayer planes, and estimating the variance from shot and read noise levels. Note that these models are only designed to work with raw images (not processed sRGB images).

Shot and read noise levels are sometimes included in image metadata, and may go by different names, so we recommend refering to the specification for your camera's metadata. The shot noise level is a measurement of how much variance is proportional to the input signal, and the read noise level is a measurement of how much variance is independent of the image. We calculate an approximation of variance using the input noisy image as variance = shot_noise * noisy_img + read_noise, and pass it as an additional input both during training and evaluation.

If shot and read noise levels are not provided in your camera's metadata, it is possible to empirically measure these noise levels by calibrating your sensor or inferring from a single image. Here is a great overview of shot and read noise, and here is an example noise calibration script.

Unprocessing Images in Other Training Pipelines

If you are only looking to generate realistic raw data, you are welcome to use the unprocessing portion of our code separately from the rest of this project. Feel free to copy our unprocess.py file for all of your unprocessing needs!

unprocessing's People

Contributors

timothybrooks avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.