Generating Training Data for Denoising Real RGB Images
via Camera Pipeline Simulation


Camille Biscarrat


(a) Input real noisy JPEG (32.4dB)
(b) N3Net [1] trained with
AWGN (32.3dB)
(c) N3Net [1] trained with
our pipeline (35.2dB)
(d) Ground truth (demosaicked)

(a) Real noise in cellphone-processed JPEG pictures is very different from uncorrelated Gaussian noise widely assumed. (b) A blind denoiser trained on additive white Gaussian noise (AWGN) is unable to recognize the noise pattern resulting and denoise the image. (c) In contrast, the network trained on realistic noise model generated by our pipeline was able to denoise properly, resulting in over 3dB improvement.

Abstract

Image reconstruction techniques such as denoising often need to be applied to the RGB output of cameras and cellphones. Unfortunately, the commonly used additive white noise (AWGN) models do not accurately reproduce the noise and the degradation encountered on these inputs. This is particularly important for learning-based techniques, because the mismatch between training and real world data will hurt their generalization. This paper aims to accurately simulate the degradation and noise transformation performed by camera pipelines. This allows us to generate realistic degradation in RGB images that can be used to train machine learning models. We use our simulation to study the importance of noise modeling for learning-based denoising. Our study shows that a realistic noise model is required for learning to denoise real JPEG images. A neural network trained on realistic noise outperforms the one trained with AWGN by 3 dB. An ablation study of our pipeline shows that simulating denoising and demosaicking is important to this improvement and that realistic demosaicking algorithms, which have been rarely considered, is needed. We believe this simulation will also be useful for other image reconstruction tasks, and we will distribute our code publicly.

Reference

Publication

Under review.

Acknowledgment

The authors would like to thank Toyota Research Institute for their generous support of this project. We also thank Tzu-mao Li for his insightful comments, and Luke Anderson for his helpful comments in revising the draft.