Skip to content

Latest commit

 

History

History
54 lines (29 loc) · 3.41 KB

deep-image-prior.md

File metadata and controls

54 lines (29 loc) · 3.41 KB

Deep Image Prior

by Dmitry Ulyanov(Skoltech, Yandex), Andrea Vedaldi(Oxford), Victor Lempitsky(Skoltech)

ArXiv:1711

This paper proposes a new method for standard inverse problems such as denoising, super-resolution, and inpainting. Surprisingly, the proposed networks do not learn any knowledge for image restoration from image database. They only train the network with a single degraded image, which is wanted to be restored.

Before reading this paper, skimming the network architecture and training concept in supmat are recommended for faster understanding. They use generator networks which have fixed random-noise(inputs) and degraded images(targets). This is against conventional restoration networks which have degraded images(inputs) and ground truth images(targets).

Approach

Image Prior

Image restoration problems, which aim to recover original image $x$ having corrupted image $x_0$, are open formulated as an optimization tasks:

eq1

where $E(x;x_0)$ is data term and $R(x)$ is an image prior. They model the prior $R(x)$ with a ConvNet $f_{\theta}(z) = x$ where $\theta$ is parameters of the ConvNet and $z$ is a fixed input. They assert that the prior term can be removed by selecting a good mapping $f_{\theta}$, so the formulation changes to

eq3

They said, this means "instead of searching for the answer in the image space we now search for it in the space of neural network's parameters".

High Noise Impedance

The proposed method is using a randomly-initialized neural network as a handcrafted prior. Here is a key experiment from the paper to catch the meaning of the method.

fig1

The experiment points out that natually-looking images are converged faster than the image plus noise for reconstruction tasks. So they could stop training the network before overfiting to noise. This means that the generator network is sufficient to capture low-level image statistics prior. In contrast, previous restoration networks learn realistic image priors from a large number of example images.

Experiment

For key image restoration tasks(blind denoising, super-resolution, and inpainting), they used GCF-BM3D, Set5, Set14, and some other images for comparison. Empathize again, they do not use database to train, but use only a single degraded image to restore.

For each task, they handcrafted networks and energy functions to find efficient image prior. All results are impressive but PSNR is higher than learning based ConvNets at only inpainting applications.

My Thoughts

  • Although some restoration performances are lower than learning based approaches which are overfited to a specific inverse function, the results are very impressive.

  • In image super-resolution, recently there is another blind SR approach, called Zero Shot SRpaper. I'm curious about the comperision between ZSSR and deep image prior for unkown degraded images.

  • I think there are lots of future researches for different image priors. For example, in inverse text inpainting problem which is removing background pictures rather than text, which architectures make proper image priors for the non-naturally-looking images?


Jan. 18, 2018 Note by Heewon