Comparing the histograms against the ones for the
initial test set scores in Figure 3, we can see that
our model performend better at improving the SSIM
scores on the damaged test images. Additionally, we
found that the VAE-based approach was highly dam-
aging to clean images compared to ours. This is ex-
plained partially by the colour shifts, but also through
the loss of fine detail and inability to remove larger
scratches, as shown in Figure 5 (first row). Addition-
ally, the Wan et al. method introduced checkerboard
artifacts to some examples Figure 5 (third row) and
struggled to reconstruct faces Figure 5 (second row).
In the case of signs and handwriting, no meaning-
ful differentiation is made between artifacts and lines
forming the letters, which results in failure to restore
writing Figure 5 (fourh and fifth row). On the other
hand, our model has successfully targeted only exist-
ing artifacts, and minimised the introduction of new
damage or loss of information.
5 CONCLUSION
The work presented in this paper demonstrates that
our approach achieves improved quality of restoration
at the task of automated dust and scratch removal for
analogue film scans when compared to state-of-the-
art. We adapt an architecture and training techniques
from the literature, and use those along with our per-
ceptual loss comprising of both exracted VGG16 fea-
ture activations and SSIM-based terms. By combin-
ing the learned natural prior of a pre-trained CNN-
based architecture with a perceptual quality metric
which targets image degradation in our loss formu-
lation, we allow the network to meaningfully differ-
entiate between dust and scratches and useful high-
frequency image features. Our model achieved bet-
ter SSIM scores compared to the VAE-based method
of Wan et al.; while this can be attributed to our ap-
proach explicitly optimising for SSIM during train-
ing, our qualitative results demonstrate that our ap-
proach is much more reliable in both restoring dust
and scratches, and preserving high frequency image
detail.
Additionally, we discuss a more comprehensive
approach to evaluating restoration quality, which also
includes measuring the information loss or new arti-
facts introduced by the restoration networks. We also
provide a data set of synthetically damaged slide film
scans to be used for benchmarking of the specific task
of dust and scratch removal for film.
As future work, we plan to collect a data set of
wild damaged film scans to evaluate our approach and
other existing approaches on real damaged input. Ad-
ditionally, when training on synthetic data where the
ground truth clean scan is available, we plan to ex-
plicitly incorporate our requirement that the network
should not damage clean inputs in the loss formula-
tion.
ACKNOWLEDGEMENTS
We thank Gerardo Aragon-Camarasa for his valuable
comments and feedback while preparing this paper.
This work was supported by the Engineering and
Physical Sciences Research Council (EPSRC) [grant
number EP/RS13222/1].
REFERENCES
Aitken, A., Ledig, C., Theis, L., Caballero, J., Wang, Z., and
Shi, W. (2017). Checkerboard artifact free sub-pixel
convolution: A note on sub-pixel convolution, resize
convolution and convolution resize. arXiv preprint
arXiv:1707.02937.
Antic, J. (2020). Deoldify: A deep learning based project
for colorizing and restoring old images (and video!).
Bousquet, O., Gelly, S., Tolstikhin, I., Simon-Gabriel, C.-J.,
and Schoelkopf, B. (2017). From optimal transport to
generative modeling: the vegan cookbook.
Chambah, M. (2019). Digital film restoration and image
quality. In ICA-BELGIUM Colour Symposium, Ghent,
Belgium.
Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-
Fei, L. (2009). ImageNet: A Large-Scale Hierarchical
Image Database. In CVPR09.
Fan, Z., Wu, H., Fu, X., Huang, Y., and Ding, X. (2018).
Residual-guide feature fusion network for single im-
age deraining. CoRR, abs/1804.07493.
Fielding, G. (2008). Digital ice: Defect detection and cor-
rection using infrared-enabled scanners.
Johnson, J., Alahi, A., and Fei-Fei, L. (2016). Perceptual
losses for real-time style transfer and super-resolution.
In European conference on computer vision, pages
694–711. Springer.
Kingma, D. P. and Ba, J. (2014). Adam: A
method for stochastic optimization. arXiv preprint
arXiv:1412.6980.
Ledig, C., Theis, L., Husz
´
ar, F., Caballero, J., Cunningham,
A., Acosta, A., Aitken, A., Tejani, A., Totz, J., Wang,
Z., et al. (2017). Photo-realistic single image super-
resolution using a generative adversarial network. In
Proceedings of the IEEE conference on computer vi-
sion and pattern recognition, pages 4681–4690.
Mao, X.-J., Shen, C., and Yang, Y.-B. (2016). Im-
age restoration using convolutional auto-encoders
with symmetric skip connections. arXiv preprint
arXiv:1606.08921.
VISAPP 2022 - 17th International Conference on Computer Vision Theory and Applications
134