PSNR SSIM LPIPS PSNR SSIM LPIPS
×2
73.8620 0.9032 0.0460 77.2794 0.9347 0.0377
×3
73.0892 0.9097 0.1131 73.0973 0.9146 0.1136
×4
69.4125 0.8130 0.2633 72.0820 0.8824 0.1971
Table 3 shows that content loss adopts the Resnet
assessment format, which considerably raises the
standard of picture creation. This research argues that
Resnet-based loss is preferable than VGG-based loss
because Resnet makes it simple for the network to
remember certain shallow feature information and
combine it with deep feature information through skip
connections.
4 CONCLUSION
Although the adversarial loss in ESRGAN may
enhance the super-resolution impact of the images, it
is also simple to produce the issue of network training
collapse since the task's complexity rises. As a result,
the loss function of the model is improved by this
study. First off, the training is stable and the picture
super-resolution effect is better based on the
adversarial loss suggested by WGAN. Second, while
the addition of TV Loss can enhance the effects of
picture super-resolution, its impact diminishes with
increasing job complexity and can even make it more
difficult to train a GAN network. Third, the content
loss that results from employing the Resnet network
for feature extraction as opposed to the VGG can
further enhance the effects of picture super-resolution.
The impact of Resnet content loss on picture super-
resolution weakens as the task's complexity rises,
although it still has a favorable impact. The effect of
image super-resolution degrades as the task's zoom
factor increases. In order to accomplish the effect of
magnifying the resolution by 4 times, a cascade
approach will be used to increase the resolution by 2
times sequentially twice. The experimental findings
demonstrate that the suggested enhancement approach
can successfully increase model performance. The
results of this study will be used to deploy the model
in the field in the future.
REFERENCES
Z. H. Wang, J. Chen, and C. H. Steven, “Deep learning for
image super-resolution: A survey, ” IEEE
transactions on pattern analysis and machine
intelligence, vol. 43, 2020, pp. 3365-3387.
Gribbon, T. Kim, and G. D. Bailey, “A novel approach to
real-time bilinear interpolation,” Proceedings.
Second IEEE international workshop on electronic
design, test and applications, IEEE, 2004, pp. 126-
131.
C. Ledig, et al. “Photo-realistic single image super-
resolution using a generative adversarial network,”
Proceedings of the IEEE conference on computer
vision and pattern recognition. 2017, pp. 4681-4690.
X. T. Wang, et al. “Esrgan: Enhanced super-resolution
generative adversarial networks,” Proceedings of the
European conference on computer vision (ECCV)
workshops, 2018.
Y. Choi, and P. Hanhoon, “Improving ESRGAN with an
additional image quality loss,” Multimedia Tools and
Applications, vol. 82, 2023, pp. 3123-3137.
Dataset https://data.vision.ee.ethz.ch/cvl/DIV2K/
M. Arjovsky, and B. Léon, “Towards principled methods
for training generative adversarial networks,” arXiv
2017, unpublished.
M. Arjovsky, S. Chintala, and B. Léon, “Wasserstein
generative adversarial networks,” International
conference on machine learning. PMLR, 2017, pp.
214-223.
T. Chan, et al. “Recent developments in total variation
image restoration,” Mathematical Models of
Computer Vision, vol. 17, 2005, pp. 17-31.
Wang, Zhou, et al. “Image quality assessment: from error
visibility to structural similarity,” IEEE transactions
on image processing, vol. 13, 2004, pp. 600-612.
R. Zhang, et al. “The unreasonable effectiveness of deep
features as a perceptual metric,” Proceedings of the
IEEE conference on computer vision and pattern
recognition, 2018, pp. 586-595.