
The addition of an encoding and decoding part with
latent space increases memory needs. In compari-
son, the original SinGAN structure, composed of only
convolution layers, needs less memory Shaham et al.
(2019). To reduce the memory requirement, a future
work is to lighten the encoding architecture and op-
timise the latent space. Instead of using the maximal
resolution in the second encoding convolutional layer,
an idea could be to update the shape according to the
level of resolution, which may accurate the results.
In future directions, other noise inputs may be
evaluated for the generative part. Indeed, in previous
research, we show an influence of the noise injection
according to a kind of acquisition Br
´
emond Martin
et al. (2022). Thus, it could be interesting to update
the noise injection according to the microscopic ac-
quisition considered and test if the result from a par-
ticular injection is still linked with the acquisition.
For the polyp dataset, the update of noise input may
reproduce particular saturation, the over/under expo-
sure of polyp topology during the imaging Ali et al.
(2020). Additionally, the generated image speculari-
ties may be evaluated to enhance the generated light
reflection on smooth objects. The loss function may
also be improved. As shown previously, it can im-
prove the contrast between the background and the
researched structureBr
´
emond Martin et al. (2022).
6 CONCLUSION
In this article we present Pair-GAN:
• A generative architecture based upon patch-
pyramidal auto-encoders;
• Taking in input a single pair of raw and GT
biomedical images;
• Which synthesise natural images, similar and in
the same statistical space as original pairs and
compared with state-of-the-art methods.
Such approach may be interesting to increase minimal
datasets to automate for instance the diagnosis grade
of a disease from a single image with deep learning
methods. An interesting perspective may be to verify
the grade of each generation from a single input pair
of images annotated with the grade.
REFERENCES
Ali, S., Zhou, F., Braden, B., Bailey, A., Yang, S., Cheng,
G., Zhang, P., Li, X., Kayser, M., Soberanis-Mukul,
R. D., et al. (2020). An objective comparison of detec-
tion and segmentation algorithms for artefacts in clin-
ical endoscopy. Scientific reports, 10(1):2748.
Arantes, R. B., Vogiatzis, G., and Faria, D. R. (2020). Rc-
gan: learning a generative model for arbitrary size
image generation. In Advances in Visual Comput-
ing: 15th International Symposium, ISVC 2020, San
Diego, CA, USA, October 5–7, 2020, Proceedings,
Part I 15, pages 80–94. Springer.
Borji, A. (2019). Pros and cons of gan evaluation measures.
Computer Vision and Image Understanding, 179:41–
65.
Br
´
emond-Martin, C., Simon-Chane, C., Clouchoux, C., and
Histace, A. (2023). Brain organoid data synthesis and
evaluation. Frontiers in Neuroscience, 17.
Br
´
emond Martin, C., Simon Ch
ˆ
ane, C., Clouchoux, C.,
and Histace, A. (2022). Aaegan loss optimizations
supporting data augmentation on cerebral organoid
bright-field images. In VISIGRAPP (4: VISAPP),
pages 307–314.
Chen, J., Xu, Q., Kang, Q., and Zhou, M. (2021a).
Mogan: Morphologic-structure-aware generative
learning from a single image. arXiv preprint
arXiv:2103.02997.
Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever,
I., and Abbeel, P. (2016). Infogan: Interpretable repre-
sentation learning by information maximizing genera-
tive adversarial nets. Advances in neural information
processing systems, 29.
Chen, X., Zhao, H., Yang, D., Li, Y., Kang, Q., and Lu,
H. (2021b). Sa-singan: self-attention for single-image
generation adversarial networks. Machine Vision and
Applications, 32:1–14.
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B.,
Warde-Farley, D., Ozair, S., Courville, A., and Ben-
gio, Y. (2014). Generative adversarial nets. Advances
in neural information processing systems, 27.
Gur, S., Benaim, S., and Wolf, L. (2020). Hierarchical patch
vae-gan: Generating diverse videos from a single sam-
ple. Advances in Neural Information Processing Sys-
tems, 33:16761–16772.
Hinz, T., Fisher, M., Wang, O., and Wermter, S. (2021). Im-
proved techniques for training single-image gans. In
Proceedings of the IEEE/CVF Winter Conference on
Applications of Computer Vision, pages 1300–1309.
Hong, D., Yao, J., Meng, D., Xu, Z., and Chanus-
sot, J. (2020). Multimodal gans: Toward cross-
modal hyperspectral–multispectral image segmenta-
tion. IEEE Transactions on Geoscience and Remote
Sensing, 59(6):5103–5113.
Iqbal, A., Sharif, M., Yasmin, M., Raza, M., and Aftab, S.
(2022). Generative adversarial networks and its appli-
cations in the biomedical image segmentation: a com-
prehensive survey. International Journal of Multime-
dia Information Retrieval, 11(3):333–368.
Jetchev, N., Bergmann, U., and Vollgraf, R. (2016). Texture
synthesis with spatial generative adversarial networks.
arXiv preprint arXiv:1611.08207.
Lindner, L., Narnhofer, D., Weber, M., Gsaxner, C.,
Kolodziej, M., and Egger, J. (2019). Using synthetic
training data for deep learning-based gbm segmenta-
tion. In 2019 41st Annual International Conference of
Pair-GAN: A Three-Validated Generative Model from Single Pairs of Biomedical and Ground Truth Images
47