Generating Videos from Stories Using Conditional GAN

Takahiro Kozaki, Fumihiko Sakaue, Jun Sato

2024

Abstract

In this paper, we propose a method for generating videos that represent stories described in multiple sentences. While research on generating images and videos from single sentences has been advancing, the generation of videos from long stories written in multiple sentences has not been achieved. In this paper, we use adversarial learning to train pairs of multi-sentence stories and videos to generate videos that replicate the flow of the stories. We also introduce caption loss for generating more contextually aligned videos from stories.

Download


Paper Citation


in Harvard Style

Kozaki T., Sakaue F. and Sato J. (2024). Generating Videos from Stories Using Conditional GAN. In Proceedings of the 19th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications - Volume 2: VISAPP; ISBN 978-989-758-679-8, SciTePress, pages 671-677. DOI: 10.5220/0012405300003660


in Bibtex Style

@conference{visapp24,
author={Takahiro Kozaki and Fumihiko Sakaue and Jun Sato},
title={Generating Videos from Stories Using Conditional GAN},
booktitle={Proceedings of the 19th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications - Volume 2: VISAPP},
year={2024},
pages={671-677},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0012405300003660},
isbn={978-989-758-679-8},
}


in EndNote Style

TY - CONF

JO - Proceedings of the 19th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications - Volume 2: VISAPP
TI - Generating Videos from Stories Using Conditional GAN
SN - 978-989-758-679-8
AU - Kozaki T.
AU - Sakaue F.
AU - Sato J.
PY - 2024
SP - 671
EP - 677
DO - 10.5220/0012405300003660
PB - SciTePress