Generating Scenery Images with Larger Variety According to User Descriptions

In this paper, a framework based on generative adversarial networks is proposed to perform nature-scenery generation according to descriptions from the users. The desired place, time and seasons of the generated scenes can be specified with the help of text-to-image generation techniques. The framew...

Full description

Saved in:
Bibliographic Details
Main Authors: Hsu-Yung Cheng, Chih-Chang Yu
Format: article
Language:EN
Published: MDPI AG 2021
Subjects:
T
Online Access:https://doaj.org/article/d4917a6f9c914dad8f72d117519c86a5
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In this paper, a framework based on generative adversarial networks is proposed to perform nature-scenery generation according to descriptions from the users. The desired place, time and seasons of the generated scenes can be specified with the help of text-to-image generation techniques. The framework improves and modifies the architecture of a generative adversarial network with attention models by adding the imagination models. The proposed attentional and imaginative generative network uses the hidden layer information to initialize the memory cell of the recurrent neural network to produce the desired photos. A data set containing different categories of scenery images is established to train the proposed system. The experiments validate that the proposed method is able to increase the quality and diversity of the generated images compared to the existing method. A possible application of road image generation for data augmentation is also demonstrated in the experimental results.