Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2023, Vol. 44 ›› Issue (4): 710-717.DOI: 10.11996/JG.j.2095-302X.2023040710

Previous Articles     Next Articles

Landscape image generation based on conditional residual generative adversarial network

SHAO Jun-qi(), QIAN Wen-hua(), XU Qi-hao   

  1. Department of Computer Science Engineering, School of Information Science and Engineering, Yunnan University, Kunming Yunnan 650504, China
  • Received:2022-11-17 Accepted:2023-02-21 Online:2023-08-31 Published:2023-08-16
  • Contact: QIAN Wen-hua (1980-), professor, Ph.D. His main research interests cover graphic image processing and computer vision, etc. E-mail:whqian@ynu.edu.cn
  • About author:

    SHAO Jun-qi (1997-), master student. His main research interests cover computer vision and image generation. E-mail:619702616@qq.com

  • Supported by:
    National Natural Science Foundation of China(62162065);Key Project of Applied Basic Research Plan of Yunnan Provincial Department of Science and Technology(2019FA044);Yunnan Young and Middle-Aged Academic and Technical Leaders Reserve Talents Project(2019HB121);Postgraduate Research and Innovation Foundation of Yunnan University(ZC-22222502)

Abstract:

The semantic segmentation map of landscape image encompasses a large number of categorical information such as the sky, white clouds, mountains, rivers, and trees. In view of the challenges presented by the numerous information categories in the semantic segmentation map and the subtle color transformations between different regions, the landscape images generated by current methods are deficient in terms of both clarity and authenticity. Consequently, a method based on conditional residual generation adversarial network (CRGAN) was proposed to generate landscape images with a higher resolution and more realistic content. Firstly, the proposed method involved the upsampling and downsampling structures of the generator network to enhance the feature extraction effect of the generator on the semantic segmentation graph. Secondly, skip connections were utilized between the encoder and decoder to transmit the feature information from the semantic segmentation graph, ensuring the integrity of such information was retained, and not lost in the encoder. Finally, a residual module was added between the encoder and decoder of the network, facilitating better extraction, transmission, and retention of semantic information. In addition, the mean square error (MSE) was employed to enhance the similarity between semantically segmented graphs and generated images. The experimental results demonstrated that compared with pix2pix and cyclegan methods, the FID index of images generated by CRGAN increased by 26.769 and 119.333, respectively. This improvement effectively enhanced the clarity and authenticity of landscape images. The universality and validity of CRGAN were also validated using a common dataset.

Key words: generative adversarial network, landscape image, image generation, deep learning, clarity

CLC Number: