A Two-Stage Emotion Generation Model Combining CGAN and pix2pix

A Two-Stage Emotion Generation Model Combining CGAN and pix2pix

Yuanqing Wang (Zhoukou Normal University, China & City University, Malaysia), Dahlan Abdul Ghani (Malaysian Institute of Information Technology, University Kuala Lumpur, China), and Bingqian Zhou (Zhoukou Normal University, China & Universiti Malaysia Sarawak, Malaysia)
Copyright: © 2023 |Pages: 21
DOI: 10.4018/JOEUC.330647
Article PDF Download
Open access articles are freely available for download

Abstract

Computer vision has made significant advancements in emotional design. Designers can now utilize computer vision to create emotionally captivating designs that deeply resonate with people. This article aims at enhancing emotional design selection by separating appearance and color. A two-stage emotional design method is proposed, which yields significantly better results compared to classical single-stage methods.. In the Radboud face dataset (RaFD), facial expressions primarily rely on appearance, while color plays a relatively smaller role. Therefore, the two-stage model presented in this article can focus on shape design. By utilizing the SSIM image quality evaluation index, our model demonstrates a 31.63% improvement in generation performance compared to the CGAN model. Additionally, the PSNR image quality evaluation index shows a 10.78% enhancement in generation performance. The proposed model achieves superior design results and introduces various design elements.This article exhibits certain improvements in design effectiveness and scalability compared to conventional models.
Article Preview
Top

1. Introduction

AI-Generated Content (AIGC) (Liu, 2023; Kang, 2023) has made significant progress in content generation. Designers can now use AIGC to create emotionally captivating designs that resonate deeply with people. This paper utilizes deep learning technology to enhance the selection of emotional design through the separation of appearance and color. Deep learning technology can solve many problems, such as object detection, image classification, image segmentation (Ranjbarzadeh & Tirkolaee, 2023), and image generation, and has been widely applied in various industries (Weber, Arabnia, Aydın, Tirkolaee, 2023; Song, Li, et al., 2022; Wu, 2022; Morales & Suárez-Rocha, 2022; Dayyala, 2022). Deep learning (Miao & Ruomu, 2023; Zhou & Fang, 2022) is a machine learning subfield originating from the study of artificial neural networks. Currently, common neural networks used in deep learning include CNN, RNN, DNN, PNN, and others. These algorithms effectively model complex relationships within data through multi-layer representations. Unsupervised learning (Kumar, et al., 2021) has always posed a significant challenge for researchers. However, the rapid progress in unsupervised learning can be attributed to extensive research on generative models. These models can generate new samples based on the high-dimensional data distribution. In 2014, Ian Goodfellow proposed Generative Adversarial Networks (GANs) (Dalva, et al., 2022; Yang, et al., 2022; Kang, et al., 2023; Goodfellow, 2016; Li, et al., 2021; Zou & XiuFang, et al., 2022) which outperformed other generative models, producing better samples owing to the generation of countermeasures network. Compared to Pixel RNN (GANs) (van den Oord, et al., 2016; Salimans & Kingma, 2016), GAN generates a sample faster. In contrast to Variational Auto-Encode (VAE) (Sahu & Majumdar, 2021; Zhu, et al., 2021), a GAN does not have a lower bound of change, and its various adversarial generation networks gradually converge, while VAE exhibits some bias. Furthermore, a GAN does not possess a lower variation bound or a complex partition function like deep Boltzmann machines (Liu, et al., 2021; Luo, et al., 2021). GANs can generate samples simultaneously, eliminating the need for repeatedly applying the Markov chain solver (Karimi, et al., 2021; López-Suárez & Urriza, 2021). By comparison, it was found that GAN is more suitable for the application of this paper. Hence, this paper mainly focuses on generating content design, which requires diversification.

During the training process of a GAN, the generator network learns the distribution of real samples and determines whether an input image or real data is generated. The optimization of the entire model involves solving a “binary minimax game” problem. Currently, GANs represent the state-of-the-art in generative models. As researchers delve deeper into the realm of GANs, they have proposed several improved algorithms and innovative applications. For instance, the WGAN (Arjovsky, et al., 2017; Gulrajani, et al., 2017; Li, et al., 2021; Dong, et al., 2020; Xu, et al., 2021; Wang, et al., 2021) addresses the instability issues encountered during GAN training. DCGAN (Radford, et al., 2015; Deleforge, et al., 2021; Kim, et al., 2021) utilizes CNN in GAN to produce images of better quality. EBGAN (Chen, et al., 2016) treats the discriminator as an energy function, assigning low energy to actual data and high energy to fake data. On the other hand, InfoGAN (Liu & Lu, 2021; Gao, et al., 2021; Chen, et al., 2021) achieves decomposable feature representation through unsupervised learning. It combines a GAN with maximizing the mutual information between the generated image and the input code.

Complete Article List

Search this Journal:
Reset
Volume 37: 1 Issue (2025)
Volume 36: 1 Issue (2024)
Volume 35: 3 Issues (2023)
Volume 34: 10 Issues (2022)
Volume 33: 6 Issues (2021)
Volume 32: 4 Issues (2020)
Volume 31: 4 Issues (2019)
Volume 30: 4 Issues (2018)
Volume 29: 4 Issues (2017)
Volume 28: 4 Issues (2016)
Volume 27: 4 Issues (2015)
Volume 26: 4 Issues (2014)
Volume 25: 4 Issues (2013)
Volume 24: 4 Issues (2012)
Volume 23: 4 Issues (2011)
Volume 22: 4 Issues (2010)
Volume 21: 4 Issues (2009)
Volume 20: 4 Issues (2008)
Volume 19: 4 Issues (2007)
Volume 18: 4 Issues (2006)
Volume 17: 4 Issues (2005)
Volume 16: 4 Issues (2004)
Volume 15: 4 Issues (2003)
Volume 14: 4 Issues (2002)
Volume 13: 4 Issues (2001)
Volume 12: 4 Issues (2000)
Volume 11: 4 Issues (1999)
Volume 10: 4 Issues (1998)
Volume 9: 4 Issues (1997)
Volume 8: 4 Issues (1996)
Volume 7: 4 Issues (1995)
Volume 6: 4 Issues (1994)
Volume 5: 4 Issues (1993)
Volume 4: 4 Issues (1992)
Volume 3: 4 Issues (1991)
Volume 2: 4 Issues (1990)
Volume 1: 3 Issues (1989)
View Complete Journal Contents Listing