簡易檢索 / 詳目顯示

研究生: 陳致廷
Zhi-Ting Chen
論文名稱: 以循環一致性和條件投影生成對抗網路進行之人臉年齡轉換
Cycle Consistency and Conditional Projection GAN for Facial Age Transformation
指導教授: 徐繼聖
Gee-Sern Hsu
口試委員: 賴尚宏
Shang-Hong Lai
王鈺強
Yu-Chiang Wang
邱維辰
Wei-Chen Chiu
郭景明
Jing-Ming Guo
徐繼聖
Gee-Sern Hsu
學位類別: 碩士
Master
系所名稱: 工程學院 - 機械工程系
Department of Mechanical Engineering
論文出版年: 2021
畢業學年度: 109
語文別: 中文
論文頁數: 67
中文關鍵詞: 生成對抗網路人臉年齡轉換人臉識別
外文關鍵詞: Generative Adversarial Network, Facial Age Transformation, Face Recognition
相關次數: 點閱:289下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報

我們提出了用於人臉年齡轉換的Cycle Consistency and Conditional Projection GAN (CCCP-GAN)。提出的 CCCP-GAN 由編碼器、解碼器和判別器組成,其中編碼器將帶有目標年齡組標籤的輸入圖像轉換為身份潛在編碼和年齡潛在編碼,並將最新合成網絡StyleGAN2的解碼器進行修改,將身份潛在代碼和年齡潛在代碼作為輸入,生成保留輸入身份並處於目標年齡的人臉。判別器由提出的條件多任務投影 (CMP) 組成,以更好地在生成的圖像上生成所需的年齡特徵,從而在單一模型中實現年齡進展和回歸。最後,我們利用三元組損失和循環損失來保留輸入身份並在輸出處獲得所需的年齡分佈。在本論文的新穎之處包括: 1. 將年齡特徵與所需年齡組標籤融合的 CMP 判別器,能夠有效監督單個生成器執行年齡進展和回歸; 2. 對最新風格轉移網路施加循環一致性,以在單一框架內實現保持身分的臉部年齡轉換; 3. 與最先進的方法相比,生成具有更好性能的高分辨率年齡轉移圖像; 4. 提出了一種新的協議,用於使用新的指標和基準來評估跨大年齡差距的年齡轉變。


We propose the Cycle Consistency and Conditional Projection GAN for facial age transformation. The CCCP-GAN is composed of an encoder, a decoder and a discriminator. The encoder converts an input image with a target age group label to an identity latent code and an age latent code. The decoder, developed on the StyleGAN2 synthesis network, takes the identity latent code and the age latent code as inputs to generate a face which preserves the input identity and is in the target age group. The discriminator is made of the proposed Conditional Multitask Projection (CMP) to better generate the desired age traits on the generated images, enabling both age progression and regression in a unified model. We exploit the triplet loss and cycle loss to preserve the input identity and attain the desired age distribution at the output. The novelties of this study include the following: 1. The CMP discriminator that fuses age features with the desired age group labels able to effectively supervise a single generator to perform age progression and regression; 2. A cycle consistency imposed on a latest style transfer network for achieving identity-preserving facial age transformation in a unified framework; 3. Generation of high-resolution age-transferred images with a better performance compared to state-of-the-art approaches; 4. A new protocol proposed for evaluating the age transformation across large age gaps with a novel metrics and benchmarks.

摘要 II Abstract III 誌謝 IV 目錄 V 圖目錄 VII 表目錄 IX 第1章 介紹 1 1.1 研究背景和動機 1 1.2 方法概述 2 1.3 論文貢獻 3 1.4 論文架構 4 第2章 文獻回顧 5 2.1 Progressive Growing of GANs 5 2.2 A Style-based GAN Architecture 6 2.3 Analyzing and Improving the Image Quality of StyleGAN 8 2.4 Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks 12 2.5 FaceNet: A Unified Embedding for Face Recognition and Clustering 13 2.6 cGANs with Projection Discriminator 13 第3章 主要方法 15 3.1 生成器設計 16 3.2 判別器設計 17 3.3 整體網路架構 18 第4章 實驗設置與分析 21 4.1 資料庫介紹 21 4.1.1 MORPH Database 21 4.1.2 Cross-Age Celebrity Database 22 4.1.3 Flickr-Faces-HQ-Aging 23 4.1.4 FG-Net Dataset 24 4.1.5 Ageing-Celebrity-Benchmark 25 4.2 實驗設置 27 4.2.1 資料劃分、設置 27 4.2.2 網路架構設置 28 4.2.3 效能評估方式 30 4.3 實驗結果與分析 32 4.3.1 不同的身分編碼設置比較 32 4.3.2 判別器和分類器設置比較 35 4.3.3 損失函數設置比較 37 4.3.4 三元組損失函數設置比較 39 4.4 與相關文獻之效能比較 41 4.4.1 一般年齡數據集之比較 42 4.4.2 跨大年齡數據集之比較 46 第5章 結論與未來研究方向 51 第6章 參考文獻 52

[1] Zhang, Zhifei, Yang Song, and Hairong Qi. "Age progression/regression by conditional adversarial autoencoder." Proceedings of the IEEE conference on computer vision and pattern recognition. 2017.
[2] Li, Peipei, et al. "Global and local consistent age generative adversarial networks." 2018 24th International Conference on Pattern Recognition (ICPR). IEEE, 2018.
[3] Yang, Hongyu, et al. "Learning face age progression: A pyramid architecture of gans." Proceedings of the IEEE conference on computer vision and pattern recognition. 2018.
[4] Liu, Yunfan, Qi Li, and Zhenan Sun. "Attribute-aware face aging with wavelet-based generative adversarial networks." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019.
[5] Fu, Yun, Guodong Guo, and Thomas S. Huang. "Age synthesis and estimation via faces: A survey." IEEE transactions on pattern analysis and machine intelligence 32.11 (2010): 1955-1976.
[6] Shen, Yujun, et al. "Interpreting the latent space of gans for semantic face editing." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020.
[7] Yang, Hongyu, et al. "Learning continuous face age progression: A pyramid of gans." IEEE transactions on pattern analysis and machine intelligence (2019).
[8] Karras, Tero, et al. "Analyzing and improving the image quality of stylegan." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020.
[9] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in ICCV, 2017
[10] F. Schroff, D. Kalenichenko, and J. Philbin, “Facenet: A unified embedding for face recognition and clustering,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 815–823
[11] Huang, Xun, and Serge Belongie. "Arbitrary style transfer in real-time with adaptive instance normalization." Proceedings of the IEEE International Conference on Computer Vision. 2017.
[12] Karras, Tero, Samuli Laine, and Timo Aila. "A style-based generator architecture for generative adversarial networks." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019.
[13] A. Karnewar and O. Wang, “Msg-gan: Multi-scale gradients for generative adversarial networks,” in CVPR, 2020.
[14] E. L. Denton, S. Chintala, R. Fergus et al., “Deep generative image models using a laplacian pyramid of adversarial networks,” in NIPS, 2015
[15] A. Odena, C. Olah, and J. Shlens, “Conditional image synthesis with auxiliary classifier gans,” in ICML, 2017.
[16] S. Reed, Z. Akata, X. Yan, L. Logeswaran, B. Schiele, and H. Lee, “Generative adversarial text to image synthesis,” arXiv preprint arXiv:1605.05396, 2016
[17] H. Zhang, T. Xu, H. Li, S. Zhang, X. Wang, X. Huang, and D. N. Metaxas, “Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks,” in ICCV, 2017.
[18] T. Kim, M. Cha, H. Kim, J. K. Lee, and J. Kim, “Learning to discover cross-domain relations with generative adversarial networks,” arXiv preprint arXiv:1703.05192, 2017.
[19] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” arXiv preprint arXiv:1411.1784, 2014.
[20] V. Dumoulin, I. Belghazi, B. Poole, O. Mastropietro, A. Lamb, M. Arjovsky, and A. Courville, “Adversarially learned inference,” arXiv preprint arXiv:1606.00704, 2016.
[21] K. Sricharan, R. Bala, M. Shreve, H. Ding, K. Saketh, and J. Sun, “Semi-supervised conditional gans,” arXiv preprint arXiv:1708.05789, 2017.
[22] T. Miyato and M. Koyama, “cgans with projection discriminator,” arXiv preprint arXiv:1802.05637, 2018.
[23] Z. Wang, X. Tang, W. Luo, and S. Gao, “Face aging with identity-preserved conditional generative adversarial networks,” in CVPR, 2018.
[24] R. Or-El, S. Sengupta, O. Fried, E. Shechtman, and I. Kemelmacher-Shlizerman, “Lifespan age transformation synthesis,” in ECCV, 2020.
[25] K. Ricanek and T. Tesafaye, “Morph: A longitudinal image database of normal adult age-progression,” in FG, 2006.
[26] B.-C. Chen, C.-S. Chen, and W. H. Hsu, “Face recognition and retrieval using cross-age reference coding with cross-age celebrity dataset,” TMM, 2015.
[27] M. Inc., “Face++ research toolkit,” http://www.faceplusplus. com.
[28] H. Pan, H. Han, S. Shan, and X. Chen, “Mean-variance loss for deep age estimation from a face,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 5285–5294.
[29] A. Bulat and G. Tzimiropoulos, “How far are we from solving the 2d & 3d face alignment problem? (and a dataset of 230,000 3d facial landmarks),” in ICCV, 2017.
[30] Z. He, M. Kan, S. Shan, and X. Chen, “S2gan: Share aging factors across ages and share aging trends among individuals,” in ICCV, 2019.
[31] L. Tran, X. Yin, and X. Liu, “Disentangled representation learning gan for pose-invariant face recognition,” in CVPR, 2017.
[32] J. Deng, J. Guo, N. Xue, and S. Zafeiriou, “Arcface: Additive angular margin loss for deep face recognition,” in CVPR, 2019.
[33] Karras, Tero, et al. "Progressive growing of gans for improved quality, stability, and variation." arXiv preprint arXiv:1710.10196 (2017).
[34] LeCun, Yann, Corinna Cortes, and C. J. Burges. "MNIST handwritten digit database." (2010): 18.
[35] Panis, Gabriel, et al. "Overview of research on facial ageing using the FG-NET ageing database." Iet Biometrics 5.2 (2016): 37-46.

QR CODE