簡易檢索 / 詳目顯示

研究生: 邱仁炎
Timotius Kuncoro
論文名稱: Latent Space Explorer in StyleGAN for Attribute Editing
Latent Space Explorer in StyleGAN for Attribute Editing
指導教授: 楊傳凱
Chuan-Kai Yang
口試委員: 賴源正
Yuan-Cheng Lai
林伯慎
Bor-Shen Lin
學位類別: 碩士
Master
系所名稱: 管理學院 - 資訊管理系
Department of Information Management
論文出版年: 2023
畢業學年度: 111
語文別: 英文
論文頁數: 62
外文關鍵詞: Generative Modelling, Latent Space Exploration
相關次數: 點閱:188下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報

  • The program made in this thesis, has a goal that is to do car attribute editing
    via latent vector exploration. The attribute that will be explored is the ambience of
    an image. After the latent space exploration is finished, the resulting latent vectors
    that are with more than 1 ambience will be saved and then shown in 2D data
    distribution using the Principal Component Analysis (PCA) method. The final
    product of this project is Python programs. Because of the time needed to do the
    exploration and save the latent vector the program is divided into 2 parts. First is
    the latent vector explorer and the second one is the program to show the saved latent
    vectors in terms of a 2D distribution.
    To achieve the goal of this study, this program used an unsupervised
    approach to generate images using Generative Adversarial Network and one of its
    extended versions called StyleGAN. Current techniques of unsupervised
    Generative Adversarial Network require a lot of high quality images to train the
    network and do the attribute editing. Because of that, in this thesis, a customized
    stanford cars dataset is used. The customization is done by adding ambiences in
    the training images. The evaluation of the exploration is done manually. For a better
    visualization of the resulting latent vectors, it is done in a 2D distribution of points.
    Even though the training and exploration took so much time, the StyleGAN
    technique produces convincing qualitative results from the dataset. After training
    and exploration, because the generating process of 1 image only takes a few
    seconds, it is possible to make interactive user-interface applications that show the
    saved latent vectors.

    ABSTRACT ..................................................................................................... III Acknowledgment..............................................................................................IV Table of Content ............................................................................................... V List of Figures .................................................................................................VII List of Table ...................................................................................................... X Chapter 1. Introduction..................................................................................... 1 1.1 Background ........................................................................................... 1 1.2 GAN & StyleGAN................................................................................. 3 1.3 Research Objectives & Scope ................................................................ 3 1.4 Thesis Organization ............................................................................... 4 Chapter 2. Literature Review............................................................................ 6 2.1. Learning representations from unlabeled data ........................................ 6 2.2. Generative Modelling Problem .............................................................. 6 2.3. Generative Adversarial Networks .......................................................... 8 2.4. Progressive Growing GAN .................................................................... 9 2.5. StyleGan.............................................................................................. 11 2.6. Latent Space Exploration ..................................................................... 13 Chapter 3. Methodology .................................................................................. 16 3.1 System Overview................................................................................. 16 3.2 Input and Output of the System............................................................ 18 3.3 Complete System Architecture............................................................. 22 3.4 Data Flow of the System...................................................................... 24 3.5 Actor ................................................................................................... 26 3.6 Preprocessing....................................................................................... 30 Chapter 4. Experimental Results .................................................................... 36 4.1 StyleGAN2 Training Experiments Results........................................... 36 4.2 Latent Space Explorer Experiments Results......................................... 40 4.3 Latent Space Visualizer Experiments Results....................................... 42 Chapter 5. Conclusion and Discussion............................................................ 44 5.1 Conclusion........................................................................................... 44 5.2 Limitation and Future Work................................................................. 45 References ........................................................................................................ 46

    [1]Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen,
    Timo Aila. “Analyzing and Improving the Image Quality of StyleGAN”. In
    CVPR, 2019.
    [2] David Bau, Jun-Yan Zhu, Hendrik Strobelt, Bolei Zhou, Joshua B.
    Tenenbaum, William T. Freeman, Antonio Torralba. “GAN Dissection:
    Visualizing and Understanding Generative Adversarial Networks”. In CVPR,
    2018.
    [3] Jiajun Wu, Chengkai Zhang, Tianfan Xue, William T. Freeman, Joshua B.
    Tenenbaum. “Learning a Probabilistic Latent Space of Object Shapes via 3D
    Generative-Adversarial Modeling”. In NIPS, 2016.
    [4] Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David
    Warde-Farley, Sherjil Ozair, Aaron Courville, Yoshua Bengio. “Generative
    Adversarial Nets”. In NIPS, 2014.
    [5] Alec Radford, Luke Metz, Soumith Chintala. “Unsupervised Representation
    Learning with Deep Convolutional Generative Adversarial Networks”. In
    CVPR, 2015.
    [6] Xiaolong Wang, Abhinav Shrivastava, Abhinav Gupta. “A-fast-rcnn: Hard
    positive generation via adversary for object detection”. In CVPR, 2017.
    [7] Geoffrey Hinton, Li Deng, Dong Yu, George E. Dahl, Abdel-rahman
    Mohamed, Navdeep Jaitly, Andrew Senior, Vincent Vanhoucke, Patrick
    Nguyen, Tara N. Sainath, and Brian Kingsbury. “Deep neural networks for
    acoustic modeling in speech recognition”. IEEE Signal Processing Magazine,
    29(6), 82–97.
    [8] Xintao Wang, Ke Yu, Shixiang Wu, Jinjin Gu, Yihao Liu, Chao Dong, Chen
    Change Loy, Yu Qiao, Xiaoou Tang. “ESRGAN: Enhanced Super-
    Resolution Generative Adversarial Networks ”. The European Conference on
    Computer Vision Workshops (ECCVW), 2018.
    [9] Tero Karras, Timo Aila, Samuli Laine, Jaakko Lehtinen. “Progressive
    Growing of GANs for Improved Quality, Stability, and Variation”. In ICLR,
    2018
    [10] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, Alexei A. Efros. “Image-to-
    Image Translation with Conditional Adversarial Networks”. In CVPR, 2017.
    [11] Jun-Yan Zhu, Taesung Park, Phillip Isola, Alexei A. Efros. “Unpaired
    image-to-image translation using cycle-consistent adversarial networks”. In
    ICCV, 2017.
    [12] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. “ImageNet
    Classification with Deep Convolutional Neural Networks”. In NIPS, 2012.
    [13] Ming-Yu Liu, Oncel Tuzel. “Coupled Generative Adversarial Networks”.
    In NIPS, 2016
    [14] Antti Rasmus, Harri Valpola, Mikko Honkala, Mathias Berglund, Tapani
    Raiko. “Semi-Supervised Learning with Ladder Networks”. In NIPS, 2015.
    [15] Michael Mathieu, Camille Couprie, Yann LeCun. “Deep multi-scale video
    prediction beyond mean square error”. In ICLR, 2016.
    [16] Javier Portilla & Eero P. Simoncelli. “A Parametric Texture Model Based
    on Joint Statistics of Complex Wavelet Coefficients”. In International Journal
    of Computer Vision, 2000.
    [17] Andrew Brock, Jeff Donahue, Karen Simonyan. “Large Scale GAN
    Training for High Fidelity Natural Image Synthesis”. In ICLR, 2019.
    [18] Adam Coates and Andrew Y. Ng. “Learning Feature Representations with
    K-Means”. In Neural Networks: Tricks of the Trade, pp. 561–580. Springer,
    2012.
    [19] Donggeun Yoo, Namil Kim, Sunggyun Park, Anthony S. Paek, In So
    Kweon. “Pixel-Level Domain Transfer”. In ECCV, 2016.
    [20] Aaron van den Oord, Nal Kalchbrenner, Koray Kavukcuoglu. “Pixel
    recurrent neural networks”. In CVPR, 2016.
    [21] Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, Pierre-
    Antoine Manzagol. “Stacked denoising autoencoders: Learning useful
    representations in a deep network with a local denoising criterion”. In The
    Journal of Machine Learning Research, 2010.
    [22] Carl Vondrick, Hamed Pirsiavash, Antonio Torralba. “Generating Videos
    with Scene Dynamics”. In CVPR, 2016.
    [23] Jascha Sohl-Dickstein, Eric A. Weiss, Niru Maheswaranathan, Surya
    Ganguli. “Deep Unsupervised Learning using Nonequilibrium
    Thermodynamics”. In ICML, 2015.
    [24] Junbo Zhao, Michael Mathieu, Ross Goroshin, Yann LeCun. “Stacked
    what-where autoencoders”. In ICLR, 2016.
    [25] Takeru Miyato, Toshiki Kataoka, Masanori Koyama, Yuichi Yoshida.
    “Spectral Normalization for Generative Adversarial Networks”. In ICLR,
    2018.
    [26] Brownlee, J. (2019) A gentle introduction to generative adversarial
    networks (Gans), MachineLearningMastery.com. Available at:
    https://machinelearningmastery.com/what-are-generative-adversarial-
    networks-gans/ (Accessed: November 20, 2022).
    [27] Han Zhang, Ian Goodfellow, Dimitris Metaxas, Augustus Odena. “Self-
    attention generative adversarial networks”. In arXiv:1805.08318, 2018.
    [28] Aaron van den Oord, Nal Kalchbrenner, Oriol Vinyals, Lasse Espeholt,
    Alex Graves, Koray Kavukcuoglu. “Conditional image generation with
    PixelCNN decoders”. In CVPR, 2016.
    [29] Tero Karras, Miika Aittala, Janne Hellsten, Samuli Laine, Jaakko Lehtinen,
    Timo Aila. “Training Generative Adversarial Networks with Limited Data”.
    In CVPR, 2020.
    [30] Diederik P Kingma, Max Welling. “Auto-encoding variational bayes”. In
    CoRR, 2014.
    [31] Emily Denton, Soumith Chintala, Arthur Szlam, Rob Fergus. “Deep
    generative image models using a Laplacian pyramid of adversarial networks”.
    In NIPS, 2015.
    [32] Honglak Lee, Roger Grosse, Rajesh Ranganath, Andrew Y. Ng.
    “Convolutional Deep Belief Networks for Scalable Unsupervised Learning of
    Hierarchical Representations”. In ICML, 2009.
    [33] Susskind, J.M., Anderson, A.K. and Hinton, G.E., 2010. The toronto face
    database. Department of Computer Science, University of Toronto, Toronto,
    ON, Canada, Tech. Rep, 3, p.29.
    [34] Brownlee, J. (2019) A gentle introduction to the progressive growing gan,
    MachineLearningMastery.com. Available at:
    https://machinelearningmastery.com/introduction-to-progressive-growing-
    generative-adversarial-networks/ (Accessed: November 21, 2022).
    [35] Tero Karras, Samuli Laine, Timo Aila. “A Style-Based Generator
    Architecture for Generative Adversarial Networks”. In CVPR, 2019.
    [36] Ting-Chun Wang, Ming-Yu Liu, Jun-Yan Zhu, Guilin Liu, Andrew Tao,
    Jan Kautz, Bryan Catanzaro. “Video-to-video synthesis”. In NIPS, 2018.
    [37] Brandon Frey. (no date) Frey Faces Dataset. Available at:
    https://cs.nyu.edu/~roweis/data.html (Accessed: November 20, 2022).
    [38] Krizhevsky, A. “Learning multiple layers of features from tiny images”.
    2009.
    [39] Judy Hoffman, Eric Tzeng, Taesung Park, Jun-Yan Zhu, Phillip Isola, Kate
    Saenko, Alexei A. Efros, Trevor Darrell. “CyCADA: Cycle-Consistent
    Adversarial Domain Adaptation”. In CVPR, 2017.
    [40] Chia-Hsing Chiu, Yuki Koyama, Yu-Chi Lai, Takeo Igarashi, Yonghao
    Yue. “Human-in-the-Loop Differential Subspace Search in High-
    Dimensional Latent Space”. In ACM Transactions on Graphics, 2020.
    [41] Brownlee, J. (2020) How to explore the gan latent space when generating
    faces, MachineLearningMastery.com. Available at:
    https://machinelearningmastery.com/how-to-interpolate-and-perform-
    vector-arithmetic-with-faces-using-a-generative-adversarial-network/
    (Accessed: December 14, 2022).
    [42] Yujun Shen, Jinjin Gu, Xiaoou Tang, Bolei Zhou. “Interpreting the Latent
    Space of GANs for Semantic Face Editing”. In CVPR, 2020.
    [43] Erik Härkönen, Aaron Hertzmann, Jaakko Lehtinen, Sylvain Paris.
    “GANSpace: Discovering Interpretable GAN Controls”. In NeurIPS, 2020.
    [44] Rameen Abdal, Peihao Zhu, Niloy J. Mitra, Peter Wonka. “StyleFlow:
    Attribute-conditioned Exploration of StyleGAN-Generated Images using
    Conditional Continuous Normalizing Flows”. In ACM Transactions on
    Graphics, 2021.
    [45] Yujun Shen, Ceyuan Yang, Xiaoou Tang, Bolei Zhou. “InterFaceGAN:
    Interpreting the Disentangled Face Representation Learned by GANs”. In
    TPAMI, 2020.
    [46] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler,
    Sepp Hochreiter. “GANs Trained by a Two Time-Scale Update Rule
    Converge to a Local Nash Equilibrium”. In NIPS, 2017.
    [47] Kaipeng Zhang, Zhanpeng Zhang, Zhifeng Li, Yu Qiao. “Joint Face
    Detection and Alignment using Multi-task Cascaded Convolutional
    Networks”. In IEEE Signal Processing Letters, 2016.
    [48] Ian Goodfellow. “Tutorial: Generative Adversarial Networks”. In NIPS,
    2016.
    [49] Square root of a matrix (2022) Wikipedia. Wikimedia Foundation.
    Available at: https://en.wikipedia.org/wiki/Square_root_of_a_matrix
    (Accessed: December 16, 2022).
    [50] Covariance matrix (2022) Wikipedia. Wikimedia Foundation. Available
    at: https://en.wikipedia.org/wiki/Covariance_matrix (Accessed: December
    16, 2022).
    [51] Bluewidz (no date) Fréchet Inception Distance. Available at:
    https://bluewidz.blogspot.com/2017/12/frechet-inception-distance.html
    (Accessed: December 16, 2022).
    [52] D. C. Dowson and B. V. Landau. “The Fréchet distance between
    multivariate normal distributions”. In Journal of Multivariate Analysis,
    1982.
    [53] Fréchet distance (2022) Wikipedia. Wikimedia Foundation. Available at:
    https://en.wikipedia.org/wiki/Fr%C3%A9chet_distance (Accessed:
    December 16, 2022).
    [54] Ali Borji. “Pros and Cons of GAN Evaluation Measures”. In CVPR, 2018.
    [55] Mario Lucic, Karol Kurach, Marcin Michalski, Sylvain Gelly, Olivier
    Bousquet. “Are GANs Created Equal? A Large-Scale Study”. In NIPS,
    2018.
    [56] SKLEARN.DECOMPOSITION.PCA (no date) scikit. Available at:
    https://scikit-
    learn.org/stable/modules/generated/sklearn.decomposition.PCA.html
    (Accessed: December 18, 2022).
    [57] Sklearn.manifold.TSNE (no date) scikit. Available at: https://scikit-
    learn.org/stable/modules/generated/sklearn.manifold.TSNE.html (Accessed: December 18, 2022).

    QR CODE