簡易檢索 / 詳目顯示

研究生: Maynard John C. Si
Maynard John C. Si
論文名稱: Perspective Preserving Style Transfer for Interior Portraits
Perspective Preserving Style Transfer for Interior Portraits
指導教授: 花凱龍
Kai-Lung Hua
口試委員: 楊傳凱
Chuan-Kai Yang
陳駿丞
Jun-Cheng Chen
學位類別: 碩士
Master
系所名稱: 電資學院 - 資訊工程系
Department of Computer Science and Information Engineering
論文出版年: 2020
畢業學年度: 108
語文別: 英文
論文頁數: 40
中文關鍵詞: Non-photorealistic renderingStyle Transfer
外文關鍵詞: Non-photorealistic rendering, Style Transfer
相關次數: 點閱:227下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • One of the jarring limitations of existing style transfer techniques is its failure
    to capture the illusion of depth through perspective. These often result
    in flat looking images as style elements are simply distributed across the image.
    Though recent methods attempt to alleviate this by considering depth
    information for a distinct stylization between foreground and background,
    they still fail to capture the perspective of an image. When used on interior
    portraits where perspective is instinctively observed through its surfaces
    (walls, ceiling, floor), previous methods cause unwanted styling such as
    style elements distorting boundaries of surfaces and style elements not receding
    according to the perspective of the surfaces. In this paper, we developed
    a simple approach to effectively preserve the perspective of interior
    portraits during style transfer, yielding stylized images that distributes and
    warps style elements according to the perspective of the interior surfaces.
    Compared to existing methods, our approach generated images having style
    elements receding towards the vanishing point of its respective surface. We
    also observe that our approach was able to preserve depth information for
    some styles despite not extracting this from the content.


    One of the jarring limitations of existing style transfer techniques is its failure
    to capture the illusion of depth through perspective. These often result
    in flat looking images as style elements are simply distributed across the image.
    Though recent methods attempt to alleviate this by considering depth
    information for a distinct stylization between foreground and background,
    they still fail to capture the perspective of an image. When used on interior
    portraits where perspective is instinctively observed through its surfaces
    (walls, ceiling, floor), previous methods cause unwanted styling such as
    style elements distorting boundaries of surfaces and style elements not receding
    according to the perspective of the surfaces. In this paper, we developed
    a simple approach to effectively preserve the perspective of interior
    portraits during style transfer, yielding stylized images that distributes and
    warps style elements according to the perspective of the interior surfaces.
    Compared to existing methods, our approach generated images having style
    elements receding towards the vanishing point of its respective surface. We
    also observe that our approach was able to preserve depth information for
    some styles despite not extracting this from the content.

    Abstract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iii Acknowledgements . . . . . . . . . . . . . . . . . . . . . . . . . . iv Contents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . v List of Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii 1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 2 Review of Related Literature . . . . . . . . . . . . . . . . . . . 5 3 Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 3.1 Interior Surface Keypoint Extraction . . . . . . . . . . . . 8 3.1.1 Keypoint Extraction Network . . . . . . . . . . . 9 3.1.2 Selecting New Surface Keypoints . . . . . . . . . 11 3.2 Interior Surface Map Construction . . . . . . . . . . . . . 13 3.3 Style Transfer . . . . . . . . . . . . . . . . . . . . . . . . 15 3.3.1 Perspective Image Reconstruction . . . . . . . . . 17 4 Results and Discussion . . . . . . . . . . . . . . . . . . . . . . 18 4.1 Implementation Details . . . . . . . . . . . . . . . . . . . 18 4.2 Perspective Preservation . . . . . . . . . . . . . . . . . . 19 4.3 Depth Map Comparison . . . . . . . . . . . . . . . . . . . 20 4.4 Stylization Quality Comparison . . . . . . . . . . . . . . 22 5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 5.1 Future Work . . . . . . . . . . . . . . . . . . . . . . . . . 27 References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

    [1] L. A. Gatys, A. S. Ecker, and M. Bethge, “Image style transfer using convolutional neural networks,”
    in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2414–2423,
    2016.
    [2] X. Huang and S. Belongie, “Arbitrary style transfer in realtime
    with adaptive instance normalization,”
    in Proceedings of the IEEE International Conference on Computer Vision, pp. 1501–1510, 2017.
    [3] K. Kozlovtsev and V. Kitov, “DepthPreserving
    RealTime
    arbitrary style transfer,” June 2019.
    [4] F. Yu, Y. Zhang, S. Song, A. Seff, and J. Xiao, “Lsun: Construction of a largescale
    image dataset
    using deep learning with humans in the loop,” arXiv preprint arXiv:1506.03365, 2015.
    [5] C.Y.
    Lee, V. Badrinarayanan, T. Malisiewicz, and A. Rabinovich, “Roomnet: Endtoend
    room layout
    estimation,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 4865–4874,
    2017.
    [6] Facebook, Inc., “Instagram: A simple, fun & creative way to capture, edit & share photos, videos &
    messages with friends & family.” https://www.instagram.com/.
    [7] Snap, Inc., “Snapchat: Snapchat lets you easily talk with friends, view live stories from around the
    world, and explore news in discover. life’s more fun when you live in the moment!.” https://www.
    snapchat.com/.
    [8] R. Poplin and A. Prins, “Behind the scenes with stadia’s style transfer ML.” https://stadia.dev/
    blog/behind-the-scenes-with-stadias-style-transfer-ml/.
    [9] J. A. Paradiso and J. A. Landay, “Guest editors’ introduction: Crossreality
    environments,” IEEE
    Pervasive Computing, vol. 8, no. 3, pp. 14–15, 2009.
    [10] J. Johnson, A. Alahi, and L. FeiFei,
    “Perceptual losses for RealTime
    style transfer and SuperResolution,”
    in Computer Vision – ECCV 2016, pp. 694–711, Springer International Publishing, 2016.
    [11] L. A. Gatys, A. S. Ecker, M. Bethge, A. Hertzmann, and E. Shechtman, “Controlling perceptual factors
    in neural style transfer,” in Proceedings of the IEEE Conference on Computer Vision and Pattern
    Recognition, pp. 3985–3993, 2017.
    [12] X. Wang, G. Oxholm, D. Zhang, and Y.F.
    Wang, “Multimodal transfer: A hierarchical deep convolutional
    neural network for fast artistic style transfer,” 2017.
    [13] X.C.
    Liu, M.M.
    Cheng, Y.K.
    Lai, and P. L. Rosin, “Depthaware
    neural style transfer,” in Proceedings
    of the Symposium on NonPhotorealistic
    Animation and Rendering, NPAR ’17, (New York, NY,
    USA), pp. 4:1–4:10, ACM, 2017.
    [14] Y. Li, C. Fang, J. Yang, Z. Wang, X. Lu, and M.H.
    Yang, “Universal style transfer via feature transforms,”
    in Advances in neural information processing systems, pp. 386–396, 2017.
    [15] Y. Jing, Y. Liu, Y. Yang, Z. Feng, Y. Yu, D. Tao, and M. Song, “Stroke controllable fast style transfer
    with adaptive receptive fields,” in Proceedings of the European Conference on Computer Vision
    (ECCV), pp. 238–254, 2018.
    [16] Y. Yao, J. Ren, X. Xie, W. Liu, Y.J.
    Liu, and J. Wang, “Attentionaware
    multistroke
    style transfer,”
    in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1467–1475,
    2019.
    [17] K. Andersen, The Geometry of an Art: The History of the Mathematical Theory of Perspective from
    Alberti to Monge. Springer Science & Business Media, Nov. 2008.
    [18] A. A. Efros and T. K. Leung, “Texture synthesis by nonparametric
    sampling,” in Proceedings of the
    Seventh IEEE International Conference on Computer Vision, vol. 2, pp. 1033–1038 vol.2, Sept. 1999.
    [19] V. Kwatra, I. Essa, A. Bobick, and N. Kwatra, “Texture optimization for examplebased
    synthesis,”
    ACM Trans. Graph., vol. 24, pp. 795–802, July 2005.
    [20] A. Hertzmann, C. E. Jacobs, N. Oliver, B. Curless, and D. H. Salesin, “Image analogies,” in Proceedings
    of the 28th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH
    ’01, (New York, NY, USA), pp. 327–340, ACM, 2001.
    [21] Y. Li, N. Wang, J. Liu, and X. Hou, “Demystifying neural style transfer,” Jan. 2017.
    [22] A. Selim, M. Elgharib, and L. Doyle, “Painting style transfer for head portraits using convolutional
    neural networks,” ACM Trans. Graph., vol. 35, pp. 129:1–129:18, July 2016.
    [23] J. Yaniv, Y. Newman, and A. Shamir, “The face of art: Landmark detection and geometric style in
    portraits,” ACM Trans. Graph., vol. 38, pp. 60:1–60:15, July 2019.
    [24] J. D’Amelio, Perspective Drawing Handbook. Courier Corporation, May 2004.
    [25] J. M. Coughlan and A. L. Yuille, “The manhattan world assumption: Regularities in scene statistics
    which enable bayesian inference,” in Advances in Neural Information Processing Systems 13 (T. K.
    Leen, T. G. Dietterich, and V. Tresp, eds.), pp. 845–851, MIT Press, 2001.
    [26] K. Simonyan and A. Zisserman, “Very deep convolutional networks for LargeScale
    image recognition,”
    Sept. 2014.
    [27] W. Chen, Z. Fu, D. Yang, and J. Deng, “Singleimage
    depth perception in the wild,” in Advances in
    neural information processing systems, pp. 730–738, 2016.

    無法下載圖示 全文公開日期 2025/02/06 (校內網路)
    全文公開日期 2025/02/06 (校外網路)
    全文公開日期 2025/02/06 (國家圖書館:臺灣博碩士論文系統)
    QR CODE