論文

査読有り 責任著者 国際誌
2022年1月9日

Extending HoloGAN by Embedding Image Content into Latent Vectors for Novel View Synthesis

Proceedings of 2022 IEEE/SICE International Symposium on System Integration (SII 2022)
  • Jing Wang
  • ,
  • Lotfi El Hafi
  • ,
  • Akira Taniguchi
  • ,
  • Yoshinobu Hagiwara
  • ,
  • Tadahiro Taniguchi

開始ページ
383
終了ページ
389
記述言語
英語
掲載種別
研究論文(国際会議プロシーディングス)
DOI
10.1109/SII52469.2022.9708823
出版者・発行元
Institute of Electrical and Electronics Engineers (IEEE)

This study aims to further develop the task of novel view synthesis by generative adversarial networks (GAN). The goal of novel view synthesis is to, given one or more input images, synthesize images of the same target content but from different viewpoints. Previous research showed that the unsupervised learning model HoloGAN achieved high performance in generating images from different viewpoints. However, HoloGAN is less capable of specifying the target content to generate and is difficult to train due to high data requirements. Therefore, this study proposes two approaches to improve the current limitations of HoloGAN and make it suitable for the task of novel view synthesis. The first approach reuses the encoder network of HoloGAN to get the corresponding latent vectors of the image contents to specify the target content of the generated images. The second approach introduces an auto-encoder architecture to HoloGAN so that more viewpoints can be generated correctly. The experiment results indicate that the first approach is efficient in specifying a target content. Meanwhile, the second approach method helps HoloGAN to learn a richer range of viewpoints but is not compatible with the first approach. The combination of these two approaches and their application to service robotics are discussed in conclusion.

リンク情報
DOI
https://doi.org/10.1109/SII52469.2022.9708823
ID情報
  • DOI : 10.1109/SII52469.2022.9708823

エクスポート
BibTeX RIS