2018年2月20日
Extracting key frames from first-person videos in the common space of multiple sensors
Proceedings - International Conference on Image Processing, ICIP
- ,
- ,
- ,
- ,
- 巻
- 2017-
- 号
- 開始ページ
- 3993
- 終了ページ
- 3997
- 記述言語
- 英語
- 掲載種別
- 研究論文(国際会議プロシーディングス)
- DOI
- 10.1109/ICIP.2017.8297032
- 出版者・発行元
- IEEE Computer Society
Selecting authentic scenes about activities of daily living (ADL) is useful to support our memory of everyday life. Key-frame extraction for first-person vision (FPV) videos is a core technology to realize such memory assistant. However, most existing key-frame extraction methods have mainly focused on stable scenes not related to ADL and only used visual signals of the image sequence even though the activities usually associate with our visual experience. To deal with dynamically changing scenes of FPV about daily activities, integrating motion and visual signals are essential. In this paper, we present a novel key-frame extraction method for ADL, which integrates multi-modal sensor signals to temper noise and detect salient activities. Our proposed method projects motion and visual features to a shared space by a probabilistic canonical correlation analysis and selects key frames there. The experimental results using ADL datasets collected in a house suggest that our key-frame extraction technique running in the shared space improves the precision of extracted key frames and the coverage of the entire video.
- リンク情報
-
- DOI
- https://doi.org/10.1109/ICIP.2017.8297032
- DBLP
- https://dblp.uni-trier.de/rec/conf/icip/LiKAMK17
- URL
- http://dblp.uni-trier.de/db/conf/icip/icip2017.html#conf/icip/LiKAMK17
- Scopus
- https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85045327144&origin=inward
- Scopus Citedby
- https://www.scopus.com/inward/citedby.uri?partnerID=HzOxMe3b&scp=85045327144&origin=inward
- ID情報
-
- DOI : 10.1109/ICIP.2017.8297032
- ISSN : 1522-4880
- DBLP ID : conf/icip/LiKAMK17
- SCOPUS ID : 85045327144