Loading [a11y]/accessibility-menu.js
Extracting key frames from first-person videos in the common space of multiple sensors | IEEE Conference Publication | IEEE Xplore

Extracting key frames from first-person videos in the common space of multiple sensors


Abstract:

Selecting authentic scenes about activities of daily living (ADL) is useful to support our memory of everyday life. Key-frame extraction for first-person vision (FPV) vid...Show More

Abstract:

Selecting authentic scenes about activities of daily living (ADL) is useful to support our memory of everyday life. Key-frame extraction for first-person vision (FPV) videos is a core technology to realize such memory assistant. However, most existing key-frame extraction methods have mainly focused on stable scenes not related to ADL and only used visual signals of the image sequence even though the activities usually associate with our visual experience. To deal with dynamically changing scenes of FPV about daily activities, integrating motion and visual signals are essential. In this paper, we present a novel key-frame extraction method for ADL, which integrates multi-modal sensor signals to temper noise and detect salient activities. Our proposed method projects motion and visual features to a shared space by a probabilistic canonical correlation analysis and selects key frames there. The experimental results using ADL datasets collected in a house suggest that our key-frame extraction technique running in the shared space improves the precision of extracted key frames and the coverage of the entire video.
Date of Conference: 17-20 September 2017
Date Added to IEEE Xplore: 22 February 2018
ISBN Information:
Electronic ISSN: 2381-8549
Conference Location: Beijing, China

Contact IEEE to Subscribe

References

References is not available for this document.