Abstract
Object proposals have contributed significantly to recent advances in object understanding in images. Inspired by the success of this approach, we introduce Deep Action Proposals (DAPs), an effective and efficient algorithm for generating temporal action proposals from long videos. We show how to take advantage of the vast capacity of deep learning models and memory cells to retrieve from untrimmed videos temporal segments, which are likely to contain actions. A comprehensive evaluation indicates that our approach outperforms previous work on a large scale action benchmark, runs at 134 FPS making it practical for large-scale scenarios, and exhibits an appealing ability to generalize, i.e. to retrieve good quality temporal proposals of actions unseen in training.
Original language | English (US) |
---|---|
Title of host publication | Lecture Notes in Computer Science |
Publisher | Springer Nature |
Pages | 768-784 |
Number of pages | 17 |
ISBN (Print) | 9783319464862 |
DOIs | |
State | Published - Sep 17 2016 |
Bibliographical note
KAUST Repository Item: Exported on 2020-10-01Acknowledgements: Research in this publication was supported by the King Abdullah University of Science and Technology (KAUST) Office of Sponsored Research, the Stanford AI Lab-Toyota Center for Artificial Intelligence Research and a Google Faculty Research Award (2015).