TY - GEN
T1 - SST: Single-Stream Temporal Action Proposals
AU - Buch, Shyamal
AU - Escorcia, Victor
AU - Shen, Chuanqi
AU - Ghanem, Bernard
AU - Niebles, Juan Carlos
N1 - KAUST Repository Item: Exported on 2020-10-01
Acknowledgements: This research was sponsored, in part, by the Stanford AI Lab-Toyota Center for Artificial Intelligence Research, Toyota Research Institute (TRI), and by the King Abdullah University of Science and Technology (KAUST) Office of Sponsored Research. This article reflects the opinions and conclusions of its authors and not TRI or any other Toyota entity. We thank our anonymous reviewers, De-An Huang, Oliver Groth, Fabian Caba, Joseph Lim, Jingwei Ji, and Fei-Fei Li for helpful comments and discussion.
PY - 2017/11/9
Y1 - 2017/11/9
N2 - Our paper presents a new approach for temporal detection of human actions in long, untrimmed video sequences. We introduce Single-Stream Temporal Action Proposals (SST), a new effective and efficient deep architecture for the generation of temporal action proposals. Our network can run continuously in a single stream over very long input video sequences, without the need to divide input into short overlapping clips or temporal windows for batch processing. We demonstrate empirically that our model outperforms the state-of-the-art on the task of temporal action proposal generation, while achieving some of the fastest processing speeds in the literature. Finally, we demonstrate that using SST proposals in conjunction with existing action classifiers results in improved state-of-the-art temporal action detection performance.
AB - Our paper presents a new approach for temporal detection of human actions in long, untrimmed video sequences. We introduce Single-Stream Temporal Action Proposals (SST), a new effective and efficient deep architecture for the generation of temporal action proposals. Our network can run continuously in a single stream over very long input video sequences, without the need to divide input into short overlapping clips or temporal windows for batch processing. We demonstrate empirically that our model outperforms the state-of-the-art on the task of temporal action proposal generation, while achieving some of the fastest processing speeds in the literature. Finally, we demonstrate that using SST proposals in conjunction with existing action classifiers results in improved state-of-the-art temporal action detection performance.
UR - http://hdl.handle.net/10754/626797
UR - http://ieeexplore.ieee.org/document/8100158/
UR - http://www.scopus.com/inward/record.url?scp=85035098459&partnerID=8YFLogxK
U2 - 10.1109/cvpr.2017.675
DO - 10.1109/cvpr.2017.675
M3 - Conference contribution
SN - 9781538604571
SP - 6373
EP - 6382
BT - 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
PB - Institute of Electrical and Electronics Engineers (IEEE)
ER -