pith. machine review for the scientific record. sign in

arxiv: 1704.00389 · v4 · submitted 2017-04-02 · 💻 cs.CV · cs.LG· cs.MM

Recognition: unknown

Hidden Two-Stream Convolutional Networks for Action Recognition

Authors on Pith no claims yet
classification 💻 cs.CV cs.LGcs.MM
keywords actionapproachframesrecognitionapproachescnnsend-to-endflow
0
0 comments X
read the original abstract

Analyzing videos of human actions involves understanding the temporal relationships among video frames. State-of-the-art action recognition approaches rely on traditional optical flow estimation methods to pre-compute motion information for CNNs. Such a two-stage approach is computationally expensive, storage demanding, and not end-to-end trainable. In this paper, we present a novel CNN architecture that implicitly captures motion information between adjacent frames. We name our approach hidden two-stream CNNs because it only takes raw video frames as input and directly predicts action classes without explicitly computing optical flow. Our end-to-end approach is 10x faster than its two-stage baseline. Experimental results on four challenging action recognition datasets: UCF101, HMDB51, THUMOS14 and ActivityNet v1.2 show that our approach significantly outperforms the previous best real-time approaches.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.