pith. machine review for the scientific record. sign in

arxiv: 1701.07368 · v2 · submitted 2017-01-25 · 💻 cs.CV

Recognition: unknown

Deep Local Video Feature for Action Recognition

Authors on Pith no claims yet
classification 💻 cs.CV
keywords featureslocalglobalvideolabelsproblemactionaggregate
0
0 comments X
read the original abstract

We investigate the problem of representing an entire video using CNN features for human action recognition. Currently, limited by GPU memory, we have not been able to feed a whole video into CNN/RNNs for end-to-end learning. A common practice is to use sampled frames as inputs and video labels as supervision. One major problem of this popular approach is that the local samples may not contain the information indicated by global labels. To deal with this problem, we propose to treat the deep networks trained on local inputs as local feature extractors. After extracting local features, we aggregate them into global features and train another mapping function on the same training data to map the global features into global labels. We study a set of problems regarding this new type of local features such as how to aggregate them into global features. Experimental results on HMDB51 and UCF101 datasets show that, for these new local features, a simple maximum pooling on the sparsely sampled features lead to significant performance improvement.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.