Conference item icon

Conference item : Conference-proceeding

Two-stream convolutional networks for action recognition in videos

Abstract:
We investigate architectures of discriminatively trained deep Convolutional Networks (ConvNets) for action recognition in video. The challenge is to capture the complementary information on appearance from still frames and motion between frames. We also aim to generalise the best performing hand-crafted features within a data-driven learning framework. Our contribution is three-fold. First, we propose a two-stream ConvNet architecture which incorporates spatial and temporal networks. Second, we demonstrate that a ConvNet trained on multi-frame dense optical flow is able to achieve very good performance in spite of limited training data. Finally, we show that multi-task learning, applied to two different action classification datasets, can be used to increase the amount of training data and improve the performance on both. Our architecture is trained and evaluated on the standard video actions benchmarks of UCF-101 and HMDB-51, where it is competitive with the state of the art. It also exceeds by a large margin previous attempts to use deep nets for video classification.

Actions


Access Document


Files:

Authors


More by this author
Institution:
University of Oxford
Oxford college:
Brasenose College
Role:
Author


Publisher:
NIPS Proceedings
Host title:
Neural Information Processing Systems (NIPS) conference
Journal:
Neural Information Processing Systems (NIPS) conference More from this journal
Volume:
27
Pages:
1-11
Publication date:
2014-12-31
Acceptance date:
2014-06-06
Event location:
Montreal, Canada
Event start date:
2016-12-08
Event end date:
2016-12-11


Subtype:
conference-proceeding
Pubs id:
pubs:679026
UUID:
uuid:1dd0bcd0-39ca-48a1-9c20-5341d6c49251
Local pid:
pubs:679026
Source identifiers:
679026
Deposit date:
2017-02-09

Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP