Trajectory-Based 3D Convolutional Descriptors for Human Action Recognition

This article presents a new method for video representation, called trajectory based 3D convolutional descriptor (TCD), which incorporates the advantages of both deep learned features and hand-crafted features. We utilize deep architectures to learn discriminative convolutional feature maps, and con...

Full description

Saved in:
Bibliographic Details
Published inJournal of Information Science and Engineering Vol. 35; no. 4; pp. 851 - 870
Main Authors SHEERAZ ARIF, 王晶(JING WANG), FIDA HUSSAIN, 费泽松(ZESONG FEI)
Format Journal Article
LanguageEnglish
Published Taipei 社團法人中華民國計算語言學學會 01.07.2019
Institute of Information Science, Academia Sinica
Subjects
Online AccessGet full text
ISSN1016-2364
DOI10.6688/JISE.201907_35(4).0009

Cover

More Information
Summary:This article presents a new method for video representation, called trajectory based 3D convolutional descriptor (TCD), which incorporates the advantages of both deep learned features and hand-crafted features. We utilize deep architectures to learn discriminative convolutional feature maps, and conduct trajectory constrained pooling to aggregate these convolutional features into effective descriptors. Firstly, valid trajectories are generated by tracking the interest points within co-motion super-pixels. Secondly, we utilize the 3D ConvNet (C3D) to capture both motion and appearance information in the form of convolutional feature maps. Finally, feature maps are transformed by using two normalization methods, namely channel normalization and spatiotemporal normalization. Trajectory constrained sampling and pooling are used to aggregate deep learned features into descriptors. The proposed (TCD) contains high discriminative capacity compared with hand-crafted features and is able to boost the recognition performance. Experimental results on benchmark datasets demonstrate that our pipeline obtains superior performance over conventional algorithms in terms of both efficiency and accuracy.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:1016-2364
DOI:10.6688/JISE.201907_35(4).0009