Learning Deep Representations of Appearance and Motion for Anomalous Event Detection
Dan Xu, Elisa Ricci, Yan Yan, Jingkuan Song and Nicu Sebe
Abstract
We present a novel unsupervised deep learning framework for anomalous event detection in complex video scenes. While most existing works merely use hand-crafted appearance and motion features, we propose Appearance and Motion DeepNet (AMDN) which utilizes deep neural networks to automatically learn feature representations. To exploit the complementary information of both appearance and motion patterns, we introduce a novel double fusion framework, combining both the benefits of traditional early fusion and late fusion strategies. Specifically, stacked denoising autoencoders are proposed to separately learn both appearance and motion features as well as a joint representation (early fusion). Based on the learned representations, multiple one-class SVM models are used to predict the anomaly scores of each input, which are then integrated with a late fusion strategy for final anomaly detection. We evaluate the proposed method on two publicly available video surveillance datasets, showing competitive performance with respect to state of the art approaches.
Dan Xu, Elisa Ricci, Yan Yan, Jingkuan Song and Nicu Sebe. Learning Deep Representations of Appearance and Motion for Anomalous Event Detection. In Xianghua Xie, Mark W. Jones, and Gary K. L. Tam, editors, Proceedings of the British Machine Vision Conference (BMVC), pages 8.1-8.12. BMVA Press, September 2015.
Bibtex
@inproceedings{BMVC2015_8,
title={Learning Deep Representations of Appearance and Motion for Anomalous Event Detection},
author={Dan Xu and Elisa Ricci and Yan Yan and Jingkuan Song and Nicu Sebe},
year={2015},
month={September},
pages={8.1-8.12},
articleno={8},
numpages={12},
booktitle={Proceedings of the British Machine Vision Conference (BMVC)},
publisher={BMVA Press},
editor={Xianghua Xie, Mark W. Jones, and Gary K. L. Tam},
doi={10.5244/C.29.8},
isbn={1-901725-53-7},
url={https://dx.doi.org/10.5244/C.29.8}
}