Human Action Segmentation using 3D Fully Convolutional Network
Pei Yu, Jiang Wang and Ying Wu
Abstract
Detailed action analysis, such as action detection, localization and segmentation, has
received more and more attention in recent years. Compared to action classification, action segmentation and localization are more useful in many practical applications that
require precise spatio-temporal information of the actions. However, performing action
segmentation and localization is more challenging, because determining the pixel-level
locations of action not only requires a strong spatial model that captures the visual appearances for the actions, but also calls for a temporal model that characterizes the dy-
namics of the actions. Most existing methods either use hand-crafted spatial models, or
can only extract short-term motion information. In this paper, we propose a 3D fully
convolutional deep network to jointly exploit spatial and temporal information in a unified framework for action segmentation and localization. The proposed deep network is
trained to combine both information in an end-to-end fashion.
Session
Posters
Files
Paper (PDF)
Supplementary (PDF)
DOI
10.5244/C.31.164
https://dx.doi.org/10.5244/C.31.164
Citation
Pei Yu, Jiang Wang and Ying Wu. Human Action Segmentation using 3D Fully Convolutional Network. In T.K. Kim, S. Zafeiriou, G. Brostow and K. Mikolajczyk, editors, Proceedings of the British Machine Vision Conference (BMVC), pages 164.1-164.12. BMVA Press, September 2017.
Bibtex
@inproceedings{BMVC2017_164,
title={Human Action Segmentation using 3D Fully Convolutional Network},
author={Pei Yu, Jiang Wang and Ying Wu},
year={2017},
month={September},
pages={164.1-164.12},
articleno={164},
numpages={12},
booktitle={Proceedings of the British Machine Vision Conference (BMVC)},
publisher={BMVA Press},
editor={Tae-Kyun Kim, Stefanos Zafeiriou, Gabriel Brostow and Krystian Mikolajczyk},
doi={10.5244/C.31.164},
isbn={1-901725-60-X},
url={https://dx.doi.org/10.5244/C.31.164}
}