Human Action Segmentation using 3D Fully Convolutional Network

Pei Yu, Jiang Wang and Ying Wu

Abstract

Detailed action analysis, such as action detection, localization and segmentation, has received more and more attention in recent years. Compared to action classification, action segmentation and localization are more useful in many practical applications that require precise spatio-temporal information of the actions. However, performing action segmentation and localization is more challenging, because determining the pixel-level locations of action not only requires a strong spatial model that captures the visual appearances for the actions, but also calls for a temporal model that characterizes the dy- namics of the actions. Most existing methods either use hand-crafted spatial models, or can only extract short-term motion information. In this paper, we propose a 3D fully convolutional deep network to jointly exploit spatial and temporal information in a unified framework for action segmentation and localization. The proposed deep network is trained to combine both information in an end-to-end fashion.

Session

Posters

Files

PDF iconPaper (PDF)
PDF iconSupplementary (PDF)

DOI

10.5244/C.31.164
https://dx.doi.org/10.5244/C.31.164

Citation

Pei Yu, Jiang Wang and Ying Wu. Human Action Segmentation using 3D Fully Convolutional Network. In T.K. Kim, S. Zafeiriou, G. Brostow and K. Mikolajczyk, editors, Proceedings of the British Machine Vision Conference (BMVC), pages 164.1-164.12. BMVA Press, September 2017.

Bibtex

            @inproceedings{BMVC2017_164,
                title={Human Action Segmentation using 3D Fully Convolutional Network},
                author={Pei Yu, Jiang Wang and Ying Wu},
                year={2017},
                month={September},
                pages={164.1-164.12},
                articleno={164},
                numpages={12},
                booktitle={Proceedings of the British Machine Vision Conference (BMVC)},
                publisher={BMVA Press},
                editor={Tae-Kyun Kim, Stefanos Zafeiriou, Gabriel Brostow and Krystian Mikolajczyk},
                doi={10.5244/C.31.164},
                isbn={1-901725-60-X},
                url={https://dx.doi.org/10.5244/C.31.164}
            }