Temporal Convolutional Networks for Action Segmentation and Detection
Autor: | Michael D. Flynn, Gregory D. Hager, Colin Lea, René Vidal, Austin Reiter |
---|---|
Rok vydání: | 2017 |
Předmět: |
FOS: Computer and information sciences
Computer science business.industry Computer Vision and Pattern Recognition (cs.CV) Speech recognition Feature extraction Computer Science - Computer Vision and Pattern Recognition ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION 020207 software engineering Pattern recognition 02 engineering and technology Image segmentation Object detection Upsampling Recurrent neural network Video tracking 0202 electrical engineering electronic engineering information engineering 020201 artificial intelligence & image processing Segmentation Artificial intelligence business Classifier (UML) |
Zdroj: | CVPR |
DOI: | 10.1109/cvpr.2017.113 |
Popis: | The ability to identify and temporally segment fine-grained human actions throughout a video is crucial for robotics, surveillance, education, and beyond. Typical approaches decouple this problem by first extracting local spatiotemporal features from video frames and then feeding them into a temporal classifier that captures high-level temporal patterns. We introduce a new class of temporal models, which we call Temporal Convolutional Networks (TCNs), that use a hierarchy of temporal convolutions to perform fine-grained action segmentation or detection. Our Encoder-Decoder TCN uses pooling and upsampling to efficiently capture long-range temporal patterns whereas our Dilated TCN uses dilated convolutions. We show that TCNs are capable of capturing action compositions, segment durations, and long-range dependencies, and are over a magnitude faster to train than competing LSTM-based Recurrent Neural Networks. We apply these models to three challenging fine-grained datasets and show large improvements over the state of the art. |
Databáze: | OpenAIRE |
Externí odkaz: |