Publications

Detailed Information

Multi-scale Recurrent Encoder-Decoder Network for Dense Temporal Classification

Cited 16 time in Web of Science Cited 16 time in Scopus
Authors

Choo, Sungkwon; Seo, Wonkyo; Jeong, Dong-Ju; Cho, Nam Ik

Issue Date
2018-08
Publisher
IEEE
Citation
2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), pp.103-108
Abstract
The temporal events in video sequences often have long-term dependencies which are difficult to be handled by a convolutional neural network (CNN). Especially, the dense pixel-wise prediction of video frames is a difficult problem for the CNN because huge memories and a large number of parameters are needed to learn the temporal correlation. To overcome these difficulties, we propose a recurrent encoder-decoder network which compresses the spatiotemporal features at the encoder and restores them to the original sized results at the decoder. We adopt a convolutional long short-term memory (LSTM) into the encoder-decoder architecture, which successfully learns the spatiotemporal relation with relatively a small number of parameters. The proposed network is applied to one of the dense pixel-prediction problems, specifically, the background subtraction in video sequences. The proposed network is trained with limited duration of video frames, and yet it shows good generalization performance for different videos and time duration. Also, by additional video specific learning, it shows the best performance on a benchmark dataset (CDnet 2014).
ISSN
1051-4651
URI
https://hdl.handle.net/10371/186884
DOI
https://doi.org/10.1109/ICPR.2018.8545597
Files in This Item:
There are no files associated with this item.
Appears in Collections:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share