@inproceedings{c577ec129a1641518e05d1200b32ce37,
title = "Efficient Human Activity Classification from Egocentric Videos Incorporating Actor-Critic Reinforcement Learning",
abstract = "In this paper, we introduce a novel framework to significantly reduce the computational cost of human temporal activity recognition from egocentric videos while maintaining the accuracy at the same level. We propose to apply the actor-critic model of reinforcement learning to optical flow data to locate a bounding box around region of interest, which is then used for clipping a sub-image from a video frame. We also propose to use one shallow and one deeper 3D convolutional neural network to process the original image and the clipped image region, respectively. We compared our proposed method with another approach using 3D convolutional networks on the recently released Dataset of Multimodal Semantic Egocentric Video. Experimental results show that the proposed method reduces the processing time by 36.4% while providing comparable accuracy at the same time.",
keywords = "activity classification, actor critic, reinforcement learning",
author = "Yantao Lu and Yilan Li and Senem Velipasalar",
note = "Publisher Copyright: {\textcopyright} 2019 IEEE.; 26th IEEE International Conference on Image Processing, ICIP 2019 ; Conference date: 22-09-2019 Through 25-09-2019",
year = "2019",
month = sep,
doi = "10.1109/ICIP.2019.8803823",
language = "English (US)",
series = "Proceedings - International Conference on Image Processing, ICIP",
publisher = "IEEE Computer Society",
pages = "564--568",
booktitle = "2019 IEEE International Conference on Image Processing, ICIP 2019 - Proceedings",
address = "United States",
}