Home
Scholarly Works
Efficient Action Recognition Using Confidence...
Conference

Efficient Action Recognition Using Confidence Distillation

Abstract

Modern neural networks are powerful predictive models. However, when it comes to recognizing that they may be wrong about their predictions, they perform poorly. For example, for one of the most common activation functions, the ReLU and its variants, even a well-calibrated model can produce incorrect but high confidence predictions. Most current action recognition methods are based on clip-level classifiers that densely sample a given video for non-overlapping, same-sized clips and aggregate the results using an aggregation function - typically averaging - to achieve video level predictions. While this approach has shown to be effective, it is sub-optimal in recognition accuracy and has a high computational overhead. To mitigate both these issues, we propose the confidence distillation framework to teach a student model how to select less ambiguous clips for the teacher, and divide the task of prediction between the two. We conduct extensive experiments on three action recognition datasets and demonstrate that our framework achieves significant improvements in action recognition accuracy (up to 20%) and computational efficiency (more than 40%).

Authors

Shalmani SM; Chiang F; Zheng R

Volume

00

Pagination

pp. 3362-3369

Publisher

Institute of Electrical and Electronics Engineers (IEEE)

Publication Date

August 25, 2022

DOI

10.1109/icpr56361.2022.9956432

Name of conference

2022 26th International Conference on Pattern Recognition (ICPR)
View published work (Non-McMaster Users)

Contact the Experts team