Abstract
Deep learning approaches have been established as the main methodology for video classification and recognition. Recently, 3-dimensional convolutions have been used to achieve state-of-the-art performance in many challenging video datasets. Because of the high level of complexity of these methods, as the convolution operations are also extended to an additional dimension in order to extract features from it as well, providing a visualization for the signals that the network interpret as informative, is a challenging task. An effective notion of understanding the network’s innerworkings would be to isolate the spatio-temporal regions on the video that the network finds most informative. We propose a method called Saliency Tubes which demonstrate the foremost points and regions in both frame level and over time that are found to be the main focus points of the network. We demonstrate our findings on widely used datasets for thirdperson and egocentric action classification and enhance the set of methods and visualizations that improve 3D Convolutional Neural Networks (CNNs) intelligibility.
Original language | English |
---|---|
Title of host publication | Proceedings of the IEEE International Conference on Image Processing (ICIP) |
Publisher | IEEE |
Pages | 1830-1834 |
DOIs | |
Publication status | Published - 2019 |
Event | International Conference on Image Processing 2019 - Taipei, Taiwan, Province of China Duration: 22 Sept 2019 → 25 Sept 2019 |
Conference
Conference | International Conference on Image Processing 2019 |
---|---|
Abbreviated title | ICIP19 |
Country/Territory | Taiwan, Province of China |
City | Taipei |
Period | 22/09/19 → 25/09/19 |
Keywords
- Visual Explanations
- Explainable Convolutions
- Spatio-temporal feature representation