ModalDrop: Modality-aware Regularization for Temporal-Spectral Fusion in Human Activity Recognition
Xin Zeng (Institute of Computing Technology, Chinese Academy of Sciences); Yiqiang Chen (Institute of Computing Technology, Chinese Academy of Sciences); Benfeng Xu (University of Science and Technology of China); Tengxiang Zhang (institute of computing technology, Chinese academy of sciences)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Although most of existing works for sensor-based HAR rely on the temporal view, we argue that the spectral view also provides complementary prior and accordingly benchmark a standard multi-view framework with extensive experiments to demonstrate its consistent superiority over single-view opponents. We then delve into the intrinsic mechanism of the multi-view representation fusion and propose ModalDrop as a novel modality-aware regularization method to learn and exploit representations of both views effectively. We demonstrate its advantage over existing representation fusion alternatives with comprehensive experiments and ablations. The improvements are consistent for various settings and are orthogonal with different backbones. We also discuss its potential application for other related tasks regarding representation or modality fusion.
The source code is available on https://github.com/studyzx/ModalDrop.git.