Advances in Human Activity Recognition and Sensor Fusion

The field of human activity recognition is rapidly advancing, with a focus on developing innovative methods for sensor-based activity recognition, robust inertial navigation, and effective sensor fusion. Recent studies have explored the use of wearable devices, such as smartwatches and earphones, to recognize human activities and track user behavior. The integration of multiple sensors and modalities has shown significant promise in improving the accuracy and robustness of activity recognition systems. Notably, the development of transformer-based architectures and attention mechanisms has enabled the effective modeling of complex temporal and spatial relationships in human activity data.

Some noteworthy papers in this area include: The CMD-HAR paper, which proposes a spatiotemporal attention modal decomposition alignment fusion strategy to tackle the problem of mixed distribution of sensor data. The Suite-IN++ paper, which presents a deep learning framework for flexiwear bodynet-based pedestrian localization, integrating motion data from wearable devices on different body parts. The MultiTSF paper, which proposes a transformer-based sensor fusion method for human-centric multi-view and multi-modal action recognition, leveraging a human detection module to generate pseudo-ground-truth labels.

Sources

CMD-HAR: Cross-Modal Disentanglement for Wearable Human Activity Recognition

Data-driven worker activity recognition and picking efficiency estimation in manual strawberry harvesting

Redundant feature screening method for human activity recognition based on attention purification mechanism

Suite-IN++: A FlexiWear BodyNet Integrating Global and Local Motion Features from Apple Suite for Robust Inertial Navigation

Correlation-Attention Masked Temporal Transformer for User Identity Linkage Using Heterogeneous Mobility Data

LSC-ADL: An Activity of Daily Living (ADL)-Annotated Lifelog Dataset Generated via Semi-Automatic Clustering

MultiTSF: Transformer-based Sensor Fusion for Human-Centric Multi-view and Multi-modal Action Recognition

EmbodiedSense: Understanding Embodied Activities with Earphones

Multi-Head Adaptive Graph Convolution Network for Sparse Point Cloud-Based Human Activity Recognition

Built with on top of