Mamba-MHAR: An efficient multimodal framework for human action recognition

Author affiliations

Authors

  • Trung-Hieu Le School of Electrical and Electronic Engineering, Hanoi University of Science and Technology, 01 Dai Co Viet Street, Bach Mai Ward, Ha Noi, Viet Nam https://orcid.org/0000-0001-6323-6959
  • Khanh-Nguyen Thai School of Electrical and Electronic Engineering, Hanoi University of Science and Technology, 01 Dai Co Viet Street, Bach Mai Ward, Ha Noi, Viet Nam https://orcid.org/0009-0009-2028-9554
  • Tuan-Anh Le Dai Nam University, 01 Pho Xom, Phu Luong Ward, Ha Noi, Viet Nam
  • Mathieu Delalandre Polytechnic University of Tours, France
  • Trung-Kien Tran Institute of Information Technology, AMST, 17 Hoang Sam, Nghia Do Ward, Ha Noi, Vietnam https://orcid.org/0000-0001-5466-0539
  • Thanh-Hai Tran School of Electrical and Electronic Engineering, Hanoi University of Science and Technology, 01 Dai Co Viet Street, Bach Mai Ward, Ha Noi, Viet Nam https://orcid.org/0000-0003-3133-3361
  • Cuong-Pham Posts and Telecommunications Institute of Technology, Nguyen Trai Street, Mo Lao Ward, Ha Noi, Viet Nam https://orcid.org/0000-0003-0973-0889

DOI:

https://doi.org/10.15625/1813-9663/22770

Keywords:

Mamba, selective state space model, selection mechanism, HAR, multimodal fusion, visual sensor, inertial sensor.

Abstract

Human Action Recognition (HAR) has emerged as an active research domain in recent years with wide-ranging applications in healthcare monitoring, smart home systems, and hu- man–robot interaction. This paper introduces a method, namely Mamba-MHAR (Mamba based Multimodal Human Action Recognition), a lightweight multimodal architecture aimed at improv- ing HAR performance by effectively integrating data from inertial sensors and egocentric videos. Mamba-MHAR consists of double Mamba-based branches, one for visual feature extraction - VideoMamba, and the other for motion feature extraction - MAMC. Both branches are built upon recently introduced Selective State Space Models (SSMs) to optimize the computational cost, and they are lately fused for final human activity classification. Mamba-MHAR achieves significant efficiency gains in terms of GPU usage, making it highly suitable for real-time deployment on edge and mobile devices. Extensive experiments were conducted on two challenging multimodal datasets UESTC-MMEA-CL and MuWiGes, which contain synchronized IMU and video data recorded in natural settings. The proposed Mamba-MHAR achieves 98.00% accuracy on UESTC-MMEA-CL and 98.58% on MuWiGes, surpassing state-of-the-art baselines. These results demonstrate that a simple yet efficient fusion of multimodal lightweight Mamba-based models provides a promising solution for scalable and low-power applications in pervasive computing environments.

Downloads

Published

27-09-2025

How to Cite

[1]T.-H. Le, “Mamba-MHAR: An efficient multimodal framework for human action recognition”, J. Comput. Sci. Cybern., vol. 41, no. 3, p. 245–264, Sep. 2025.

Issue

Section

Articles

Similar Articles

<< < 3 4 5 6 7 8 9 10 11 12 > >> 

You may also start an advanced similarity search for this article.