← 返回
基于自适应神经模糊分类的高精度动作识别:先进生物信号与RGB融合技术
Advanced Biosignal-RGB Fusion With Adaptive Neurofuzzy Classification for High-Precision Action Recognition
| 作者 | Iqra Aijaz Abro · Haifa F. Alhasson · Shuaa S. Alharbi · Mohammed Alatiyyah · Dina Abdulaziz AlHammadi · Ahmad Jalal |
| 期刊 | IEEE Access |
| 出版日期 | 2025年1月 |
| 技术分类 | 储能系统技术 |
| 技术标签 | 储能系统 |
| 相关度评分 | ★★★★★ 5.0 / 5.0 |
| 关键词 | 动作识别 多感官数据 特征提取 神经模糊分类器 准确率 |
语言:
中文摘要
在使用多传感器数据进行动作识别的领域中,生物信号与RGB模态的融合为提升动作分类系统精度提供了新途径。本文提出一种自适应神经模糊分类框架,融合肌电信号、加速度计数据和视觉信息,通过模糊逻辑优化多模态数据的特征融合。
English Abstract
In the domain of action recognition using multisensory data, the integration of RGB and signal-based modalities offers a promising approach to enhance the accuracy of action classification systems. Our system was developed through experimentation on three benchmark datasets: UTD-MHAD (University of Texas at Dallas Multimodal Human Action Dataset), HWU-USP and LaRa. Initially, the data undergoes preprocessing, where gaussian and butterworth filters are applied to the RGB and signal data, respectively. Following this, windowing/segmentation is applied to signals and RGB data. After that, features are extracted from the signal data, including auto-regression, MFCC (Mel-frequency Cepstral Coefficients), and transient detection principle, while the RGB (Red Green Blue) was processed as a combined input to extract features such as angles, velocity, full-body elliptical modeling, fiducial points, and a 2.5D point cloud of the entire body. These features are then fused, followed by the application of the Yeo-Johnson power optimizer to refine the data. The optimized data is subsequently classified using a Neurofuzzy classifier to recognize different actions. This classifier is chosen for its ability to adapt to the heterogeneous nature of multimodal data, where features are spread across different domains, making traditional classifiers less effective. The Neurofuzzy model employs cross-validation for training and testing to ensure reliable results. The results also suggest that the proposed model yields a higher accuracy than the existing models. More specifically, in the HWU-USP dataset, the accuracy amounts to mean 89%, in the LaRa, to mean 91% and 88% over the UTD-MHAD dataset. The system under study effectively distinguishes related actions, but its efficiency is hindered by the complexity of individual actions and the increased noise in the dataset.
S
SunView 深度解读
该多传感器融合技术可应用于阳光电源储能系统的人机交互和安全监控。通过融合视觉和生物信号数据,实现储能电站运维人员的行为识别和异常动作检测,提升工业现场的安全管理水平,为智能运维系统提供人机协同支持。