基于注意力机制的多模态人体行为识别算法
DOI:
作者:
作者单位:

1.西北工业大学 电子信息学院;2.河北师范大学 计算机与网络空间安全学院

作者简介:

通讯作者:

中图分类号:

TP391.41

基金项目:

陕西省重点研发项目(2021ZDLGY15-01, 2021ZDLGY09-04, 2021GY-004和2020GY-050);深圳市国际合作研究项目(GJHZ20200731095204013);国家自然基金(61772419)


Multi-modal human behavior recognition algorithm based on attention mechanism
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    提出了基于注意力机制的多模态人体行为识别算法。针对多模态特征的有效融合问题,设计基于注意力机制的双流特征融合卷积网络TAM3DNet(Two-stream Attention Mechanism 3D Network)。主干网络采用结合注意力机制的AM3DNet(Attention Mechanism 3D Network),将特征图与注意力图进行加权后得到加权行为特征,从而使网络聚焦于肢体运动区域的特征,减弱背景和肢体静止区域的影响。将RGB-D数据的颜色和深度两种模态数据分别作为双流网络的输入,从两条分支网络得到彩色和深度行为特征,然后将融合特征进行分类得到人体行为识别结果。

    Abstract:

    A multi-modal human behavior recognition algorithm based on attention mechanism is proposed. Aiming at the problem of effective fusion of multimodal features, a two-stream feature fusion convolutional network TAM3DNet (Two-stream Attention Mechanism 3D Network) based on attention mechanism is designed. The backbone network adopts AM3DNet (Attention Mechanism 3D Network), which combines the attention mechanism, and weights the feature map and the attention map to obtain the weighted behavior characteristics, so that the network focuses on the characteristics of the limb movement area and reduces the influence of the background and the static area of the limb. The color and depth modal data of the RGB-D data are respectively used as the input of the dual-stream network, and the color and depth behavior features are obtained from the two branch networks, and then the fusion features are classified to obtain the human behavior recognition results.

    参考文献
    相似文献
    引证文献
引用本文

宋真东,杨国超,马玉鹏,冯晓毅.基于注意力机制的多模态人体行为识别算法计算机测量与控制[J].,2022,30(2):276-283.

复制
分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2022-01-07
  • 最后修改日期:2022-01-11
  • 录用日期:2022-01-11
  • 在线发布日期: 2022-02-22
  • 出版日期: