基于视觉的无标记运动学分析方法
DOI:
CSTR:
作者:
作者单位:

1.福州大学电气工程与自动化学院福州350108; 2.福建省医疗器械和医药技术重点实验室福州350108

作者简介:

通讯作者:

中图分类号:

TP391.4TH77

基金项目:

国家自然科学基金(62373108)、福建省技术创新重点攻关及产业化项目(校企联合类)(2024XQ001)资助


Markerless vision-based kinematic analysis method
Author:
Affiliation:

1.College of Electrical Engineering and Automation, Fuzhou University, Fuzhou 350108, China; 2.Fujian Provincial Key Laboratory of Medical Instrument and Pharmaceutical Technology, Fuzhou 350108, China

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    针对基于标记的光学系统成本高、耗时长、专业性强的问题,提出了一种采用两个视角的视觉无标记运动学分析方法,旨在实现便捷、低成本的运动学评估。首先,集成Swin Transformer的全局上下文建模能力、坐标注意力的精准位置感知能力、双向特征金字塔网络的多尺度特征融合能力,构建二维特征提取架构,克服自遮挡、关键点小目标检测问题,有效提取二维特征。其次,提出以关键点位置合理性和肢体长度一致性为关节上下文约束的三角测量方法,并结合人体参数化模型对三维关键点进行重构,提高关键点估计精度。最后,构建关键点增强模型,获取解剖标记集并结合肌骨模型进行运动学分析。公开数据集上的运动学评估实验表明,所提方法的平均关节角度误差为8.59°,平均关节位置误差为42.02 mm,优于现有的高性能方法。同时,为验证方法在真实场景下的适用性,以商用动作捕捉系统Xsens作为评估标准,并与当前主流方法OpenCap进行比较,分别对肩关节和步态运动学展开分析。实验结果表明,在肩关节和步态运动学上,所提方法与Xsens的相关系数分别为0.92和0.86,较OpenCap的相关系数分别提高9.52%和7.40%;角度误差分别为13.97°和3.12°,较OpenCap的误差分别下降27.01%和25.18%。综上所述,在公开数据集和真实场景下,所提方法可实现比当前主流方法更准确的运动学分析,对促进运动学分析相关应用的推广具有重要意义。

    Abstract:

    To address the high cost, time-consuming nature, and specialized expertise required by marker-based optical systems, this article proposes a visual kinematic analysis method utilizing two viewpoints to achieve convenient, low-cost kinematic evaluation. First, a two-dimensional feature extraction architecture is established by integrating the global context modelling capability of the Swin Transformer, the precise positional awareness of coordinate attention, and the multi-scale feature fusion capability of the bidirectional feature pyramid network. It overcomes challenges such as occlusion and small target detection for keypoints, enabling effective extraction of two-dimensional features. Secondly, a triangulation method is proposed, employing joint contextual constraints based on keypoints position plausibility and limb length consistency. This is combined with a parametric human model to reconstruct 3D keypoints, enhancing estimation accuracy. Finally, a keypoint augmentation model is formulated to obtain an anatomical label set, which is then integrated with a musculoskeletal model for kinematic analysis. Kinematic evaluation on public datasets demonstrates an average joint angular error of 8.59° and average joint positional error of 42.02 mm, outperforming existing high-performance methods. To validate real-world applicability, commercial motion capture system Xsens serves as the evaluation benchmark against the mainstream OpenCap method, with analyses conducted on shoulder joint and gait kinematics, respectively. Experimental results show that for shoulder joint and gait kinematics, the proposed method achieves correlation coefficients of 0.92 and 0.86, respectively, with Xsens, representing improvements of 9.52% and 7.40% over OpenCap. Angular errors are reduced to 13.97° and 3.12°, respectively, marking decreases of 27.01% and 25.18% compared to OpenCap. In summary, the proposed method achieves more accurate kinematic analysis than current mainstream approaches on both public datasets and in real-world scenarios, holding significant implications for advancing applications related to kinematic analysis.

    参考文献
    相似文献
    引证文献
引用本文

黄高华,李玉榕,姜海燕,陈建国.基于视觉的无标记运动学分析方法[J].仪器仪表学报,2026,47(1):312-324

复制
分享
相关视频

文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:
  • 最后修改日期:
  • 录用日期:
  • 在线发布日期: 2026-03-30
  • 出版日期:
文章二维码