![]() Flowchart of our system.
|
[基於光學雷達技術的3D骨架測量] Motion capture technology was originally designed to simplify the process of animation production by eliminating the time-consuming and laborious task of drawing character animation frame by frame. In prior years, motion capture technology requires expensive equipment, that could only be afforded by the film industry. Furthermore, the collected data is frequently inaccurate, which needs to be manually fine-tuned in the post-production stage. However, thanks to recent advances, motion capture devices gradually become more reliable, and affordable for other applications such as entertainment, logistics and sport training. The vast majority of motion capture systems require the user to measure bone lengths before capturing the motion, except in some rare cases. As a result, the measurement of skeletal length is a crucial step in motion tracking, as it affects the overall accuracy of motion capturing. The most intuitive and obvious way is to measure the bone length is with a ruler. However, this conventional process is an unstable factor in the overall quality of motion tracking, since there is a discrepancy between the bone length measured on the skin surface and the actual bone length. In addition, the measuring process costs extra time and labor, both of which can be reduced by automation. This research aims to propose a fast, effortless and robust system to measure the skeletal lengths of the human body, using a single LiDAR camera. Our main goal is to provide an affordable yet reliable device that can be used in medical rehabilitation and other related applications. In order to achieve our objectives, we adopted multiple neural network architectures, including HigherHRNet and DeepLab v3, and an ellipse approximation method based on arc-support vectors. Our experimental result shows that our method has an average error of 3% compared to bone lengths measured by X-ray, which is more accurate than the conventional ruler method.
SUMMARY (中文總結):
Compare our method with X-ray photos and manual measurement.
REFERENCES: [1] Bowen Cheng, Bin Xiao, Jingdong Wang, Honghui Shi, Thomas S Huang, and Lei Zhang. Higherhrnet: Scale-aware representation learning for bottom-up human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5386–5395, 2020. [2] Changsheng Lu, Siyu Xia, Ming Shao, and Yun Fu. Arc-support line segments revisited: An efficient high-quality ellipse detection. IEEE Transactions on Image Processing, 29:768–781, 2019. [3] Bolei Zhou, Hang Zhao, Xavier Puig, Tete Xiao, Sanja Fidler, Adela Barriuso, and Antonio Torralba. Semantic understanding of scenes through the ade20k dataset. International Journal of Computer Vision, 127(3):302–321, 2019. |