News Release

Motion in action: Scientists develop method to track human movements more accurately

Automatic feature point detection and tracking of human action in time-of-flight videos

Peer-Reviewed Publication

Chinese Association of Automation

Scientists have developed a data-driven method to better detect and track human movements for use in technologies such as at-home personal training videos or monitoring at-risk elderly patients.

The scientists published their results in IEEE/CAA Journal of Automatica Sinca (JAS), a joint publication of the IEEE and Chinese Association of Automation. The collaborative research team includes scientists from the University of North Texas, the North China Institute of Aerospace Engineering and Hefei University of Technology.

"Our goal in this paper [was] to address tracking of human subject movement with high accuracy and consistency," said Xiaohui Yuan, an associate professor at the University of North Texas and a visiting professor at the China University of Geosciences."This is beyond tracking a person or a car in a surveillance video or tracking the pose of a person to estimate his or her actions."

Yuan points to conventional tracking methods in which the most important identification is the center of the body. In those methods, once the body's core is identified, the location of extremities can be approximated. Yuan's method inverts these traditional methods.

Through the use of a time-of-flight camera, the entire range of view is scanned by each laser, rather than each laser only scanning in its pre-determined boundary. The camera measures depth of objects from the camera, and can isolate background, stationary objects from the human subject. Then, using a three-dimensional point cloud created by pulling the subject's silhouette from the human, it identifies five extreme points: the head, hands, and feet. Once those points are marked, the joints can be identified.

"In computer vision-based physical evaluation and in training, the accuracy of body part position must be ensured," Yuan said. "In addition, given that our depth imaging device can acquire only surface data of a 3D volume, a detected extreme point could become invisible after an action, [such as] rotation, which makes the consistency of detection critical."

In previous iterations of this motion capture technology, since the camera focused on the body's core, changing an arm's position from side to front could cause incorrect feedback or even a system error.

"The extreme points and joints in a depth image of a human body can be identified with a satisfactory accuracy and consistency by optimizing the geodesic distance and geometrical properties," Yuan said. "No template is required in this process, which makes the proposed method validated to any human figures and poses."

The researchers are now working to improve the robustness and efficiency of their method to better facilitate real-time analysis.


Fulltext of the paper is available:

IEEE/CAA Journal of Automatica Sinica (JAS) is a joint publication of the Institute of Electrical and Electronics Engineers, Inc (IEEE) and the Chinese Association of Automation. The objective of JAS is high quality and rapid publication of articles, with a strong focus on new trends, original theoretical and experimental research and developments, emerging technologies, and industrial standards in automation. The coverage of JAS includes but is not limited to: Automatic control,Artificial intelligence and intelligent control, Systems theory and engineering, Pattern recognition and intelligent systems, Automation engineering and applications, Information processing and information systems, Network based automation, Robotics, Computer-aided technologies for automation systems, Sensing and measurement, Navigation, guidance, and control. JAS is indexed by IEEE, ESCI, EI, Inspec, Scopus, SCImago, CSCD, CNKI. We are pleased to announce the new 2016 CiteScore (released by Elsevier) is 2.16, ranking 26% among 211 publications in Control and System Engineering category.

To learn more about JAS, please visit:

Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.