Skip to main content

ORIGINAL RESEARCH article

Front. Bioeng. Biotechnol., 10 August 2022
Sec. Biomechanics
Volume 10 - 2022 | https://doi.org/10.3389/fbioe.2022.857975

The reliability and validity of gait analysis system using 3D markerless pose estimation algorithms

  • 1CAS Key Laboratory of Human-Machine Intelligence-Synergy Systems, Research Center for Neural Engineering, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
  • 2Shenzhen College of Advanced Technology, University of Chinese Academy of Sciences, Shenzhen, China
  • 3Guangdong-Hong Kong-Macao Joint Laboratory of Human-Machine Intelligence-Synergy Systems, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China

Quantifying kinematic gait for elderly people is a key factor for consideration in evaluating their overall health. However, gait analysis is often performed in the laboratory using optical sensors combined with reflective markers, which may delay the detection of health problems. This study aims to develop a 3D markerless pose estimation system using OpenPose and 3DPoseNet algorithms. Moreover, 30 participants performed a walking task. Sample entropy was adopted to study dynamic signal irregularity degree for gait parameters. Paired-sample t-test and intra-class correlation coefficients were used to assess validity and reliability. Furthermore, the agreement between the data obtained by markerless and marker-based measurements was assessed by Bland–Altman analysis. ICC (C, 1) indicated the test–retest reliability within systems was in almost complete agreement. There were no significant differences between the sample entropy of knee angle and joint angles of the sagittal plane by the comparisons of joint angle results extracted from different systems (p > 0.05). ICC (A, 1) indicated the validity was substantial. This is supported by the Bland–Altman plot of the joint angles at maximum flexion. Optical motion capture and single-camera sensors were collected simultaneously, making it feasible to capture stride-to-stride variability. In addition, the sample entropy of angles was close to the ground_truth in the sagittal plane, indicating that our video analysis could be used as a quantitative assessment of gait, making outdoor applications feasible.

Introduction

Gait parameters have been proposed as an index of overall gait pathology for elderly people. It uses kinematic and kinetic variables to be taken as a cleaner reflection of gait quality (Brach et al., 2001). Kinematic analysis is often performed in laboratory research using three-dimensional motion capture or wearable sensors, which are expensive, immobile, data-limited, and require expertise (Cronin, 2021). Recently, video-based pose estimation suggests the potential for analyzing gait kinematic parameters (Andriluka et al., 2014).

Video-based 2D body pose estimation is a well-studied problem in computer vision, with state-of-the-art methods being based on deep networks (Toshev and Szegedy, 2014; Fang et al., 2017; Güler et al., 2018). With the advent of deep neural networks, it is now possible to estimate joint angles without the need for reflective markers. One of the more popular approaches, CMU’s OpenPose enables key body landmarks to be tracked from multiple humans in a video in real-time (Cao et al., 2017). Yagi et al. (2020) used OpenPose to detect multiple individuals and their joints in images to estimate step positions, stride length, step width, walking speed, and cadence, in comparison with multiple infrared camera motion capture system OptiTrack (Lénárt et al., 2018). Kidziński et al. (2020) designed machine learning models (e.g., convolutional neural networks, random forest, and ridge regression models) to predict clinical gait metrics based on trajectories of 2D body poses extracted from videos using OpenPose. Similarly, Stenum et al. (2021) used OpenPose to compare spatiotemporal and sagittal kinematic gait parameters of healthy adults against recorded optical marker–based motion captured from walking simultaneously. These previous studies have been performed in comparison between markerless and marker-based methods; however, they only learned to infer joint angles or joint locations in the sagittal plane.

Some researchers have detected 3D skeletons by existing 2D human pose detectors from images/video (Martinez et al., 2017; Moreno-Noguer, 2017; Nie et al., 2017) by directly using image features (Zhou et al., 2016; Zhou et al., 2018). Martinez et al. (2017) used a relatively simple deep feed-forward network to lift 2D pose to 3D pose efficiently based on given high-quality 2D joint information. Nakano et al. (2020) compared joint positions estimated from the 3D markerless motion capture technique based on OpenPose with multiple synchronized cameras against recorded three-dimensional motion capture. Multicamera systems are not easy to deploy in real-life environments. Instead, we try to adopt the OpenPose to computer the 2D pose input to our 3DPoseNet based on a single-camera system and then estimate gait parameters in sagittal, coronal, and transverse planes.

The main contributions of this work include 1) novelty: unlike previous studies, we did not directly extract the trajectories of 2D body poses to predict gait metrics using machine learning models. Instead, we try to estimate the 3D human pose from video to measure gait parameters. Then, we clarify associations and agreements of motion analysis using markerless and marker-based systems and confirm the reliability and validity of 3D human pose from video. 2) Dataset: we have collected the synchronized motion capture cameras and a single-camera video dataset of movement sequences for elderly and young people. 3) Application: our video-based gait analysis workflow is freely available, involves minimal user input, and does not require prior gait analysis expertise.

Methods

In the laboratory workflow, participants were marked with a total of 39 markers placed on bony landmarks. The Body39 joints were labeled by the Plug-In Gait full body model in Nexus software. Then, participants were required to perform the clear step on the treadmill, keeping 3 s and repeating three times. The positions of these markers were tracked by several optical cameras, which were later reconstructed into 3D position time series. Measures derived from the Vicon data served as ground_truth labels. In our proposed workflow, we used a camera to record the participant’s movement. The open-source OpenPose algorithm was adopted to extract trajectories of 2D key points from continuous images. The 2D joint positions from OpenPose were used as input, and we adopted a 3D keypoint detection algorithm (3DPoseNet) to estimate body joint locations in 3-dimensional space. It was important to note that these two workflows were synchronized by hardware. Finally, these signals from two workflows were converted to joint angles as a function of time.

Experimental setting

Figure 1 showed the laboratory environment and setup, allowing us to capture data from seven sensors (six Vicon MX motion capture cameras and one Vue video camera). The designated laboratory area was about 5 m × 8 m × 3 m, where participants were fully visible in all cameras. The motion capture cameras were rigged on the wall shelf, four on each left and right edge and two roughly mid-way on the horizontal edges, for recoding three-dimensional marker trajectories at 60 Hz (Liang et al., 2021). The video camera was also rigged on the wall shelf, for recording images of the walking sequences at 60 Hz. The camera images were RGB files with a 1920 × 1080 pixel resolution. Multiple infrared cameras and digital video camera recording were used for hardware synchronization so that each time point of the motion capture data point corresponded to the time of every video frame.

FIGURE 1
www.frontiersin.org

FIGURE 1. Overview of the experimental environment and setup.

Moreover, 15 healthy elders [mean (SD) age: 56.6 (2.53) years] and 15 healthy young people [mean (SD) age: 27.27 (4.31) years] wore minimal, close-fitting clothes and participated in walking on the treadmill. The speed was set at 1.5 km/h for the younger and 0.8 km/h for the elderly. Motion capture was recorded by tracking 39 markers, Figure 2A showed the Body39 joints labeled by the Plug-In Gait full body model in Nexus software (Vicon®, 2002). The standard protocols of setting markers were designed to match the skeletal configuration of the Human3.6M dataset (Ionescu et al., 2014).

FIGURE 2
www.frontiersin.org

FIGURE 2. Body39 joints are based on the Plug-In Gait full body model (A) and OpenPose Body25 keypoint model (B).

Extracting 2D keypoints with OpenPose

The OpenPose algorithm first estimates features from each image using a 10-layer VGG19 network (Cao et al., 2017). In addition, the obtained feature map is put into two convolutional neural networks for calculating the confidence and affinity vectors for each key point. The heat maps with confidence and with affinity fields are obtained. Then, the 2D body joint locations are clustered, according to dichotomy matching in graph theory. The nonparametric representations called part affinity fields are used to regress joint position and body segment connections between the joints. Finally, the output has the confidence of prediction and X and Y pixel coordinates. Figure 2B shows the Body25 joints labeled by the OpenPose body model.

Extracting 3D keypoints with 3DPoseNet

After obtaining 2D detections using OpenPose, our goal is to estimate 3D body joint locations. Formally, our input is a series of 2D points xϵ2n, and our output is a series of points in 3D space  yϵ3n. We aim to learn a function  f:2n3n that minimizes the prediction error over a dataset of N poses:

f=minf1Ni=1N(f(xi)yi),(1)

where xi is 2D joint obtained from the output of OpenPose. f is a deep and multilayer neural network with batch normalization (Ioffe and Christian, 2015), dropout (Srivastava et al., 2014), rectified linear units (RELUs) (Nair and Hinton, 2010), and residual connections (He et al., 2016).

Figure 3 shows a diagram with the basic building blocks of 3DPoseNet. The network is a multilayer convolutional neural network which inputs an array of 2D joint positions and outputs a series of joint positions in 3D. First, the linear layer applies directly to the input, which increases its dimensionality to 1024. Then, there are two same residual blocks. Each block includes two linear layers, Batch Norm, RELUs, and Dropout, with residual connections. Before the final prediction, the linear layer can be applied to produce outputs of size 3n. Initially, the weights of our linear layers are set using Kaiming initialization (He et al., 2015). The 3DPoseNet is trained for 200 epochs using Adam (Kingma and Ba, 2014), a starting learning rate of 0.001 and exponential decay, using mini-batches of size 64. In addition, we first train the network using the Human3.6M dataset (Ionescu et al., 2014).

FIGURE 3
www.frontiersin.org

FIGURE 3. Diagram with the basic building blocks of 3DPoseNet.

Gait parameter extraction

First of all, we filled gaps in keypoint trajectories using linear interpolation and smoothed trajectories using a one-dimensional unit-variance Gaussian filter. It should be noted that the markerless-based system returns 3D coordinates resolved in a local system around the middle of the hip joint. The 3D coordinates provided by the marker-based system came from a 3D global coordinate reference system fixed on the ground. These two different coordinate systems were moved to a new, coincident human-based local coordinate system {O}, as shown in Figure 4A. We centered each univariate time series by subtracting the coordinates of the pelvis and scaling all values by dividing them by the Euclidean distance between the right hip and the right shoulder.

FIGURE 4
www.frontiersin.org

FIGURE 4. Illustration of the human body coordinate system (A) and extracted joint angle features for the subject (B).

The definition of the coordinate system was as follows: X was anterior/posterior, Y was lateral/medial, and Z was inferior/superior. The alignment procedure was taken from the study by Kabsch, (1976) and involved the initial rotation of the measurement systems, followed by the translation toward the desired origin. The origin of the coordinate system {O} at the time t was denoted as o(t)=1/2(PR(t)+PL(t)), where PR(t) and PL(t) were the 3D position of the RASI and LASI joints at time t, respectively. Two unit vectors can be determined by V1=(PR(t)PL(t))/PR(t)PL(t) and V2=(o(t)o(t1))/o(t)o(t2). Accordingly, {O}=[xh,yh,zh] can be denoted by

xh=V1,(2)
yh=V1×V2V1×V2,(3)
zh=xh×yhxh×yh.(4)

Commonly, the 3D limb skeleton can be represented by three link segments: Upper Trunk, Thigh, and Shank. The 3D position vector between the right and left shoulder was denoted with  u. The 3D position of the right hip, knee, and ankle as joints was presented as  δ, α, and β, respectively. The upper trunk, right thigh, and shank as three links were represented as U=uo, T=αδ, and S=αβ, respectively, as shown in Figure 4B. The knee angle was illustrated in the equation as follows:

RTS=arccosTST×S.(5)

Furthermore, we calculated the joint angles between each link segment {U,T,S} with respect to the normal vectors of the sagittal, coronal, and transverse planes {xh,yh,zh} and indicated as RUX, RTX,  RSX,  RUY, RTY,  RSY,  RUZ, RTZ, and  RSZ.

It should be noted that we estimated aforementioned features within each gait cycle, and the gait cycle was defined as the time interval between two consecutive heel-strike events. A heel-strike was the contact points between the heel and surface. We calculated gait events of right heel-strikes in marker motion capture and camera data by independently applying the different methods to each set of data. For marker data, we found one derived time series helpful for improving the detection of the gait cycle. The time series was the x-coordinates of the right ankles. Heel-strikes were defined by the time points of positive peaks in the anterior–posterior ankle trajectories. For the markerless-based system, heel-strike events were detected by visual inspection. The process was greatly aided by the identification of ankle key points obtained from the deep neural network model. In this letter, the RUX, RTX,  RSX,  RUY, RTY,  RSY,  RUZ, RTZ, and  RSZ were extracted at 2% increments throughout the entire cycle Φ(t)K(k=30), for representing the gait pattern at each time stamp.

Moreover, we used sample entropy to measure the time series of joint angles in a gait cycle. Sample entropy is a nonlinear measurement way to analyze time series signals and is proposed by Richman and Moorman (2000). A higher sample entropy value indicates more randomness in time series, and lower value shows more self-similarity. The computational procedures of sample entropy (SampEn) are as follows (Pham, 2010):

Given a standardized (with zero mean and unit variance) time series {x(j);1jN}, N is the total number of data points.

Step 1. Construct subsequences of length m: Xm(1), Xm(2),,  Xm(Nm), where Xm(i)={x(i+k);0km1} and m is called as embedding dimension.

Step 2. Compute the distance between Xm(i) and Xm(j), represented by d(Xm(i),Xm(j)), as:

d(Xm(i),Xm(j))=max{|x(i+k)x(j+k)|;0km1, 1i, jNm, ij}(6)

Step 3. Calculate the probability that any vector Xm(j) which is similar to Xm(i) within T as follows:

Ci(m,T)=ni(m,T)Nm+1,i=1,, Nm+1,(7)

where ni(m,T) is the number of vectors Xm(j) that are similar to Xm(i) subject to the criterion of similarity: d(Xm(i),Xm(j))T.

Step 4. Calculate

(m,T)=1Nm+1i=1Nm+1Ci(m,T).(8)

Step 5. Set m=m+1 and repeat steps 1–4.

Step 6. Calculate the sample entropy as

SampEn(N,m,T)=ln(m+1,T)(m,T).(9)

Hence, sample entropy is the negative natural logarithm of condition probability, without allowing self-matches. To calculate sample entropies of those time series, it is important to determine the appropriate values of the parameters m and T. Usually, the constant value of m is 1 or 2, T value ranging from 0.1 SD to 0.25 SD (SD is the standard deviation of time series) (Lake et al., 2002). In this experiment, we selected m=2 and T=0.2SD for each gait cycle data.

Statistical analysis

Reliability and validity are central characteristics that define the quality of measurement methods and the test result potential for application in research and clinical practice (Michelini et al., 2020). First, we needed to assess the test–retest reliability within markerless and marker-based motion analysis systems, using intra-class correlation coefficients [ICC (C, 1)]. ICCs were determined as follows: almost perfect, 0.81–1.0; substantial, 0.61–0.80; moderate, 0.41–0.60; fair, 0.21–0.40; slight, 0.00–0.20 (Koo and Li, 2016).

Next, to assess the potential difference in joint angles among different measurement systems, an independent-sample t-test was used, when the variable conformed to the normality and homogeneity of variance simultaneously. If the variable did not conform to the normality or homogeneity assumption, the Wilcoxon nonparametric test was used for the difference analysis (Liang et al., 2021). In the event of a statistically significant main effect, we performed post-hoc pairwise comparisons with Bonferroni corrections.

Furthermore, we calculated standard errors of measurements (SEM) and intra-class correlation coefficients [ICC (A, 1)] with 95% confidence intervals (CI) of each joint angle to assess correlations and consistency. The smallest detectable change (SDC) was calculated as 1.962SEM. SDC can be regarded as the smallest change between any two steps that cannot be attributed to measurement errors (Furlan and Sterr, 2018). In addition, agreements between the joint angles at maximum flexion obtained from the markerless and marker-based systems were assessed by Bland–Altman analysis, which permits the delineation of systematic analysis (Bland and Altman, 1986). p-values less than 0.05 were deemed statistically significant. Data preprocessing, algorithms, and statistical analyses were implemented using Python (version 3.5).

Results

Participant characteristics

The data for gender and age did not conform to the normality (p < 0.05); therefore, the Wilcoxon nonparametric tests were used for difference analysis. The data of mass, height, and BMI were consistent with the normality and homogeneity assumptions (p > 0.05); therefore, an independent-sample t-test was used for the difference analysis. As shown in Table 1, the elderly group demonstrated a larger BMI than the young group (elderly: 24.17 ± 2.81, young: 21.53 ± 2.11, and p = 0.009). Moreover, age showed significant differences (elderly: 56.60 ± 2.53, young: 27.27 ± 4.31, and p = 0.000). It was indicated that strategies used for the control of gait deviation related to age between healthy young and elderly adult play an important role (Mehdizadeh, 2018).

TABLE 1
www.frontiersin.org

TABLE 1. Comparison of characteristics between elderly and young groups.

The reliability of motion analysis system

Mean values ±standard deviation and intra-class correlation coefficients for sample entropy of each joint angle test-retested using camera and Vicon are described in Table 2. There were moderate ICCs of the data obtained by marker motion capture (ICCs = 0.466–0.741 and p < 0.05). Similarly, ICCs of the data obtained by markerless motion capture were moderate (ICCs = 0.506–0.734 and p < 0.05). The reliability was confirmed on the joint angles measured using a single camera.

TABLE 2
www.frontiersin.org

TABLE 2. Statistic property for marker and markerless motion analysis during test–retest.

The validity of motion analysis system

All gait parameters were normally distributed and showed homogeneity of variance. Criterion validity, analyzed by the comparisons of joint angles results extracted from video recordings and results generated by the marker motion capture system, is shown in Table 3 and illustrated by the Bland–Altman plot (Figure 5).

TABLE 3
www.frontiersin.org

TABLE 3. Comparison of characteristics between marker and markerless motion analysis.

FIGURE 5
www.frontiersin.org

FIGURE 5. Bland–Altman plot of RTS (A), RUX (B), RTX (C), RSX (D), RUY (E), RTY (F), RSY (G), RUZ (H), RTZ (I), and RSZ (J) at maximum flexion and comparison between the markerless- and marker-based systems.

We calculated the sample entropy of joint angles across the stride cycle that was averaged for the walking bout of each individual participant, shown in Table 3. The SDC showed small values, with ranges between 0.040 and 0.071. It was indicated that small real differences between measurements could be detected by our method. Bias for sample entropy of joint angles between measurements was also shown to be over small, 0.034 for RTS, 0.018 for RUX, 0.081 for RTX, and 0.047 for RSX, and they were not statistically different (p > 0.05). This is supported by intra-class correlation coefficients that were substantial and the ICC estimates exceeded 0.61, indicating excellent reliability. Only reliability for RSY [95% CI = (0.291, 0.699)] and RTZ [95% CI = (0.378, 0.7360)] was lower, but still on a level of good-to-excellent reliability (p < 0.05). The Bland–Altman plots for each joint angle at maximum flexion obtained from the markerless and marker-based systems are provided in Figure 5. The x-axis and y-axis represent the average and difference between the outputs of the two methods, respectively. Thick and dotted lines denote the mean and zero of difference, respectively. The green and red lines represent the upper and lower limits of 95%, respectively, indicating most data are distributed within this range. The blue line shows that the difference between the two methods increases/decreases as the angle increases. The thick line was close to the dotted line on RTS, RUX, RTX, RSX, RTY, and RTZ, indicating higher consensus among the results of the two methods, otherwise on RUY, RSY, RUZ, and RSZ. This was consistent with the results given in Table 3.

Discussion

In this study, the markerless pose estimation algorithm, OpenPose combined with 3DPoseNet, and signal processing techniques were used to evaluate a non-invasive method capable of capturing gait parameters. The main findings show that the proposed methods for extracting gait angles possess their own good validity and reliability. Several previous studies have used markerless-based analysis to study gait patterns of walking or other human movements (Michelini et al., 2020; Nakano et al., 2020; Ota et al., 2020). Our findings were consistent with these reports in that 3D markerless pose estimation in providing quantitative information about human movement is very promising.

First, ICC (C, 1) indicates the test–retest reliability within markerless- and marker-based systems was in almost complete agreement. At joint angles RTS, RUX, RTX, and RSX, there was no significant difference between the markerless- and marker-based motion analysis systems and the ICC (A, 1) was high enough. However, RUY, RTY, RSY, RUZ, RTZ, and RSZ were fair. One of the reasons is that there are differences in measurement methods of sample entropy of angles between markerless- and marker-based systems. While a marker-based system consisted of several cameras to form three-dimensional motion data, a markerless-based motion analysis system combined with deep algorithms to provide three-dimensional coordinates based on one camera. Therefore, the rotation motion has not been accurately measured. Another reason is that RUY, RTY, RSY, RUZ, RTZ, and RSZ were defined as the coronal and transverse plane angle of the normal vector relative to the upper trunk, right thigh, and shank. In the experiment, we acquired a dataset from stationary video camera recordings of healthy human gait, with sagittal plane views. It is possible that camera angles would likely affect results.

Some limitations of this study should be noted. First, some sources of error may be intrinsic to 3D markerless pose estimation. First and most obvious, it is difficult to track human movements frame-by-frame perfectly from the video. For example, the left and right segments could interchange or disappear in OpenPose. In this situation, our 3DPoseNet algorithm cannot predict 3D pose from a failed detector output. Second, the markerless-based identification system is unlikely to be equivalent to the marker landmarks. While marker placement depends on manual palpation of bony landmarks, a markerless-based system relies on visually labeled generalized key points. The placement of motion capture markers also owns some degree of error. These errors can affect the validity of the precision evaluation of a markerless-based system. Despite these limitations, the marker motion capture has been recognized as the ground truth (Nakano et al., 2020). As a result, the proposed method for evaluating a markerless-based system can be considered to be reasonable.

We did not pre-estimate the sample size. However, it exceeded what is commonly required for reliability studies (Koo and Li, 2016), since 30 subjects were instructed to walk on three times for the sake of collecting enough information to perform the analysis. Our approach also includes manual marking of the contact points between the heel and surface by visual inspection, which we register in order to accumulate a database to refine gait events. In the next work, it may be possible to obtain more accurate video-based analyses by training gait-specific networks from coronal and transverse views. It may be beneficial to train networks that are specific to each population, such as elderly people who experience accidental falls or abnormal gait. The markerless-based analysis described in the current study is promising for future applications. Such a method can classify different gait types and automatically extract quantitative gait information from the video.

Conclusion

This study demonstrated the potential for combining OpenPose and 3DPoseNet markerless pose estimation algorithms to identify gait pathology. Given economic and time constraint problems, we have gained several insights from this exercise: 1) laboratory-based optical motion capture is a reasonable baseline predictor, while 3D markerless pose estimation networks were close to the ground_truth statistically significantly; 2) quantitative evaluations indicate that our proposed workflow trained on experimental movements can be generalized to non-experimental-specific poses; 3) correlation between the quantified results of network convergence support our initial hypothesis that learning a mapping from images to predict kinematics gait parameters is feasible; 4) the test–retest reliability within the device was in almost complete agreement. It was indicated that our video analysis could be used as a quantitative assessment of gait outside of a clinic. For predicting abnormal gait patterns or fall risk, future work should also include elderly people who experienced a fall.

Data availability statement

The raw data supporting the conclusion of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving human participants were reviewed and approved by the Shenzhen Institute of Advanced Technology Institutional Review Board. The patients/participants provided their written informed consent to participate in this study.

Author contributions

GZ: conceptualization, methodology, and funding acquisition; SL: data curation and writing—original draft preparation; YZ: formal analysis, visualization, and investigation; YD: software and validation; GL: writing—reviewing and editing and funding acquisition.

Funding

This work was supported in part by the National Key R&D Program of China (2018YFC2001400/04, 2019YFB1311400/01), in part by Shandong Key R&D Program (2019JZZY011112), in part by the Innovation Talent Fund of Guangdong Tezhi Plan (2019TQ05Z735), in part by High Level-Hospital Program, Health Commission of Guangdong Province (HKUSZH201901023), in part by Guangdong-Hong Kong-Macao Joint Laboratory of Human-Machine Intelligence-Synergy Systems (2019B121205007).

Acknowledgments

The authors would like to thank all the volunteers for participating in this study.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Andriluka, M., Pishchulin, L., Gehler, P., and Schiele, B. (2014). “2d human pose estimation: new benchmark and state of the art analysis,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 3686–3693.

CrossRef Full Text | Google Scholar

Bland, J. M., and Altman, D. G. (1986). Statistical methods for assessing agreement between two methods of clinical measurement. lancet 327 (8476), 307–310. doi:10.1016/s0140-6736(86)90837-8

CrossRef Full Text | Google Scholar

Brach, J. S., Berthold, R., Craik, R., VanSwearingen, J. M., and Newman, A. B. (2001). Gait variability in community-dwelling older adults. J. Am. Geriatrics Soc. 49 (12), 1646–1650. doi:10.1111/j.1532-5415.2001.49274.x

CrossRef Full Text | Google Scholar

Cao, Z., Simon, T., Wei, S. E., and Yaser, S. (2017). “Realtime multi-person 2d pose estimation using part affinity fields,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 7291–7299.

CrossRef Full Text | Google Scholar

Cronin, N. J. (2021). Using deep neural networks for kinematic analysis: challenges and opportunities. J. Biomechanics 123, 110460. doi:10.1016/j.jbiomech.2021.110460

PubMed Abstract | CrossRef Full Text | Google Scholar

Fang, H. S., Xie, S., Tai, Y. W., and Lu, C. (2017). “Rmpe: Regional multi-person pose estimation,” in Proceedings of the IEEE international conference on computer vision, 2334–2343.

CrossRef Full Text | Google Scholar

Furlan, L., and Sterr, A. (2018). The applicability of standard error of measurement and minimal detectable change to motor learning research—a behavioral study. Front. Hum. Neurosci. 12, 95. doi:10.3389/fnhum.2018.00095

PubMed Abstract | CrossRef Full Text | Google Scholar

Güler, R. A., Neverova, N., and Kokkinos, I. (2018). “Densepose: dense human pose estimation in the wild,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 7297–7306.

Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 770–778.

CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2015). “Delving deep into rectifiers: Surpassing human-level performance on imagenet classification,” in Proceedings of the IEEE international conference on computer vision, 1026–1034.

CrossRef Full Text | Google Scholar

Ioffe, S., and Christian, S. (2015). “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in International conference on machine learning, 448–456.

Google Scholar

Ionescu, C., Papava, D., Olaru, V., and Sminchisescu, C. (2014). Human3. 6m: large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE Trans. Pattern Anal. Mach. Intell. 36 (7), 1325–1339. doi:10.1109/tpami.2013.248

PubMed Abstract | CrossRef Full Text | Google Scholar

Kabsch, W. (1976). A solution for the best rotation to relate two sets of vectors. Acta Cryst. Sect. A 32 (5), 922–923. doi:10.1107/s0567739476001873

CrossRef Full Text | Google Scholar

Kidziński, Ł., Yang, B., Hicks, J. L., Rajagopal, A., Delp, S. L., Schwartz, M. H., et al. (2020). Deep neural networks enable quantitative movement analysis using single-camera videos. Nat. Commun. 11 (1), 4054. doi:10.1038/s41467-020-17807-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Kingma, D. P., and Ba, J. (2014). Adam: a method for stochastic optimization. Comput. Sci.

Google Scholar

Koo, T. K., and Li, M. Y. (2016). A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J. Chiropr. Med. 15 (2), 155–163. doi:10.1016/j.jcm.2016.02.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Lake, D. E., Richman, J. S., Griffin, M. P., and Moorman, J. R. (2002). Sample entropy analysis of neonatal heart rate variability. Am. J. Physiology-Regulatory, Integr. Comp. Physiology 283 (3), R789–R797. doi:10.1152/ajpregu.00069.2002

CrossRef Full Text | Google Scholar

Lénárt, Z., Nagymáté, G., and Szabó, A. (2018). Validation process of an upper limb motion analyzer using optitrack motion capture system. Biomech. Hung. 11 (2), 93–99. doi:10.17489/2018/2/07

CrossRef Full Text | Google Scholar

Liang, S., Zhao, G., Zhang, Y., Diao, Y., and Li, G. (2021). Stability region derived by center of mass for older adults during trivial movements. Biomed. Signal Process. Control 69, 102952. doi:10.1016/j.bspc.2021.102952

CrossRef Full Text | Google Scholar

Martinez, J., Hossain, R., and Romero, J. (2017). “A simple yet effective baseline for 3d human pose estimation,” in Proceedings of the IEEE international conference on computer vision, 2640–2649.

CrossRef Full Text | Google Scholar

Mehdizadeh, S. (2018). The largest lyapunov exponent of gait in young and elderly individuals: a systematic review. Gait Posture 60 (2), 241–250. doi:10.1016/j.gaitpost.2017.12.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Michelini, A., Eshraghi, A., and Andrysek, J. (2020). Two-dimensional video gait analysis: A systematic review of reliability, validity, and best practice considerations. Prosthet. Orthot. Int. 44 (4), 245–262. doi:10.1177/0309364620921290

PubMed Abstract | CrossRef Full Text | Google Scholar

Moreno-Noguer, F. (2017). “3d human pose estimation from a single image via distance matrix regression,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2823–2832.

CrossRef Full Text | Google Scholar

Nair, V., and Hinton, G. E. (2010). “Rectified linear units improve restricted boltzmann machines,” in International conference on machine learning, 807–814.

Google Scholar

Nakano, N., Sakura, T., Ueda, K., Omura, L., Kimura, A., Iino, Y., et al. (2020). Evaluation of 3D markerless motion capture accuracy using OpenPose with multiple video cameras. Front. Sports Act. Living 2, 50. doi:10.3389/fspor.2020.00050

PubMed Abstract | CrossRef Full Text | Google Scholar

Nie, B., Wei, P., and Zhu, S. C. (2017). “Monocular 3d human pose estimation by predicting depth on joints,” in IEEE international conference on computer vision, 3447–3455.

CrossRef Full Text | Google Scholar

Ota, M., Tateuchi, H., Hashiguchi, T., Kato, T., Ogino, Y., Yamagata, M., et al. (2020). Verification of reliability and validity of motion analysis systems during bilateral squat using human pose tracking algorithm. Gait posture 80, 62–67. doi:10.1016/j.gaitpost.2020.05.027

PubMed Abstract | CrossRef Full Text | Google Scholar

Pham, T. D. (2010). GeoEntropy: a measure of complexity and similarity. Pattern Recognit. 43 (3), 887–896. doi:10.1016/j.patcog.2009.08.015

CrossRef Full Text | Google Scholar

Richman, J. S., and Moorman, J. R. (2000). Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiology-Heart Circulatory Physiology 278 (6), H2039–H2049. doi:10.1152/ajpheart.2000.278.6.h2039

CrossRef Full Text | Google Scholar

Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R. (2014). Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15 (1), 1929–1958.

Google Scholar

Stenum, J., Rossi, C., and Roemmich, R. T. (2021). Two-dimensional video-based analysis of human gait using pose estimation. PLoS Comput. Biol. 17 (4), e1008935. doi:10.1371/journal.pcbi.1008935

PubMed Abstract | CrossRef Full Text | Google Scholar

Toshev, A., and Szegedy, C. (2014). “Deeppose: human pose estimation via deep neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 1653–1660.

CrossRef Full Text | Google Scholar

Vicon®, (2002). Vicon® manual, Vicon®612 motion systems. Oxford, UK: Oxford Metrics Ltd. Plug-in-Gait modelling instructions

Google Scholar

Yagi, K., Sugiura, Y., Hasegawa, K., and Saito, H. (2020). Gait measurement at home using a single RGB camera. Gait posture 76, 136–140. doi:10.1016/j.gaitpost.2019.10.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, X., Zhu, M., and Leonardos, S. (2016). “Sparseness meets deepness: 3d human pose estimation from monocular video,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 4966–4975.

CrossRef Full Text | Google Scholar

Zhou, X., Zhu, M., Pavlakos, G., Leonardos, S., Derpanis, K. G., Daniilidis, K., et al. (2018). Monocap: monocular human motion capture using a cnn coupled with a geometric prior. IEEE Trans. Pattern Anal. Mach. Intell. 41 (4), 901–914. doi:10.1109/tpami.2018.2816031

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: single-camera video, 3D markerless pose estimates, 3D marker-based motion analysis, validity, reliability

Citation: Liang S, Zhang Y, Diao Y, Li G and Zhao G (2022) The reliability and validity of gait analysis system using 3D markerless pose estimation algorithms. Front. Bioeng. Biotechnol. 10:857975. doi: 10.3389/fbioe.2022.857975

Received: 19 January 2022; Accepted: 13 July 2022;
Published: 10 August 2022.

Edited by:

Rezaul Begg, Victoria University, Australia

Reviewed by:

Susanna Summa, Bambino Gesù Children’s Hospital (IRCCS), Italy
João Paulo Morais Ferreira, Superior Institute of Engineering of Coimbra (ISEC), Portugal

Copyright © 2022 Liang, Zhang, Diao, Li and Zhao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Guoru Zhao, gr.zhao@siat.ac.cn

Download