ORIGINAL RESEARCH article
Volume 7 - 2019 | https://doi.org/10.3389/fbioe.2019.00473
Sensor Data Required for Automatic Recognition of Athletic Tasks Using Deep Neural Networks
- School of Human Kinetics, Faculty of Health Sciences, University of Ottawa, Ottawa, ON, Canada
Movement screens are used to assess the overall movement quality of an athlete. However, these rely on visual observation of a series of movements and subjective scoring. Data-driven methods to provide objective scoring of these movements are being developed. These currently use optical motion capture and require manual pre-processing of data to identify the start and end points of each movement. Therefore, we aimed to use deep learning techniques to automatically identify movements typically found in movement screens and assess the feasibility of performing the classification based on wearable sensor data. Optical motion capture data were collected on 417 athletes performing 13 athletic movements. We trained an existing deep neural network architecture that combines convolutional and recurrent layers on a subset of 278 athletes. A validation subset of 69 athletes was used to tune the hyperparameters and the final network was tested on the remaining 70 athletes. Simulated inertial measurement data were generated based on the optical motion capture data and the network was trained on this data for different combinations of body segments. Classification accuracy was similar for networks trained using the optical and full-body simulated inertial measurement unit data at 90.1 and 90.2%, respectively. A good classification accuracy of 85.9% was obtained using as few as three simulated sensors placed on the torso and shanks. However, using three simulated sensors on the torso and upper arms or fewer than three sensors resulted in poor accuracy. These results for simulated sensor data indicate the feasibility of classifying athletic movements using a small number of wearable sensors. This could facilitate objective data-driven methods that automatically score overall movement quality using wearable sensors to be easily implemented in the field.
Movement screens are used to assess the overall movement quality of an athlete. Typically, the athlete will perform a series of movements while a trained rater visually observes and scores the movements. The goals of movement screens are to predict injury risk and identify performance deficits that can be targeted in training. While interrater and intrarater reliabilities for movement screens such as the Functional Movement Screen (FMS™) are good (Minick et al., 2010; Teyhen et al., 2012), interrater reliability for subtest components can be poor and dependent on rater experience (Smith et al., 2013; Gulgin and Hoogenboom, 2014; Bonazza et al., 2017). Furthermore, concerns have been raised that grading criteria can be somewhat ambiguous (Frost et al., 2015; Bonazza et al., 2017) and scores may not be sensitive enough to detect movement abnormalities (Clifton et al., 2013). Recent work has aimed to develop objective scoring methods for movement screens (Ross et al., 2018). Data-driven approaches have the potential to improve the repeatability of scoring and increase the ability to detect subtle differences in movement patterns. However, current methods require manual processing of motion capture data before scoring can be performed, including cropping trials to isolate each movement. Additionally, the reliance on optical motion capture could be a barrier to implementation of these methods in the field.
Wearable sensors are an attractive alternative to optical motion capture for motion analysis applications. They are cost-effective and portable, allowing for the collection of motion data outside of a laboratory and over large capture volumes. Furthermore, wearable sensors have the potential to be less cumbersome than optical markers depending on the number and placement of sensors. Previous work investigated optimal placement and number of sensors to classify activities of daily living (Pannurat et al., 2017), everyday activities (Kern et al., 2003; Olguin and Pentland, 2006; Atallah et al., 2011; Cleland et al., 2013), and fall detection (Gjoreski et al., 2011). However, which sensors are necessary to best classify movement screening tasks remains unclear.
Human activity recognition is an area of research that seeks to automatically identify human activities by applying machine learning techniques to motion data. Methods have been developed to classify movements including hand gestures (Kim and Toomajian, 2016), activities of daily living (Hammerla et al., 2016), and movements typical in various sports (Nguyen et al., 2015; Kautz et al., 2017). Previously, activity recognition methods employed techniques that required hand-selected features as input (Bulling et al., 2014). However, convolutional neural networks (CNNs), a type of deep neural network (DNN), are now commonly used to automatically generate features through deep learning (Zeng et al., 2014; Yang et al., 2015; Lee et al., 2017). CNNs have shown promising results in activity recognition; however, they are unable to capture time dependencies. Recurrent neural networks are a type of neural network that include a memory component that allows them to model temporal dependencies. The combination of CNNs to extract features with long-short-term memory (LSTM) recurrent networks to capture temporal dependences has provided improved classification performance over CNNs alone (Ordóñez and Roggen, 2016).
The use of deep neural networks in movement screens would allow for a continuous data collection during a movement screen. Individual movements could then be automatically identified and segmented as a preparation for further analysis or scoring. This would decrease the manual effort required for the analysis process and increase the utility of these objective measurement techniques. The ability to perform the movement classification and scoring based on data from a minimal set of wearable sensors would further increase the applicability of data-driven movement screens. Therefore, our first aim was to use a deep neural network to identify when movements typical of movement screens occur within motion data. Our second aim was to compare networks trained using optical motion capture data with those trained using data available from wearable sensors.
Data Collection and Processing
Optical motion capture data were collected from 417 athletes performing a series of movement tests by Motus Global (Rockville Center, NY). The athletes competed in a variety of sports, including baseball, basketball, soccer, golf, tennis, track and field, squash, cricket, lacrosse, football, and volleyball. They ranged in skill level from recreational athletes to those playing in major professional sports leagues (e.g., NBA, MLB, PGA, etc.). Participants provided informed consent for future use of their data for research before completing the protocol. The secondary use of the data was approved by the University of Ottawa Research Ethics Board. Forty-five retroreflective markers were placed on the athlete for motion tracking (Ross et al., 2018) and data were recorded at 120 Hz using an eight-camera Raptor-E (Motion Analysis, Santa Rosa, CA) motion capture system. Each athlete performed a series of movement tests consisting of 21 unique movements. The 13 movements most likely to challenge mobility and stability were selected for analysis in this study, including hop down right/left (HDR, HDL), bird-dog right/left (BDR, BDL), drop jump (DJ), T-balance right/left (TBR, TBL), step-down right/left (SDR, SDL), L-hop right/left (LHR, LHL), and lunge right/left (LR, LL) (Ross et al., 2018). Individual trials were collected for each movement.
Start and end time points were manually identified for each trial (Ross et al., 2018) for use as a ground truth of when each activity was performed. The optical motion data used in the analysis (OPT) included global x, y, z coordinates for 45 markers. To simulate data that can be obtained using inertial measurement units (sIMU), marker trajectories were processed in Visual3D (C-Motion, Inc., Germantown, MD) and global angular orientation Euler angles and the Euclidean norm of the center of mass linear acceleration and angular velocity for each body segment were calculated. The Euclidean norm of the velocity and acceleration was used to reduce the reliance on accurate sensor alignment. Accelerations and velocities were low-pass filtered at 15 Hz with a zero-lag second order Butterworth filter.
Deep Neural Network
Athletes were randomly separated into training (67%, n = 278), validation (33%, n = 69), and test (33%, n = 70) subsets. A single matrix was created for each subset by concatenating data from all movement trials performed by all athletes in the subset. Each variable was normalized by subtracting the mean and dividing by the standard deviation of all data frames across athletes and movements in the training set for that variable. A sliding window approach was used to divide the subset data into data segments containing an equal number of data frames. The stride for the sliding window was 1/4 the window size. Each data segment was assigned a label according to the movement that was performed for the majority of the data segment. A “Null” label was included to describe times when none of the movements were being performed for a total of 14 classes.
A deep neural network based on the work of Ordóñez and Roggen (2016) was implemented in PyTorch (Paszke et al., 2017). The architecture combines convolutional layers to extract features with recurrent layers to model the temporal dynamics. The network includes four convolutional layers, two long-short-term memory (LSTM) recurrent layers, a linear fully connected layer, and a softmax classifier (Figure 1). The input to the network is the windowed time series data. The length of the input data was the sliding window size and the number of columns depended on the data used: 3*45 for OPT (x, y, and z component of each trajectory) and 5 * number of body segments for the sIMU data (3 Euler angles + 1 angular velocity norm + 1 linear acceleration norm).
Figure 1. Architecture of the deep neural network used to classify athletic movements. The network combines convolutional and recurrent layers (Ordóñez and Roggen, 2016). Tensor sizes and function inputs based on the OPT data and final architecture parameters are shown. PyTorch functions and inputs are shown for each layer. SWS, sliding window size; C, number of CNN channels; N, number of columns in the input data; k, CNN kernel size; L, LSTM cells; Nclasses, number of movements classified.
For network training, a mini-batch size of 100 was used. A stochastic gradient descent (SGD) optimizer with momentum was used for training with a cross-entropy loss criterion. The DNN was trained to classify the movement performed during a given windowed data segment.
Hyperparameter tuning was performed using a grid search with the validation set of the optical motion data (OPT). The learning parameters were tuned first as these have a larger impact on classifier performance (Hammerla et al., 2016). The learning parameters were the learning rate and momentum of the SGD optimizer. Five values of each were explored (Table 1) resulting in 25 DNNs trained on the OPT training set. The micro-averaged F1 score was calculated for the validation set to assess the performance of each DNN. The F1 score is a measure of classification accuracy that is the harmonic mean of precision and recall (). The micro-averaged F1 score calculates the mean across the classes by considering all individual predictions, which is suitable for classes of different sizes. The micro-averaged F1 score is equivalent to the micro-averaged precision, micro-averaged recall, and classification accuracy.
After selecting the learning rate and momentum that produced the best F1 score, the architecture parameters were tuned. Two to three values were tested for each of the following parameters: sliding window size, CNN kernel filter size, CNN channels, and LSTM cells (Table 1). Note that CNN kernels of size 5 and 6 were used with window size 24 and CNN kernels of size 6 and 8 were used with window size 48. Models were assessed based on the micro-averaged F1 score.
Comparison of Simulated IMU Sensor Data
Once the final learning and architecture parameters were determined, the final model was used to identify movements in the test set. In this case, the DNN was used individually on each athlete. All trials of athlete's data were combined and then segmented using sliding windows and the DNN was used to classify each window. Then for each frame of data, the class probabilities from each window containing that frame were averaged, and a final classification was made for that frame of data.
DNNs using the final learning and architecture parameters were also trained for the simulated IMU data on the training subset. Different combinations of simulated sensor locations were examined (Table 2). DNNs were evaluated on the test subset following the procedure outlined above. For each DNN, the confusion matrix, accuracy, precision, recall, and F1 score were calculated. Micro and macro averages and metrics for each class were produced.
The learning rate and momentum of the SGD optimizer both had a large effect on the micro-averaged F1 score for the OPT validation set (Figure 2A). The best F1 score was obtained for a learning rate of 0.001 and momentum of 0.9, and these values were used for all subsequent models. The DNN parameters had a relatively small effect on the F1 scores, with values ranging from 0.895 to 0.911 (Figure 2B). The best results were obtained for a sliding window size of 48 (0.04 s), CNN kernel size of 6 frames, 32 CNN channels, and 64 LSTM cells. These parameters were selected for use in the final DNN.
Figure 2. Hyperparameter tuning was performed in two steps: learning parameters (A) and architecture parameters (B). (A) Effect of learning rate and momentum on micro-averaged F1 score. (B) Effect of sliding window size, CNN kernel size, CNN channels, and LSTM cells on micro-averaged F1 score. Mean and standard deviation of all DNNs at each parameter level are shown.
Comparison of Simulated IMU Sensor Data
Deep neural networks trained using optical data (OPT) and all 13 body segments of sIMU data (sIMU13) had similar micro-averaged F1 scores (0.901 and 0.902, respectively). In general, including more body segments improved performance (Figure 3), although only small improvements were obtained by including more than four body segments. Bird-dog (BDR/L) movements were predicted well (F1 score > 0.76) for all networks, while drop jumps (DJ) tended to be more poorly identified in general.
Figure 3. (A) Micro-averaged F1 score for DNNs trained using simulated IMU (sIMU) data from various combinations of body segments. (B) F1 score for each movement for a selection of DNNs trained on sIMU data. Scores were calculated on the test set based on classification of individual data frames. Movements are HDR/L, hop down right/left; BDR/L, bird-dog right/left; SDR/L, step-down right/left; LHR/L, L-hop right/left; DJ, drop jump; LR/L, lunge right/left; TBR/L, T-balance right/left.
The effect of including upper or lower limb data can be observed in the confusion matrices for the sIMU3U and sIMU3L models (Figure 4). With the torso and upper arms included (sIMU3U), the DNN frequently confuses left and right versions of tasks. Tasks involving jumping were also confused. The network using the torso and shanks (sIMU3L) is better able to distinguish between left and right, but occasionally confuses the T-balance (TBR/L) and lunge tasks (LR/L). L-hops (LHR/L) are sometimes classified as hop downs (HDR/L) in both three-segment networks (sIMU3L, sIMU3U).
Figure 4. Confusion matrices for the sIMU3U (A) and sIMU3L (B) DNNs. Values are percentage of the frames of the true movement classified as the predicted movement.
The true and predicted movements over time for the OPT, sIMU1, sIMU3L, and sIMU13 models are shown in Figure 5 for a representative athlete. OPT, sIMU13, and sIMU3L were better able to predict the entire duration of movements. Networks with fewer body segments tended to switch between predictions. The misclassification between movements and Null largely occurs at the beginning and end of a movement.
Figure 5. Example of movement classification for a representative athlete using DNNs trained using different sets of input data. Data collected in separate trials for each movement have been concatenated and displayed continuously.
The complete set of precision, recall, F1 scores, and confusion matrices are included in the Supplementary Material.
The deep neural network (DNN) combining convolutional and recurrent layers was able to successfully identify athletic movements for both optical motion capture trajectories and simulated inertial measurement unit (sIMU) data. DNNs trained using optical motion capture data (OPT) and full body simulated IMU (sIMU13) data had similar performance with F1 scores of approximately 0.90. Classification accuracy was poor (<70%) if fewer than three body segments were included or the lower limbs were not included in the sIMU data.
There was minimal difference between micro-averaged F1 scores for the DNNs trained using five or 13 body segments. This indicates that it is not necessary to include measurements from the head or more than one segment from each upper or lower limb. This is encouraging as the use of fewer sensors would simplify the set-up before a movement screen and would be less of a hindrance to the athlete's motion. The F1 score for sIMU3L, which used the torso and shanks, was only ~0.04 less than for the full body DNN. Therefore, depending on the desired accuracy, classification rates may be sufficient using only three sensors.
Some movements, such as the bird-dogs, were more easily identified by the DNN, even for networks trained on sIMU data from one or two segments. This is likely because trunk motion for these movements is substantially different from that of the other movements, with the trunk horizontal and relatively stationary throughout the motion. Including arm segments prevented confusion between T-balances and lunges. The drop jump was classified particularly poorly when few segments were used, often being classified as the null condition. This may in part be due to the way the start and end of the L-hop motion was defined. The L-hop involved the athlete jumping horizontally forward, landing on one foot, then jumping laterally and landing on the opposite foot. This movement was defined to begin when the athlete had reached their maximum height during the initial jump. Without sufficient data, the DNN was unable to differentiate between the end of the drop jump movement, which involved a vertical jump, and the initial jump of the L-hop which was included in the null condition.
The F1 score we achieved in classifying athletic movements is similar to previously reported human activity classification results. The architecture of the DNN used in this study was based on the work of Ordóñez and Roggen (2016), who achieved an F1 score of 0.895 on a dataset including various modes of locomotion. Other work has reported classification accuracies ranging from 83 to 100% for everyday activities (Pärkkä et al., 2006; Yeoh et al., 2008; Attal et al., 2015; Yang et al., 2015) and 79–93% for movements involved in various sports (Schuldhaus et al., 2015; Groh et al., 2016; Anand et al., 2017; Cust et al., 2019).
Previous work on classification of everyday activities, such as walking, jogging, sitting, stair climbing, etc., has identified one sensor placed at the waist as producing the best classification accuracy (Cleland et al., 2013; Pannurat et al., 2017). In the current study, we found that a single simulated torso sensor resulted in a poor classification accuracy of 48%. This discrepancy can likely be attributed to the differences in activities included, as optimal sensor placement depends on the activity (Atallah et al., 2011; Attal et al., 2015). The activities classified in the previous studies involve activities that are repetitive and take place over a relatively long period of time. The athletic movements included in our study, however, are short single movements. Furthermore, the need to differentiate right and left versions of the movements makes classification with a single torso-mounted sensor more challenging.
The sIMU DNNs relied on simulated IMU data generated based on optical motion tracking markers. Therefore, these results likely represent a best-case scenario for classification of these athletic movements using wearable sensors. Sensor drift is a common issue with IMUs and therefore it is possible that misclassification rates would be larger using real sensor data, particularly for long data collections as drift increases over time. Care would also need to be taken to standardize sensor placement on each body segment. While we have used the Euclidean norm of the angular velocity and linear acceleration, error would be introduced into the angular orientation of the body segments by misaligned sensors. Additionally, it may be possible to mitigate sensor misalignment issues using a static or dynamic calibration at the beginning of the data collection. Despite the reliance on simulated sensor data, the results presented here highlight the potential for movement classification using wearable sensors and provide guidance for sensor placement in future work.
In this study, separate data trials were recorded for each motion and these were combined for the classification. As a result, the amount of null data frames included was relatively small. It may be necessary to included more null condition training data, including transitions between movements, for the DNN to be used successfully on continuously collected data.
Accurate classification of movements is critical for this DNN approach to be used with no manual intervention in combination with data-driven assessments of movement quality, as the quality could only be assessed on properly identified movements. Some errors may be possible to correct with additional processing, such as when the classification jumps to another movement for a few frames in the middle of an otherwise accurately classified movement. We observed that a large source of error was over- or under-estimating the start or end points of a movement with misclassification between the movement and the null condition. It is possible that movement quality could still be quantified with these slight errors in start and end points, but future work will be required to verify this. Alternately, a small amount of manual intervention could be used to verify task identification before proceeding to quantification of movement quality.
The favorable classification rates obtained in this work using simulated sensor data demonstrates the feasibility of classifying athletic tasks typical of movement screens using wearable sensors. Using simulated IMU data, we observed the best classification accuracy by including data from all body segments; however, we obtained good results using as few as three simulated sensors. This indicates that classification of these athletic movements using real IMU data would require at least three sensors and should include the torso and legs. Implementation of a movement classification DNN with wearable sensor data would facilitate automatic data-driven assessment of movement quality, eliminating subjective scoring, and increasing the ability to detect subtle differences.
Data Availability Statement
Code, sample data, and trained DNNs weights are available at doi: 10.5281/zenodo.3546204.
The studies involving human participants were reviewed and approved by University of Ottawa Research Ethics Board. Written informed consent to participate in this study was provided by the participant or their legal guardian/next of kin.
AC, GR, and RG conceived of the study, interpreted the results, and critically revised the manuscript. GR collected and processed the data. AC implemented the neural network, analyzed the results, and prepared the manuscript.
This work was funded by the Natural Sciences and Engineering Research Council (NSERC) of Canada (RGPIN-2014-05560) and the Ontario Early Researcher Award Program (ER17-13-007).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The data used in this study were collected by Motus Global.
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/article/10.3389/fbioe.2019.00473/full#supplementary-material.
Anand, A., Sharma, M., Srivastava, R., Kaligounder, L., and Prakash, D. (2017). “Wearable motion sensor based analysis of swing sports,” in Proceedings - 16th IEEE International Conference on Machine Learning and Applications, ICMLA 2017 (Cancun), 261–267. doi: 10.1109/ICMLA.2017.0-149
Atallah, L., Lo, B., King, R., and Yang, G. Z. (2011). “Sensor positioning for activity recognition using wearable accelerometers,” in IEEE Transactions on Biomedical Circuits and Systems (IEEE), 320–329. doi: 10.1109/TBCAS.2011.2160540
Attal, F., Mohammed, S., Dedabrishvili, M., Chamroukhi, F., Oukhellou, L., and Amirat, Y. (2015). Physical human activity recognition using wearable sensors. Sensors 15, 31314–31338. doi: 10.3390/s151229858
Bonazza, N. A., Smuin, D., Onks, C. A., Silvis, M. L., and Dhawan, A. (2017). Reliability, validity, and injury predictive value of the functional movement screen. Am. J. Sports Med. 45, 725–732. doi: 10.1177/0363546516641937
Bulling, A., Blanke, U., and Schiele, B. (2014). A tutorial on human activity recognition using body-worn inertial sensors. ACM Comput. Surv. 46, 1–33. doi: 10.1145/2499621
Cleland, I., Kikhia, B., Nugent, C., Boytsov, A., Hallberg, J., Synnes, K., et al. (2013). Optimal placement of accelerometers for the detection of everyday activities. Sensors 13, 9183–9200. doi: 10.3390/s130709183
Clifton, D. R., Harrison, B. C., Hertel, J., and Hart, J. M. (2013). Relationship between functional assessments and exercise-related changes during static balance. J. Strength Cond. Res. 27, 966–972. doi: 10.1519/JSC.0b013e318260b723
Cust, E. E., Sweeting, A. J., Ball, K., and Robertson, S. (2019). Machine and deep learning for sport-specific movement recognition: a systematic review of model development and performance. J. Sports Sci. 37, 568–600. doi: 10.1080/02640414.2018.1521769
Frost, D. M., Beach, T. A. C., Callaghan, J. P., and McGill, S. M. (2015). FMS scores change with performers' knowledge of the grading criteria - Are general whole-body movement screens capturing “dysfunction”? J. Strength Cond. Res. 29, 3037–3043. doi: 10.1097/JSC.0000000000000211
Gjoreski, H., Luštrek, M., and Gams, M. (2011). “Accelerometer placement for posture recognition and fall detection,” in Proceedings - 2011 7th International Conference on Intelligent Environments, IE 2011 (Nottingham), 47–54. doi: 10.1109/IE.2011.11
Groh, B. H., Fleckenstein, M., and Eskofier, B. M. (2016). “Wearable trick classification in freestyle snowboarding,” in BSN 2016 - 13th Annual Body Sensor Networks Conference (San Francisco, CA: IEEE), 89–93. doi: 10.1109/BSN.2016.7516238
Gulgin, H., and Hoogenboom, B. (2014). The functional movement screening (FMS)TM: an inter-rater reliability study between raters of varied experience. Int. J. Sports Phys. Ther. 9, 14–20.
Hammerla, N. Y., Halloran, S., and Plötz, T. (2016). “Deep, convolutional, and recurrent models for human activity recognition using wearables,” in IJCAI International Joint Conference on Artificial Intelligence (New York, NY), 1533–1540.
Kautz, T., Groh, B. H., Hannink, J., Jensen, U., Strubberg, H., and Eskofier, B. M. (2017). Activity recognition in beach volleyball using a deep convolutional neural network: leveraging the potential of deep learning in sports. Data Min. Knowl. Discov. 31, 1678–1705. doi: 10.1007/s10618-017-0495-0
Kern, N., Schiele, B., and Schmidt, A. (2003). “Multi-sensor activity context detection for wearable computing,” in Lecture Notes in Computer Science, eds E. Aarts, R. W. Collier, E. van Leonen, and B. de Ruyter (Berlin; Heidelberg: Springer), 220–232.
Kim, Y., and Toomajian, B. (2016). Hand gesture recognition using micro-Doppler signatures with convolutional neural network. IEEE Access 4, 7125–7130. doi: 10.1109/ACCESS.2016.2617282
Lee, S.-M., Yoon, S. M., and Cho, H. (2017). “Human activity recognition from accelerometer data using Convolutional Neural Network,” in 2017 IEEE International Conference on Big Data and Smart Computing (BigComp) (Jeju), 131–134. doi: 10.1109/BIGCOMP.2017.7881728
Minick, K. I., Kiesel, K. B., Burton, L., Taylor, A., Plisky, P., and Butler, R. J. (2010). Interrater reliability of the Functional Movement Screen. J. Strength Cond. Res. 24, 479–486. doi: 10.1519/JSC.0b013e3181c09c04
Nguyen, L. N. N., Rodríguez-Martín, D., Català, A., Pérez-López, C., Samà, A., and Cavallaro, A. (2015). “Basketball activity recognition using wearable inertial measurement units,” in Proceedings of the XVI International Conference on Human Computer Interaction (Vilanova i la Geltrú), 1–6. doi: 10.1145/2829875.2829930
Olguin, D. O., and Pentland, A. (2006). “Human activity recognition: accuracy across common locations for wearable sensors,” in IEEE 10th International Symposium on Wearable Computers (Montreaux, 5–7.
Ordóñez, F. J., and Roggen, D. (2016). Deep convolutional and LSTM recurrent neural networks for multimodal wearable activity recognition. Sensors 16:115. doi: 10.3390/s16010115
Pannurat, N., Thiemjarus, S., Nantajeewarawat, E., and Anantavrasilp, I. (2017). Analysis of optimal sensor positions for activity classification and application on a different data collection scenario. Sensors 17:774. doi: 10.3390/s17040774
Pärkkä, J., Ermes, M., Korpipää, P., Mäntyjärvi, J., Peltola, J., and Korhonen, I. (2006). “Activity classification using realistic data from wearable sensors,” in IEEE Transactions on Information Technology in Biomedicine, 119–128. doi: 10.1109/TITB.2005.856863
Paszke, A., Gross, S., Chintala, S., Chanan, G., Yang, E., DeVito, Z., et al. (2017). “Automatic differentiation in PyTorch,” in NIPS Autodiff Workshop (Long Beach, CA).
Ross, G. B., Dowling, B., Troje, N. F., Fischer, S. L., and Graham, R. B. (2018). Objectively differentiating movement patterns between elite and novice athletes. Med. Sci. Sports Exerc. 50, 1457–1464. doi: 10.1249/MSS.0000000000001571
Schuldhaus, D., Zwick, C., Körger, H., Dorschky, E., Kirk, R., and Eskofier, B. M. (2015). “Inertial sensor-based approach for shot/pass classification during a soccer match,” in 21st ACM KDD Workshop on Large-Scale Sports Analytics (Sydney, NSW), 1–4.
Smith, C. A., Chimera, N. J., Wright, N. J., and Warren, M. (2013). Intrarater reliability of the Functional Movement Screen. J. Strength Cond. Res. 27, 982–987. doi: 10.1519/JSC.0b013e31825c32a8
Teyhen, D. S., Shaffer, S. W., Lorenson, C. L., Halfpap, J. P., Donofry, D. F., Walker, M. J., et al. (2012). The functional movement screen: a reliability study. J. Orthop. Sports Phys. Ther. 42, 530–540. doi: 10.2519/jospt.2012.3838
Yang, J. B., Nguyen, M. N., San, P. P., Li, X. L., and Krishnaswamy, S. (2015). “Deep convolutional neural networks on multichannel time series for human activity recognition,” in International Joint Conference on Artificial Intelligence (Buenos Aires), 3995–4001.
Yeoh, W. S., Pek, I., Yong, Y. H., Chen, X., and Waluyo, A. B. (2008). Ambulatory monitoring of human posture and walking speed using wearable accelerometer sensors. Conf. IEEE Eng. Med. Biol. Soc. 2008, 5184–5187. doi: 10.1109/IEMBS.2008.4650382
Zeng, M., Nguyen, L. T., Yu, B., Mengshoel, O. J., Zhu, J., Wu, P., et al. (2014). “Convolutional neural networks for human activity recognition using mobile sensors,” in International Conference on Mobile Computing, Applications and Services, MobiCASE 2014 (Austin, TX), 197–205. doi: 10.4108/icst.mobicase.2014.257786
Keywords: human activity recognition, wearable sensors, machine learning, neural network, movement screens
Citation: Clouthier AL, Ross GB and Graham RB (2020) Sensor Data Required for Automatic Recognition of Athletic Tasks Using Deep Neural Networks. Front. Bioeng. Biotechnol. 7:473. doi: 10.3389/fbioe.2019.00473
Received: 11 October 2019; Accepted: 23 December 2019;
Published: 21 January 2020.
Edited by:Peter A. Federolf, University of Innsbruck, Austria
Reviewed by:Christian Maurer-Grubinger, Independent Researcher, Straßwalchen, Austria
Rezaul Begg, Victoria University, Australia, Australia
Copyright © 2020 Clouthier, Ross and Graham. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Ryan B. Graham, firstname.lastname@example.org