ORIGINAL RESEARCH article
Front. Neurorobot.
Volume 19 - 2025 | doi: 10.3389/fnbot.2025.1631998
This article is part of the Research TopicAdvancing Neural Network-Based Intelligent Algorithms in Robotics: Challenges, Solutions, and Future Perspectives - Volume IIView all 5 articles
Dynamic Graph Neural Networks for UAV-Based Group Activity Recognition in Structured Team Sports
Provisionally accepted- 1Air University, Islamabad, Islamabad, Pakistan
- 2Guodian Nanjing Automation Co., LTD, Nanjing, Liaoning Province, China
- 3King Khalid University, Abha, Saudi Arabia
- 4Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia
- 5University of Bremen, Bremen, Bremen, Germany
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Introduction: Understanding group actions in real-world settings is essential for the advancement of applications in surveillance, robotics, and autonomous systems. Group activity recognition, particularly in sports scenarios, presents unique challenges due to dynamic interactions, occlusions, and varying viewpoints. To address these challenges, we develop a deep learning system that recognizes multi-person behaviors by integrating appearance-based features (HOG, LBP, SIFT), skeletal data (MediaPipe, MOCON), and motion features. Our approach employs a Dynamic Graph Neural Network (DGNN) and Bi-LSTM architecture, enabling robust recognition of group activities in diverse and dynamic environments. To further validate our framework's adaptability, we include evaluations on Volleyball and SoccerTrack UAV-recorded datasets, which offer unique perspectives and challenges.Method: Our framework integrates YOLOv11 for object detection and SORT for tracking to extract multi-modal features-including HOG, LBP, SIFT, skeletal data (MediaPipe), and motion context (MOCON). These features are optimized using genetic algorithms and fused within a Dynamic Graph Neural Network (DGNN), which models players as nodes in a spatio-temporal graph, effectively capturing both spatial formations and temporal dynamics.We evaluated our framework on three datasets: a volleyball dataset, SoccerTrack UAVbased soccer dataset, and NBA basketball dataset. Our system achieved 94.5% accuracy on the volleyball dataset (mAP: 94.2%, MPCA: 93.8%) with an inference time of 0.18 seconds per frame. On the SoccerTrack UAV dataset, accuracy was 91.8% (mAP: 91.5%, MPCA: 90.5%) with 0.20 seconds inference, and on the NBA basketball dataset, it was 91.1% (mAP: 90.8%, MPCA: 89.8%) with the same 0.20 seconds per frame. These results highlight our framework's high performance and efficient computational efficiency across various sports and perspectives.Discussion: Our approach demonstrates robust performance in recognizing multi-person actions across diverse conditions, highlighting its adaptability to both conventional and UAV-based video sources.
Keywords: unmanned aerial vehicles, neural network models, machine learning, Body pose, group action recognition, feature extraction, deep Learning Unmanned aerial vehicles, deep learning
Received: 20 May 2025; Accepted: 15 Jul 2025.
Copyright: © 2025 Zahra, Wu, Alshehri, Alqahtani, Aljuaid, Jalal and Liu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Ahmad Jalal, Air University, Islamabad, 44000, Islamabad, Pakistan
Hui Liu, University of Bremen, Bremen, 28359, Bremen, Germany
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.