ORIGINAL RESEARCH article
Front. Robot. AI
Sec. Industrial Robotics and Automation
Volume 12 - 2025 | doi: 10.3389/frobt.2025.1612392
This article is part of the Research TopicInnovations in Industry 4.0: Advancing Mobility and Manipulation in RoboticsView all 6 articles
Adaptive Emergency Response and Dynamic Crowd Navigation for Mobile Robot using Deep Reinforcement Learning
Provisionally accepted- 1Birla Institute of Technology and Science, Dubai, United Arab Emirates
- 2Rochester Institute of Technology Dubai, Dubai, Dubai, United Arab Emirates
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Mobile robots have emerged as a reliable solution for dynamic navigation in real-world applications. Effective deployment in high-density crowds and emergency scenarios requires not only accurate path planning but also rapid adaptation to changing environments. However, autonomous navigation in such environments remains a significant challenge, particularly in time-sensitive applications such as emergency response. Existing path planning and reinforcement learning approaches often lack adaptability to uncertainties and time-varying obstacles, thereby making them less suitable for unstructured real-world scenarios. To address these limitations, a Deep Reinforcement Learning (DRL) framework for dynamic crowd navigation using three algorithms, Deep Deterministic Policy Gradient (DDPG), Twin Delayed Deep Deterministic Policy Gradient (TD3), and Deep Q-Network (DQN), is proposed. A context-aware state representation that combines Light Detection and Ranging (LiDAR)-based obstacle perception, goal orientation, and robot kinematics to enhance situational awareness is developed. The proposed framework is implemented in a ROS2 Gazebo simulation environment using the TurtleBot3 platform and tested in challenging scenarios to identify the most effective algorithm. Extensive simulation analysis demonstrates that TD3 outperforms the other approaches in terms of success rate, path efficiency, and collision avoidance. This study contributes a reproducible, constraint-aware DRL navigation architecture suitable for real-time, emergency-oriented mobile robot applications.
Keywords: deep reinforcement learning, mobile robot, Deep Q-network, Deep Deterministic Policy Gradient, Twin Delayed DeepDeterministic Policy Gradient, Crowd navigation
Received: 15 Apr 2025; Accepted: 01 Sep 2025.
Copyright: © 2025 Alexander, Suchir Vangaveeti, Venkatesan, Mounsef and Ramanujam. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Kalaichelvi Venkatesan, Birla Institute of Technology and Science, Dubai, United Arab Emirates
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.