Skip to main content

ORIGINAL RESEARCH article

Front. Robot. AI, 04 March 2024
Sec. Computational Intelligence in Robotics
Volume 11 - 2024 | https://doi.org/10.3389/frobt.2024.1362294

Collaborative robots (cobots) for disaster risk resilience: a framework for swarm of snake robots in delivering first aid in emergency situations

www.frontiersin.orgSyed Kumayl Raza Moosavi1 www.frontiersin.orgMuhammad Hamza Zafar2 www.frontiersin.orgFilippo Sanfilippo2,3*
  • 1School of Electrical Engineering and Computer Sciences, National University of Sciences and Technology, Islamabad, Pakistan
  • 2Department of Engineering Sciences, University of Agder, Grimstaad, Norway
  • 3Department of Software Engineering, Kaunas University of Technology, Kaunas, Lithuania

Cobots are robots that are built for human-robot collaboration (HRC) in a shared environment. In the aftermath of disasters, cobots can cooperate with humans to mitigate risks and increase the possibility of rescuing people in distress. This study examines the resilient and dynamic synergy between a swarm of snake robots, first responders and people to be rescued. The possibility of delivering first aid to potential victims dispersed around a disaster environment is implemented. In the HRC simulation framework presented in this study, the first responder initially deploys a UAV, swarm of snake robots and emergency items. The UAV provides the first responder with the site planimetry, which includes the layout of the area, as well as the precise locations of the individuals in need of rescue and the aiding goods to be delivered. Each individual snake robot in the swarm is then assigned a victim. Subsequently an optimal path is determined by each snake robot using the A* algorithm, to approach and reach its respective target while avoiding obstacles. By using their prehensile capabilities, each snake robot adeptly grasps the aiding object to be dispatched. The snake robots successively arrive at the delivering location near the victim, following their optimal paths, and proceed to release the items. To demonstrate the potential of the framework, several case studies are outlined concerning the execution of operations that combine locomotion, obstacle avoidance, grasping and deploying. The Coppelia-Sim Robotic Simulator is utilised for this framework. The analysis of the motion of the snake robots on the path show highly accurate movement with and without the emergency item. This study is a step towards a holistic semi-autonomous search and rescue operation.

1 Introduction

Societies encounter unforeseeable crisis situations such as earthquakes, fires, floods, hazardous spills, hurricanes/typhoons, tsunamis, terrorist attacks, refugee crises, and more. These crises can arise from natural causes or human activities and result in substantial loss of life, injuries, displacement of people, and damage to property. In retrospect, humans have learned to adapt and manage such calamities on a global scale. However, due to an increase in the impact of the ever increase disasters, exacerbated partly due to the climate change, management of such disasters have become complex in this socio-ecological landscape. According to the latest data from insurer Munich Re (Munich, 2023), the average losses due to natural catastrophes over 5 years (2017–2021), adjusted to inflation, was approximately $270bn while the statistics for 2022 alone was over $270bn. Moreover, the impact of disasters are not always recorded in detail when disasters do occur.

1.1 Disaster management

The need to manage such disaster situations is apparent. Therefore, it is increasingly important for disaster managers to assume an expanding role in safeguarding their communities through the formulation of effective management strategies. Emergency management processes are commonly categorised into distinct stages, although there is no universally agreed-upon model. On the aftermath of an event, disaster management typically involves four phases (Cova, 1999): to mitigate devastation of effective areas of potential disasters; have preparedness by incorporating trained personnel and shelter facilities; effective response during search and rescue (SAR) operations; recovery during the aftermath of the disaster. Conventional approaches, including field monitoring, physics-based models, expert surveys, and multi-criteria decision-making techniques, are utilised to identify hazards and risk factors. However, these methods often require extensive human effort and involve very high risk. One of the technological tools that can be used in these dangerous environments without adding risk to the life of humans, who are in the process of SAR operation, are robots. By integrating human skills with automation, a harmonious blend can be achieved, leveraging the adaptability of manual processes and the effectiveness and consistency of machines. This synthesis enables the realisation of benefits such as accessing hazardous or hard-to-reach areas, remote operation, adaptability, reducing human fatigue and risk. Particularly, in emergency scenario, there is a need for very close collaboration between first responders and robots. Therefore, collaborative robots, also known as cobots (Jeyson et al., 2022), represent a promising solution.

Many natural and man-made events have occurred in history that have prompted the use of robots for post disaster conditions. In the highly radioactive environment at Chernobyl after the 1986 nuclear blast, the KOMATSU, the TELEROB and many other significant ground vehicles were used for multiple tasks in place of humans (Tochilin et al., 2021; Sparkes, 2022). In 2007, the I-35 bridge collapsed into the fast moving currents, inaccessible to human divers, of the Minnesota river. Remote marine vehicles were then used for search, reconnaissance and mapping underwater (Murphy et al., 2011). In 2012, on the aftermath of the Finale Emilia earthquake in Italy, a combination of unmanned aerial vehicles (UAVs) and unmanned ground vehicles (UGVs) were used for the structural inspection of buildings (Kruijff et al., 2012). The fallout information was provided to the Italian National Fire Corps and state archeologists.

1.2 Virtual framework for disaster resilience

To enhance the definition of disaster recovery management strategy, researchers have proposed the integration of collaborative systems in a virtual space (Bertolino and Tanzi, 2019). An example of addressing this issue can be found in a recent study (Magid et al., 2019), where a novel framework and diverse control strategies were introduced for enhancing the collaborative performance of heterogeneous robotic swarms in the context of sensing, monitoring, and mapping flood and landslide disaster zones. The research presents a foundation of virtual simulators that demonstrate various robot interaction protocols and system modelling concepts within the Gazebo environment of the Robot Operating System (ROS). The use of digital twins to enhance human-robot collaboration (HRC) in complex production systems was explored by Malik and Brem (2021), presenting a case study, highlighting the potential advantages, and building blocks of digital twins in the field of collaborative robotics. In Burke et al. (2004), a field study on human-robot interaction during an urban SAR training exercise in Miami was presented, focusing on the challenges and dynamics that arise in this context. In Wagner (2021), an analysis on emergency evacuation as a valuable paradigm for studying human-robot interaction was provided, emphasising the need for ethical considerations. Different design approaches are discussed together with ethical implications, outlines, and a roadmap for the development and evaluation of emergency evacuation robots. In Sanfilippo (2022), the author proposed a simulation case study for SAR operation combining modular robot, grasping, and locomotion capabilities. Thus, computer based virtual models of physical systems can be used to test and validate complex strategies and scenarios prior to their implementation in real world applications.

Building upon virtual models for disaster scenarios, a case study for a SAR simulation environment with HRC using snake robotics is presented in this study. The main contribution of this work includes the development of the control framework for HRC between an unmanned aerial vehicle (UAV), multiple snake robots, a first responder, and multiple victims of a disaster. The proposed architecture distributes the control scheme of the HRC into four phases simulated in a virtual environment: deployment (HRC), path planning and sharing (robot to robot collaboration), grasping and locomotion, releasing first aid item and return. Deployment involves initialising the HRC system by the first responder by positioning the team of snake robots and the UAV; Path Planning and Sharing focuses on optimising collaborative robot paths to prevent collisions; Grasping and Locomotion addresses precise object manipulation and robot movement; Releasing First Aid Item and Return entails placing a first aid item and safely returning to a designated location. The CoppeliaSim - formerly known as Vrep (Rohmer et al., 2013) is adopted as the virtual simulator. Regarding visual perception, a camera is placed at the bottom of the UAV, which is deployed on top of a maze/disaster scenario, for image keypoints extraction, while another camera is attached on the head of the snakes for path following.The respective paths are obtained at runtime by each of the first responders’ snake robot. Multiple models of snake robots (Liljebäck et al., 2014) are used for locomotion on a dynamic path. To prove the efficacy of the design, a case study featuring a maze is presented, which combines human and multi-robot collaboration, locomotion, object grasping and dropping, path following and obstacle avoidance.

1.3 Collaborative capabilities of humans and unmanned vehicles in shared spaces

In today’s dynamic and fast-paced world, the integration of unmanned vehicles into shared spaces is becoming increasingly common. Collaborative unmanned systems have emerged to meet our society’s wide-ranging grand challenges, with their advantages including high performance, efficiency, flexibility, and inherent resilience (Zhang et al., 2022). The essence of collaborative capabilities lies in the synergy between humans and unmanned vehicles. It entails a multifaceted approach aimed at harmonising the efforts of both entities to ensure their coexistence in shared spaces. At its core, this collaborative framework is driven by a profound need for seamless and secure interaction, which is underpinned by a range of technological advancements and innovative strategies.

One of the primary challenges that collaborative capabilities address is the navigation of completely unknown dynamic and unstructured environments. These spaces, whether they be urban jungles with constantly changing traffic patterns or uncharted wilderness with rugged terrains, pose significant obstacles. Humans rely on their cognitive abilities and prior experience to navigate these landscapes, but unmanned vehicles depend on a different set of skills. They harness the power of multi-sensor data fusion (Khaleghi et al., 2013), integrating information from various sources such as cameras, LiDAR, radar, and GPS. This fusion not only enhances their perception of the environment but also empowers them to make informed decisions. Through this synergy, unmanned vehicles can navigate through unknown terrains with greater precision, safety, and adaptability.

In the context of disaster environments, enhanced planning and sequential decision-making become even more critical components of collaborative capabilities (Unhelkar et al., 2020). These environments are characterised by extreme unpredictability, chaos, and rapidly changing conditions, making efficient interaction between humans and unmanned vehicles a matter of life and death. In such high-stress scenarios, a proactive approach to planning and decision-making is essential.

Multi-agent coordination (Queralta et al., 2020) is another pivotal component of collaborative capabilities. Shared spaces often involve a multitude of agents, both human and unmanned, operating in tandem. For instance, in a disaster settings, unmanned delivery vehicles may need to coordinate with one another to optimise routes and minimise debris collision, while also interacting seamlessly with victims and other vehicles. Collaborative algorithms, backed by real-time communication and negotiation mechanisms, ensure that these interactions are conducted smoothly and safely. This ability to coordinate with other agents makes shared spaces more efficient and less congested, benefiting both human and unmanned vehicle operators.

Collaborative capabilities extend beyond the realm of autonomous vehicles and encompass human-vehicle interaction. In shared spaces, it is essential that unmanned vehicles communicate with humans in a user-friendly and intuitive manner. Trust and understanding are critical factors in ensuring the success of various applications. Collaborative interfaces may include informative displays, natural language communication, and even gestures to establish a clear and mutual understanding between humans and unmanned vehicles. This level of communication helps build trust and confidence in the technology, ultimately leading to more widespread acceptance.

This paper is organised as follows. A review of the related research work is given successively in Section 2. Then, the model of a snake robot is provided in Section 3. Subsequently, the proposed framework architecture is outlined in Section 4. Simulation results are presented in Section 5. A Discussion is provided on swarm robots and HRC in Section 6. Finally, conclusion and future work are discussed in Section 7.

2 Related work

Snake robots have gained significant attention in the field of robotics due to their unique locomotion capabilities and potential applications in disaster scenarios. These robots, inspired by the flexibility and maneuverability of biological snakes, offer a great potential for navigating complex and challenging environments. In rough terrains, steered vehicles often provide non feasible solutions between waypoints due to the kinematic and dynamic limitations, specially on curvatures, even when vehicles can rotate and turn in place (Eskandarian et al., 2019). Snake robots exhibit a range of physical configurations and purposes, although their movement is often inspired by snakes. These robots can differ in terms of redundancy, wheel usage, and even their ability to operate in both land and water environments. Their slender, elongated bodies with thin cross-sections make them particularly well-suited for exploring narrow spaces or pipes. The distribution of mass and the presence of multiple ground contact points contribute to their stability, especially when compared to other robotic designs like wheeled or multipedal systems (Hopkins et al., 2009). Snake robots promise impressive adaptability to various terrains, primarily relying on the roughness of the ground or obstacles to gain sufficient traction and move forward without slipping (Webster et al., 2006). This adaptability and stability in different terrains make them robust to mechanical failure, enabling exploration in uncertain and challenging environments. In terms of gait patterns, snake movement can be categorised into four categories (Seeja et al., 2022): (a) lateral undulation; (b) concertina; (c) rectilinear progression; and (d) side winding.

There are two distinct approaches to snake robot locomotion based on the understanding of the environment: obstacle avoidance locomotion and obstacle accommodation/exploitation locomotion. In cluttered environments, snake robots exploit obstacles as an aid for propulsion purposes. This is known as “obstacle-aid Locomotion” (OAL) (Holden et al., 2014). Snakes utilise a strategy of pushing against unevenness or irregularities in the environment, creating bends in their body. This bending pattern is propagated from the head to the tail, allowing for smoother locomotion. However, this method is heavily reliant on the friction present in the environment, and collisions with obstacles can hinder further motion, potentially causing mechanical stress or damage to the equipment. The significance of environment perception, mapping, and representation cannot be overstated. In fact, our research group has introduced the term “perception-driven obstacle-aided locomotion” (POAL) to underscore this concept. POAL refers to a locomotion approach where a snake robot leverages its sensory-perceptual system to utilise the surrounding operational space. It identifies walls, obstacles, or external objects as means of propulsion. Our group’s work on POAL has been documented in several publications (Sanfilippo et al. (2016a; b, 2017a)). To facilitate the design and simulation of POAL, our research group has developed SnakeSIM, a virtual rapid-prototyping framework. SnakeSIM enables researchers to engage in the safer, faster, and more efficient design and simulation of POAL (Sanfilippo et al. (2017b; 2018)). In terms of control, attaining POAL necessitates precisely identifying possible push-points and properly determining feasible contact response forces. Because of the lack of compliance, achieving this with typical rigidly-actuated robots is exceedingly difficult. To address this challenge, our research group has developed Serpens, a novel modular snake robot equipped with series elastic actuators (SEA). Serpens is notable for its low cost, open-source nature, and high compliance, making it suitable for various applications. We recently introduced Serpens in our research publications (Sanfilippo et al., 2019; Duivon et al., 2022). Regarding guidance, a biologically inspired steering controller was presented in (Rañó et al., 2018). With respect to navigation, a local path planning algorithm for snake robots was introduced in (Hanssen et al., 2020).

Various investigations in literature have concentrated on motion dealing with obstacle avoidance. In this context, the environment perception, mapping, and representation play a crucial role in the overall model. These elements are fundamental for the successful functioning and decision-making of the snake robot in its environment. An example of snake robot motion is the artificial potential field (APF) based locomotion, as described by Davy (2002). This approach involves creating an artificial field around objects, and the robot’s motion is designed to avoid this force field. Another algorithm utilised in snake robot locomotion is the central pattern generator (CPG), mentioned by Nor and Ma (2014). CPG enables the robot to navigate around obstacles or barriers by adjusting the turning of its body from its intended trajectory.

In the field of snake robotics, most of the past literature focuses on specific, static scenarios rather than the possibility of exploring dynamically changing and unpredictable scenarios. This represents a significant research gap. While numerous studies have investigated the locomotion and control mechanisms of snake robots in controlled environments, there is a lack of comprehensive research considering real-world situations where the environment and task requirements dynamically evolve. Such scenarios, which involve navigating through complex and unpredictable terrains, pose unique challenges that need to be addressed to enhance the adaptability and robustness of snake robots. Furthermore, the potential for collaboration between snake robots and other robots or humans is almost untapped. Investigating how snake robots can efficiently work together and engage collaboratively with other entities opens up new opportunities for applications in fields such as SAR, exploration, and HRC. Bridging these research gaps will help to the advancement of snake robot capabilities and their practical deployment in real-world circumstances.

3 Modelling

To derive a kinematic model for the locomotion of the snake robot on a horizontal and flat surface, Liljebäck et al. (2013) proposed a linearisation of the model due to the many degrees of freedom and the dynamical couplings between links of the robot. A snake robot consists of N rigid links of length 2l joined by N − 1 joints. Each link is assumed to have the same mass m, thereby the center of mass of each rigid link is at the center point and the total mass of the snake comes out to be N × m. The mathematical model is described in terms of the kinematic parameters of the snake robot illustrated in Figure 1B.

FIGURE 1
www.frontiersin.org

FIGURE 1. (A) Force profile of individual link in a snake robot (B) Kinematic and Force/Torque parameters of a snake robot.

The snake robot moves on a horizontal surface with N + 2 degrees of freedom. The heading (orientation) of the snake robot is denoted by θ̄ and is defined as the average of link angles as described in Eq. 1.

θ̄=1N×i=1Nθi(1)

where link angles (θi) are defined as the angle that the link forms with the global x-axis. On the other hand, joint angles, denoted as ϕi, is different such that it defines the difference between the link angles between each link i.e., θiθi+1. The global position of the snake robot is given in eq. 2 as:

p=pxpy=1Nmi=1Nmxi1Nmi=1Nmyi(2)

Following this, the forward velocity of the snake robot movement is defined as a component of the center of mass velocity (ṗ) and the current heading as shown in eq. 3.

v̄t=ṗxcosθ̄+ṗysinθ̄(3)

Figure 1A shows the joint forces and the friction forces acted upon link i. Using the first principle of motion a dynamic model can be described for the whole snake robot in matrix form as shown in eq.4.

mẌ=fR,x+DThxmŸ=fR,y+DThy(4)

where Ẍ=[xï,xi+1̈xN̈], Ÿ=[yï,yi+1̈yN̈], fR,x, fR,y are the ground friction forces. hx = [hx,1hN,1] and hy = [hy,1hN,1] are defined as the matrix for the joint constraint forces hx,i and hy,i respectively. The torque balance equation for the link i is given in eq. 5.

Jθ̈i=uiui1lsinθihx,i+hx,i1+lcosθihy,i+hy,i1,(5)

where ui is defined as the torque forces exerted on the link from the next link in the chain of links of the snake robot. By using matrix form and introducing state variables the dynamic model of the snake robot can be compactly described in a state space form as shown in eq. 6.

ẋ=θ̇ṗθ̈p̈=Fx,u(6)

where elements of the F(x, u) can be found in the mathematical breakdown provided by Liljebäck et al. (2013).

4 Framework architecture

As introduced by Sanfilippo (2022), the selected control framework is organised hierarchically, as shown in Figure 2. The input layer enables the robot to be guided by a human operator to achieve teleoperation or by other external systems (for example, an external planner) to reach higher levels of autonomy.

FIGURE 2
www.frontiersin.org

FIGURE 2. The proposed control framework architecture.

The core layer is the only layer required to perform the standard functions and capabilities of guidance, navigation, and control (GNC). These characteristics of the unmanned snake robot operate independently of one another and in parallel:

• Guidance: this concerns the process of identifying the desired course or trajectory for the snake robot to follow. It includes the decision-making process that specifies the robot’s objectives and constraints. To generate commands that direct the robot along a desired path, the guidance system considers the robot’s environment, mission requirements, and any other relevant parameters;

• Navigation: it involves the decision-making process regarding the optimal movement of the snake robot, including determining the appropriate location, timing, and method of locomotion. This decision-making process takes into account both external system commands and the sensory data gathered by the snake robot. The desired outcome of the navigation process is to generate a trajectory that includes path and velocity information for the robot to follow;

• Control: it serves as the central component of the presented control framework, offering researchers the flexibility to develop alternative control methods. The inputs to the control module include the desired trajectory and pertinent information obtained from the guidance level, such as perception data. The objective is to determine the necessary setpoints for the robot’s actuators, enabling it to accurately track the desired trajectory.

In this work, a planner is added to the previously presented framework architecture (Sanfilippo, 2022). The planner considers four phases of HRC: 1) deployment (HRC): a first responder deploys the snake robots, a UAV, and first aid items; 2) path planning and sharing (robot to robot collaboration): the UAV captures the planimetry of the area of interest and extracts keypoints of the maze to share with the snake robots. The first responder assigns victims to each snake robot which then form their paths (e.g., shortest path) to their respective targets.

3) grasping and locomotion: the snake robot grasps the first aid item and locomotes to reach the victim to be rescued; 4) releasing first aid item and return: the snake robot delivers and drops the first aid item to the victim to be rescued. Successively, the snake robot returns to the first responder.

The task layer encompasses a variety of tasks that have clearly defined objectives to accomplish. The following tasks have been implemented:

• Line follow: in this task, the robot employs its visual sensor to track a designated line (Kelasidi et al., 2017). Through the utilisation of a proportional integral derivative (PID) controller, the snake robot computes the required adjustments to its locomotion parameters, ensuring that the line remains within the camera’s field of view.

• Line search: in case the vision sensor loses the path following line, for example, during sharp turns, this task performs the operation of rotating and exploring to re-adjust its position to get back on track of the path. The task makes the robot head rotate left and right while making small steps forwards or backwards to find the line.

• Object search: once the Emergency item is detected by the snake robot vision sensor, the line following control scheme is replaced by the object tracking and grasping scheme.

• Locate drop zone: once the drop zone near the victim’s position is detected by the snake robot vision sensor, the line following control scheme is replaced by the object drop sequence.

• Track object: upon detection of an object with a specific color, this particular task employs a PID controller to compute the precise adjustments required for the snake robot’s parameters. These adjustments are aimed at maintaining the object within the camera’s field of view, thus ensuring continuous tracking.

• Pregrasping: in this task, it is assumed that the object to be grasped is positioned in front of the snake robot and is visible to the camera prior to initiating the execution. The snake robot begins by executing a continuous bending motion until the object to be grasped is no longer within the camera’s field of view. Subsequently, a series of forward steps are performed by the snake robot, followed by a continuous bending motion of the head, aiming to detect a collision with the object to be grasped. This sequence is repeated until the object is determined to be in an optimal position for grasping. In the event that the object to be detected does not experience collision, the snake robot moves backwards until the object to grasp is back into the field of view of the vision sensor.

• Grasping: after the pregrasp task is completed, a series of bending maneuvers is then executed to transition the snake robot into the whole-body grasping posture. The snake robot adapts its shape to the object to grasp by determining minimum number of modules (nmin) needed to accomplish grasping according to Eq. 7.

nmin=Cobjlm(7)

where Cobj minimum circular length of object and lm is the length of the module. To ensure the object remains correctly positioned throughout this sequence, torque sensing is implemented at the joint level. Once the bending procedure concludes, collision detection is employed to verify that the object is correctly positioned within the body of the snake. With the object grasped, the snake robot rotates itself until the path line is in front of the vision sensor, thereby, the line following locomotion begins.

• Drop zone tracking: this task ensures the drop zone is in front of the snake and within the camera’s field of view via PID controller for precise adjustments.

• Dropping: the snake robot advances towards the drop zone, continuing until it is no longer visible within the camera’s field of view. At this point, the snake robot will proceed to move forward for a predetermined duration, based on empirical observations. Following this, the snake robot will come to a stop and release the object by straightening its body, with the center section of the snake pushing the object onto the drop zone.

• Return to line: after the snake robot drops an item, it moves backwards for a predefined number of steps and then performs a sinusoidal rotatory movement to get back on track of the path line.

At the aftermath of a disaster event, first responders arrive at the scene to provide assistance. Due to obstructions from the fallout of the disastrous event, the victims are unreachable by the first responders. In the first phase of the HRC, the rescue responders deploy the snake robots and emergency items (e.g., a water tank or an oxygen tank) in close proximity to themselves, while a UAV is launched on top of the area of interest, which is cluttered with obstacles. Since the environment is dynamic and unknown to the snake robots, assistance from the rescue responder is required. The rescue responder provide the telemetry of the map to their deployed snake robots by the use of a UAV that is equipped with a vision sensor to create a map of the environment. The UAV is responsible for mapping out positions of key elements in the area of interest which include obstacles, positions of multiple victims, first aid items and positions of the snake robots. The control framework for the mapping task procedure of the second phase is itemised as follows:

• Obstacles Map: this task uses the static image taken from the UAV to mark positions of the obstacles in the maze. It also marks the boundary regions creating a bound space of the map for the snake robot traversal.

• Snake robots’ position: this task compares the pre-defined template of a snake robot with every area in the map to determine the position and orientation of the robot. The line path points to be made start from head of the snake robot.

• Drop zone positions: this task determines the position of the drop zone on the maze nearest to the victim. The line following path will end near this position.

• Path planner: once the requisite positions are determined. The snake robots creates path points with small increments on the simulation environment. The path starts from each snake’s position and ends near their drop zone. The control scheme involves A* algorithm (Wang et al., 2015) with the obstacles as the heuristics of the algorithm for optimal path planning.

• Path creation: this task takes the path points and converts them according to the simulation world environment. These points are then embedded on the map maze. A line is then joined between each point to create the path on the maze which the snake robot will follow.

With this, the system enters the third phase of the HRC process. The snake robot grabs the item from the first responder and positions itself to get on track of the path line. It is important to note that in this work, a swarm of snake robots is considered, operating collaboratively to fulfill the mission objectives. Each snake robot within the swarm contributes to the overall retrieval and delivery process. The robot follows the path line to the position of the victim. When the robot’s vision sensor detects the drop zone near the victim, the fourth phase of the HRC system initiates. In this phase, the robot performs the tasks of dropping the object on the drop zone. After task completion, the snake robots re-adjusts themselves back onto the line and follow the path line back to the first responder, consequently, ending the simulation when the final snake robot has returned. The state machine flowchart of the four-state HRC phases is shown in Figure 3.

FIGURE 3
www.frontiersin.org

FIGURE 3. Simulation flowchart.

5 Simulation results

Due to the complexity of the environment, real-world development of control algorithms for snake robots can be challenging. Testing novel control approaches can potentially damage both the environment and the snake robot and can be time consuming. A realistic simulator is much more efficient for the development of control strategy. Coppelia Sim (Rohmer et al., 2013) is chosen as the simulation environment in this study because it is a flexible simulation framework that supports multiple operating systems. Each module can be controlled via embedded script, plugins, a remote application programming interface (API) client or a user-defined solution. Lua lightweight, multi-paradigm programming language (Ierusalimschy et al., 2012) - created in 1993 - is used within the Coppelia Sim simulator.

To demonstrate the potential of the proposed framework for Search and Rescue (SAR) operations, a case study presented in this work offers simulation results. The purpose of the operation is for the multi-snake robot swarm to retrieve a distinctly colored object (marked as a green cube) from a first responder and bring it to the victims in need (at the drop zone). It is crucial to highlight that within this simulation, there are multiple victims requiring assistance, and a number of first responders who interact with the swarm of snake robots.

The simulated autonomous planner for this operation is divided into two main parts, each contributing to the successful accomplishment of this complex objective:

• A remote API - via Jupyter Notebook (Kluyver et al., 2016) - initialises the simulation, performs all the image recognition tasks and determines the path points, using the A* algorithm, from the snake robots to the drop zone positions. These path points are then converted to the simulation cartesian coordinate system and sent to the Coppelia Sim simulator. A child script written embedded in each snake robot model in the simulator receives these points and draws their path onto the maze.

• The snake robots starts their locomotion simultaneously (rectilinear progression) by first performing the pregrasping/grasping operation to get the emergency item and place the robot where the path line is in front of the snake head vision sensor. By avoiding obstacles, the snake robots traverse toward the drop zone following the path line. The snake robot then drops the object onto the drop zone, readjusts itself back to the path line and returns to the original position near the first responder, consequently ending the simulation.

A sequence of successive screenshots for the selected four phase HRC system case is shown in Figure 4.

FIGURE 4
www.frontiersin.org

FIGURE 4. A sequence of successive screenshots for the selected four phase HRC system case study. The screenshots contain the simulated environment, raw and processed video streams of the snake vision sensor and robot joint’s (7 in total) force / torque measurements. (A) Disaster Scenario (B) HRC Phase 1 (C) HRC Phase 2 (D−F) HRC Phase 3 (G−I) HRC Phase 4.

6 Discussion

A disaster scenario, man-made or natural, consists of many unknown variables and dangers. Human intervention becomes difficult if the risk of falling debris is a factor. In such situations a swarm of robots being deployed would be considered as an ideal scenario for SAR operations. Bio-inspired snake robots can traverse a variety of challenging terrains, such as narrow paths, uneven surfaces, as well as gravel, and debris, among others, where robots with different mobility systems may encounter significant difficulties. The snake like structure of the robot aids in locomotion into tight spaces. The collaborative synergy between snake robots and human responders amplifies the impact of SAR efforts, enabling the possibility to reach locations that otherwise might be unapproachable. Furthermore, the modular nature of snake robots allows for customisation and adaptability. Different modules can be attached or detached enhancing their versatility. This modularity aligns with collaborative efforts, as snake robots can be equipped with various sensors, cameras, or tools to aid in data collection, assessment, and interaction with the environment. The path to the victim followed by the snakes, grabbing the emergency item in the process and their subsequent return back following the defined path line, is showcased in Figure 5. The figure illustrates that the slender design of the snake aids in maintaining adherence to the path even with the emergency item in grasp. Figure 6 displays the torque profiles of the snake’s joints during different phases of the HRC process. A maximum torque value of 15 kgm2/s2 is needed during the grasping process while for ungrasping the joint force of 25 kgm2/s2 is needed. Joints 3 and 4 require the highest torque during ungrasping because the snake robot has to unclench the item and push it forward towards the victim drop zone. The head of the snake is responsible for turning and rotating to grab the item, therefore, the tail, which is Joint 7, shows the least amount of torque required during the grasping phase.

FIGURE 5
www.frontiersin.org

FIGURE 5. Navigational trajectory of autonomous snake robots and adaptive grasping/ungrasping movements executed upon locating emergency item/victims.

FIGURE 6
www.frontiersin.org

FIGURE 6. Torque profiles of seven joint actuators in a snake robot (Joint 7 is at the tail of the snake), showing the variations in mechanical stress during grasping, path-following, and ungrasping phases over time.

Snake robots, while innovative and adaptable, come with inherent limitations. Adaptability to terrains with different friction properties is still an open challenge (Chitikena et al., 2024). Energy efficiency poses a concern due to the energy-intensive movement. Additionally, snake robots typically have limited payload capacity. Their slender, elongated design prioritises agility and flexibility over the ability to carry heavy loads. This limitation affects their utility in applications requiring the transport of significant equipment or materials, restricting their roles to surveillance, data collection, and light manipulation tasks. Furthermore, operating these robots requires specialised knowledge and skills, as their movement patterns are complex and vastly different from more conventional robotic systems. Moreover, communication reliability, slower speeds, and ethical considerations (Chitikena et al., 2023) must also be addressed for effective collaboration. Balancing these drawbacks with their benefits necessitates ongoing research and tailored deployment strategies.

The use of multi-snake robots also offers a solution to a common limitation in disaster response: limited human resources. In disaster scenarios, the availability of trained human responders can be scarce or overwhelmed. Snake robots can operate autonomously and require minimal human intervention, thereby alleviating the pressure on the limited pool of responders and allowing them to focus on tasks that require human expertise and decision-making. However, there are several challenges to be addressed in the implementation of a swarm of snake robots in disaster environments. Firstly, ensuring effective communication and coordination among the robots is crucial. The development of robust communication systems that can withstand interference and transmit critical data in real-time is a significant technological challenge. The integration of advanced sensors and perception capabilities is another area of focus. These snake robots must be equipped with state-of-the-art sensors to provide comprehensive situational awareness.

7 Conclusion and future work

The adoption of collaborative robots (cobots) by disaster managers heavily relies on their capabilities, reliability, and robustness during field deployments. The extent of autonomy exhibited by robotic systems not only influences the manpower needed for their operation but also determines the complexity and adaptability of the system. However, achieving full autonomy in real-world rescue scenarios is currently challenging and not readily applicable in practical situations. Nevertheless, there is a clear inclination towards incorporating semi-autonomous behaviours instead of relying solely on manual control. This approach aims to alleviate the cognitive burden on the operator, enabling them to multitask or operate multiple systems concurrently. However, it is crucial to involve humans in the decision-making loop to guide the robot’s actions, particularly in tasks that involve dynamic changes throughout the mission, such as search and rescue operations during disaster scenarios. This paper proposes a control framework for human-robot collaboration (HRC) in an environment for disaster scenarios. A simulation using the Coppelia Sim (Rohmer et al., 2013) interface is used, in which a swarm of robots is emulated using an unmanned aerial vehicle (UAV) and multiple snake robots. The swarm of snake robots is deployed to navigate through a dynamic disaster scenario cluttered with obstacles, with the objective of retrieving an emergency item from the first responder, such as an oxygen tank or a water tank, and delivering it to various victims scattered throughout the provided map. The UAV provides data of the site planimetry to the snake robots. Using the A* algorithm the snake robot follows the optimal path towards the victim on the maze. The control framework considers four phases of HRC: deployment (HRC), path planning and sharing (robot to robot collaboration), grasping and locomotion, releasing first aid item and return. Simulation results show the efficacy of devising such a system in a physical environment for the future. Torque plots and navigational trajectory of the autonomous snake robots demonstrate the practicality of implementing such design frameworks in real world scenarios. In future, multiple robots can be assigned distinct roles, such as search and rescue, hazard assessment, structural assessment, or overlapped tasks for a common goal and they can execute these tasks concurrently. Physical obstacles may also be replaced with synthetic fire (in mixed reality) for obstacle avoidance.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

SM: Conceptualization, Methodology, Software, Writing–original draft, Writing–review and editing. MZ: Data curation, Formal Analysis, Investigation, Validation, Writing–review and editing. FS: Funding acquisition, Project administration, Resources, Supervision, Visualization, Writing–review and editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work is supported by the Biomechatronics and Collaborative Robotics research group at the Top Research Center Mechatronics (TRCM), University of Agder (UiA), Norway.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Bertolino, M., and Tanzi, T. J. (2019). “Advanced robot 3d simulation interface for disaster management,” in 2019 Kleinheubach Conference, 1–4.

Google Scholar

Burke, J. L., Murphy, R. R., Coovert, M. D., and Riddle, D. L. (2004). Moonlight in miami: Field study of human-robot interaction in the context of an urban search and rescue disaster response training exercise. Human–Computer Interact. 19, 85–116. doi:10.1207/s15327051hci1901and2_5

CrossRef Full Text | Google Scholar

Chitikena, H., Gravdahl, I., Pettersen, K. Y., Mohammadi, A., Sanfilippo, F., Stavdahl, y., et al. (2024). “Adaptive manoeuvring control for planar snake robots in uncertain friction environments,” in Accepted for publication to the Proc. of the American Control Conference (ACC) (Toronto, Ontario, Canada: American Automatic Control Council (AACC)).

Google Scholar

Chitikena, H., Sanfilippo, F., and Ma, S. (2023). Robotics in search and rescue (sar) operations: an ethical and design perspective framework for response phase. Appl. Sci. 13, 1800. doi:10.3390/app13031800

CrossRef Full Text | Google Scholar

Cova, T. J. (1999). Gis in emergency management. Geogr. Inf. Syst. 2, 845–858.

Google Scholar

Davy, A. J. (2002). Establishment and manipulation of plant populations and communities in terrestrial systems, 1. Cambridge University Press, 223–241.

CrossRef Full Text | Google Scholar

Duivon, A., Kirsch, P., Mauboussin, B., Mougard, G., Woszczyk, J., and Sanfilippo, F. (2022). The redesigned serpens, a low-cost, highly compliant snake robot. Robotics 11, 42. doi:10.3390/robotics11020042

CrossRef Full Text | Google Scholar

Eskandarian, A., Wu, C., and Sun, C. (2019). Research advances and challenges of autonomous and connected ground vehicles. IEEE Trans. Intelligent Transp. Syst. 22, 683–711. doi:10.1109/tits.2019.2958352

CrossRef Full Text | Google Scholar

Hanssen, K. G., Transeth, A. A., Sanfilippo, F., Liljebäck, P., Stavdahl, Ø., et al. (2020). Path planning for perception-driven obstacle-aided snake robot locomotion,” in 2020 IEEE 16th International Workshop on Advanced Motion Control (AMC), Kristiansand, Norway, 98–104. doi:10.1109/AMC44022.2020.9244366

CrossRef Full Text | Google Scholar

Holden, C., Stavdahl, Ø., and Gravdahl, J. T. (2014). “Optimal dynamic force mapping for obstacle-aided locomotion in 2d snake robots,” in 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, 321–328.

CrossRef Full Text | Google Scholar

Hopkins, J. K., Spranklin, B. W., and Gupta, S. K. (2009). A survey of snake-inspired robot designs. Bioinspiration Biomimetics 4, 021001. doi:10.1088/1748-3182/4/2/021001

PubMed Abstract | CrossRef Full Text | Google Scholar

Ierusalimschy, R., Celes, W., and de Figueiredo, L. H. (2012). Lua: the programming language.

Google Scholar

Jeyson Carpio, A., Samuel Luque, C., Juan Chambi, C., Jesus Talavera, S., et al. (2022). Analysis of the intuitive teleoperated system of the TxRob multimodal robot. Int. J. Adv. Comput. Sci. Appl. 13 (10). doi:10.14569/ijacsa.2022.0131093

CrossRef Full Text | Google Scholar

Kelasidi, E., Liljebäck, P., Pettersen, K. Y., and Gravdahl, J. T. (2017). Integral line-of-sight guidance for path following control of underwater snake robots: theory and experiments. IEEE Trans. Robotics 33, 610–628. doi:10.1109/tro.2017.2651119

CrossRef Full Text | Google Scholar

Khaleghi, B., Khamis, A., Karray, F. O., and Razavi, S. N. (2013). Multisensor data fusion: a review of the state-of-the-art. Inf. Fusion 14, 28–44. doi:10.1016/j.inffus.2011.08.001

CrossRef Full Text | Google Scholar

Kluyver, T., Ragan-Kelley, B., Pérez, F., Granger, B., Bussonnier, M., Frederic, J., et al. (2016). Jupyter Notebooks-a publishing format for reproducible computational workflows., vol. 2016

Google Scholar

Kruijff, G.-J. M., Pirri, F., Gianni, M., Papadakis, P., Pizzoli, M., Sinha, A., et al. (2012). “Rescue robots at earthquake-hit mirandola, Italy: a field report,” in 2012 IEEE international symposium on safety, security, and rescue robotics (SSRR) (IEEE), 1–8.

CrossRef Full Text | Google Scholar

Liljebäck, P., Pettersen, K. Y., Stavdahl, Ø., and Gravdahl, J. T. (2013). Snake robots: modelling, mechatronics, and control. Springer.

Google Scholar

Liljebäck, P., Stavdahl, Ø., Pettersen, K. Y., and Gravdahl, J. T. (2014). Mamba - a waterproof snake robot with tactile sensing,” in 2014 IEEE/RSJ Int. Conf. Intelligent Robots Systems, Chicago, IL, USA, 294–301. doi:10.1109/IROS.2014.6942575

CrossRef Full Text | Google Scholar

Magid, E., Pashkin, A., and Simakov, N. o. (2019). “Artificial intelligence based framework for robotic search and rescue operations conducted jointly by international teams,” in Proceedings of 14th International Conference on Electromechanics and Robotics “Zavalishin’s Readings” ER (ZR) 2019, Kursk, Russia, 17-20 April 2019 (Springer), 15–26.

CrossRef Full Text | Google Scholar

Malik, A. A., and Brem, A. (2021). Digital twins for collaborative robots: a case study in human-robot interaction. Robotics Computer-Integrated Manuf. 68, 102092. doi:10.1016/j.rcim.2020.102092

CrossRef Full Text | Google Scholar

Munich, Re (2023). Natural disasters in 2022. Accessed: 2023-05-29

Google Scholar

Murphy, R. R., Steimle, E., Hall, M., Lindemuth, M., Trejo, D., Hurlebaus, S., et al. (2011). Robot-assisted bridge inspection. J. Intelligent Robotic Syst. 64, 77–95. doi:10.1007/s10846-010-9514-8

CrossRef Full Text | Google Scholar

Nor, N. M., and Ma, S. (2014). “Cpg-based locomotion control of a snake-like robot for obstacle avoidance,” in 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 347–352. doi:10.1109/ICRA.2014.6906634

CrossRef Full Text | Google Scholar

Queralta, J. P., Taipalmaa, J., Pullinen, B. C., Sarker, V. K., Gia, T. N., Tenhunen, H., et al. (2020). Collaborative multi-robot search and rescue: planning, coordination, perception, and active vision. Ieee Access 8, 191617–191643. doi:10.1109/access.2020.3030190

CrossRef Full Text | Google Scholar

Rañó, I., Eguíluz, A. G., and Sanfilippo, F. (2018). “Bridging the gap between bio-inspired steering and locomotion: a braitenberg 3a snake robot,” in 2018 15th International Conference on Control, Automation, Robotics and Vision (ICARCV) (IEEE), 1394–1399.

Google Scholar

Rohmer, E., Singh, S. P. N., and Freese, M. (2013). “V-rep: a versatile and scalable robot simulation framework,” in IEEE International Conference on Intelligent Robots and Systems, 1321–1326.

CrossRef Full Text | Google Scholar

Sanfilippo, F. (2022). Combining grasping with adaptive path following and locomotion for modular snake robots. Int. J. Mech. Eng. Robotics Res. 11, 59–65. doi:10.18178/ijmerr.11.2.59-65

CrossRef Full Text | Google Scholar

Sanfilippo, F., Azpiazu, J., Marafioti, G., Transeth, A. A., Stavdahl, Ø., and Liljebäck, P. (2016a). “A review on perception-driven obstacle-aided locomotion for snake robots,” in Proc. of the IEEE 14th International Conference on Control. Phuket, Thailand: Automation, Robotics and Vision (ICARCV), 1–7. doi:10.1109/ICARCV.2016.7838565

CrossRef Full Text | Google Scholar

Sanfilippo, F., Azpiazu, J., Marafioti, G., Transeth, A. A., Stavdahl, Ø., and Liljebäck, P. (2017a). Perception-driven obstacle-aided locomotion for snake robots: the state of the art, challenges and possibilities. Appl. Sci. 7, 336. doi:10.3390/app7040336

CrossRef Full Text | Google Scholar

Sanfilippo, F., Helgerud, E., Stadheim, P. A., and Aronsen, S. L. (2019). “Serpens, a low-cost snake robot with series elastic torque-controlled actuators and a screw-less assembly mechanism,” in Proc. of the IEEE 5th International Conference on Control (Automation and Robotics (ICCAR), Beijing, China, 133–139. doi:10.1109/ICCAR.2019.8813482

CrossRef Full Text | Google Scholar

Sanfilippo, F., Stavdahl, Ø., and Liljebäck, P. (2017b). “Snakesim: a ros-based rapid-prototyping framework for perception-driven obstacle-aided locomotion of snake robots,” in Proc. of the IEEE International Conference on Robotics and Biomimetics (ROBIO), Macau, Macao, 1226–1231. doi:10.1109/ROBIO.2017.8324585

CrossRef Full Text | Google Scholar

Sanfilippo, F., Stavdahl, Ø., and Liljebäck, P. (2018). SnakeSIM: a ros-based control and simulation framework for perception-driven obstacle-aided locomotion of snake robots. Artif. Life Robotics 23, 449–458. doi:10.1007/s10015-018-0458-6

CrossRef Full Text | Google Scholar

Sanfilippo, F., Stavdahl, Ø., Marafioti, G., Transeth, A. A., and Liljebäck, P. (2016b). “Virtual functional segmentation of snake robots for perception-driven obstacle-aided locomotion?,” in Proc. of the IEEE International Conference on Robotics and Biomimetics (ROBIO), Qingdao, China, 1845–1851. doi:10.1109/ROBIO.2016.7866597

CrossRef Full Text | Google Scholar

Seeja, G., Selvakumar Arockia Doss, A., and Hency, V. B. (2022). A survey on snake robot locomotion. IEEE Access 10, 112100–112116. doi:10.1109/access.2022.3215162

CrossRef Full Text | Google Scholar

Sparkes, M. (2022). Call for robots at chernobyl. New Sci. 254, 8. doi:10.1016/S0262-4079(22)00774-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Tochilin, A., Voronkov, I., and Alabin, A. (2021). “Experience and prospects of using robotics in the nuclear power industry,” in IOP Conference Series: Materials Science and Engineering, 1047. IOP Publishing, 012193. doi:10.1088/1757-899X/1047/1/012193

CrossRef Full Text | Google Scholar

Unhelkar, V. V., Li, S., and Shah, J. A. (2020). “Decision-making for bidirectional communication in sequential human-robot collaborative tasks,” in Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction, 329–341.

CrossRef Full Text | Google Scholar

Wagner, A. R. (2021). Robot-guided evacuation as a paradigm for human-robot interaction research. Front. Robotics AI 8, 701938. doi:10.3389/frobt.2021.701938

CrossRef Full Text | Google Scholar

Wang, C., Wang, L., Qin, J., Wu, Z., Duan, L., Li, Z., et al. (2015). “Path planning of automated guided vehicles based on improved a-star algorithm,” in 2015 IEEE International Conference on Information and Automation (IEEE), 2071–2076.

CrossRef Full Text | Google Scholar

Webster, R. J., Okamura, A. M., and Cowan, N. J. (2006). “Toward active cannulas: miniature snake-like surgical robots,” in 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2857–2863.

CrossRef Full Text | Google Scholar

Zhang, F., Yu, J., Lin, D., and Zhang, J. (2022). Unic: towards unmanned intelligent cluster and its integration into society. Engineering 12, 24–38. doi:10.1016/j.eng.2022.02.008

CrossRef Full Text | Google Scholar

Keywords: cobots, search and rescue operation, human robot collaboration, snake robots, path planning, disaster scenarios, swarm robots

Citation: Moosavi SKR, Zafar MH and Sanfilippo F (2024) Collaborative robots (cobots) for disaster risk resilience: a framework for swarm of snake robots in delivering first aid in emergency situations. Front. Robot. AI 11:1362294. doi: 10.3389/frobt.2024.1362294

Received: 28 December 2023; Accepted: 21 February 2024;
Published: 04 March 2024.

Edited by:

Armin Wedler, Institute of Robotics and Mechatronics, Germany

Reviewed by:

Jinjin Yan, Harbin Engineering University, China
Riccardo Minto, University of Padua, Italy

Copyright © 2024 Moosavi, Zafar and Sanfilippo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Filippo Sanfilippo, filippo.sanfilippo@uia.no

Download