ORIGINAL RESEARCH article
Sec. Robotic Control Systems
Volume 8 - 2021 | https://doi.org/10.3389/frobt.2021.550644
A Holistic Approach to Human-Supervised Humanoid Robot Operations in Extreme Environments
- 1Institute for Experiential Robotics, Northeastern University, Boston, MA, United States
- 2Irish Manufacturing Research, National Science Park, Mullingar, Ireland
Nuclear energy will play a critical role in meeting clean energy targets worldwide. However, nuclear environments are dangerous for humans to operate in due to the presence of highly radioactive materials. Robots can help address this issue by allowing remote access to nuclear and other highly hazardous facilities under human supervision to perform inspection and maintenance tasks during normal operations, help with clean-up missions, and aid in decommissioning. This paper presents our research to help realize humanoid robots in supervisory roles in nuclear environments. Our research focuses on National Aeronautics and Space Administration (NASA’s) humanoid robot, Valkyrie, in the areas of constrained manipulation and motion planning, increasing stability using support contact, dynamic non-prehensile manipulation, locomotion on deformable terrains, and human-in-the-loop control interfaces.
As the worldwide energy demand is expected to increase by 50% within the next 3 decades, nuclear energy will play a critical role in meeting clean energy targets worldwide. At the same time, many of the world’s nuclear reactors are aging; from Japan to the United Kingdom to the United States, scientists, engineers and regulators are counting on new innovative technologies that will make decommissioning and clean-up missions safe for humans, environmentally-friendly and cost-effective. Furthermore, industrializing countries are investing in building new nuclear power plants to meet their growing energy demands.
Nuclear energy operations and nuclear disasters have great international impact with no boundaries. Ensuring safe, efficient, and productive operations of facilities and improving response to unplanned emergencies at any location around the globe is in the best interest of the international community. Moreover, the urgency and scale of the problems identified in high-consequence situations, such as the Fukushima (Japan) clean up and waste tank decommissioning in Savannah River Site (United States), require an interdisciplinary team of scientists, engineers, and technologists to solve similar yet sufficiently distinct challenges. For example, since 1989, the United States Department of Energy has spent over $250 billion of public funds on cleanup. The cleanup is less than half complete and the remaining mission scope is estimated to cost at least another $250 billion more over a 40–50- year period. Similarly, the Japanese government estimated the total costs for the Fukushima cleanup at $188 billion for the next 40 years.
It is now clear that robotics will play a key role in accelerating these cleanup timelines and reducing the costs, by addressing operational needs and challenges in nuclear facilities. Robotics technologies are needed to remotely access nuclear and other highly hazardous facilities under human supervision to perform inspection and maintenance tasks during normal operations. During decommissioning, robots will become the eyes and hands of human operators in places no human has gone for more than 50 years. However, using robots in such environments is not without challenges, the Fukushima disaster has shown that conventional robots must be significantly modified, Nagatani et al. (2011b), to cope with highly radioactive environments. This is typically achieved by equipping the system with lead plates to protect electrical components, Nagatani et al. (2011a), which in-turn add weight and may impair functionality. Alternatively, the number of electronic components can be minimized and used in conjunction with specifically hardened parts, however high levels of radiation still can destroy such systems in a matter of days, Funakoshi (2016), Urabe and Stapczynski (2017), Yirmibeşoğlu et al. (2019). Nevertheless, robots have already been deployed in less active nuclear environments in commissioning and waste disposal task, Bogue (2011), and on-going research is demonstrating promising applications for human-supervised robotics in a range of different tasks in the future, Marturi et al. (2016). They will improve our ability to respond to and recover from unplanned events or operational emergencies in such critical and safety-significant applications. Nuclear environments are dangerous for humans to operate in due to the presence of highly radioactive materials. They are typically distant as the facilities separate these dangerous environments from humans by thick walls. And they present daring operational conditions by size and configuration with tight passages, debris accumulated over the years and cluttered internals.
Our research plan is motivated by the need for general purpose robots in routine and emergency operations in nuclear facilities for: 1) Disaster response and environmental clean-up, such as more than 175 waste tanks in Hanford (WA) or the F-canyon in Savannah River (SC). These processes rely heavily on accurate remote sampling and characterization before permanently grouting the facilities, by collecting samples from different spots for analyzing what and how much radioactive material remains. The operating environment is unknown and cluttered with vertical and horizontal piping, fallen debris and puddles and muck-like material on the floors (in the waste tanks). 2) Operational efficiencies—federal laws require nuclear facilities to develop and maintain emergency preparedness plans to protect the public. These emergencies include unusual events during normal operations to black swan events such as Fukushima, Chernobyl and Three Mile Island accidents. Furthermore, the aging workforce in the energy sector requires the adoption of technology to keep up with day-to-day operations. As a result, human-supervised robot assets with robust manipulation capabilities in these challenging environments and situations are needed. 3) Worker safety—Before any work can begin, human workers must enter a facility to characterize radioactive hazards, such as type of radiation, dose rates, and location of sources. This data is then used to determine the proper protective clothing and stay time limits for personnel. Replacing personnel with robots would be highly desirable in radioactive environments.
This paper discusses our approach to move towards utilizing humanoid robots in nuclear energy operations and nuclear disasters by furthering the development of human-supervised robot control and manipulation capabilities. Section 2 presents a method capable of handling manipulation and motion planning in constrained environments, such as gloveboxes. Section 3 presents work in utilizing support contacts to increase stability of a standing humanoid robot operating inside a glovebox. Section 4 presents a method to plan dynamic non-prehensile manipulation behaviors in a highly-constrained environment, with focus on gloveboxes. Section 5 presents a way to estimate in-situ deformable terrains in order to navigate in unknown environments. Section 6 presents a human-in-the-loop user interface for operating humanoid robots.
2 Constrained Manipulation and Motion Planning
In hazardous environments it is crucial to perform manipulation tasks effectively. Additionally, such environments often provide constraints on the motions allowed, such as when operating through glovebox ports. In order to accomplish effective manipulation, different robot configurations should be evaluated. This is particularly important for redundant systems such as humanoid robots. A robot performance measure can be classified as local, e.g., manipulability (Yoshikawa, 1984) or global such as workspace analysis (Vahrenkamp and Asfour, 2015). Local indexes are advantageous as they provide a more generic solution and may be utilized in control frameworks without workspace knowledge. Hence, they can be used to choose a configuration based on a robot’s inherent capability. However, local indexes study the system’s kinematic transformations from configuration to task space, ignoring environmental constraints that have significant effects on the robot’s admissible motions. In particular, for operations in hazardous environments, the workspace is often unstructured and uncontrolled. Moreover, unscheduled contacts may lead to catastrophic results. For these reasons, it is important to transmit an accurate measure of what the robot can or cannot do in its current pose to a remote operator/supervisor.
The following section recalls the work presented in (Long and Padır, 2018; Long and Padir, 2019; Long et al., 2019) in which a new measure called the constrained manipulability polytope (CMP) that considers the system’s kinematic structure, including closed chains, composite sub-mechanisms, joint limits, and the presence of obstacles is developed.
2.1 Related Work
The manipulability ellipsoid first defined (Yoshikawa, 1984) measures the capabilities of a robot manipulator based on its kinematic structure. Extensions to include positional joint limits are proposed in (Tsai, 1986) using penalty functions and (Abdel-Malek et al., 2004) using augmented jacobian matrices. Robots with heterogeneous joint velocity limits are examined in (Lee, 1997), while dynamic constraints are considered (Bowling and Khatib, 2005; Zollo et al., 2008). For humanoid robots, improvements on local measures can be obtained by including the effects of contact while evaluating the dynamic manipulability of a humanoid’s center of mass (Gu et al., 2015; Azad et al., 2017). Moreover in (Azad et al., 2017) either joint torque and/or acceleration limits are accounted for by using a scaling matrix. In (Vahrenkamp et al., 2012, 2013) the manipulability ellipsoid is augmented to include environmental constraints. The authors include joint position limits and the detrimental effects of nearby obstacles using a spatial decomposition referred to as the hyperoctants approach. Manipulability polytopes (Kokkinis and Paden, 1989) provide a more elegant and indeed exact method for representing velocity limits in the Cartesian space. There are several examples where diverse constraints, defined by a set of inequality or equality equations, have been incorporated into polytopes. For instance, mobile robot toppling constraints have been integrated into the available wrench set for a cable-driven parallel robot in (Rasheed et al., 2018). Alternatively, friction constraints can be added after linearization (Caron et al., 2017).
Consider an n degree-of-freedom (DOF) manipulator in m dimensional space. Let
and its volume, denoted as
2.3 Constrained Manipulability
The manipulability polytope does not give a true picture of the robot’s capabilities as they may be reduced due to environmental or joint limit constraints. Thus in our previous work (Long and Padır, 2018; Long and Padir, 2019), a method of considering obstacle and joint position limits is given. To do so the kineostatic danger field (Ragaglia et al., 2014) as an input which limits the maximum attainable velocity in the direction of a potential collision. The robot’s velocity is reduced until the danger-field value is below a predefined threshold. The kineostatic danger field divides the robot’s links into l control points (CPs) and the workspace into c cells. The danger field for the
Finally, by introducing
The robot’s velocity is reduced until the danger field value at the obstacle location, denoted as o is below a threshold, i.e., a desired danger value. Eq. 11 is re-written as
rewritten, for the
Eq. 15 considers the reduced performance capabilities due to nearby obstacles. It is similarly convenient to consider the effects of joint limit proximity in the polytope before transformation to the Cartesian space, thus avoiding improper penalization due to redundancy (Tsai, 1986; Abdel-Malek et al., 2004). For the
2.4 Applications of Constrained Manipulability Polytope for Humanoid Robots
In the following, applications for the performance measure are demonstrated using NASA’s humanoid robot Valkyrie (Radford et al., 2015) interacting with a glovebox for nuclear decommissioning task.
2.4.1 Experiment 1: Right Arm Insertion
In the first experiment the right arm of Valkyrie is inserted into the glovebox. The glovebox is considered as an obstacle that reduces the manipulator’s performance, as unwanted collision could be dangerous. Figure 1 shows the right arm insertion task, demonstrating how the CMP changes with time. The initial reduction in manipulability is due to positional joint limits. As the right hand passes through the glovebox port the system experiences a reduction of velocity capacity due to the constrained space signifying that the hand cannot move quickly without increasing the likelihood of a collision. A partial recovery can be observed as the right hand is fully inserted, meaning the system can manipulate objects within the space.
FIGURE 1. Valkyrie inserting right arm into a glovebox and the evaluation of the manipulability polytope (red) and constrained manipulability polytope (blue) of the right arm. A video of the task is available here: https://www.youtube.com/watch?v=FzlhsLH5IPU.
2.4.2 Experiment 2: Dual Arm Insertion
Convex polytopes are geometric objects, thus can be combined through standard geometric operations. These combinations can be used to represent composite robotic chains both serial and parallel. In (Long and Padir, 2019), we have shown how the manipulability of mechanisms in series can be obtained from the minkowski sum of sub-mechanisms, while manipulability mechanisms in parallel can be obtained by a straightforward concatenation of inequality constraints.
In this experiment, we demonstrate the former, as Valkyrie inserts both arms into the glovebox. It is assumed that the two arms form a closed chain, in a scenario where the arms carry a common object or tool. The goal is to show how the CMP can be combined to obtain that of a closed chain system. To model the closed chain, we use the virtual object procedure, i.e., a rigid straight link extending from the left to the right hand (Long et al., 2015). Figure 2 shows the motion associated with the dual-arm insertion task at four time instants and the CMP for the closed-chain system evaluated at the right-arm end effector. This is obtained by first calculating
FIGURE 2. Valkyrie inserting both arms into a glovebox shown at four timesteps along with the coordinating polytopes,
2.4.3 Experiment 3: Reachability Analysis
Finally, a reachability study/workspace analysis is presented in Figure 3. The environment is discretized into voxels. At each voxel, an optimization procedure obtains a feasible IK solution while trying to maximize the robot’s distance to obstacles. The CMP is calculated in this configuration. The workspace discretization is shown in Figure 3. The voxel’s color is defined by the volume of
FIGURE 3. The space is discretized into 3D voxels. (A) At each voxel, the IK solution is obtained for the left arm (left images) and the right arm (right images). The corresponding volume of the CMP is calculated for each voxel, giving a good understanding of the robot’s workspace. A video is available here: https://youtu.be/jc7X4WakdoE(B) The MP’s volume
3 Increasing Stability Using Support Contacts
As nuclear facilities reach the end of their life cycle they must be decommissioned in a safe and efficient manner. A particularly dangerous task is the decontamination of gloveboxes that have been previously used to manipulate radioactive material. Although a robotic system that is specifically designed for glovebox operations may be the best solution, humanoid robots are an attractive option since they can operate in a variety of environments and use tools that are designed for humans. While conducting operations within the glovebox, the constraints imposed by the ports, gloves, and the external structure, which effectively fix the arms at the entry points, must be considered. The inability to alter body configuration greatly diminishes the robot’s capacity to take steps in arbitrary directions. This in turn leads to a danger of toppling during task execution as the system cannot easily change the support polygon’s location.
Toppling occurs when the ZMP leaves the support polygon (SP) (Vukobratović and Borovac, 2004). If the SP cannot be displaced, alternative methods to maintain stability must be employed. For example, in (Rasheed et al., 2018), the ZMP position for a cable-driven mobile robot is modified online by a tension distribution algorithm. In (Khatib and Chung, 2014), it is shown that the SP size can be increased by using supplementary contact points. Similar to these approaches, we propose to exploit the contacts in the glovebox (i.e., leaning on the entry ports) in order to shift the ZMP towards the center of the SP while performing manipulation tasks.
This section presents our work on planning kinematic motions with support contacts without a predefined contact schedule that maintains the stability of the ZMP. To accomplish this, we model rigid-body contacts using complementarity constraints and solve a nonlinear constrained optimization for joint velocities and contact forces. Owing to the differentiable contact model, gradient-based optimization can reason about contacts between the robot arms and the glovebox ports. This optimization also respects constraints that ensure an object is grasped by the end effectors, the ZMP is in a safe region, and the deviation of the object’s position from a desired position is acceptable. Furthermore, we present a null-space-based torque controller that prioritizes the stability, i.e., generating the support forces, and projects the torques needed for the manipulation task onto the null space of the support forces. The proposed methodology is tested through 2.5D, quasi-static simulations by considering a humanoid robot with two planar arms manipulating a relatively heavy object on an elevated plane representing the glovebox.
3.1 Related Work
For planning a motion with contact interactions, both the discrete contact events (e.g., making/breaking contacts at certain locations) and the continuous variables (e.g., joint positions, contact forces, stability constraints) must be considered. One approach is to use a contact-before-motion planner such as those presented in (Hauser et al., 2005; Escande et al., 2013). In this case, first, a sequence of contacts at predefined locations is determined, then a continuous motion is planned subject to contact constraints. In contrast, in a motion-before-contact planner, the contacts are obtained as a result of the motion planning (Escande et al., 2013). Alternatively, contact-implicit motion planning (also known as, motion planning through contacts) can be used to plan for smooth motions and contact events at the same level.
In contact-implicit motion planning, a differentiable contact model is used to enable gradient-based optimization to reason about contacts. Complementarity constraints are widely used to model rigid-body contacts with friction, as proposed in (Stewart and Trinkle, 1996; Anitescu and Potra, 1997). (Yunt and Glocker, 2005; Posa et al., 2014) use complementarity constraints to model contacts in a trajectory optimization problem. The main idea here is to consider the contact-related parameters as additional optimization variables such that the contact events evolve along with continuous motion variables. Such an optimization problem can be solved locally through constrained nonlinear optimization algorithms such as sequential quadratic programming (Anitescu and Potra, 1997; Fletcher and Leyffer, 2004).
Once a kinematic motion with contact interactions is planned, it can be executed using a null-space-based controller. Park and Khatib (Park and Khatib, 2006) proposed a torque control framework for humanoid robots with multiple contacts and verified the method experimentally in (Park and Khatib, 2008). Moreover, they extended this method to a unified hierarchical whole-body control framework for humanoid robots in (Khatib et al., 2008). In this framework, tasks are hierarchically ranked. Thus, the torques required for a lower-priority task are projected onto the null-space of the Jacobian matrix associated with a higher-priority task. In (Henze et al., 2016a), a whole-body torque controller for humanoid robots is proposed that combines passivity-based balancing proposed in (Henze et al., 2016b) with a hierarchical null-space-based control that is similar to (Khatib et al., 2008).
3.2.1 Static Equilibrium
For being balanced, the robot needs to be in static equilibrium (Vukobratović and Borovac, 2004). In this case, the static equilibrium of the system can be evaluated considering the following wrenches: the wrench due to the robot’s mass, the wrenches at the end effectors due to the object wrench, and the wrenches at the support contact points. Henceforward, we enumerate the left and right arms as the first and second arms, respectively.
In the static equilibrium, the net force must be zero:
where m is the total mass of the robot,
Additionally, the projection of the net moment,
using the wrench matrix
3.2.2 Motion Planning
In this work, we ignore the dynamic effects and investigate the quasi-static case for dual-arm manipulation of an object in a confined space, i.e., a glovebox. In the following, robot’s joint positions and velocities are denoted by
As a result, the following optimization problem is solved for the joint displacements
3.2.3 Torque Control
Using this optimization procedure, the robot configuration and the support forces’ magnitude are obtained. Nevertheless, a torque controller is necessary to execute the planned motions.
The torques necessary to generate the desired object wrench
The support forces are oriented normal to the contacting robot geometry. Hence, using the contact angle
Similarly, the joint torques required to generate these forces, denoted as
For the glovebox task, the support forces are crucial to maintain the stability of the robot, while generating the desired object wrench has a lower priority. Thus, we compose the joint torques,
is the null space projector of the support forces, and
To test the proposed framework, we run simulation experiments in which a humanoid robot that has two planar 4-DOF arms with revolute joints manipulates a relatively heavy rigid bar on an elevated plane. The robot’s arms pass through two ports representing the glovebox. We neglect the dynamics (i.e., velocities and accelerations) and assume point contacts without friction. The weights are selected as
We investigate the task of moving the object 40 cm forward on a straight path that consists of nine equally spaced waypoints. The results are depicted in Figure 4. Each step of the motion is indicated by a color from blue to red. In the initial configuration (indicated by blue), the robot grasps the object from both ends. During the simulation, the position of the ZMP is calculated with and without the effect of the support contacts on the glovebox frame. The latter is known as the fictitious ZMP (FZMP) (Vukobratović and Borovac, 2004) and may fall outside of the SP.
FIGURE 4. (A) Planned motion for carrying an object on a straight path. (B) Zoomed in change of the ZMP and the FZMP (i.e., fictitious ZMP without supporting contacts). throughout the task.
The results show that the contact-implicit motion planning method can increase the stability of the robot using support contacts while performing the manipulation task. The robot makes contacts with the glovebox to maintain its balance while moving the object on the desired path. As soon as the FZMP leaves the SR in the second step, the right arm makes a contact with the left end of the port to push the ZMP into the SR. As the object moves further away from the base, the contact angle is varied so that the magnitude of the support force in
A zoomed-in version of the SP area is depicted in Figure 4B to show the change of the ZMP and the FZMP throughout the simulation. Even though the FZMP moves forward along with the object’s position and leaves the SP eventually, the ZMP does not leave the SR owing to the support forces. It is also noted that, in Step 7 (indicated by yellow), the ZMP is more centralized with respect to the y axis compared to the other steps due to the symmetry of the support forces.
Figure 5 show the magnitudes of the support forces and the joint torques with respect to the distance between the object and the robot’s base. The magnitudes of the support forces are much larger than the magnitude of the object wrench, and therefore the torques are much more affected by the support forces than the object wrench. This is why force and torque vs. distance characteristics are quite similar—i.e., the torque is dominated by the support forces (especially after Step 5). Moreover, the changes of
FIGURE 5. The magnitudes of (A) the support forces vs. the distance of the object from the base, and (B) the joint torques vs. the distance of the object from the base.
4 Dynamic Non-prehensile Manipulation
There is an increasing need to carry out decontamination and decommissioning tasks in safe and effective manner. A particularly dangerous task is glovebox decontamination and decommissioning that typically involves transporting debris and objects from the interior of the glovebox to an exit port, where they are bagged and removed (Long et al., 2018; Önol et al., 2018). Such tasks do not always require dexterous manipulation behaviors and instead simply require objects to be push from the interior to the exit port of a glovebox.
Contact-implicit trajectory optimization (CITO) is a promising method to generate contact-rich behaviors given only a high-level definition of a task. In this approach, a differentiable contact model is used to enable gradient-based optimization to reason about contacts such that discrete contact events and continuous trajectories are found simultaneously as a result of smooth optimization.
In this section, we present a CITO method based on a variable smooth contact model to plan dynamic non-prehensile manipulation behaviors for a 7-DOF robot arm in a highly-constrained environment. We demonstrate that the proposed method can solve complex tasks despite tight constraints imposed by the environment by exploiting the smooth virtual forces. Moreover, we experimentally verify that the physical inaccuracy introduced by the residual virtual forces is admissible and the motions found by this framework are realistic enough to be run on the hardware.
4.1 Related Work
Complementarity constraints are widely used to model rigid-body contacts in trajectory optimization (Yunt and Glocker, 2005; Posa et al., 2014; Gabiccini et al., 2018). This approach can find complex motions, but it typically suffers from poor convergence speed. Thus (Tassa et al., 2012; Mordatch et al., 2015; Mastalli et al., 2016; Neunert et al., 2016; Manchester and Kuindersma, 2017), use smoother fragments of the complementarity constraints. (Neunert et al., 2017; Giftthaler et al., 2017; Marcucci et al., 2017; Neunert et al., 2018), on the other hand, define contact forces as a smooth function of distance, i.e., a smooth contact model. Using such a contact model, highly-dynamic complex motions for a quadruped robot are planned and executed in real-time in (Neunert et al., 2018). However, it is difficult to tune smooth contact models (Carius et al., 2018), and the resulting motions may be physically inaccurate due to the non-physical contact forces that act from distance. In order to address these problems, we have recently proposed a variable smooth contact model (VSCM) (Önol et al., 2018) that injects virtual forces to the underactuated dynamics with frictional rigid-body contact mechanics, such that the states of the manipulator and the objects are coupled in a smooth way. Furthermore, the smoothness of the contact model is adjusted by optimization such that large virtual forces are permitted in the initial phases of optimization but vanish as the optimization converges. As a result, the VSCM improves the convergence of CITO without compromising the physical fidelity of resulting motions.
CITO has been used for animated characters (Mordatch et al., 2021a; Mordatch et al., 2012b) and in robotics (Tassa et al., 2012; Posa et al., 2014; Mordatch et al., 2015; Mastalli et al., 2016; Manchester and Kuindersma, 2017; Neunert et al., 2017; Carius et al., 2018; Neunert et al., 2018; Winkler et al., 2018). Although this method is task independent and can be generalized to both locomotion and manipulation problems, the majority of the related literature considers only the former. On the other hand, in (Mordatch et al., 2012a; Posa et al., 2014; Gabiccini et al., 2018), manipulation tasks are investigated but their analyses are either limited to a planar case or based on animated characters where physical fidelity is not critical. Recently (Önol et al., 2018, 2019, 2020; Sleiman et al., 2019), used CITO for non-prehensile manipulation tasks. Yet, they consider only tabletop pushing scenarios. Moreover, in general, experimental results in this domain are very limited, albeit with some notable exceptions (Mordatch et al., 2015; Mastalli et al., 2016; Neunert et al., 2017, 2018; Giftthaler et al., 2017; Carius et al., 2018; Winkler et al., 2018; Sleiman et al., 2019). Nonetheless, to the best of our knowledge, there is no experimental verification of CITO for constrained dynamic manipulation.
4.2.1 Dynamic Model
The dynamics of an underactuated system consisting of an
In this study,
4.2.2 Contact Model
The virtual forces generated by the contact model acts upon the unactuated DOF in addition to the external rigid-body contacts. Consequently, the robot’s and objects’ dynamics are related through the virtual forces. We assume an exponential relationship between the magnitude of the normal contact force γ and the signed distance between paired contact geometries ϕ, as depicted in Figure 6. While frictional forces are not considered in this contact model, the rigid-body contact mechanics [i.e.,
FIGURE 6. The relationship between the virtual force and the distance.
This model is analogous to a spring model and (Marcucci et al., 2017) lists several reasons for not using damping (i.e., a velocity term) in such a contact model. The corresponding virtual force effective at the center of mass of the free body associated with the contact candidate
In the variable smooth contact model, the virtual stiffness k for each time step and contact pair is a decision variable of optimization and initialized with a large value such that there is a non-zero virtual force on each contact candidate. Nonetheless, the virtual forces are penalized as an integrated cost, so that they vanish as the optimization converges, see Figure 6. This approach helps to discover contact candidates that are initially distant.
4.2.3 Trajectory Optimization
In this study, the optimal control problem is transcribed into a finite-dimensional nonlinear constrained optimization by assuming constant control inputs over N discretization intervals. Final cost terms penalize the deviations of the objects’ poses from desired poses,
The following optimization problem is solved by a sequential quadratic programming (SQP) algorithm by rolling out the dynamics:
The lower and upper bounds for the control variables
4.3.1 Application: Non-Prehensile Manipulation in a Glovebox
Our overall objective is to enable a robot to carry out such manipulation tasks with only high-level commands such as desired object poses. Figure 7D shows Sawyer, a 7-DOF robot arm, and a mock-up glovebox environment. Non-prehensile manipulation is advantageous in this case, as the highly-constrained environment means that grasp configurations are difficult to obtain. Thus, we consider non-prehensile manipulation tasks in a glovebox.
FIGURE 7. Snapshots from the hardware experiments: pushing an object on a table (A–C), pushing an object within a glovebox (D–F), manipulating an unreachable object by exploiting inter-object contacts (G–I).
The proposed method is tested in three different scenarios of increasing complexity: 1) pushing an object on a table, Figure 7A-C 2) pushing an object in a glovebox, Figure 7D-F, and 3) ejecting an unreachable object from the glovebox by exploiting physical interactions in the environment, Figure 7G-I. In the first case, there is a (red) box on a table and the task is to move it 20 cm along the x axis, see Figure 7A for the reference frame. In the second case, the task is to move the object 10 cm along the
4.3.2 Experimental Setup
In the experiments, standard and commercially-available hardware2 is used in order to facilitate reproducibility. The dynamics is modeled using MuJoCo physics engine (Todorov et al., 2012) with time steps of 5 ms while the control sampling period is 50 ms. The SQP solver SNOPT (Gill et al., 2005) is used to solve the optimization problem. As an interface between MuJoCo and SNOPT, IFOPT (Winkler et al., 2018) is employed. The planned position, velocity, and acceleration trajectories are interpolated with 10 ms steps and executed on the robot by using the built-in inverse dynamics feed-forward controller of Sawyer. For detecting the poses of the objects and the glovebox through the head camera of Sawyer, AprilTag 2 algorithm (Wang and Olson, 2016; Malyuta, 2017) is used.
For all cases, the weights are
Figure 7 demonstrates initial, intermediate, and final snapshots from the experiments. Table 1 shows the position and orientation deviations for the object (
Despite the trivial initial guess and no additional tuning for different tasks, the proposed method is capable of finding a motion that successfully completes each task in simulation. That is, the object is moved to within 1-cm radius of the desired position while the change of orientation is negligible (i.e., smaller than 15°). It is noteworthy that in the second and third cases, the glovebox port imposes tight constraints on the motion and the workspace of the robot, yet still our method can handle this without any additional penalties or constraints for collisions with the glovebox or tuning. Moreover, the last task requires the blue box to have a high velocity when the contact between it and the red box is broken because when the red box collides with the yellow box that is under the blue box, the red box cannot apply a force on the blue box anymore and the blue box would still be in the glovebox. Thus, using a non-dynamic planner or running the resulting motions through a position controller without velocities and accelerations would not work in this case.
A more detailed numerical analysis of the experiments is given in the following. In the simulation for Task 1, the box is moved 19.8 cm along the x axis and 0.7 cm along the y axis; whereas, in the hardware experiment, the box is pushed only 15 cm along the x axis and 0.5 cm along the y axis. For Task 2, the displacements of the box along the −y and x axes are 9.8 and 0.8 cm in the simulation. In the experiment, the box is moved 8 cm along the −y axis, which is satisfactory, but also 7 cm along the −x axis due to the relatively large rotation (ca. 20°) about the z axis. In Task 3, the blue box is ejected from the glovebox, namely the task is completed, both in the simulation and the hardware experiment. Since the final position of the blue box could not be detected in the experiment, only the final positions of the red box are compared here. In the simulation, the red box is moved 10.8 cm along the −y axis and 0.4 cm along the −x axis, and these quantities are 7.5 and 0.3 cm for the hardware experiment.
On average, the position and orientation discrepancies between the simulation and hardware experiments are 5 cm and 12°. Such differences can be deemed reasonable since we playback the planned trajectories on the robot using a naive joint-level controller, i.e., without a closed-loop controller that tracks and compensates for the deviation of the object’s trajectory from the planned one. The deviations of the executed motions from the planned motions are expected considering errors caused by modeling and perception. The main goal of this study is to show that the proposed method can solve for complex tasks by exploiting the smooth virtual forces and the residual non-physical forces do not hinder the task performance.
5 In-situ Terrain Classification and Estimation
Robust locomotion on deformable terrains is necessary for biped humanoid robots to perform tasks effectively in unstructured environments. The knowledge of deformable terrain properties, particularly the stiffness, has major implications in modeling the robot walking dynamics, which is the key to achieve stable gait patterns. Prior studies on walking stabilization chose to model such walking dynamics using pre-identified stiffness or damping constants (Wittmann et al., 2016; Wu et al., 2018; Hashimoto et al., 2012). However, it is unlikely for robots to access such terrain properties in advance when deployed to unknown environments.
In this section, we present an in-situ ground classification and estimation method that can be used to improve the stability of the robot while traversing unknown terrain, utilizing NASA’s humanoid robot Valkyrie (Radford et al., 2015). The terrain estimation works in two steps: i) The robot tries to identify the terrain type from a database. If the terrain is recognized, all needed data can be retrieved and used by the controller. ii) If the terrain is classified as an unknown type, the robot then estimates its stiffness by using Bernstein-Goriatchkin (Ding et al., 2013; Caurin and Tschichold-Gurman, 1994) pressure-sinkage model. The estimated stiffness can then be fed to stabilizers such as the one proposed in (Hashimoto et al., 2012).
5.1 Related Work
Our study on robot foot-terrain interaction is inspired by (Skonieczny et al., 2014), where the interaction between soil and single wheel is analyzed using optical flow techniques. Computer vision techniques have been widely used for terrain classification in the past (Weiss et al., 2008; Brandão et al., 2016). However, due to the poor lighting conditions in the outer space, it is desirable to augment vision-based techniques with a terrain classification approach that relies on physical contacts between the robot foot and the terrain. We thus aim at providing a “sense-of-walking” to the robot by using on-board sensors. In (Walas et al., 2016; Otte et al., 2016), ankle mounted force/torque sensors and accelerometers are used, respectively, to achieve terrain classification. Our approach is comparable to (Otte et al., 2016) as Recurrent Neural Networks (RNNs) are used but they differ in the aspect that we perform terrain classification with a bipedal robot while (Otte et al., 2016) uses a wheeled mobile robot. To describe terrains’ properties under pressure, various pressure-sinkage models have been studied in the past (Komizunai et al., 2010; Ding et al., 2013). There is no common opinion on which model is better than others. We choose Bernstein-Goriatchkin model considering it is one of the most commonly used models and it is relatively easy to implement. The method we developed for terrain estimation can be viewed as an extension of (Will Bosworth1 and Hogan, 2016), where spring model is used to estimate the ground stiffness by measuring the force and leg displacement during the interaction between a quadrupedal robot, Super Mini Cheetah (SMC), and the ground. Since SMC has spherical shaped feet, point contacts are used when modeling the foot-terrain interaction. This is not applicable for bipedal robots with flat soles because the support force on sole is distributed unevenly. To overcome this challenge (Caurin and Tschichold-Gurman, 1994), develops a special sensor system consisting of multiple sensor cells to measure the force distribution. Instead of using additional sensors, our approach achieves estimation for stiffness using a single load cell which provides only a net force measurement.
5.2 Terrain Classification
To study the foot-ground interaction, we designed a set of experiments involving Valkyrie interacting with four types of terrains, including wood mulch, rubber mulch, mason sand, and gym foam mats. A custom sandbox, as seen in Figure 8, was built for Valkyrie to stand in and perform a set of pre-designed motions. Half of the sandbox is covered by a wood plate, where Valkyrie can stand stably, and the other half is filled with specified deformable ground materials to be tested. The sandbox is not used when testing on the gym foam mat, which Valkyrie can stand on directly. The data for terrain classification and estimation is collected by commanding Valkyrie to perform a touch-land motion.
FIGURE 8. Valkyrie is performing the touch-land motion. The red arrow represents CoM and yellow arrow represents support force on each foot.
Valkyrie starts from standing on the wood plate of the sandbox and then moves the right foot onto the terrain. During this motion, Valkyrie slowly moves the center of mass (CoM) from the left foot to the center of the two feet (Figure 8). The support force on the right foot increases from zero to approximately 650 N, which is about half the weight of Valkyrie, in this process. Valkyrie then moves the right foot back onto the wood plate and the motion is finished. The right foot force and torque changes during this motion are measured by the load cell. Meanwhile, the right foot sinkage in the terrain is also calculated using joint positions.
One method to calculate the foot sinkage, δ, is to use the kinematic chain of the robot. Transformation matrix from coordinate frame RightFoot1 to RightFoot2 is denoted with
We use RNNs, particularly Long Short-Term Memory (LSTM) network, to classify the terrain. Comparing to other classification methods such as Support Vector Machine (SVM), the use of RNNs gains us two advantages. First, our method can be applied to a raw data stream rather than accumulated data over time, thus, we can achieve terrain classification in real-time. Second, it has been shown that RNNs have better performance than SVM (Otte et al., 2016) and we achieved 95% accuracy on average during experiments.
Fifteen features are selected as input for the LSTM network, which are: right foot force in X, Y, Z directions, right foot torque in X, Y, Z directions, pitch/roll angular displacement/acceleration of the right ankle, pitch/roll angular torque of the right ankle, and the right foot sinkage, represented by the first three elements on third row of the transformation
5.3 Terrain Estimation
Although the RNN based classification method works well, there are chances that Valkyrie may encounter with terrains of unknown type. To handle such situations, a method that can estimate an unknown terrain’s stiffness, which governs the foot sinkage behavior under load, is desired.
5.3.1 Terrain Stiffness Model
FIGURE 9. Contact between the foot and ground terrain. (A) Even sinkage when the contact is flat. (B) Non-even sinkage when the contact is oblique.
In this case, it is impractical to identify the sinkage value corresponding to the pressure calculated from the measured force on ankle. To handle this mismatch, we calculate the net force acting on the sole resulting from different levels of sinkage using the following equation
where F is the net force acting on the foot,
Theoretically, if two sets of F and
5.3.2 Experiments and Results
To test the proposed estimation method, we run four experiments: two with gym foam mats (Gym Foam Mat one and two), Rubber Mulch, and Mason Sand. The same touch-land motion as described in Section 5.2 is performed by Valkyrie to collect data for terrain estimation. The support force acting on Valkyrie’s right foot and the frame transformation
TABLE 2. Estimated stiffness parameters k and n for tested terrains using Bernstein-Goriatchkin Model.
To validate the estimation results for the foam mats and the rubber mulch, we compare ankle torques measured from the load cell with ankle torques calculated using the estimated parameters. We choose torque comparison as the validation approach for one practical reason: many bipedal robots walking algorithms keep the robot balanced by tuning torques at the ankle (Kim et al., 2007; Kang et al., 2012). We thus posit that a set of terrain parameters that can correctly predict torque changes can be useful in walking algorithms for keeping robots stable on deformable terrain. When Valkyrie’s foot lands on the terrain, the torque measured at the ankle should be equal to the torque caused by the force acting on the sole around the ankle axis. Considering the torque caused by gravity, we have.
Figure 11 shows the comparison between the calculated foot pitch and roll torques using estimated parameters and measured foot pitch and roll torques from the ankle load cell. The blue curves represent the measured results while the red curves stand for the estimated results. The plots show that the estimated parameters can predict the foot pitch torque well. As for the foot roll torque, the estimated results match the trend of torque changes with an offset of 1.5 Nm on average. One possible source of this offset is the friction between the foot and the terrain. The friction coefficient between the foot and the terrain is currently not estimated and is planned as the future work.
FIGURE 11. Foot pitch and roll torque comparison. Blue curves stand for measured torques and red curves represent the estimated torques.
To numerically examine how well the estimated parameters predict the foot pitch torque, we calculate the Normalized Root-Mean-Square Error (NRMSE) between the estimated pitch torques and the measured pitch torques. The Gym Foam Mat one and two show a NRMSE of 15.9 and 15.7%, respectively, which is the equivalent to an accuracy of 84.1 and 84.3%. While the Rubber Mulch showed an NRMSE of 24.2%, which is an accuracy of 75.8%. Since the estimated parameters from the Gym Foam Mat one and two and Rubber Mulch experiments can well predict foot pitch torque (an average accuracy of 80%) and roll torque (same trend but with an offset), we assert that the proposed method can be used to measure the stiffness of an unknown deformable terrain. Future work will focus on improving the estimation accuracy, including estimation of friction coefficient for a thorough knowledge of unknown terrains.
6 User Interface
Supervisory-control interfaces are an important aspect to robots in extreme environments as you often want a human-in-the-loop to help evaluate and make decisions for critical objectives of a task at hand. In any supervisory-control interface, it is important to find the balance of control between robot and human. Too much control or autonomy on the robot side leaves the human unable to potentially provide valuable feedback to the robot. While too much control on the human side can leave the operator overwhelmed with information and decision making, leaving less focus for the critical elements of a given task. Additionally, too much control on the human side tends to increase the interface complexity and therefore increase the amount of required operator training to utilize the interface.
In this section, we present our work on a computer-based, supervisory-control interface for NASA’s humanoid robot Valkyrie (Radford et al., 2015), that aims to balance the supervisory-control to allow for relatively untrained operators. Our interface was originally designed to accomplish the tasks put forward by the Space Robotics Challenge, a challenge using a simulated Valkyrie robot in a virtual, mock-up Mars environment with tasks such as turning valves, moving equipment, and inspection. However, our interface was purposely designed to be generic enough to accomplish a wide variety of tasks and not require tasks to be pre-programmed into the interface and is therefore extendable to a wide variety of domains.
6.1 Related Work
The DARPA Robotics Challenge (DRC) was a two-year challenge with the goal of accelerating progress in human-supervised robotic technology for disaster-response operations. The DRC helped support development for a variety of robotic research, with one key area being the design of human-robot interaction with focus on supervisory control methodologies (Yanco et al., 2015; Norton et al., 2017). The DRC allowed for a variety of supervisory-control interfaces for humanoid robots to be developed furthering the state-of-the-art. However, most of the developed interfaces required multiple well-trained operators in order to successfully complete tasks (Norton et al., 2017). Several interfaces aimed to reduce the amount of input required on the human operator, but still allowed operators to take over low-level commands of the robot when needed and provided a wide range of robot sensor data (Stentz et al., 2015; DeDonato et al., 2017).
Our Valkyrie interface was designed to mimic several of the humanoid interfaces designed for the DRC finals, but with a reduced number of capabilities, inputs, and sensor data to allow for both a single operator and to reduce operator training time. The goal was to create a comparable interface to ones used at the DRC, but user-friendly enough to allow for non-experts to quickly learn and operate.
Our interface, as seen in Figure 12, utilizes RViz, a 3D visualization tool for ROS, and custom Qt Widgets, also known as Panels in RViz. RViz was chosen as the interface backbone since ROS is widely used in the robotics community making it familiar to many and for its built-in data visualization tools. RViz was also used for several of the interfaces designed for DRC (DeDonato et al., 2015; Kohlbrecher et al., 2015; Stentz et al., 2015). The interface can be broken down into four custom panels: Arm Motion Planner, Step Planner, Neck Interface, and Hand Interface. Additionally, there are three built-in RViz panels: the main visualization window, a display panel, and a camera feed panel.
The main window consists of the built in 3D data visualization of RViz and for our specific interface includes a point-cloud from Valkyrie’s Multisense SL sensor located in the head, a robot model of Valkyrie’s current joint state and planned joint state from the arm motion planner, and interactive markers to interface with the arm motion and step planners. The display panel allows users to select what is actively visualized in the main window and the camera feed panel for our interface is a camera feed from one of the stereo cameras in the Multisense SL.
6.2.1 Arm Motion Planner Panel
The Arm Motion Planner panel utilizes a whole body inverse kinematics solver that finds motion plans based on a set of cost and constraints (Long et al., 2016, 2017). Cost and constraints can be categorized as either kinematic, collision avoidance, or ZMP. To reduce the complexity of the interface, operators only need to define the desired end-effector positions and a predefined set of cost and constraints are used, i.e., balancing constraint or velocity cost, etc. To define the end-effector positions, operators can add waypoints for the end-effector to traverse though. This is accomplished through the use of interactive markers that represent a waypoint and are visually represented as a single end-effector. Operators are able to translate and rotate the waypoint by using arrows and circular scroll bars that surround the marker visual. The waypoints are ordered to allow for a variety of paths. A plan can be requested after the waypoints have been placed in a satisfactory manner. Operators are provided a planning indicator and are informed when the planning is complete and whether the planner was successful in finding a plan or unsuccessful. If a plan returns unsuccessful, then the returned plan either violated one of the default set of costs or it was unable to traverse through all the placed waypoints. Operators are notified of the reason the planned returned unsuccessful and informed on how to improve the waypoints to results in a successful plan. For example, a waypoint may be out of the robot’s workspace and need to be moved closer to the robot. After a plan is returned, the operator can visualize the robot’s movement through the entire plan before allowing the robot to execute it to ensure that the plan both provides the desired outcome and is collision free. Figure 12B demonstrates an example of placed waypoints and the final robot configuration of the returned plan. After visualizing the plan, operators can request the robot execute it.
6.2.2 Step Planner Panel
The Step Planner panel allows operators to designate a goal position for the robot to navigate to. It works in a similar strategy as the Arm Motion Planner panel, except rather than multiple waypoints, there is only a single waypoint that represents the final goal pose for feet with a visual of a pair of feet. Additionally, the interactive marker that represents the waypoint only allows for 2D interaction on a plane. After this single waypoint is placed, operators can request a plan which is generated using an A* search-based footstep planner. Same as the Arm Motion Planner panel, operators are provided a planning indicator and informed when the plan returns. As soon as a successful plan is found the planner will return. However, the planner will infinitely plan until a successful plan is found, therefore, there is a 10 s timeout in place to force a return and the operator is notified of this timeout and to try again. As soon as a successful plan is returned, colored interactive markers that represent each footstep in the returned plan are visualized, red for the right foot and blue for the left foot. These interactive markers can be adjusted by the operator by clicking on the footstep marker to activate the interactive marker controls and then adjusting the footstep in the same method as the waypoints, the interactive marker controls can also be switched off by clicking the footstep marker. Figure 12C shows an example of a placed goal marker, a returned footstep plan, and selected footstep in adjustment mode. When finished making adjustments, operators are then able to send the plan to the robot for execution.
6.2.3 Neck Interface and Hand Interface Panels
The Neck Interface and Hand Interface panels both work in the same manner with a series of joint sliders and buttons with some predefined joint positions. The Neck Interface panel allows operators to move the head of the robot right-left and up-down with sliders and a predefined neutral position, where the robot is looking straight forward. The Hand Interface panel has sliders for each individual finger to open-close along with two predefined open hand and closed fist grasp. Sliders were added over additional grasp types to allow the operator to position the hand at any intermediate pose if needed.
Humanoid robots are equipped with the necessary combination of location and manipulation capacities to fulfill a variety of roles in man-made hazardous environments. However, there are numerous challenges before this potential can be realized. In this paper, we presented our research to help further the development of utilizing humanoid robots in supervisory roles for extreme environments, with emphasis on operation in nuclear facilities.
We first examined ways to improve manipulation within constrained environments, with a focus on glovebox operation, by looking into three different areas. First, we aimed to better understand a robot’s feasible workspace in constrained environments by developing a method called the constrained manipulability polytope (CMP) that considers both the robot’s capabilities and the constraints imposed by an environment. The CMP methodology assists in both autonomous motion planning and human-in-the-loop teleoperation by providing feedback of the robot’s workspace to the operator and indeed by providing a representation of both robot joint and Cartesian space constraints as a bounded limits in tool motion. Second, we investigated using contact supports to increase stability and we presented our theoretical methodology to accomplish contact-implicit motion planning. The objective of this work was to increase the robot’s workspace by using support contacts to reach areas that are not within the robot’s workspace without support. Third, we considered using non-prehensile manipulation over dexterous manipulation and presented a methodology capable of pushing objects in a constrained environment. The purpose of this work was to add the capability of moving objects around that might not be within reach of the robot or easily manipulated with dexterous hands.
We also detailed methods to improve locomotion for humanoid robots by developing a technique to accomplish in-situ terrain classification and estimation that can be used to better inform a walking controller. Finally, we presented our generic humanoid robot interface that allows for robot operation from operators with limited training and/or knowledge of robotics. When operating in high-risk environments, such as nuclear facilities, it will be necessary to keep humans-in-the-loop especially individuals with expertise in the environment the robot is operating in. Future work should consider further developing all these areas and combining them into one robotic platform with a supervisory-control interface for a single operator. Human supervised robots will be a key role in the future operation, maintenance, and decommissioning of nuclear facilities.
Data Availability Statement
The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.
PL authored the Constrained Manipulation and Motion Planning section. AO authored the Increasing Stability Using Support Contacts section and Dynamic Non-Prehensile Manipulation section. MaW authored the in-Situ Terrain Classification and Estimation. MuW authored the Supervisory-Control Interface and compiled the manuscript. TP supervised the work.
This research is supported by the Department of Energy under Award Number DE-EM0004482, by the National Aeronautics and Space Administration under Grant No. NNX16AC48A issued through the Science and Technology Mission Directorate and by the National Science Foundation under Award No. 1451427, 1944453.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
2Information regarding the glovebox (https://www.belart.com/bel-art-h50026-0000-sidentry-glove-box-30-x-24-x-24.html): and the objects (http://www.melissaanddoug.com/deluxe-jumbo-cardboard-blocks---40-pieces/2784.html): are available online.
Anitescu, M., and Potra, F. A. (1997). Formulating dynamic multi-rigid-body contact problems with friction as solvable linear complementarity problems. Nonlinear Dynamics 14, 231–247. doi:10.1023/a:1008292328909
Azad, M., Babič, J., and Mistry, M. (2017). Dynamic manipulability of the center of mass: A tool to study, analyse and measure physical ability of robots. In IEEE Int. Conf. on Robotics and Automation. IEEE, 3484–3490.
Brandão, M., Shiguematsu, Y. M., Hashimoto, K., and Takanishi, A. (2016). Material recognition cnns and hierarchical planning for biped robot locomotion on slippery terrain. In 2016 IEEE-RAS 16th International Conference on Humanoid Robots (Humanoids). 81–88. doi:10.1109/HUMANOIDS.2016.7803258
Caurin, G., and Tschichold-Gurman, N. (1994). The development of a robot terrain interaction system for walking machines. Proceedings of the 1994 IEEE International Conference on Robotics and Automation 2, 1013–1018. doi:10.1109/ROBOT.1994.351224
DeDonato, M., Dimitrov, V., Du, R., Giovacchini, R., Knoedler, K., Long, X., et al. (2015). Human-in-the-loop control of a humanoid robot for disaster response: A report from the darpa robotics challenge trials. J. Field Robotics 32, 275–292. doi:10.1002/rob.21567
DeDonato, M., Polido, F., Knoedler, K., Babu, B. P. W., Banerjee, N., Bove, C. P., et al. (2017). Team wpi-cmu: Achieving reliable humanoid behavior in the darpa robotics challenge. Journal of Field Robotics. doi:10.1002/rob.21685
Ding, L., Gao, H., Deng, Z., Song, J., Liu, Y., Liu, G., et al. (2013). Foot-terrain interaction mechanics for legged robots: Modeling and experimental validation. The International Journal of Robotics Research 32, 1585–1606. doi:10.1177/0278364913498122
Fletcher*, R., and Leyffer,‡, S. (2004). Solving mathematical programs with complementarity constraints as nonlinear programs. Optimization Methods and Software 19, 15–40. doi:10.1080/10556780410001654241
Gabiccini, M., Artoni, A., Pannocchia, G., and Gillis, J. (2018). A computational framework for environment-aware robotic manipulation planning. In Robotics Research. Springer, 363–385. doi:10.1007/978-3-319-60916-4_21
Hashimoto, K., Kang, H.-j., Nakamura, M., Falotico, E., Lim, H.-o., Takanishi, A., et al. (2012). Realization of biped walking on soft ground with stabilization control based on gait analysis. In 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems. 2064–2069. doi:10.1109/IROS.2012.6385684
Henze, B., Dietrich, A., and Ott, C. (2016a). An approach to combine balancing with hierarchical whole-body control for legged humanoid robots. IEEE Robot. Autom. Lett. 1, 700–707. doi:10.1109/lra.2015.2512933
Henze, B., Roa, M. A., and Ott, C. (2016b). Passivity-based whole-body balancing for torque-controlled humanoid robots in multi-contact scenarios. The International Journal of Robotics Research 35, 1522–1543. doi:10.1177/0278364916653815
Kang, H.-j., Nishikawa, K., Falotico, E., Lim, H.-o., Takanishi, A.H., et al. (2012). Biped walking stabilization on soft ground based on gait analysis. In 2012 4th IEEE RAS EMBS International Conference on Biomedical Robotics and Biomechatronics. BioRob, 669–674. doi:10.1109/BioRob.2012.6290870
Khatib, O., and Chung, S.-Y. (2014). Suprapeds: Humanoid contact-supported locomotion for 3d unstructured environments. In Robotics and Automation (ICRA), 2014 IEEE International Conference on. IEEE, 2319–2325.
Kohlbrecher, S., Romay, A., Stumpf, A., Gupta, A., Von Stryk, O., Bacim, F., et al. (2015). Human-robot Teaming for Rescue Missions: Team ViGIR's Approach to the 2013 DARPA Robotics Challenge Trials. J. Field Robotics 32, 352–377. doi:10.1002/rob.21558
Komizunai, S., Konno, A., Abiko, S., and Uchiyama, M. (2010). Development of a static sinkage model for a biped robot on loose soil. In 2010 IEEE/SICE International Symposium on System Integration. 61–66. doi:10.1109/SII.2010.5708302
Long, P., Keleştemur, T., Önol, A. Ö., and Padir, T. (2019). optimization-based human-in-the-loop manipulation using joint space polytopes. In 2019 International Conference on Robotics and Automation (ICRA). IEEE, 204–210.
Long, X., Wonsick, M., Dimitrov, V., and Padır, T. (2016). Task-oriented planning algorithm for humanoid robots based on a foot repositionable inverse kinematics engine. In 2016 IEEE-RAS 16th International Conference on Humanoid Robots (Humanoids). IEEE, 1114–1120.
Long, X., Wonsick, M., Dimitrov, V., and Padır, T. (2017). IEEE), 4452–4459. Anytime multi-task motion planning for humanoid robots. 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
Marcucci, T., Gabiccini, M., and Artoni, A. (2017). A two-stage trajectory optimization strategy for articulated bodies with unscheduled contact sequences. IEEE Robot. Autom. Lett. 2, 104–111. doi:10.1109/lra.2016.2547024
Marturi, N., Rastegarpanah, A., Takahashi, C., Adjigble, M., Stolkin, R., Zurek, S., et al. (2016). Towards advanced robotic manipulation for nuclear decommissioning: A pilot study on tele-operation and autonomy. In 2016 International Conference on Robotics and Automation for Humanitarian Applications (RAHA). IEEE, 1–8.
Mastalli, C., Havoutis, I., Focchi, M., Caldwell, D. G., and Semini, C. (2016). Hierarchical planning of dynamic movements without scheduled contact sequences. IEEE International Conference on Robotics and Automation (ICRA) (IEEE), 4636–4641.
Mordatch, I., Lowrey, K., and Todorov, E. (2015). Ensemble-cio: Full-body dynamic motion planning that transfers to physical humanoids. In 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 5307–5314.
Mordatch, I., Popović, Z., and Todorov, E. (2012a). Contact-invariant optimization for hand manipulation. In Proceedings of the ACM SIGGRAPH/Eurographics Symposium on Computer Animation. Eurographics Association), 137–144.
Nagatani, K., Kiribayashi, S., Okada, Y., Otake, K., Yoshida, K., Tadokoro, S., et al. (2011a). Gamma-ray irradiation test of electric components of rescue mobile robot quince. In 2011 IEEE International Symposium on Safety, Security, and Rescue Robotics. IEEE, 56–60.
Nagatani, K., Kiribayashi, S., Okada, Y., Tadokoro, S., Nishimura, T., Yoshida, T., et al. (2011b). Redesign of rescue mobile robot quince. In 2011 IEEE International Symposium on Safety, Security, and Rescue Robotics. IEEE, 13–18.
Neunert, M., Farshidian, F., and Buchli, J. (2016). Efficient whole-body trajectory optimization using contact constraint relaxation. In 2016 IEEE-RAS 16th International Conference on Humanoid Robots (Humanoids). IEEE, 43–48.
Neunert, M., Farshidian, F., Winkler, A. W., and Buchli, J. (2017). Trajectory optimization through contacts and automatic gait discovery for quadrupeds. IEEE Robot. Autom. Lett. 2, 1502–1509. doi:10.1109/lra.2017.2665685
Neunert, M., Stäuble, M., Giftthaler, M., Bellicoso, C. D., Carius, J., Gehring, C., et al. (2018). Whole-body nonlinear model predictive control through contacts for quadrupeds. IEEE Robot. Autom. Lett. 3, 1458–1465. doi:10.1109/lra.2018.2800124
Norton, A., Ober, W., Baraniecki, L., McCann, E., Scholtz, J., Shane, D., et al. (2017). Analysis of human-robot interaction at the DARPA Robotics Challenge Finals. The International Journal of Robotics Research 36, 483–513. doi:10.1177/0278364916688254
Önol, A. Ö., Long, P., and Padır, T. (2018). A comparative analysis of contact models in trajectory optimization for manipulation. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE.
Önol, A. Ö., Long, P., and Padır, T. (2019). Contact-implicit trajectory optimization based on a variable smooth contact model and successive convexification. In 2019 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2447–2453.
Otte, S., Weiss, C., Scherer, T., and Zell, A. (2016). Recurrent neural networks for fast and robust vibration-based ground classification on mobile robots. IEEE International Conference on Robotics and Automation (ICRA), 5603–5608. doi:10.1109/ICRA.2016.7487778
Posa, M., Cantu, C., and Tedrake, R. (2014). A direct method for trajectory optimization of rigid bodies through contact. The International Journal of Robotics Research 33, 69–81. doi:10.1177/0278364913506757
Radford, N. A., Strawser, P., Hambuchen, K., Mehling, J. S., Verdeyen, W. K., Donnan, A. S., et al. (2015). Valkyrie: NASA's First Bipedal Humanoid Robot. J. Field Robotics 32, 397–419. doi:10.1002/rob.21560
Ragaglia, M., Bascetta, L., Rocco, P., and Zanchettin, A. M. (2014). Integration of perception, control and injury knowledge for safe human-robot interaction. In IEEE Int. Conf. on Robotics and Automation. IEEE, 1196–1202.
Rasheed, T., Long, P., Marquez-Gamez, D., and Caro, S. (2018). Tension distribution algorithm for planar mobile cable-driven parallel robots. In Cable-Driven Parallel Robots. Springer, 268–279. doi:10.1007/978-3-319-61431-1_23
Skonieczny, K., Moreland, S. J., Asnani, V. M., Creager, C. M., Inotsume, H., and Wettergreen, D. S. (2014). Visualizing and analyzing machine-soil interactions using computer vision. J. Field Robotics 31, 820–836. doi:10.1002/rob.21510
Sleiman, J.-P., Carius, J., Grandia, R., Wermelinger, M., and Hutter, M. (2019). Contact-implicit trajectory optimization for dynamic object manipulation. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE.
Stewart, D. E., and Trinkle, J. C. (1996). An implicit time-stepping scheme for rigid body dynamics with inelastic collisions and coulomb friction. Int. J. Numer. Meth. Engng. 39, 2673–2691. doi:10.1002/(sici)1097-0207(19960815)39:15<2673::aid-nme972>3.0.co;2-i
Tassa, Y., Erez, T., and Todorov, E. (2012). and stabilization of complex behaviors through online trajectory optimization. In 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE), 4906–4913.
Walas, K., Kanoulas, D., and Kryczka, P. (2016). Terrain classification and locomotion parameters adaptation for humanoid robots using force/torque sensing. In 2016 IEEE-RAS 16th International Conference on Humanoid Robots (Humanoids). 133–140. doi:10.1109/HUMANOIDS.2016.7803265
Wang, J., and Olson, E. (2016). AprilTag 2: Efficient and robust fiducial detection. In 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 4193–4198. doi:10.1109/IROS.2016.7759617
Weiss, C., Tamimi, H., and Zell, A. (2008). A combination of vision- and vibration-based terrain classification. In 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems. 2204–2209. doi:10.1109/IROS.2008.4650678
Will Bosworth1, S. K., Jonas, W., and Hogan, N. (2016). Robot locomotion on hard and soft ground: measuring stability and ground properties in-situ. In IEEE International Conference on Robotics and Automation (ICRA). 3582–3589.
Winkler, A. W., Bellicoso, C. D., Hutter, M., and Buchli, J. (2018). Gait and trajectory optimization for legged systems through phase-based end-effector parameterization. IEEE Robot. Autom. Lett. 3, 1560–1567. doi:10.1109/lra.2018.2798285
Wittmann, R., Hildebrandt, A.-C., Wahrmann, D., Sygulla, F., Rixen, D., and Buschmann, T. (2016). Model-based predictive bipedal walking stabilization. In 2016 IEEE-RAS 16th International Conference on Humanoid Robots (Humanoids). IEEE, 718–724.
Yanco, H. A., Norton, A., Ober, W., Shane, D., Skinner, A., and Vice, J. (2015). Analysis of human-robot interaction at the darpa robotics challenge trials. J. Field Robotics 32, 420–444. doi:10.1002/rob.21568
Yirmibeşoğlu, OD, Oshiro, T, Olson, G, Palmer, C, and Mengüç, Y (2019). Evaluation of 3d printed soft robots in radiation environments and comparison with molded counterparts. Front. Robot. AI 6, 40. doi:10.3389/frobt.2019.00040
Zollo, L., Accoto, D., Torchiani, F., Formica, D., and Guglielmelli, E. (2008). Design of a planar robotic machine for neuro-rehabilitation. In IEEE Int. Conf. on Robotics and Automation. IEEE, 2031–2036.
Keywords: humanoid robots, motion planning, supervised autonomy, nuclear, glovebox
Citation: Wonsick M, Long P, Önol AÖ, Wang M and Padır T (2021) A Holistic Approach to Human-Supervised Humanoid Robot Operations in Extreme Environments. Front. Robot. AI 8:550644. doi: 10.3389/frobt.2021.550644
Received: 09 April 2020; Accepted: 10 May 2021;
Published: 18 June 2021.
Edited by:Chie Takahashi, University of Cambridge, United Kingdom
Reviewed by:Manuel Ferre, Polytechnic University of Madrid, Spain
William R. Hamel, The University of Tennessee, United States
Copyright © 2021 Wonsick, Long, Önol, Wang and Padır. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Taşkın Padır, email@example.com