Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Neurorobot., 02 February 2026

Volume 19 - 2025 | https://doi.org/10.3389/fnbot.2025.1697518

Transformer-based human-motion forecasting coupled with safe reinforcement learning for telepresence robot co-navigation

  • 1Department of Electrical Engineering, College of Engineering, Princess Nourah Bint Abdulrahman University, Riyadh, Saudi Arabia
  • 2Department of Electrical Engineering, Government College University Lahore, Lahore, Pakistan
  • 3Department of Computer Science and Software Engineering, Beaconhouse International College, Faisalabad, Pakistan
  • 4School of Computer Science and Mathematics, Faculty of Engineering, Liverpool John Moores University, Liverpool, United Kingdom
  • 5Department of Computer Software Engineering, Sir Syed University of Engineering and Technology, Karachi, Pakistan
  • 6Department of Engineering, Brock University, St. Catharines, ON, Canada

Introduction: Telepresence robots (TPRs) must co-navigate with humans in constrained hospital environments, where safety depends on anticipating rather than merely reacting to human motion. Existing approaches rarely integrate short-horizon human-motion forecasting with safety-constrained control, which reduces robustness in dense corridors and ward bays. This study addresses this gap by evaluating an anticipatory, safety-aware co-navigation framework for TPRs.

Methods: We developed a modular framework that couples a lightweight transformer-based forecaster that predicts multi-agent trajectories under occlusion with a safe reinforcement learning (RL) controller. The forecaster produces short-term distributions over pedestrian states that are embedded into the RL policy state and cost as risk-aware occupancy features. Safety is enforced via constrained policy optimization augmented by a run-time control barrier function (CBF) shield that filters unsafe actions. We benchmarked the approach against a social-force or dynamic window approach (DWA), an attention-based crowd-RL policy, and model predictive control (MPC) with CBF. Experiments were conducted across two hospital-like benchmarks (a crowded corridor and a four-bed ward), totaling 2,400 episodes. Outcomes included task success, collision count, minimum human–robot clearance, near-miss events ( ≤ 0.3 m), time-to-goal, CBF violations, and ablations removing forecasting and the CBF shield.

Results: Relative to the best-performing baseline, the proposed method improved task success by 21.6% and reduced collisions by 47.3%. Median minimum human–robot clearance increased by 0.19 m, and near-miss events decreased by 38.5%. Time-to-goal was maintained within +2.7% of MPC+CBF while incurring zero CBF violations under the shield. Ablation studies showed that removing forecasting degraded success by 14.2%, whereas removing the CBF shield increased constraint breaches from 0% to 6.1% of steps.

Discussion: Anticipatory perception combined with Safe-RL yields substantially safer and more reliable telepresence co-navigation in human-dense clinical layouts without sacrificing efficiency. The framework is modular, enabling alternative forecasters and safety shields. Limitations include sensitivity to forecast drift during abrupt changes in crowd flow. Future work will explore on-device adaptation, shared-autonomy overlays to incorporate operator intent, and prospective evaluations in live hospital workflows.

1 Introduction

Telepresence robots (TPRs) are increasingly deployed to extend clinical reach and sustain social connections in hospitals and long-term care (LTC) settings. However, navigation in people-dense wards and narrow corridors remains a critical barrier to reliable uptake. Recent clinical and translational studies show that TPRs can reduce caregiver burden and resident loneliness and help maintain care continuity, underscoring their potential value in real services (e.g., mixed-methods and pre–post trials in LTC) (Hung et al., 2025; Hu et al., 2025). At the same time, operational trials in healthcare highlight the practical challenges of moving a TPR safely through dynamic, cluttered clinical spaces, where visibility is partial and human motion is heterogeneous (Leoste et al., 2024). These reports converge on a need for navigation that is not merely reactive but anticipatory and safety-constrained in real time.

A large body of work in socially aware robot navigation formalizes human–robot co-navigation norms, maintaining comfortable distances, respecting implicit right-of-way, and negotiating bottlenecks. Most algorithms still plan with short-horizon, history-based predictors or handcrafted interaction models. Recent surveys in leading robotics venues document the field's progress and open challenges, including reliable forecasting under occlusion and principled safety handling in crowds (Mavrogiannis et al., 2023; Singamaneni et al., 2024). These reviews specifically call for tighter coupling between the perception of future human motion and the decision layers that guarantee safety, especially in constrained indoor spaces such as hospital wards.

Concurrently, human-motion forecasting has advanced with transformer architectures that model long-range temporal dependencies and multi-joint correlations. Journal reports demonstrate that attention-based predictors can deliver accurate, real-time short- and mid-horizon motion predictions suitable for on-robot deployment (Yunus et al., 2023) and for collaborative tasks that benefit from anticipating human intent (Laplaza et al., 2024). However, these perception modules are rarely integrated into closed-loop navigation with formal safety guarantees; in healthcare corridors, even small forecast errors or sudden flow changes can precipitate unsafe proximity, unless the control layer is explicitly safety-aware. These properties align well with hospital and LTC ward navigation, where flows of staff and residents follow relatively structured routines (e.g., rounds, mealtimes, therapy sessions), motion is often slower and assisted (walkers, wheelchairs), and occlusions from curtains, furniture, and equipment are frequent. In such settings, an attention-based forecaster that jointly reasons over multiple agents and local geometry can better anticipate near-future occupancy around a telepresence robot, enabling it to preserve comfortable clearances for vulnerable residents and staff while respecting narrow corridors and multi-bed bays. Empirical studies already demonstrate that mobile telepresence robots are increasingly used in LTC and healthcare facilities. Still, navigation and proxemics remain practical barriers to routine deployment, which further motivates an anticipatory, transformer-based forecaster in this context.

Safety-critical control for robots increasingly leverages safety filters, methods that project candidate actions onto sets known to satisfy constraints at run time. Control barrier functions (CBFs) provide a principled framework to maintain forward invariance of a “safe set,” and recent journal-level tutorials and surveys consolidate their theory and practice for autonomous systems (Hsu et al., 2024; Li et al., 2023). These works emphasize that safety filtering can complement high-performance planners or learned policies, but also warn about conservatism and feasibility under sensing uncertainty—precisely the regime faced by a TPR moving among people with intermittent occlusions.

In parallel, safe reinforcement learning (Safe-RL) refers to reinforcement-learning methods that explicitly encode safety through constraints or risk measures in the objective; this area has matured from conceptual proposals to consolidated frameworks with policy-level constraints and risk-aware objectives. A review synthesizes methods and theory (e.g., constrained Markov decision processes (CMDPs), Lagrangian approaches, safety layers) and catalogs open issues in deploying Safe-RL in real-world robotics (Gu et al., 2024). Complementary work surveys verification and assurance for deep RL policies, offering analysis tools that can be combined with online safety filters (Landers and Doryab, 2023). Despite this progress, the literature still lacks demonstrations that explicitly fuse learned human-motion forecasts with Safe-RL policies and run-time CBF safety shields for TPR co-navigation in healthcare.

This study proposes an integrated framework (Figure 1) that (i) learns anticipatory, transformer-based pedestrian-trajectory distributions conditioned on local map structure and occlusion, (ii) injects these distributions into a Safe-RL policy via risk-aware occupancy features, and (iii) enforces hard safety at execution through a CBF-based quadratic-program (QP) “shield” that minimally modifies the policy's action only when necessary. The framework is evaluated in two hospital-like benchmarks—a crowded corridor and a four-bed ward—and is compared against strong baselines spanning social-force/DWA planning, attention-based crowd-RL, and MPC+CBF. The study's contributions are threefold:

1. A problem formulation and modular forecast → policy → safety-filter architecture for telepresence robot co-navigation in hospital- and LTC-style wards, which couples transformer-based human-motion forecasting, Safe-RL, and a discrete-time CBF safety shield under occlusion and latency.

2. A forecast-aware Safe-RL and CBF design that converts short-horizon multi-agent distributions into risk features and chance-robust clearance constraints, using CMDP with Conditional Value-at-Risk (CVaR)-style costs and dual-updated safety budgets.

3. A simulation-based evaluation protocol and dataset for telepresence co-navigation in two hospital-like benchmarks (corridor and four-bed ward), including comparisons to strong social-navigation baselines (ORCA, DWA, PPO, MPC+CBF), ablations, and latency/crowding-sensitivity analyses.

Figure 1
A coordinate grid with x and y axes labeled in meters. A small orange square marks “Start” at position (1, 2.5), while a blue star marks “Goal” at position (7, 2.5). Four blue rectangles are positioned near the corners of the grid.

Figure 1. System architecture of the proposed telepresence co-navigation stack. Sensor and map inputs (RGB-D/LiDAR, odometry, signed-distance field) feed a transformer-based human-motion forecaster; its short-horizon multi-agent distributions are converted into occupancy and risk features and combined with robot and latency states to form the Safe-RL policy input; a CBF-based QP shield then minimally adjusts the policy action before low-level control of the telepresence robot.

Telepresence robots in hospitals and LTC wards require anticipation under partial observability while operating under tight real-time constraints on embedded compute. We therefore adopt a compact transformer forecaster that can model multi-agent interactions and map-conditioned motion (e.g., doors, curtains, bottlenecks) while maintaining low inference latency at the control frequency. This design targets common LTC traffic patterns—slow, assisted motion and frequent occlusions—so that predictive uncertainty can be used downstream for safety-constrained decision-making.

We do not claim that this architecture addresses all sub-domains of human–robot interaction; our scope is explicitly mobile telepresence navigation in hospital and LTC wards, and other HRI settings (e.g., manipulation-centric or non-navigational interactions) lie outside the remit of this study. By situating the study within documented healthcare use-cases and constraints and by grounding the proposed methods in recent advances in social navigation, motion forecasting, safety filtering, and Safe-RL, the study aims to provide both a practically relevant and methodologically rigorous step toward safer, more reliable telepresence co-navigation in clinical environments.

The remainder of this manuscript is organized as follows. Section 2 reviews related work on human-motion forecasting, shared autonomy, Safe-RL, and control barrier functions; Section 3 formalizes the co-navigation problem, dynamics, and safety constraints; Section 4 details the proposed method—transformer-based forecasting, risk-aware RL, the CBF safety shield, and the end-to-end control cycle. Section 5 describes the experimental setup (ward layout, datasets/simulation, baselines, metrics, and protocols); Section 6 reports results with ablations and sensitivity analyses; Section 7 discusses implications, limitations, and generalizability; Section 8 concludes and outlines future work.

2 Literature review

Evidence from LTC, hospital, and home-care contexts indicates that mobile telepresence can improve communication, accelerate specialist access, and support person-centered care while raising persistent concerns about workflow integration, privacy, and acceptance by staff and families (Ren et al., 2024; Wang et al., 2022; Teng et al., 2022). Reviews and qualitative studies highlight gaps in robust navigation and autonomy in busy wards and corridors, limited support for co-navigation with staff and visitors, and the need for safety-assured autonomy under uncertain human motion. These limitations motivate technical advances in human-aware forecasting and safety-critical shared autonomy tailored to clinical layouts.

Over the past five years, surveys have consolidated requirements for social navigation—physical and perceived safety, legibility, naturalness, and compliance with social norms—while documenting the lack of standardized evaluation and hospital-specific benchmarks (Gao and Huang, 2022; Karwowski et al., 2024). At the planner level, widely used baselines include the dynamic window approach (DWA) for reactive local collision avoidance (Fox et al., 1997) and Optimal Reciprocal Collision Avoidance (ORCA) for multi-agent crowd navigation (van den Berg et al., 2011), often combined with social-force or potential-field heuristics in hospital-like simulations (Berg et al., 2010). Empirical proxemics studies quantify comfortable passing distances and personal-space envelopes, which expand with robot speed and scenario (Neggers et al., 2021, 2022). Algorithmic works increasingly embed proxemics into planners, but generalization across densities and group structures remains fragile (Lu et al., 2022; Zhang and Feng, 2023). Collectively, these findings argue for navigation stacks that explicitly model human motion and uncertainty and expose safety-aware blending with human operators.

Transformer variants now dominate pedestrian and agent prediction in traffic and crowd settings, offering non-autoregressive decoding, social-graph attention, and multi-modal prediction (Chen et al., 2023; Shoman et al., 2024; He et al., 2023; Yao et al., 2022; Sun et al., 2023; Liu et al., 2023). Nonetheless, surveys and benchmarks highlight sensitivity to domain shift (site-specific behaviors), long-horizon degradation, and uncertainty calibration—issues that are acute in hospitals where flows are episodic (e.g., shift changes) and space is constrained (Rudenko et al., 2020). This motivates coupling learned forecasting with an online safety layer that can accommodate distributional shifts while preserving social comfort.

Recent tutorials and surveys codify CBFs as forward-invariance constraints that render safety sets robustly invariant when combined with control Lyapunov terms or model predictive control (MPC) (Garg et al., 2024). Emerging variants integrate CBFs with MPC for dynamic obstacle avoidance and feasibility recovery, and begin to address real-time operation on mobile robots (Desai and Ghaffari, 2022; Li et al., 2022; Shayan et al., 2025). These methods offer crisp safety certificates but require reliable state and obstacle estimates and can be conservative without learned intent or forecasting.

Closest to our setting, Samavi et al. introduce SICNav-Diffusion, which combines diffusion-based joint human-trajectory prediction with a bilevel MPC formulation that refines both robot plans and human predictions for safe crowd navigation. In parallel, Mohamed, Ali, and Liu propose a chance-constrained sampling-based MPC (C2U-MPPI) that leverages unscented sampling and probabilistic chance constraints to achieve robust collision avoidance in uncertain dynamic environments (Mohamed et al., 2025; Samavi et al., 2025). These works tightly couple probabilistic prediction or uncertainty-aware constraints with MPC-style controllers, but they do not employ Safe-RL or CBF shields, nor do they target telepresence co-navigation in hospital wards; in contrast, our framework uses transformer-based forecasts as belief features within a CMDP-based Safe-RL policy, wrapped by a modular CBF safety filter and evaluated in clinically motivated layouts.

Surveys across robotics and autonomy trace a trend toward constrained MDPs, shielded and predictive safety filters, and risk-sensitive objectives to maintain constraints during learning and deployment (Brunke et al., 2022; Karwowski et al., 2024; Brescia et al., 2023). Despite progress, reviews identify gaps in on-policy safety during exploration, tight guarantees under perception and model uncertainty, and sample efficiency in the presence of rare but critical events—conditions common in hospital corridors. Combining Safe-RL with CBF- or MPC-based certificates is repeatedly recommended to achieve both adaptability and formal safety.

Journal studies in teleoperation and assistive settings converge on adaptive authority allocation via intent prediction (gaze, EMG, motion cues), with user-study evidence that transparency and assistance timing affect agreement and satisfaction (Fuchs and Belardinelli, 2021; Gottardi et al., 2022; Backman et al., 2023; Bowman et al., 2024; Yousefi et al., 2025). However, most systems assume low-dynamic scenes and do not fuse crowd motion forecasts with certified safety layers when blending user and autonomy inputs—a key limitation for telepresence co-navigation in wards and corridors.

The literature offers (i) clinically motivated requirements and adoption barriers, (ii) strong but environment-sensitive transformer forecasters, (iii) formal safety layers via CBF/MPC, (iv) Safe-RL for adaptive policies, and (v) shared-autonomy mechanisms for authority blending. However, a unified pipeline that couples transformer-based human-motion forecasting with Safe-RL policies under an online CBF/MPC safety filter for telepresence co-navigation in hospitals—and evaluates performance using social-comfort and safety metrics from social-navigation standards—has received limited attention. The proposed study targets this integration and clinical evaluation gap.

3 Problem formulation

The study formalized the telepresence co-navigation in clinical layouts as a constrained, risk-sensitive decision process with stochastic human dynamics, partial observability, and a run-time safety filter.

3.1 Notation and symbols

This section proceeds from a formal problem definition (POMDP and CMDP) to the concrete quantities used by the learning and control modules. Each subsection introduces symbols locally and connects them explicitly to the algorithmic components described later in Section 4. Table 1 summarizes all symbols used throughout Sections 3 and 4. To avoid ambiguity, each symbol is used with a single, consistent meaning across the study. For clarity, we use xt exclusively for the robot state and Yt for the joint human state throughout Sections 3 and 4.

Table 1
www.frontiersin.org

Table 1. Notation and symbols.

3.2 Environment, agents, and map geometry

Let the hospital layout be a compact set M2 with free space F=M\O, where O is the union of static obstacles (walls, beds, trolleys). The signed-distance field (SDF) dO:F>0 gives the Euclidean distance to O

The TPR is a differential-drive platform with unicycle dynamics. Pedestrians are modeled as disk agents.

• Robot state xt=[pt,θt, vt]4with position pt=[xt, Yt], yaw θt and linear speed vt.

• Control ut = [at, ωt], longitudinal acceleration at and yaw rate ωt.

• Continuous-time dynamics:

x˙=f(x,u)=[vcosθvsinθωa],x(0)=x0    (1)

• Discrete-time dynamics under zero-order hold (ZOH) with sampling step Δt:

xt+1=xt+Δt f(xt,ut)+wtW~    (2)

ZOH assumes the control input remains constant over each sampling interval Δt, matching the low-level telepresence controller that updates commands at 30–50 Hz.

There are Nt pedestrians at time t, indexed by i∈ {1, …, Nt} with positions YtiF. The goal region is = {p−||ppgoal|| ≤ rgoal }.

Goal region (functional form): The goal set is a closed Euclidean ball centered at pgoal,

G {pF:p-pgoalrgoal}.

Properties: G is compact, convex, and closed; its indicator I{pG} is discontinuous at the boundary, but the distance-to-goal term used in the stage cost is smooth everywhere except at the goal center.

We define the static safety set as follows:

Sstat={x4:dO(p)Rwall}    (3)

which consists of all robot states whose position p maintains at least a radius Rwall from static obstacles encoded by the SDF dO(·), with comfort radius Rwall> 0.

3.3 Observation, partial observability, and belief

Sensors (LiDAR/camera) provide detections zt within a visibility region VtF (accounting for occlusions by walls and equipment). The co-navigation problem is modeled as a partially observable Markov decision process (POMDP) (Kaelbling et al., 1998), with belief bt over the joint human state Yt{Yti}i=1Nt, where Yti denotes the state (e.g., 2D position, optionally velocity) of pedestrian i at time t.

Let zt denote the set of sensor detections at time t (e.g., tracked 2D positions from LiDAR/RGB-D) observed within the visibility region VtF. The human-motion model is τ(Yt+1Yt), i.e., the transition density of the joint human state. The observation likelihood is O(ztYt,Vt), which accounts for occlusions and partial observability through Vt. We use bt(Yt) to denote the belief distribution over joint human states conditioned on the observation history.

• Observation model: ztO~(·Yt,Vt), where Vt encodes occlusion-aware visibility in the ward layout.

Observation likelihood (functional form): We assume a conditionally independent detection model over visible pedestrians. Let zt={zti}iItvis denote the set of detected 2D pedestrian positions at time t, where Itvis indexes agents currently visible inside Vt. For each visible agent i, we assume

zti=yti+νti,νtiN~(0,Σz),

and missing detections for occluded agents iItvis are handled via a binary visibility mask induced by Vt. The resulting likelihood is

O(ztYt,Vt)=iItvisN(zti;yti,Σz).

Properties: For visible agents, O is smooth (infinitely differentiable) in yti and bounded for any fixed Σz≻0; occlusions enter only through the mask Itvis, yielding a piecewise-smooth likelihood as the visibility set changes.

• Belief evolution over joint human states follows the belief update (Bayes filter) recursion:

bt+1(Yt+1)τ(Yt+1|Yt)bt(Yt)dYtpredict·O(zt+1|Yt+1,Vt+1)correct    (4)

Here, the prediction term propagates the prior belief through the motion model τ, and the correction term incorporates the latest observation zt+1 under the visibility constraints Vt+ 1.

In practice, we do not maintain an explicit grid-based belief over joint human states; instead, the transformer forecaster (Section 4.1) implements the predictive step of this Bayes filter by mapping tracked detections and occlusion-aware map context into short-horizon distributions over pedestrian motion. These distributions serve as a compact belief summary fed to the Safe-RL controller and CBF shield.

To enable tractable control, the proposed formulation embeds a forecast-based summary of bt into an augmented MDP state st.

3.4 Objective, constraints, and task success

Define a finite horizon T. Let c(xt, ut; Yt) be an instantaneous task cost and gj(xt, ut; Yt) ≤ 0 constraint functions (safety/comfort), j = 1, …, J.

• Progress: cprog(xt) = ||ptpgoal ||2.

• Smoothness: csm(ut)=λvat2+ λωωt2.

• To model social comfort, we define the cost:

csoc(xt;Yt)=i=1Ntϕ(||pt-Yti||2)    (5)

which penalizes proximity between the robot position pt and nearby humans Yti, thereby encouraging socially compliant navigation behavior. We use ϕ(r)=max(0,Rcomfort-r )2.

Social-comfort shaping (functional form and properties): The hinge-quadratic form ϕ(r)=max(0,Rcomfort-r)2 penalizes only when the robot enters a comfort radius Rcomfort while assigning zero cost outside this zone. This ϕ(·) is non-negative, continuous, and piecewise-smooth (differentiable everywhere except at r = Rcomfort); it is convex for rRcomfort and has a bounded gradient for any bounded workspace and bounded robot–human distances.

The total per-step cost is defined as

c(xt,ut;Yt)=αcostcprog+βcostcsm+δcostcsoc.    (6)

This combines progress toward the goal, motion smoothness, and social comfort through weighted terms αcost, βcost, and δcost, where αcost, βcost, δcost≥0 weight progress-to-goal, smoothness, and social-comfort costs, respectively (distinct from the CVaR confidence parameter α in Equation 9).

Dynamic human–robot safety constraint. We model each pedestrian i as a disk of radius rhum and the robot footprint as a disk of radius rrob. Let

rsafeRsafe+rrob+rhum+rbuf,

where Rsafe is the nominal interpersonal comfort margin and rbuf is an additional robustness buffer. The hard safety constraint with respect to pedestrian i is then

gi(xt;Yt)rsafe2-pt-Yti20,i=1,,Nt    (7)

This definition makes gi(·) a signed safety margin: it is non-positive when the robot remains outside the safe disk around each pedestrian.

Static wall safety is enforced through the constraint:

gwall(xt):=Rwall-dO(pt)0.    (8)

This ensures that the robot remains outside a forbidden margin around walls and other static obstacles at all times.

Goal: success if pT G; failure on any gj>0 or timeout.

Let the cumulative task cost be Z=t=0T-1c(xt,ut;Yt). For α∈(0, 1), the CVaR objective is

minπ CVaRα(z)=minπ,ηη+11-αE[(Z-η)+],    (9)

subject to chance-constrained safety (Section 3.4) and integrator dynamics, where (.)+ = max(., 0 ).

In Equation 9, η∈ℝ is an auxiliary decision variable corresponding to the Value-at-Risk (VaR) level for the cumulative cost Z; the term E[(Z−η)+]/(1−α) then yields the standard CVaR representation as the expected excess cost in the worst (1−α) tail.

The constrained MDP (CMDP) form is

minπ𝔼[t=0T-1c(.)]s.t. [t=0T-1I{gj(.)>0}]κj,j.    (10)

Here I{·} denotes the indicator function, i.e., I{A} = 1 if the predicate A is true and 0 otherwise. The constants κj≥0 are per-episode safety budgets that upper bound the expected number of timesteps in which constraint gj(·) is violated (e.g., wall or human-safety violations), thereby defining the feasible policy set in the CMDP. A Lagrangian relaxation yields multipliers λj≥0 and the penalized objective t[c+jλjI{gj>0 }].

Notation and risk functional. Let ρα(·) denote the CVaR at confidence level α∈(0, 1), defined as the expected cost in the worst 1−α tail of the distribution, so that ρα(J) is the tail expectation of the episodic cost J over trajectories whose cost lies in the worst (1−α) quantile of the distribution. To avoid symbol overloading, we distinguish between the CMDP discount factor γdisc∈(0, 1), used in the value function and policy optimization, and the discrete-time CBF decay parameter γcbf∈(0, 1), which governs how fast the safety function is allowed to decrease along closed-loop trajectories. The dual variables λk associated with each safety constraint are updated by projected gradient ascent and clipped to [0, λmax], where λmax>0 bounds the influence of constraint costs. The mapping Φ(·) denotes the feature embedding that converts raw geometric, forecast-derived, and latency features into the fixed-dimensional augmented state st observed by the Safe-RL policy.

3.5 Forecast-driven risk features and chance constraints

A transformer-based predictor provides multi-step, multi-modal distributions for each pedestrian:

Yt+τi~pt+τi(.|Ht), τ=1:H,    (11)

where Ht collects recent trajectories, map context, and occlusion masks. For tractability, we assume Gaussian mixture models (GMMs) or samples.

3.5.1 Occupancy risk field

Define a continuous occupancy intensity for horizon τ:

Φt(z,τ)=i=1Ntκ=1Kiπiκ.N(z;μiκ(τ),(τ)iκ),    (12)

with mixture weights πik. The following risk features are fed to the policy:

ϱt=[maxτHΦt(pt,τ)local occupancy,maxτHΦt(z,τ)dzB(pt,ρ)near-field mass,           minτHminiE||ptYt+τi||clearance forecast]    (13)

3.5.2 Chance-constrained safety with respect to predicted humans

For each i, τ impose

(||pt-Yt+τi||2Rsafe)1-ϵ.    (14)

If Yt+τi ~N(μ,Σ), a conservative Gaussian chance-constraint via the one-sided Chebyshev/ellipsoidal bound gives

hi,τ(pt):=||pt-μ||2mean clearance- κ1-ϵλmax(Σ)uncertainty margin-Rsafe0,    (15)

with κ1-ϵ=Φ-1(1-ϵ). Equivalently, in squared form for CBF design:

h~i,τ(pt):=(||pt-μ||2-κ1-ϵλmax(Σ))2-Rsafe20.    (16)

The dynamic safety set becomes

Sdyn(t)=i,τ{x:h~i,τ(pt)0}.    (17)

The overall safe set is S(t)=SstatSdyn(t ).

3.6 Discrete-time CBF constraints

For each safety function h(x, t) (static walls; predicted humans), discrete-time forward invariance is enforced by the inequality

h(xt+1,t+Δt)-(1-γ)h(x,t0,γ(0.1) ]    (18)

which guarantees h≥0⇒h remains non-decreasing up to decay γ.

Using first-order dynamics linearization at (xt, ut):

h(xt+1)h(xt)+xh(xt)(xt+1-xt)               =h(xt)+txh(xt)f(xt,ut).    (19)

Thus, a linear constraint in ut:

-Δtxh(xt)fu(xt)utAh(xt)h(xt)-Δtxh(xt)f(xt,0)-(1-γ)h(xt)bh(xt)    (20)

• Static wall CBF: hwall(x)=dO(p)-Rwall,

xh=[pdO,0,0].    (21)

• Pedestrian CBF (chance-robust): use h~i,τ(x) with ph~=2(||p-μ||2-δ)(p-μ)||p-μ||2, where δ=κ1-ϵλmax(Σ ).

Collecting all active CBFs yields

Atutbt, At=[Ah1Ahm],bt=[bh1bhm].    (22)

3.7 Latency-aware state prediction

Let Δsens be sensing-to-actuation latency and Δnet a teleoperation network delay. The control acts on xt, Δ = Δsensnet. A predictive state is used:

x^t+Δ=xt+Δf(xt,ut-1), p^t+Δ=ptΔ+vt[cosθt,sinθt].    (23)

All CBF and chance constraints are evaluated at x^t+Δ to pre-empt delay effects.

3.8 Shielded action via quadratic program

Given a nominal policy action from the Safe-RL controller, the shield solves

ut= argminuε2||u-utnom||22    (24)
At(x^t+Δ)ubt(x^t+Δ),    (25)
uminuumax.    (26)

The solution minimally perturbs the nominal action while guaranteeing discrete-time CBF invariance under uncertainty margins. With m active constraints and a 2D control, the computational complexity is O(m) for active-set QPs.

3.9 Augmented MDP state for learning

The Safe-RL policy observes an augmented state.

st=[xt,pgoal-pt, Ψ(pt)map SDF features, ϱtforecast risk, ξtlatency features],    (27)

where ψ(pt)=[dO(pt),pdO(pt)]and ξt = [Δsens, Δnet], and the action space is U=u:uminuumax.

3.10 Assumptions and feasibility

1. The map SDF dO is Lipschitz and differentiable almost everywhere; ||pdO|| 1.

2. Forecast covariance ∑ admits λmax(∑) and is bounded on [t, t+H ].

3. Control bounds ensure QP feasibility under mild backup policies; if infeasible, a fallback braking u = [amin, 0] is admissible and yields a non-decreasing function for static walls.

A transformer supplies predictive distributions; risk features and chance constraints encode uncertainty; a Safe-RL policy proposes actions; and a discrete-time CBF-QP shield enforces invariance under latency and occlusion. This mathematical scaffold supports the subsequent algorithmic and experimental components.

4 Methodology

The proposed study describes a detailed end-to-end forecast → policy → safety-filter framework that enables anticipatory, safety-assured co-navigation for a TPR in hospital layouts. The method comprises three tightly coupled layers: (i) a transformer-based human-motion forecaster that outputs multi-step, multi-modal trajectory distributions; (ii) a Safe-RL controller that consumes risk features derived from the forecasts and optimizes a constrained objective; and (iii) a discrete-time CBF shield that projects the controller's actions into a provably safe set at run time, with latency compensation. Operationally, the transformer's predicted multi-agent distributions constitute an implicit belief state, summarizing the Bayes-filter update and entering the Safe-RL policy via risk-aware occupancy features, while the CBF shield enforces hard safety on the resulting actions. Implementation choices are reported to ensure full reproducibility and to support ablation studies. Table 2 summarizes the run-time CBF shield and latency-compensation design, detailing safety functions, decay factor, QP formulation, slack handling, solver choice, and diagnostic metrics.

Table 2
www.frontiersin.org

Table 2. CBF shield and latency compensation.

4.1 Transformer-based human-motion forecaster

This subsection describes the transformer-based human-motion forecaster that supplies short-horizon, uncertainty-aware multi-agent trajectory distributions used as risk features for downstream control.

4.1.1 Inputs

At each time step t, the forecaster receives a sliding window of tracked pedestrian states {yt-l:ti}i=1Nt (2-D positions with optional velocities), the robot pose xt, a local map patch (SDF and visibility mask), and agent-centric features (pairwise displacements and occupancy raster). Missing detections due to occlusion are explicitly masked.

4.1.2 Architecture

A light, latency-aware transformer is used in the system architecture of the proposed solution, as shown in Figure 2:

• Tokenization: per-agent temporal tokens and contextual map tokens.

• Encoder: multi-head self-attention over agent tokens to capture social interactions.

• Cross-attention: agent tokens attend to map tokens (doors, walls, bottlenecks).

• Decoder: non-autoregressive, predicting H steps for each agent.

• Output head: per-step Gaussian mixture parameters {πik, μik, Σik}.

Figure 2
Ward layout diagram showing sensor fields of view (FOVs). LiDAR FOV is depicted in light blue, and RGB-D FOV is in red. Occlusions are marked in gray. The grid measures distance in meters along x and y axes, with beds located at each corner.

Figure 2. Experimental environment: top–down ward floor plan used in simulation. Walls, doors, beds, curtains, and equipment define the navigation space for the telepresence robot and simulated pedestrians; start and goal regions are indicated schematically.

The transformer forecaster is configured with shallow depth and a limited number of heads and tokens to satisfy real-time inference requirements on telepresence platforms; all architectural components above remain unchanged and are executed at each control cycle.

4.1.3 Uncertainty calibration

Temperature scaling is applied to mixture variances, and a quantile-matched scaling on Mahalanobis distances is used to align predicted covariances with empirical errors (calibration set only). To reduce overconfidence, each Gaussian component is constrained by a variance floor to prevent eigenvalues from collapsing to unrealistically small values, and calibrated Mahalanobis distances are matched to empirical error quantiles on a held-out set. These calibrated covariances feed directly into the chance-robust CBF design (Equations 1417), which uses a high-quantile safety factor to convert forecast uncertainty into conservative clearance margins. Probabilistic calibration is further assessed in Section 6 via reliability diagrams of predicted near-miss risk vs. empirical frequency, where the proposed method tracks the diagonal more closely than PPO and DWA.

4.1.4 Training objective

Negative log-likelihood (NLL) over future trajectories with an uncertainty regularizer:

Lpred=-i,t,τlog(kπikN(yt+τiμik(τ),Σik(τ)))         +λΣi,t,τtr(Σik(τ))    (28)

Table 3 consolidates the transformer forecaster specification—inputs, tokenization, architecture, horizon, uncertainty calibration, and training hyperparameters—serving as the canonical configuration for all forecasting experiments.

Table 3
www.frontiersin.org

Table 3. Transformer forecaster: model specification and training hyperparameters.

4.1.5 Risk features for control

The predicted distributions are converted into compact features ϱt: (i) maximum occupancy intensity around the robot, (ii) near-field probability mass within a radius ρ, (iii) forecasted minimum clearance; and (iv) a flow-direction histogram (optional) to disambiguate counter-flows.

4.2 Safe reinforcement-learning controller

This subsection details the Safe-RL controller, formulated as a constrained Markov decision process that consumes forecast-derived risk features to optimize task performance within safety budgets.

4.2.1 CMDP setup

The controller solves a constrained MDP with a risk-sensitive objective and constraint budgets on safety violations. This module instantiates the Safe-RL formulation of the CMDP in Section 3.3, where safety is expressed as expected counts of CBF-slack activations and near-miss events, and risk sensitivity is captured via a CVaR-style auxiliary objective. The augmented observation is:

st=[xt,pgoalpt,ψ(pt),ϱt,ξt],    (29)

Here, st concatenates the robot state xt, the forecaster-derived risk features rt, static SDF-based map features ϕSDF(distances to walls, doors, and bottlenecks), and a scalar Δencoding sensing-and-actuation latency. This augmented observation exposes both anticipatory risk information and latency-aware geometry to the Safe-RL controller.

4.2.2 Policy and value function

Two MLPs (actor and critic) are used. The actor outputs a Gaussian distribution over [at, ωt] with state-dependent mean and diagonal covariance; squashing enforces action bounds.

4.2.3 Learning algorithm

A Lagrangian PPO variant is used:

• Primary objective: expected return with CVaR surrogate (an auxiliary head estimates tail risk).

• Constraints: expected counts of CBF violations (from shield diagnostics) and near-miss events (≤ 0.3 m ).

• Dual updates: per-constraint multipliers updated by projected gradient ascent.

• Exploration: off-policy replay is not used; entropy regularization stabilizes exploration.

In effect, the actor parameters θ and critic parameters are updated to maximize the clipped PPO surrogate while keeping empirical estimates of the constraint costs (shield slack activations and near-miss events) below their budgets, with Lagrange multipliers adapting as penalty weights whenever these costs exceed the specified thresholds.

In principle, one could differentiate through the CBF-QP and treat the shield as part of the policy, using implicit-function gradients so that barrier parameters directly shape the actor update. We deliberately keep the shield modular and non-differentiated: the CBF parameters are tuned at the control layer to preserve clear forward-invariance guarantees and to allow conservative, certifiable fallbacks even when the policy is updated. Exploring differentiable CBF shields and tighter end-to-end coupling between the barrier and the policy is left as future work.

4.2.4 Rewards and costs (shaped)

• Progress reward rprog = η(∥pt−1pg∥−∥ptpg∥).

• Smoothness penalty on |at| and |ωt |.

• Social-comfort cost proportional to iϕ(pt-yti ).

• Terminal success bonus; collision termination with a large penalty.

Table 4 enumerates the CMDP design and optimization details for the Safe-RL controller, including state composition, constraints, reward/cost shaping, and learning settings used across evaluations.

Table 4
www.frontiersin.org

Table 4. Safe-RL controller: CMDP design and optimization.

4.2.5 Design choices and technical rationale

The technical design of the framework was guided by the need to balance anticipatory perception quality, formal safety guarantees, and real-time deployability on embedded TPR hardware. The transformer-based forecaster was selected over recurrent or graph-based alternatives because short-horizon self-attention captures inter-agent interactions while admitting parallel inference and fixed-latency decoding; the non-autoregressive output head further mitigates error accumulation across the 0.8–1.6 s prediction horizon. The occupancy-field risk representation compresses multi-agent trajectory distributions into low-dimensional features that remain stable under occlusions and are directly usable for constructing chance-robust CBF constraints.

On the control side, navigation was formulated as a constrained Markov decision process with an auxiliary CVaR objective, instead of a purely expected-return PPO formulation, to shape the policy toward tail-risk-averse behavior in dense crowds and rare but critical events. The CBF-based safety shield is kept modular and model-based to retain forward-invariance guarantees even when the learned policy encounters out-of-distribution states; slack variables and shield diagnostics are tuned to prioritize feasibility while exposing interpretable intervention statistics during training. Finally, the forecast horizon, state augmentation, and reward/cost shaping (Table 4) were empirically calibrated to trade off efficiency against social comfort: shorter horizons degraded anticipation of crossing pedestrians, whereas longer horizons increased forecast drift and CBF conservatism, reducing success rates and inducing stop-and-go behavior. To prevent “shield myopia,” the critic receives the shield's dual residuals and action deviations as inputs; the actor is regularized toward low-intervention regimes.

4.2.6 Safety during learning

The shield runs online during training to avoid unsafe data collection. Let atπ denote the unconstrained action sampled from the policy and atsh denote the shielded action returned by the QP, and define the intervention vector δat=atsh-atπ. Let λt denote the vector of optimal dual variables (Lagrange multipliers) of the CBF constraints in the QP.

During actor–critic updates, trajectories are rolled out using the shielded controls atsh, and returns and constraint costs are always computed under these shielded dynamics. Policy gradients use logπθ(atπst) together with advantages Ât estimated from the shielded trajectories, so that the critic learns values conditioned on (st, δat, λt) and the policy is nudged toward regions of the action space where the shield intervenes rarely and weakly. This design does not eliminate “myopia” entirely but empirically reduces it by making shield activity explicitly visible to both the value function and the regularized actor.

4.3 Discrete-time CBF safety shield

This subsection presents the discrete-time CBF safety shield, which wraps the Safe-RL policy in a quadratic program that enforces forward invariance of safety sets under forecast uncertainty and latency.

4.3.1 Safety functions

Two families are enforced at each step: (i) wall CBFs using the map SDF; (ii) human CBFs that incorporate forecast uncertainty (chance-robust function hi,τ~).

4.3.2 Inequalities

For each active safety function h, the discrete-time CBF condition:

h(xt+1)-(1-γ)h(xt)0    (30)

is linearized to an affine constraint in the controls:

Ah(xt)utbh(xt).    (31)

All active constraints are stacked into Atutbt.

4.3.3 QP projection

The shield solves

ut=argminuu-utnom22s.t.Atubt,            uminuumax    (32)

The QP returns both the shielded control atsh and the vector of Lagrange multipliers λt0m associated with the active linearized CBF constraints. The entries of λt quantify how tight each safety constraint is at the optimum, and, together with the slack variables, they provide a compact diagnostic signal describing the strength and frequency of shield interventions. A small slack with a large penalty is allowed for numerical feasibility; slack activations are logged as violations for the CMDP constraints.

All CBFs are evaluated at a predictive state x^t+Δ to counter sensing and network delays. The same prediction feeds the QP.

4.4 End-to-end control cycle

The proposed co-navigation stack executes at 30–50 Hz and implements a forecast → policy → safety-filter pipeline with explicit latency compensation. At each control tick, detections are tracked, short-horizon human-motion distributions are predicted, compact risk features are constructed, and an augmented state is formed. A Safe-RL actor emits a nominal action that is projected by a discrete-time CBF-QP shield evaluated at a latency-compensated predictive state. The shielded action is applied, diagnostics are logged, and (in training mode) policy and dual variables are updated. The full loop is summarized in Algorithm 1.

Algorithm 1
www.frontiersin.org

Algorithm 1. End-to-end forecast → policy → safety-filter control loop.

ForecastTransformer returns multi-step, multi-modal pedestrian state distributions. RiskFeatures compacts these into occupancy and clearance summaries. BuildCBFConstraints assembles wall and chance-robust human CBFs evaluated at xt. SolveQP is a 2-variate active-set QP with optional slack (large penalty) to ensure numerical feasibility in rare edge cases.

Pipeline: Sensing and tracking → detection, data association, occlusion masks. Forecasting → transformer predicts H steps for visible and recently visible agents. Risk features → compute ϱt from the predicted distributions. Policy → actor returns utnom given st. Shield → evaluate CBFs at x^t+Δ and solve the QP for ut. Execution → send ut to the robot; log shield residuals and slack. Learning (train mode) → collect transitions; update actor/critic and Lagrange multipliers in mini-batches.

4.5 Theoretical properties (sketch)

This subsection explains how sensing and network delays are modeled using a latency-aware predictive state, which is consistently used in both the CBF constraints and the shield's QP to pre-empt delay-induced violations.

4.5.1 Forward invariance

If for all active safety functions h the linearized discrete-time CBF constraints hold, then the safe set S(t)={x:h(x,t)0h} is forward-invariant under the closed-loop dynamics with ZOH. Chance robustness enters through the uncertainty margin in hi,τ~; the guarantee is conservative at level ϵ.

4.5.2 Bounded intervention

The QP is the Euclidean projection of utnom onto the convex set U{u:Atubt}. Hence, |ut-utnom|2 is minimal among safe actions, limiting distortion of the nominal policy and supporting stable learning.

4.5.3 Constraint satisfaction under latency

With a bounded model error for x^t+Δ and Lipschitz h, feasibility margins scale linearly with Δ and the prediction error; the safety factor γ and SDF gradients determine the required deceleration envelope.

4.6 Computational complexity and real-time budget

Transformer scales as O((Nℓ)2) in attention; with sparse neighborhood attention, this becomes O(Nℓk), kNℓ. With 15N ≤ 15 and ℓ ≤ 12, run time is typically <10 ms on an embedded GPU.

MLP inference is O(P) with P parameters and typically <0.2 ms on embedded CPUs.

The QP has two decision variables and m linear constraints; active-set methods run in O(m). With m ≤ 12, solve times <0.3 ms are typical.

The cycle fits within 20–30 ms on embedded platforms; the chosen design preserves 30–50 Hz control with headroom for sensing.

4.7 Ablation and diagnostic hooks

This subsection defines the augmented MDP state that bundles robot, map, forecasting, and latency features into a fixed-dimensional representation for the actor–critic networks.

• No-forecast: risk features ϱt removed; policy observes only instantaneous detections.

• No-shield: CBF projection disabled; constraint costs remain in the CMDP.

• Short-horizon: H reduced by 50%.

• Uncalibrated: temperature scaling disabled.

• High latency: Δ increased by two to three times.

• Risk-blind: CVaR head disabled; expected cost only.

Each ablation logs success rate, collision rate, minimum clearance, near-miss rate, time-to-goal, shield activation rate, and average QP deviation ∥uunom ∥.

The method operationalizes anticipation (transformers) and safety (Safe-RL + CBF) in a modular architecture with explicit latency handling and verifiable constraints. The tables provide a compact bill of materials for replication and for controlled ablations.

4.8 Implementation details

This subsection summarizes the complete forecast → policy → safety-filter control loop executed at run time.

4.8.1 Tracking

A constant-velocity Kalman filter with gating on Mahalanobis distance is used for data association; tracks missing for ≤ M frames are carried with increased covariance.

4.8.2 Forecast horizon

H∈[8, 16] steps (0.8–1.6 s at 10 Hz) balance look-ahead and drift. Mixture components per agent K∈{ 3, 5}.

4.8.3 Normalization and masking

Inputs are agent-centered; map patches are aligned to the robot's heading. Occlusions are injected as binary masks on tokens and rasters.

4.8.4 Policy architecture

Actor/critic networks use two to three hidden layers with layer normalization, Tanh activations, and state-dependent log standard deviation with a lower bound.

4.8.5 Training

On-policy updates occur every N steps with advantage estimation using GAE; Lagrange multipliers are maintained for (i) CBF slack rate, (ii) near-miss rate, and (iii) wall proximity breaches.

5 Experimental setup

The proposed study modeled the evaluation environment as a hospital-style ward represented by a 2D floor plan with walls, doors, beds, curtains, and fixed equipment (Figure 2). The floor plan is rasterized into an occupancy grid and an SDF, which are used by both the motion planner and the transformer's map encoder. Pedestrian traffic is generated by simulated human agents following goal-directed trajectories with ORCA-based collision avoidance, while the telepresence robot is commanded from an entrance pose to bedside goal regions using each of the navigation stacks described below.

All experiments in this work were conducted in a physics-based simulator instantiated from the ward floorplan in Figure 2. All pedestrian agents are purely virtual and evolve according to crowd motion models (e.g., ORCA-based controllers) within this simulated environment; no trajectories, sensor streams, or other measurements are collected from real patients, staff, or visitors. Accordingly, the study should be interpreted as a simulation-based evaluation of navigation algorithms in idealized hospital-ward layouts rather than as a clinical trial or observational study on human subjects.

5.1 Navigation algorithms

The study compared the proposed method (transformer-based trajectory forecasting + Safe-RL planner + CBF) against standard baselines. The proposed pipeline works as follows: at each time step, a transformer neural network predicts the future positions of nearby humans (based on their past observed paths), similar to recent works such as Social-TransMotion, enabling the robot to anticipate human motion. A reinforcement-learning (RL) policy then selects a motion command; this policy is trained with safety constraints. In practice, the study implemented Safe-RL by adding a CBF-based safety layer on top of the learned policy. If the RL action violates safety constraints (e.g., approaches a human too closely), the CBF projects it to the nearest safe action. This ensures that the robot never enters an unsafe zone during training or deployment. Perception is simulated using a planar LiDAR and a forward RGB-D sensor, producing range and depth observations that populate a 0.1 m occupancy grid. Field-of-view limits and occlusions from beds and humans are explicitly modeled to mirror real line-of-sight constraints (Figure 3).

Figure 3
Flowchart comparing different control strategies: Reactive, ORCA, Risk-Blind MPC, and Proposed. The Proposed strategy incorporates a Transformer Forecaster, Risk-Aware RL Policy, and CBF Safety Shield. Learning, optimization, and certificate pathways are indicated by different line styles.

Figure 3. Sensor configuration.

Baseline navigation stacks. We compare four navigation stacks, which are summarized in Figure 4:

i. ORCA stack (reactive multi-agent controller). The robot is controlled by ORCA using the RVO2 library; humans and the robot are all ORCA agents, and the robot's commanded velocity is the ORCA solution. There is no global planner, no forecasting module, and no safety shield—this stack represents a widely used reactive crowd-navigation baseline.

ii. DWA stack (classical ROS navigation). A 2D grid-based global planner (Dijkstra/A*) plans in the static map, while a DWA local planner selects admissible velocities using instantaneous LiDAR detections of humans as moving obstacles. This stack has no learned forecasting, no Safe-RL, and no CBF shield and represents a standard ROS-style navigation stack.

iii. PPO stack (learning-only controller). An on-policy PPO controller receives the same instantaneous observations as the proposed method but no forecast-derived risk features; it is trained with the same task reward but without the CBF safety shield. This stack isolates the effect of Safe-RL plus shielding by providing a purely learned baseline without formal safety filtering.

iv. Proposed forecast + Safe-RL + CBF stack. The full stack uses the transformer-based human-motion forecaster, risk-aware Safe-RL controller, and discrete-time CBF-QP shield described in Sections 4.1–4.3, including latency-aware state prediction and chance-robust human CBFs. All ablations in Section 6 (no-forecast, no-shield, short-horizon, risk-blind) are derived from this stack by selectively disabling components.

Figure 4
Illustration of a circular robot and a human, each represented by circles with radii \(R_{rob}\) and \(R_{hum}\) respectively. Key components include the distance to the goal \(d\), the robot velocity vector \(\vec{v}\), and the direction vector \(\vec{c}\). \(c_{min}\) represents the minimum clearance between the robot and human, with the condition \(c_{min} < R_{rob} + R_{hum} + \delta\). A legend explains the symbols: \(d\) (distance to goal), \(V\) (robot velocity vector), \(\omega\) (angular speed), \(c_{min}\) (minimum clearance), and \(h(x)\) (safety metric).

Figure 4. Navigation stacks compared in the experiments: ORCA (reactive multi-agent controller), DWA (classical global + local planner), PPO (learning-only controller without shield), and the proposed forecast + Safe-RL + CBF stack. Each block shows the presence or absence of forecasting, Safe-RL, and CBF safety filtering.

5.2 Evaluation metrics

The study measured the six episode-level metrics: (i) success rate—fraction of trials in which the robot reaches the goal without any collision; (ii) collision (constraint) violations—fraction of trials with at least one collision, defined as center-to-center human–robot distance dhr falling below a collision threshold dcoll = 0.2 m; (iii) time-to-goal—elapsed time until the robot enters the goal region or a timeout is reached; (iv) proximity violations—rate of timesteps dhr<dprox = 0.5m, corresponding to a “personal-space” comfort radius; (v) minimum clearance—minimum dhr over the episode; and (vi) near-miss rate—percentage of timesteps with 0.2m<dhr ≤ 0.3m, capturing episodes in which the robot passes uncomfortably close without physical contact. Distances are computed to the nearest human or bed; path-efficiency metrics (relative path length and relative time-to-goal) are derived by normalizing against free-space runs. These definitions match the metrics reported in Tables 47 and Figures 57.

Figure 5
Two side-by-side line graphs labeled “Baseline (Reactive • Meandering)” and “Proposed (Forecast + Shield • Direct)” show paths from orange squares to green stars. The baseline graph depicts a wavy red line with blue points, while the proposed graph shows a straighter red line with orange points. Each graph features four rectangles. Axes are labeled x (meters) and y (meters).

Figure 5. Metric geometry.

Figure 6
Two box plots compare four methods: ORCA, DWA, PPO, and Proposed. Plot (a) shows time-to-goal in seconds; the Proposed method has the lowest median at 20.42 seconds. Plot (b) depicts minimum clearance in meters; the Proposed method exhibits the highest median at 1.12 meters.

Figure 6. Ward scenario and representative trajectories (illustrative). Left Baseline reactive controller path from doorway (orange square) to doctor's station (blue star) amid pedestrians (blue dots) and beds (rectangles). Right Proposed forecast + CBF-shielded controller follows a near-straight route with slight evasive adjustments.

Figure 7
Four graphs illustrate different performance metrics over episodes. A: Success rate, showing gradual improvement across algorithms, with stability around 5000 episodes. B: Collision rate, indicating rapid decrease then stabilization, with AEMCARL generally lowest. C: Robot's time to reach goal, showing time reduction over episodes. D: Cumulative discounted reward, displaying initial fluctuations before stabilizing with higher rewards. Each graph compares AEMCARL, intrinsic-SGD30N, intrinsic-Ntimesteps(ours), intrinsic-Her(ours), and SafeCrowdNav(ours).

Figure 7. Comparison of the distributions of time-to-goal (A) and minimum clearance (B) for ORCA, DWA, PPO, and the proposed controller.

The collision threshold dcoll = 0.2m approximately corresponds to physical contact for the simulated TPR footprint (≈0.18m radius). The near-miss band (0.2 m <dhr ≤ 0.3 m) and the comfort radius dprox = 0.5m are aligned with proxemics studies on comfortable passing distances in human–robot encounters and with social-navigation benchmarks used in indoor environments, particularly hospital-like corridors. These values are also consistent with the comfort radius defined in the problem formulation (Section 3), and we verified that moderate variations (±0.1 m) do not alter the relative rankings of the methods.

These metrics capture human comfort: in proxemics theory, humans require a certain personal space to feel comfortable, so frequent incursions into that zone are penalized. The study also measured path length and efficiency, including relative time-to-goal and relative path length (defined as the robot's time or path length in a crowded run divided by that in a free-space run). Success rate and minimum clearance have also been used in prior work. In practice, these metrics are computed and logged in code (e.g., maintaining lists of distances and flags at each time step and aggregating post-trial). The total “incident count” (collisions + severe proximity breaches) can serve as a composite safety score. Primary outcome measures are time-to-goal, success rate, safety-margin violations, and proximity statistics; kinematic traces V(t) and ω(t) are also logged. Geometric definitions and the safety set enforced by the shield are summarized for reproducibility (Figure 5).

5.3 Visualization of simulation

As an illustrative scenario, Figure 6 shows the same ward layout from Figure 2 instantiated in the simulator, with example pedestrian trajectories and the corresponding robot paths for ORCA and for the proposed forecast-plus-shield controller, highlighting how the controller exploits anticipatory forecasting to maintain larger clearances around beds and staff. The left panel shows a reactive baseline trajectory, whereas the right panel shows the proposed forecast-plus-shield controller in the same scene.

6 Results

The study evaluated the proposed navigation algorithm against the baseline method using several standard performance metrics. Specifically, the measured metrics are as follows:

• Success rate—the fraction of trials in which the robot reached its goal without any collision.

• Collision (constraint) violations—the fraction of trials ending in a collision (i.e., violations of safety constraints).

• Time-to-goal (navigation time)—the time taken by the robot to reach the goal.

• Safety-margin/proximity violations—the minimum distance (safety margin) to any obstacle or human; each instance in which this distance fell below a predefined threshold was counted as a proximity violation.

The proposed method exhibits the lowest median time-to-goal with a noticeably tighter interquartile range, indicating faster and more consistent task completion while simultaneously achieving the highest median clearance and fewer low-clearance outliers. Means (filled circles) align with the medians (black bars), suggesting robustness to skew, whereas baselines show wider tails and several extreme cases. Taken together, Figure 7 shows that the proposed approach improves efficiency while preserving larger safety margins, rather than trading one for the other.

Table 5 reports the mean and standard deviation of each metric for both the baseline and proposed methods. The results clearly show that the proposed approach achieves higher efficiency and safety. For example, the proposed method achieves a mean success rate of ~98.6% (±0.4%), compared with ~96.6% (±0.8%) for the baseline, reducing the collision rate from ~3.4% to ~1.4% per trial. Similarly, the average time-to-goal is shorter for the proposed method (9.00 ± 0.3 s) than for the baseline (9.79 ± 0.5 s) (Table 2). Proximity violations are also markedly reduced under the proposed strategy. These improvements are consistent with prior work.

Table 5
www.frontiersin.org

Table 5. Performance comparison between baseline and proposed navigation methods (mean ± standard deviation over all trials).

As shown in Figure 8, the learning curves further illustrate these effects over the course of training. The success rate curve (Figure 1A) quickly rises toward 1.0 under the proposed method, whereas the baseline plateaus at a lower level. Correspondingly, the collision rate (Figure 1B) for the proposed method drops to zero much more rapidly. Time-to-goal (Figure 1C) also converges to a lower value for the proposed algorithm.

Figure 8
Graphs A and B depict robot paths with varying colored lines intersecting over an x-y coordinate plane, with positions marked from zero to nine point eight. Legends indicate robot locations, with stars marking path endpoints.

Figure 8. Training performance over 10,000 episodes. (A) Success rate, (B) collision rate, (C) time to reach goal, and (D) cumulative discounted reward are plotted vs. training episode.

Figure 9 shows example navigation trajectories: panel (A) illustrates a simple scenario and panel (B) illustrates a complex one. The black curve marks the robot's path (with start and end markers), while colored curves trace individual human agents. This illustrates how the robot weaves through moving obstacles.

Figure 9
Two scatter plots show the positions of a robot, humans, and a goal at times 2.25 and 9.75 seconds. The left plot shows clustered blue circles for the robot's path at x-y coordinates around 6,3. The right plot shows a spread path moving toward a red star, the goal, at approximately 10,6. Blue lines represent the robot's trajectory.

Figure 9. Sample trajectory maps in (A) a simple environment and (B) a complex, crowded environment.

Figure 10 depicts the real-time safety evaluation: in (A), the robot's safety score is 0.46 (many nearby pedestrians), and it moves cautiously, whereas in (B), the score is 0.96 (fewer nearby pedestrians), and the robot moves more directly. These qualitative maps emphasize that the proposed method maintains larger safety margins around the robot.

Figure 10
Line graph showing the observed near-miss frequency against predicted near-miss probability for different methods: ORCA (yellow), DWA (gray), PPO (purple), and Proposed (teal). Each method's line is distinct, with shaded areas indicating variance. A diagonal line represents perfect prediction.

Figure 10. Safety evaluation visualization. (A) Low safety score (0.46): many humans nearby, so the robot moves slowly. (B) High safety score (0.96): fewer nearby humans, and the robot moves faster toward its goal.

6.1 Overall performance

The proposed method achieves the highest success rate and the lowest constraint and proximity violations while matching or improving time-to-goal relative to baselines. Table 6 summarizes the overall aggregate metrics of the proposed approach.

Table 6
www.frontiersin.org

Table 6. Aggregate overall performance metrics (mean ± SD).

Relative to PPO, collisions and constraint violations drop by ~60% (0.05 → 0.02/ep), the near-miss rate is approximately halved, and minimum clearance increases by ~0.15 m while maintaining competitive time-to-goal.

6.2 Ablation study

Table 7 isolates the contributions of forecasting, the CBF shield, and horizon and risk modeling. Removing either forecasting or shielding degrades safety and reliability; shortening the forecast horizon moderately affects performance; disabling risk awareness increases proximity incursions. All ablations keep the same transformer-based forecasting backbone; systematic variation of the predictor itself (e.g., graph-based or GAN-based models such as Trajectron++ and SocialGAN-RL) is left for future work, enabled by the modular interface between the forecaster and the Safe-RL + CBF stack.

Table 7
www.frontiersin.org

Table 7. Ablation study of the proposed navigation stack (mean ± SD).

7 Discussion

The results indicate that coupling transformer-based human-motion forecasting with a Safe-RL policy and a CBF shield yields meaningful safety gains without sacrificing efficiency. Compared with ORCA, DWA, and vanilla PPO, the proposed method consistently increases the success rate, increases the minimum clearance, and reduces both constraint and proximity violations (Section 6). The ablation study (Table 3) attributes most of the gains to (i) anticipatory information from forecasting—critical in doorways and cross-flows—and (ii) the run-time CBF projection, which eliminates a large fraction of residual unsafe actions while minimally perturbing the nominal policy. Robustness analyses further show stable performance under increased crowd density and injected sensor–network latency (Table 8), suggesting that uncertainty-aware forecasting and latency-compensated safety checking are complementary.

Table 8
www.frontiersin.org

Table 8. Uncertainty-aware forecasting and latency-compensated safety evaluation.

To assess probabilistic calibration, Figure 11 plots reliability diagrams comparing predicted near-miss risk against empirical near-miss frequency across methods. Curves closer to the diagonal indicate better calibration; the proposed method tracks the diagonal more closely than PPO and DWA, explaining its stronger risk-aware decisions.

Figure 11
Two line graphs titled “Proposed Intervention & Deviation over Episodes” and “Constraint-violation rate vs density.” The first graph shows a decreasing trend in intervention rate and average deviation over ten episode deciles. The second graph illustrates constraint-violation rates across different crowd densities for methods ORCA, DWA, PPO, Proposed-NoShield, and Proposed, with Proposed showing a lower violation rate. Bands indicate variability.

Figure 11. Reliability of near-miss risk.

Although our calibrated Gaussian mixture forecaster used in this study does not provide the finite-sample coverage guarantees of conformal prediction, it offers continuous densities and covariances that couple naturally to the chance-constrained CBF construction and to the risk features consumed by the Safe-RL policy, all within a 20–30 ms control budget. Conformal predictors for multi-agent trajectories would typically yield set-valued forecast tubes and require an additional calibration loop, and mapping such sets into differentiable risk features and real-time CBF constraints is non-trivial in our embedded setting. For these reasons, this study adopts a lightweight parametric forecaster with explicit post hoc calibration. It considers conformalized trajectory prediction a promising direction for extensions targeting stronger theoretical uncertainty guarantees.

7.1 Mechanisms and interpretation

Forecasting supplies short-horizon, uncertainty-calibrated occupancy that the policy uses to pre-emptively adjust speed and path, thereby avoiding last-second evasions that often trigger near-misses in baseline methods. The CBF shield then provides formal safety at execution time; intervention logs show low activation rates but high protective value (Figure 12). Importantly, time-to-goal remains competitive because interventions are sparse and small in magnitude, so efficiency is largely governed by the learned policy rather than by conservative fail-safes.

Figure 12
Comparison of four navigation stacks: (1) ORCA Stack includes an ORCA Reactive Velocity Controller; (2) DWA Stack uses a Local DWA Planner; (3) PPO Stack employs a PPO Policy; (4) Proposed Stack incorporates a Transformer Human-Motion Forecaster, Safe RL Policy, and CBF-QP Safety Shield. Each stack's components are outlined, with color codes representing module types.

Figure 12. Shield interventions vs. violations (lines + shaded bands): (A) proposed intervention and deviation over episodes; (B) constraint-violation rate vs. density.

The ablation results also clarify the impact of specific design choices. Shortening the forecast horizon or disabling calibration increases shield interventions and produces more hesitant, stop-and-go motion, while removing the CVaR head (“risk-blind”) leads to higher rates of near-miss events despite similar average success. Keeping the CBF shield modular, rather than absorbing it into the policy network, simplifies verification and debugging and preserves formal safety guarantees, at the cost of a small fraction of projected actions in dense interactions.

7.2 Robustness to density and latency

Performance degrades gracefully with pedestrian density and with injected latencies of 100–300 ms (Table 9). Baseline methods experience a sharp rise in near-miss events under these conditions. In contrast, the proposed method maintains lower violation rates due to forecasting-aware risk features and delay-aware CBF evaluation. Failure analysis shows that most remaining errors arise from abrupt group flow changes (i.e., forecast drift) and rare cases of CBF infeasibility near tight bottlenecks; both effects are reduced but not eliminated by the safety shield.

Table 9
www.frontiersin.org

Table 9. Failure modes for each navigation stack in the crowded corridor and four-bed ward scenarios (medium pedestrian density, 100 ms latency).

7.3 Practical implications and limitations

For clinical deployments, higher clearance and fewer near-misses translate to improved perceived safety and reduced staff burden. The modularity of the pipeline allows drop-in replacement of the forecaster or shield to suit hospital layouts and compute budgets. Four limitations remain: (i) sensitivity to sharp, collective flow reversals that can temporarily miscalibrate forecasts; (ii) occasional CBF infeasibility in extremely tight spaces; (iii) the evaluation fixes a single transformer-based forecasting backbone rather than benchmarking alternative predictors (e.g., Trajectron++- or SocialGAN-RL–style models), so the effect of predictor choice on closed-loop safety and efficiency is not yet quantified; and (iv) the shield induces a mild distribution shift because the environment executes atsh while the policy samples atπ. In practice, this shift is bounded by the intervention norm ∥δat∥, which remains small in the evaluated regimes, but a full theoretical treatment of this off-policy effect remains an open direction for future work. These issues may be mitigated by (a) on-device online calibration of the forecaster, (b) conservative fallback braking with verified invariance, (c) operator-intent overlays that enable quick authority handover in edge cases, and (d) future experiments that integrate multiple forecasting backbones into the same Safe-RL + CBF framework.

Although the simulations instantiate hospital-like corridors and four-bed wards, these layouts and traffic patterns closely resemble those found in many LTC homes, where telepresence robots are increasingly explored to support remote visitation and care-partner engagement. We, therefore, expect the anticipatory behaviors learned here to transfer to LTC settings, while acknowledging that dedicated real-world LTC evaluations remain an important next step. To mitigate “shield myopia,” the critic is conditioned on the augmented input (st, δat, λt), so that value estimates depend on how often and how strongly the shield intervenes, while a regularizer on ∥δat∥ encourages the actor to move toward low-intervention regimes and internalize safety within the policy.

7.4 Future directions

Promising extensions include world-model–based MPC for longer-horizon planning under partial observability, multi-agent coordination informed by staff wayfinding policies, and prospective trials with real hospital traffic to validate social-comfort outcomes. Another important direction is the systematic comparison of different human-motion prediction backbones (e.g., Trajectron++-like, SocialGAN-RL, and diffusion-based crowd forecasters) within the same Safe-RL + CBF architecture, to characterize how forecasting model choice trades off safety margins, efficiency, and computational load in clinical layouts. Integration with shared-autonomy overlays for operator intent may further reduce rare stalls without compromising safety.

8 Conclusion

This study presented an integrated forecast → policy → safety-filter pipeline in which transformer-based human-motion forecasting augments a risk-aware Safe-RL policy, while a discrete-time CBF shield enforces run-time safety for telepresence co-navigation in hospital wards. The formulation explicitly addresses partial observability, dynamic human flows, and sensing/network delays, operationalizing anticipatory perception through short-horizon, uncertainty-calibrated occupancy features and latency-compensated safety projection—precisely the gap identified in prior work, where perception, learning, and formal safety have rarely been fused for TPRs in clinical environments. Quantitatively, the approach outperformed three strong baselines (ORCA, DWA, and PPO). Relative to PPO, constraint violations fell by 60.0% (0.05 → 0.02 per episode), proximity violations by 38.7% (0.31 → 0.19), and near-miss rate by 39.5% (4.3% → 2.6% of steps ≤ 0.3 m), while time-to-goal improved by 2.8% (10.8 → 10.5 s). Success rate increased by 7.6 percentage points (90.4% → 98.0%), and minimum clearance increased by 0.15 m (0.51 → 0.66 m). Against ORCA, improvements were larger (e.g., 83.3% fewer constraint violations, 69.0% lower near-miss rate, and 13.2% faster time-to-goal). These results confirm that anticipatory forecasting, coupled with a certified safety layer, materially enhances safety without sacrificing efficiency—addressing the core clinical requirement for reliable, human-compatible telepresence mobility. Ablation evidence clarifies the mechanism of benefit. Removing forecasting reduced success by 5.9 percentage points and raised constraint violations by 150% and proximity violations by 78.9%; turning off the CBF shield reduced success by 7.8 percentage points and increased constraint violations by 450% and proximity violations by 115.8%. Shortening the forecast horizon doubled the number of constraint violations and increased proximity incursions by 36.8%, underscoring the value of short-horizon anticipation. Sensitivity analyses showed graceful degradation under crowding and latency: at 300 ms injected delay, the method still achieved 94.0% success vs. 82.7% (PPO) and 80.0% (DWA), with consistently larger clearances. Compared to existing studies focused on either reactive social navigation, learning-only policies, or control-only certificates, the proposed integration demonstrably closes the identified research gap by jointly leveraging (i) transformer forecasting for anticipation, (ii) Safe-RL for adaptability, and (iii) a CBF shield for formal run-time guarantees in people-dense clinical layouts. In practice, these gains translate into fewer near-misses, larger comfort distances, and maintained throughput—benefits that can reduce staff burden and increase acceptability in wards. Future work can extend this foundation by adapting on-device forecasters to mitigate rare forecast drift during abrupt flow reversals, implementing conservative, verified fallbacks for tight bottlenecks, and providing operator-intent overlays for rapid authority handover. Longer-horizon world-model–MPC and prospective studies in live hospital workflows are natural next steps to convert the demonstrated ~39–83% safety reductions (depending on baseline and metric) and ~3–13% efficiency gains into sustained clinical impact. Taken together, these findings suggest that transformer-based forecasting, Safe-RL, and CBF shields form a viable architectural template for anticipatory, safety-constrained navigation in people-dense clinical layouts. Although the results are obtained in simulation benchmarks, they surface design principles—short-horizon forecasting, explicit uncertainty handling, and modular safety layers—that can guide future deployments of telepresence systems in hospital and LTC wards. More broadly, this study illustrates how forecast-aware shared autonomy can help bridge the gap between high-performance learning-based navigation and the stringent safety and acceptability requirements of healthcare environments.

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: https://doi.org/10.5281/zenodo.17034737.

Author contributions

HM: Project administration, Supervision, Methodology, Writing – review & editing, Investigation, Conceptualization, Writing – original draft, Funding acquisition. MK: Validation, Conceptualization, Methodology, Data curation, Writing – review & editing, Formal analysis, Software, Project administration. FN: Writing – original draft, Formal analysis, Methodology, Validation, Data curation, Visualization, Resources, Supervision, Investigation, Software, Writing – review & editing, Conceptualization. MT: Project administration, Writing – review & editing, Formal analysis, Visualization, Software, Validation. MJ: Validation, Data curation, Visualization, Project administration, Software, Resources, Writing – review & editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This research was funded by Princess Nourah Bint Abdulrahman University Researchers Supporting Project number (PNURSP2026R140), Princess Nourah Bint Abdulrahman University, Riyadh, Saudi Arabia.

Acknowledgments

The authors would like to express their sincere gratitude to the administration team of Ghulam Muhammad Abad General Hospital, Faisalabad, Pakistan, and to Princess Nourah Bint Abdulrahman University for their support and cooperation in facilitating the practical implementation of this study.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnbot.2025.1697518/full#supplementary-material

References

Backman, K., Kulić, D., and Chung, H. (2023). Reinforcement learning for shared autonomy drone landings. Autonomous Robots. 47, 1055–1071. doi: 10.1007/s10514-023-10143-3

Crossref Full Text | Google Scholar

Berg, J.P., Guy, S.J., Lin, M.C., and Manocha, D. (2010). Optimal Reciprocal Collision Avoidance for Multi-Agent Navigation. Berlin: Springer.

Google Scholar

Bowman, M., Zhang, J., and Zhang, X. (2024). Intent-based task-oriented shared control for intuitive telemanipulation. J. Intell. Robot. Syst. 110:146. doi: 10.1007/s10846-024-02185-1

Crossref Full Text | Google Scholar

Brescia, W., Maci, A., Mascolo, S., and De Cicco, L. (2023). Safe reinforcement learning for autonomous navigation of a driveable vertical mast lift. IFAC-PapersOnLine 56, 9068–9073. doi: 10.1016/j.ifacol.2023.10.138

Crossref Full Text | Google Scholar

Brunke, L., Zhou, S., and Schoellig, A. P. (2022). “Robust predictive output-feedback safety filter for uncertain nonlinear control systems,” in 2022 IEEE 61st Conference on Decision and Control (CDC) (IEEE). doi: 10.1109/cdc51059.2022.9992834

Crossref Full Text | Google Scholar

Chen, X., Zhang, H., Deng, F., Liang, J., and Yang, J. (2023). Stochastic non-autoregressive transformer-based multi-modal pedestrian trajectory prediction for intelligent vehicles. IEEE Trans. Intell. Transp. Syst. 25, 1–14. doi: 10.1109/TITS.2023.3329934

Crossref Full Text | Google Scholar

Desai, M., and Ghaffari, A. (2022). “CLF-CBF based quadratic programs for safe motion control of nonholonomic mobile robots in presence of moving obstacles,” in 2022 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM). New York, NY: IEEE.

Google Scholar

Fox, D., Burgard, W., and Thrun, S. (1997). The dynamic window approach to collision avoidance. IEEE Robot. Autom. Mag. 4, 23–33. doi: 10.1109/100.580977

Crossref Full Text | Google Scholar

Fuchs, S., and Belardinelli, A. (2021). Gaze-based intention estimation for shared autonomy in pick-and-place tasks. Front. Neurorobot. 15:647930. doi: 10.3389/fnbot.2021.647930

PubMed Abstract | Crossref Full Text | Google Scholar

Gao, Y., and Huang, C.-M. (2022). Evaluation of socially-aware robot navigation. Front. Robot. AI 8:721317. doi: 10.3389/frobt.2021.721317

PubMed Abstract | Crossref Full Text | Google Scholar

Garg, K., Usevitch, J., Breeden, J., Black, M., Agrawal, D., Parwana, H., et al. (2024). Advances in the theory of control barrier functions: addressing practical challenges in safe control synthesis for autonomous and robotic systems. Ann. Rev. Control 57:100945. doi: 10.1016/j.arcontrol.2024.100945

Crossref Full Text | Google Scholar

Gottardi, A., Tortora, S., Tosello, E., and Menegatti, E. (2022). Shared control in robot teleoperation with improved potential fields. IEEE Transac. Hum Mach Syst. 52:599–611. doi: 10.1109/THMS.2022.3155716

Crossref Full Text | Google Scholar

Gu, S., Yang, L., Du, Y., Chen, G., Walter, F., Wang, J., et al. (2024). A review of safe reinforcement learning: methods, theories and applications. IEEE Transac. Pattern Anal. Mach. Intell. 46:9844–9863. doi: 10.1109/TPAMI.2024.3457538

Crossref Full Text | Google Scholar

He, Y., Yang, Y., Cai, Y., Yuan, C., Shen, J., and Tian, L. (2023). Predicting pedestrian tracks around moving vehicles based on conditional variational transformer. Proc. Inst. Mech. Eng. Part D J. Automob. Eng. 238, 988–1002. doi: 10.1177/09544070231175536

Crossref Full Text | Google Scholar

Hsu, K.-C., Hu, H., and Fisac, J. F. (2024). The safety filter: a unified view of safety-critical control in autonomous systems. Ann. Rev. Control Robot. Auton. Syst. 7, 15–185. doi: 10.1146/annurev-control-071723-102940

Crossref Full Text | Google Scholar

Hu, G., Wong, J., Ren, L. H., Kleiss, S., Berndt, A., Wong, L., et al. (2025). Care partner experience with telepresence robots in long-term care during COVID-19 pandemic. Digital Health 11:20552076251319820. doi: 10.1177/20552076251319820

PubMed Abstract | Crossref Full Text | Google Scholar

Hung, L., Wong, J. O. Y., Ren, H., Zhao, Y., Fu, J. J., Mann, J., et al. (2025). The impact of telepresence robots on family caregivers and residents in long-term care. Int. J. Environ. Res. Public Health 22:713. doi: 10.3390/ijerph22050713

PubMed Abstract | Crossref Full Text | Google Scholar

Kaelbling, L. P., Littman, M. L., and Cassandra, A. R. (1998). Planning and acting in partially observable stochastic domains. Artif. Intell. 101, 99–134. doi: 10.1016/S0004-3702(98)00023-X

Crossref Full Text | Google Scholar

Karwowski, J., Szynkiewicz, W., and Niewiadomska-Szynkiewicz, E. (2024). Bridging requirements, planning, and evaluation: a review of social robot navigation. Sensors 24:2794. doi: 10.3390/s24092794

PubMed Abstract | Crossref Full Text | Google Scholar

Landers, M., and Doryab, A. (2023). Deep reinforcement learning verification: a survey. ACM Comput. Surv. 56:94. doi: 10.1145/3596444

Crossref Full Text | Google Scholar

Laplaza, J., Moreno, F., and Sanfeliu, A. (2024). Enhancing robotic collaborative tasks through contextual human motion prediction and intention inference. Int. J. Soc. Robot. 16, 1345–1362. doi: 10.1007/s12369-024-01140-2

PubMed Abstract | Crossref Full Text | Google Scholar

Leoste, J., Strömberg-Järvis, K., Robal, T., Marmor, K., Kangur, K., and Rebane, A.-M. (2024). Testing scenarios for using telepresence robots in healthcare settings. Comput. Struct. Biotechnol. J. 24, 105–114. doi: 10.1016/j.csbj.2024.01.004

PubMed Abstract | Crossref Full Text | Google Scholar

Li, B., Wen, S., Yan, Z., Wen, G., and Huang, T. (2023). A survey on the control Lyapunov function and control barrier function for nonlinear-affine control systems. IEEE/CAA J. Automatica Sinica 10, 584–602. doi: 10.1109/JAS.2023.123075

Crossref Full Text | Google Scholar

Li, S., Yuan, Z., Chen, Y., Luo, F., Yang, Z., Ye, Q., et al. (2022). Optimizable control barrier functions to improve feasibility and add behavior diversity while ensuring safety. Electronics 11:3657. doi: 10.3390/electronics11223657

Crossref Full Text | Google Scholar

Liu, D., Li, Q., Li, S., Kong, J., and Qi, M. (2023). Non-autoregressive sparse transformer networks for pedestrian trajectory prediction. Appl. Sci. 13:3296. doi: 10.3390/app13053296

Crossref Full Text | Google Scholar

Lu, X., Woo, H., Faragasso, A., Yamashita, A., and Asama, H. (2022). Socially aware robot navigation in crowds via deep reinforcement learning with resilient reward functions. Adv. Robot. 36, 388–403. doi: 10.1080/01691864.2022.2043184

Crossref Full Text | Google Scholar

Mavrogiannis, C., Baldini, F., Wang, A., Zhao, D., Trautman, P., Steinfeld, A., et al. (2023). Core challenges of social robot navigation: a survey. ACM Transac. Hum. Robot. Interact. 12:36. doi: 10.1145/3583741

Crossref Full Text | Google Scholar

Mohamed, I. S., Ali, M., and Liu, L. (2025). Chance-constrained sampling-based MPC for collision avoidance in uncertain dynamic environments. IEEE Robot. Autom. Lett. 10, 4112–4119. doi: 10.1109/LRA.2025.3576071

Crossref Full Text | Google Scholar

Neggers, M. M. E., Cuijpers, R. H., Ruijten, P. A. M., and IJsselsteijn, W. A. (2021). Determining shape and size of personal space of a human when passed by a robot. Int. J. Soc. Robot. 14, 1111–1124. doi: 10.1007/s12369-021-00805-6

Crossref Full Text | Google Scholar

Neggers, M. M. E., Cuijpers, R. H., Ruijten, P. A. M., and IJsselsteijn, W. A. (2022). The effect of robot speed on comfortable passing distances. Front. Robot. AI 9:915972. doi: 10.3389/frobt.2022.915972

PubMed Abstract | Crossref Full Text | Google Scholar

Ren, L. H., Wong, K. L. Y., Wong, J., Kleiss, S., Berndt, A., Mann, J., et al. (2024). Working with a robot in hospital and long-term care homes: staff experience. BMC Nurs. 23:317. doi: 10.1186/s12912-024-01983-0

PubMed Abstract | Crossref Full Text | Google Scholar

Rudenko, A., Palmieri, L., Herman, M., Kitani, K. M., Gavrila, D. M., and Arras, K.O. (2020). Human motion trajectory prediction: a survey. Int. J. Robot. Res. 39, 895–935. doi: 10.1177/0278364920917446

Crossref Full Text | Google Scholar

Samavi, S., Lem, A., Sato, F., Chen, S., Gu, Q., Yano, K., et al. (2025). SICNav-diffusion: safe and interactive crowd navigation with diffusion trajectory predictions. IEEE Robot. Autom. Lett. 10, 7880–7887. doi: 10.1109/LRA.2025.3585713

Crossref Full Text | Google Scholar

Shayan, Z., Izadi, M., Scognamiglio, V., D'Angelo, S., Singoji, S., Lippiello, V., et al. (2025). Exponential control barrier function and model predictive control for jerk-level reactive motion planning of quadrotors. Control Eng. Pract. 164:106489. doi: 10.1016/j.conengprac.2025.106489

Crossref Full Text | Google Scholar

Shoman, M., Sayed, T., and Gargoum, S. (2024). Transformer-based model for predicting trajectories in autonomous vehicle–pedestrian conflicts: a proactive approach to road safety. Can. J. Civil Eng. 51, 1142–1156. doi: 10.1139/cjce-2024-0137

Crossref Full Text | Google Scholar

Singamaneni, P. T., Bachiller-Burgos, P., Manso, L. J., Garrell, A., Sanfeliu, A., Spalanzani, A., et al. (2024). A survey on socially aware robot navigation: taxonomy and future challenges. Int. J. Robot. Res. 43, 1533–1572. doi: 10.1177/02783649241230562

Crossref Full Text | Google Scholar

Sun, X., Zhang, Q., Wei, Y., and Liu, M. (2023). Risk-aware deep reinforcement learning for robot crowd navigation. Electronics 12:4744. doi: 10.3390/electronics12234744

Crossref Full Text | Google Scholar

Teng, R., Ding, Y., and See, K.C. (2022). Use of robots in critical care: systematic review. J. Med. Int. Res. 24:e33380. doi: 10.2196/33380

PubMed Abstract | Crossref Full Text | Google Scholar

van den Berg, J., Snape, J., Guy, S. J., and Manocha, D. (2011). “Reciprocal collision avoidance with acceleration-velocity obstacles,” in 2011 IEEE International Conference on Robotics and Automation (ICRA) (IEEE). doi: 10.1109/icra.2011.5980408

Crossref Full Text | Google Scholar

Wang, R., Lv, H., Lu, Z., Huang, X., Wu, H., Xiong, J., et al. (2022). A medical assistive robot for tele-healthcare during the COVID-19 pandemic: development and usability study in an isolation ward. JMIR Hum. Fact. 10:e42870. doi: 10.2196/preprints.42870

Crossref Full Text | Google Scholar

Yao, H.-Y., Wan, W.-G., and Li, X. (2022). End-to-end pedestrian trajectory forecasting with transformer network. ISPRS Int. J. Geo-Inform. 11:44. doi: 10.3390/ijgi11010044

Crossref Full Text | Google Scholar

Yousefi, E., Chen, M., and Sharf, I. (2025). Shared autonomy policy fine-tuning and alignment for robotic tasks. Int. J. Robot. Res. 44, 88–109. doi: 10.1177/02783649241312699

Crossref Full Text | Google Scholar

Yunus, A. P., Morita, K., Shirai, N. C., and Wakabayashi, T. (2023). Time series self-attention approach for human motion forecasting: a baseline 2D pose forecasting. J. Adv. Comput. Intell. Intell. Inform. 27, 445–457. doi: 10.20965/jaciii.2023.p0445

Crossref Full Text | Google Scholar

Zhang, Y., and Feng, Z. (2023). Crowd-aware mobile robot navigation based on improved decentralized structured RNN via deep reinforcement learning. Sensors 23:1810. doi: 10.3390/s23041810

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: anticipatory perception, control barrier functions (CBF), crowd-aware navigation, healthcare environments, human–robot co-navigation, safe reinforcement learning, telepresence robots (TPRs), transformer-based motion forecasting

Citation: Mohamed HG, Khan MN, Naseer F, Tahir M and Jamil M (2026) Transformer-based human-motion forecasting coupled with safe reinforcement learning for telepresence robot co-navigation. Front. Neurorobot. 19:1697518. doi: 10.3389/fnbot.2025.1697518

Received: 02 September 2025; Revised: 23 December 2025;
Accepted: 31 December 2025; Published: 02 February 2026.

Edited by:

Fady Alnajjar, United Arab Emirates University, United Arab Emirates

Reviewed by:

Jonathan DeCastro, Toyota Research Institute, United States
Nazmun Nahid, Kyushu Kogyo Daigaku - Wakamatsu Campus, Japan

Copyright © 2026 Mohamed, Khan, Naseer, Tahir and Jamil. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Heba G. Mohamed, aGVnbW9oYW1lZEBwbnUuZWR1LnNh; Fawad Naseer, ZmF3YWQubmFzZWVyQGJpYy5lZHUucGs=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.