- 1Bloustein School of Planning and Public Policy, Rutgers University, New Brunswick, NJ, United States
- 2Department of Mechanical Engineering, Rutgers University, Piscataway, NJ, United States
This paper examines the evolving landscape of mobile robotics, focusing on challenges faced by roboticists working in industry when integrating robots into human-populated environments. Through interviews with sixteen industry professionals specializing in social mobile robotics, we examined two primary research questions: (1) What approaches to person detection and representation are used in industry? and (2) How does the relationship between industry and academia impact the research process? Our findings reveal diverse approaches to human detection, ranging from basic obstacle avoidance to advanced systems that differentiate among classes of humans. We suggest that robotic system design overall and human detection in particular are influenced by whether researchers use a framework of safety or sociality, how they approach building complex systems, and how they develop metrics for success. Additionally, we highlight the gaps and synergies between industry and academic research, particularly regarding commercial readiness and the incorporation of human-robot interaction (HRI) principles into robotic development. This study underscores the importance of addressing the complexities of social navigation in real-world settings and suggests that strengthening avenues of communication between industry and academia will help to shape a sustainable role for robots in the physical and social world.
1 Introduction and related work
Mobile robots are rapidly moving from controlled settings into human-populated environments, where detecting and responding to people is central to safe and productive operation (Gartner, 2023; Rubio et al., 2019). Applications span logistics, surveillance, maintenance, transportation, and service settings (Jahn et al., 2020; Sharma et al., 2023), but development across these sectors shares common challenges, particularly around human detection and interaction (Mavrogiannis et al., 2023), integration of cutting-edge research, and commercial maturation (Carlsson et al., 2002; Hekkert et al., 2007; Matarić, 2018). While sensing, navigation, and mobility capabilities are evolving rapidly (Patle et al., 2019), enhancing robots’ abilities to detect human movement and gestures is critical for navigating complex settings facilitate co-deployment and collaboration (Navarro et al., 2021). This broader scope places human detection at the heart of human-robot interaction (HRI) as a field that integrates technical innovation with the social, economic, and governance contexts in which robots are designed and deployed (Stam, 2015; Hannigan et al., 2022; Robey et al., 2001; Kuhlmann et al., 2019; Serrano and Fischer, 2007).
HRI has expanded as an interdisciplinary field integrating AI, robotics engineering, cognitive science, and policy and design thinking (Sinapov et al., 2024) applied across domains such as healthcare (Brooks et al., 2024), farming (Cila et al., 2024), education (Byeon et al., 2024), hazardous environments (Pérez-D’Arpino et al., 2024), and industrial automation (Mukherjee et al., 2022). Robots in social spaces impact not only designated “users” but also bystanders (Dobrosovestnova and Weiss, 2024) who may be incidentally co-present (Rosenthal-von der Pütten et al., 2020) or engaged in unplanned interaction with the robot. This work emphasizes the complexities of trust, accessibility, and context-specific design (Mckenna et al., 2024), particularly as robots transition from experimental prototypes to practical tools in real-world environments (Joshi, 2023). Additionally, this literature increasingly underscores the importance of incorporating the insights of non-technical stakeholders (Joshi, 2023) to ensure commercial robots are not only technically robust but also usable and aligned with human and societal needs (Cila et al., 2024; Wang et al., 2019). Effective interaction however depends not only on social design principles but also on the robot’s perceptual and sensing capacities.
Effective environmental sensing and object detection are crucial in HRI as they enable robots to accurately perceive and interpret human presence, actions, and the surrounding environment (Bonci et al., 2021). Technical advances in sensing and computation, including low-cost high-resolution sensors (Robinson et al., 2023), improved processor efficiency (Chen et al., 2022; Cervera, 2020), and advances in classical optimization methods (Banisetty et al., 2021), machine learning methods (Samsani and Muhammad, 2021), or hybrid learning systems (Gil et al., 2021), have expanded possibilities for human-specific detection and, relatedly, deployment in complex or populated environments. Yet progress in academic research often struggles to scale into commercial systems (Gil et al., 2023), where inconsistent metrics create obstacles for comparison across navigation approaches and complicate evaluation of safety and usability (Mavrogiannis et al., 2023).
A further tension arises between the academic and industrial settings in which robot innovation is occurring under different resource and economic constraints. Academic research is often driven by novelty, controlled evaluation, and publication incentives, while industry emphasizes robust reliability, scalability under cost constraints, and market positioning (United Nation, 2015). Our findings suggest that while person detection may appear solved in controlled environments, it remains unevenly deployed in commercial settings. A small but growing, body of HRI literature, therefore, has turned to industry perspectives to understand how robots are being designed and built in practice. Tsoi et al. (2024) found that roboticists ranked collision avoidance and intimate distance violations as their highest priorities in social robot navigation. Dobrosovestnova and Weiss (2024) showed how commercial robot development draws on feedback from controlled laboratory testing as well as from feedback from deployment in the field, including planned interaction with users and unplanned encounters with bystanders.
Bridging these divergent settings requires understanding person detection not only as a technical capability but as a socio-technical problem that sits at the intersection of safety, sociality, and performance (Kuhlmann et al., 2019; Salazar and Russi-Vigoya, 2021; Bernstein et al., 2022). In industrial contexts, person detection is often framed through a logic of safety, or the imperative to prevent contact, ensure reliability, and maintain predictable operation within regulatory bounds (Mavrogiannis et al., 2023). In contrast, research on social navigation pursues a logic of sociality, in which robots must perceive, interpret, and sometimes anticipate human behavior to inhabit social environments (de Graaf and Malle, 2019; Thompson et al., 2025). These two logics coexist uneasily: emphasizing safety can limit interaction, while emphasizing sociality can introduce risk and uncertainty. From the perspective of innovation-systems and Technology Readiness Level (TRL) frameworks, this tension marks a transition from lab or pilot-grade safety to more robust social-commercial readiness (Östlund et al., 2023; Eriksson and Music, 2021), which can entail functional reliability alongside product integration into organizational, regulatory, and human-interaction contexts (Yfanti and Sakkas, 2024; Salvador-Carulla et al., 2024). By examining how engineers navigate this continuum, we situate person detection as a challenge of socio-technical coordination, where the development of a robotics system is characterized, in part, by how it mediates human–robot relations. This paper extends that line of work by examining how commercial engineers conceptualize human detection and the relationship between academic research and industrial practice. Drawing on qualitative interviews with sixteen roboticists with industry experience in the field of social mobile robots, we asked two research questions: (Gartner, 2023) What approaches to person detection are used in practice in industry, and (Rubio et al., 2019) how does the industry-academia shape the research process? Our findings suggest that engineers diverge in how they prioritize human identification for safety and performance, how contextual and environmental clues are incorporated into training social behavior, and when “acting socially” is prioritized over optimized robotic function in isolation. In discussing industry’s message to academia, we highlight the gaps as well as the complementary roles for disparate actors in the robotics ecosystem, noting how industry professionals conceptualize commercial readiness, identify metrics for social robotic success measurement, and understand the field of human-robot interaction in the context of developing commercially-deployable robots.
2 Methodology
We conducted semi-structured interviews with 16 industry professionals involved in building, designing, or deploying robots that interact with humans—a group we term “social navigation roboticists.” To align with our research focus on industry–academia gaps, all participants had current or recent industry experience.
Interviews followed an 11-question semi-structured protocol, developed to incorporate interdisciplinary concerns through feedback from university faculty in computer science, engineering, and public policy. Topics included robot use cases, navigation systems, perception and sensing, modeling approaches (classical and data-driven modeling), product development feedback channels, and challenges or bottlenecks in research, design, and commercialization, and views on the most pressing industrial challenges they felt deserved greater attention from academic researchers. Clarifying follow-ups were added as needed; the full protocol is available upon request. Interviews lasted 45–60 min, were conducted via video call, recorded with consent, and anonymized under a Rutgers University IRB-approved protocol Pro2020001632.
2.1 Participants
Participants were recruited through LinkedIn, faculty and university networks, professional associations, and referrals from prior contacts. Approximately 25 professionals were invited, and sixteen agreed to participate. Because this study targeted a small population of senior engineers and technical leads in commercial robotics, participant availability, rather than data saturation, was the principal limiting factor in recruitment, a common characteristic of expert-interview research in emergent industries. Eligible respondents were required to be active industry professionals with at least 5 years of experience in the field. Despite outreach to a gender-diverse pool, all participants were male.
To capture diversity across robot form factors and human-environment contexts, we sought participants from multiple application domains, including self-driving vehicles, warehousing and logistics, consumer electronics, inspection, and defense. Representation was necessarily uneven, reflecting our professional network and the present concentration of active commercial deployments in logistics and autonomous-vehicle sectors. The participants had an average of 12 years of industry experience (
2.2 Data analysis
As a cross-disciplinary team of researchers, we combined our engineering and social science backgrounds to investigate how mobile robots are being embedded in social spaces. Both researchers participated in each interview. Interviews were recorded and transcribed for qualitative coding. Each author open-coded a first round independently, with codes iteratively refined to establish an agreed-upon codebook (Pratt, 2023). Using an aligned codebook, we performed a second round of coding. Next, we identified clusters of codes and developed conceptual categories, again informed by concepts derived from the literature (Marshall et al., 2021). Each author wrote thematic memos to summarize key ideas signified by the open-codes. These categories were used to inform emergent themes, again informed by the prior literature review (Marshall et al., 2021). Coding reliability was ensured through joint reconciliation of code assignments and regular cross-checking between authors rather than quantitative scoring, consistent with qualitative best practice for expert-interview studies. A summary of the final codebook, including five major code categories and their emergent themes, is presented in Table 2. To complement the qualitative coding, we also conducted quantitative frequency analysis of thematic trends across participants.
3 Results
Through these interviews, we sought to develop a detailed and nuanced view of how challenges related to human-robot interaction, and social navigation in particular, are considered in industry. Industry professionals who participated in this study offered grounded and pragmatic perspectives as well as conceptual and theoretical views on best practices. Data on approaches to person detection in robotics yielded four important themes: (1) person/object differentiation; (2) context and environment as integral to detection; (3) use-case as an optimizing framework; and (4) how robots can learn to act socially in complex human environments. Responses to industry’s message to academia focused on three key themes: (1) commercial readiness; (2) emergent problems at scale; and (3) incorporating HRI into robotics development.
3.1 Person detection
3.1.1 Person detection and human-object differentiation
A central question to our study was: how are people represented in a robot’s planning environment? Respondents offered three broad approaches to human detection: first, systems in which people are not differentiated from other obstacles (humans as obstacles); second, systems in which people are detected as a distinct but uniform category (humans as a uniform category); and third, systems in which robots not only distinguish people from nonhuman obstacles but also differentiate among types of humans (humans as a variable category). Twelve unique robots were worked on by the participants in this study. Although this is far from a fully representative sample of the commercial space, it offers a telling snapshot of the current state of person detection in robots currently being deployed. All of the robotic systems represented in this study are custom-built by the companies that the participants work for. We ensure the anonymity of our participants by withholding identifying details about the robots that they are building, however Table 3 provides some context about the robotic systems being developed by the participants in this study. To better understand the current state of person detection capabilities, we identified the function of the technology used on each of the twelve robots. These results, illustrated in Figure 1, show that five robots worked on by participants currently have no person detection capability, another five robots have uniform person detection, while two robots can distinguish between different types of people.
First, some robots in development or in use rely mainly on low-level safety systems designed to prevent robots from colliding with obstacles, including people. These systems do not have to distinguish people from other obstacles to trigger safety behaviors, such as stopping. As one respondent put it, “at the moment, we do not differentiate [between] human and object. If there is someone or something in front of the robot, it will stop and wait for the obstacle to move away. And after a certain amount of time, it will find another path to its target.” Some respondents said that person detection was not necessary for safety; or, as one engineer elaborated, “right now, our main focus is to be safe. And to be safe, we just need to avoid all types of obstacles. So, that is why we do not focus on differentiating humans from other things.” Robots that are in this category include a robot vacuum, two quadrupeds, a humanoid, and one that operates in warehouse settings.
In other cases, specialized human detection is already standard and considered necessary for safety. Industrial robots that share space with other agents, one engineer said, need to be able to classify moving objects so as to “unbound their behavior more accurately, accounting for their capacity to move.” One respondent said, “We detect people and treat them specially. People get a right-of-way affordance, and at the bottom layer of the system, the reactive system causes the vehicle to freeze if an agent gets too close.” Here, the respondent points out that specialized human detection systems are layered on low-level safety systems that provide redundancy. Representative technologies in this category include two other warehouse robots, another quadruped, a person-following robot, and a self-driving vehicle.
Finally, some robots being produced at scale not only recognize persons but distinguish among classes of humans. Integrating “a semantic level of person detection,” as one respondent put it, allows robots to distinguish “a police officer from a construction worker” and to know “which people have the right of way or you have to stop for.” One respondent said, “There are all sorts of designations and modifiers for pedestrians and other agents, like crossing guards, emergency vehicles, or cyclists.” Another engineer said they focused on unusual cases in which a person may “move in un-person-like ways,” such as when “on a scooter or on a skateboard [a person] has very different dynamics than a typical pedestrian.” Self-driving cars are a hot spot of focus on human classification so that, as one engineer put it, autonomous vehicles can recognize the relative “vulnerability of a road user,” including “subdivisions within the class of ‘people,’ like children who are even less predictable than adults and harder to see.” That engineer said it was important for autonomous vehicles (AVs) to recognize, for example, that “people inside a vehicle are clearly more protected than people not inside a vehicle.” Both of the robots represented in this category are self-driving vehicles.
Though most engineers said human identification was a key research question, most framed it as an area of future development rather than standard current practice. Human detection appears to be standard practice for self-driving cars and some other AVs, but many commercial robots we learned about in this study do not yet differentiate people from other objects. The range, complexity, and unpredictability of human behaviors continue to make person-detection such a challenge that one respondent who worked in logistics and warehousing said “we’re not differentiating between types of people yet,” while another whose multipurpose surveillance and safety robot is already in use agreed human detection was not yet incorporated into their robot’s perception capabilities, but “that’s very much an area of research right now.”
3.1.2 Person detection: contextual clues and environmental factors
Person-detection is informed by context and several engineers echoed “the necessity of perceiving different types of contexts” so robots can, for example, interpret and plan around human behavior. Several respondents argued individual and group behavior must be modeled differently because, as one said: “There’s a big difference between one person and a couple of people walking together.” Another agreed that crowds produce emergent behaviors not displayed by individuals, like when “a stadium is letting out, it is probably not useful to look at any individual person and ask, what are they going to do? The crowd dynamics matter more than any individual there.” Moreover, behavior flows from groups, but also environmental context; as one engineer said, “it matters how many people are in the context. If the robot is on the street, that’s a lot of people, but a household and a workspace are shared spaces with a limited number of people.”
Sightlines and spatial scale are also important environmental factors. One roboticist who oversaw the development of robots for large logistical settings remarked how sensor constraints meant that robots would not be able to detect or respond to humans especially at a distance:
“Range is a constraining factor. RGBD sensors for indoor use cannot detect people far away. But, people can engage at a distance; and the range that people react to each other is highly context dependent. In a crowded or cluttered space, interaction might happen when people are close. But in wide open situation, people respond to each other from far away– and I’m not talking just about active communication-like speech or gesture, it’s more that people will nudge their trajectory to one side or the other as they’re walking to make it clear which side they intend to pass on.”
Another engineer in warehousing logistics said situations with limited physical sight lines may limit a robotic sensor so a machine “sees” only part of a human–a head or body part behind a stack of boxes–resulting in a “lot of false negative detections.” Taking environmental constraints into account when building and selecting sensors is therefore critical for accurate human detection.
3.2 Optimization versus integration
A key area of disagreement among respondents centered on whether sharing spaces with humans is the optimal use for robots and, if so, how robots should be taught to navigate complex environments. Some engineers argued that building for robot-only environments optimizes efficient, safe, and productive use cases. These respondents argued that instead of beginning with the question of how to refine human-detection, we should ask whether it is necessary or optimal for robots to deploy in spaces shared by humans. The degree of control that engineers have over the robot’s end-use environment (a warehouse compared to a pedestrian street) affects this decision. Two engineers in logistics and warehousing argued that robots working in isolation can be optimized for productivity and this approach will continue to have industry applications in the coming years. As one said, “ultimately, this industry will remove pedestrians from the logistics operating area, not just for safety, but for performance.” Where space is not an issue, another argued, it makes sense to simply “segregate pedestrians to the piece-picking area, and the robots to the case-picking area.” Another respondent said they had worked at companies that took “both extremes,” and noted that “in a warehouse environment, the thinking around safety is very much ‘we’ll slap a fence around it [and] nobody will interact with the robot.”’ But, the same engineer said, that only works for some types of industrial robots, “obviously you cannot do that with self-driving cars.”
Engineers who advocated for integrating robots into human environments were also internally divided about best practices. Some said humans should be part of robot development from the beginning while others said robot integration must be approached through a stepwise development pathway.
Those in favor of ramping up complexity suggested beginning with simple deployment scenarios (robots isolated from or trained to cease motion around humans) and moving toward sophisticated use cases in which robots could work alongside humans or be integrated into populated environments. One engineer said development generally begins with perfecting controlled environments and progressing from there because “interacting with humans is really hard. Generally, in the world of robotics and products, the easiest thing to deploy is a robot that works in isolation. Because you could work on that system, perfect it, send it to the customer and then it just works and no one messes with it. So that’s the easiest thing and where most products start.” Another engineer said they thought of person recognition and safety “in a hierarchical sense,” where “at a lower level you have dynamic obstacle representation where humans are not something different than any other dynamic obstacle,” but building on that, “we need at the top level some way to identify different classes of things in the environment, and people are a very important class.”
Building human recognition into an increasingly complex perception stack can also reflect regulatory requirements. In some cases, regulation shaped the development process because, as one engineer said, “if you want to release a product into an environment where it is no longer physically separated from people, you have to pass safety certifications, which is expensive and hard.” This respondent added: “at this point, the bar is: do not run into people, do not kill people,” but then if “you want to go to the next step of being able to interact with people closely, that’s more challenging” because safety and interaction have to be more tightly integrated. As the same respondent pointed out, if the robot is “meant to interact with a person, you cannot just say, hey, if there’s a person within a 1 m radius of the robot, just stop. You need to enable moving in close proximity with people. And that’s really hard.”
Other engineers approached the problem differently, with some arguing that rather than building complexity ladders, you should start with the most realistic–even if complex–use case for robots. One said, most “economic activity happens around people,” while others pointed out that there is no point in optimizing self-driving cars for isolated controlled environments; instead, one engineer stated, “it is obvious that from the get-go you need to handle humans appropriately.”
Intended use cases also shape a company’s approach to person detection. For robots engaged in a following task, one respondent pointed out, recognizing the person is the first goal. Or, if human interaction will be critical to a robot’s use, then teaching “non-experts to safely operate” a robot is critical. Importantly, interface design is key so users can unlock a robot’s potential uses, or what one engineer called its “taskability,” especially if a robot has many capabilities, from, “taking pictures and carrying payloads, to thermal scanning and vibration sensing.” In these cases a commercial robot engineer said, “what is as challenging [as safety] is taskability or how you communicate to the robot what you want it to do.” This problem can be solved with simple interfaces–domestic cleaning robots with a single button–or, as several respondents suggested, through the growing capacity of large language models to allow natural-language direction. However, the challenge remains that “the more capabilities your robot has, the more things it can do, the harder it becomes to let a non-expert user program the robot in a flexible way.” Intuitive interfaces and making clear through design how a user can direct a robot was stressed by our respondents as essential for machines designed to participate in the human world. As one engineer concluded: from an engineering perspective, “it may be complicated for the robot to do what it is supposed to do, but it must be easy for the human to ask it to do that task.”
3.2.1 Acting socially and training for human complexity
Alongside the challenges of person detection and safety optimization, respondents highlighted the steep learning curve of teaching robots how to “act socially.” For robots to interact with humans, not just freeze or avoid them, engineers have to think about how robots behave around people. One said they thought about a “robot attitude,” explaining, “robots working in environments shared with human beings need to be respectful and work around what humans are doing. Only then can we talk about whether robots can be helpful.” Another said: “There is a negotiation between how much you want to approach the person and how much to avoid them.” Neither inert nor interrupting robots are helpful and well-integrated robots. Cleaning robots, one respondent said, ideally recognize when users are “playing the piano or having a meeting.” More broadly, “there need to be mechanisms for robots to read a situation and then either end a job early or assess if there are still spaces robots can claim without disturbance to the user.”
Integrating robots into complex environments without endangering or inconveniencing humans requires, as one roboticist in the self-driving space said, huge training datasets that include fringe scenarios. While the AV industry is now big enough so some companies develop niche expertise, like “autonomous trucking companies that focus on straight land driving and fuel economy,” while ride-hailing companies need to “from the get-go, focus on environments with lots of complex interactions.” Thinking about both “the big picture and how a technology and a system will be scaled” depends on developing the technology for the right framework.
Finally, professionals in commercial robotics often reminded us that human detection and human behavior models are business decisions. Respondents spoke about building trust with customers and public representatives. “We try really hard to be good citizens,” one engineer said, and “trust, probably at the most local level, is critical.” This can rely on a “tight feedback loop [with] local government regulators” among other stakeholders. Central to building trust, this engineer argued, is demonstrating “this thing actually works.” For some companies, being responsive to customer needs meant setting aside, for now, specialized human detection to focus on, for example, tailoring a robot to warehouse settings or specialized tasks like inspecting offshore drilling rigs where human interaction was minimal. We keep in mind, one engineer said, that “the robots are there for a purpose and we want to make sure they are valuable, for example by improving efficiency, having a reliable system, and integrating well with their facilities.”
“At the end of the day,” another respondent said, “customers do not care how the technology works, they just want it to do a job, do it efficiently and be cost effective and safe.” Therefore, this engineer said, experts in the field have to decide what is safe and optimal. But, they added, while “it does not matter what technology you put in it, I think we’re entering a new era where products will be more software-defined so that we can ship a piece of hardware and then update its software” as navigation and perception best practices change over time.
3.3 Messages from industry to academia
The second research question investigated the relationship between industry and academia impact the research process. We aimed to understand what challenges and opportunities the academic research community should be focused on from an industry perspective. To this end participants were asked: “what message would you share with the academic community?” Respondents gave a broad range of answers to this open-ended question. Many prefaced their opinions by saying they were not experts on the current state of academic research, a valuable datapoint in and of itself. We grouped their responses into three broad categories, discussed below.
3.3.1 Commercial readiness
Commercial Readiness encompasses ideas of technological maturity and whether systems meet current safety standards. Respondents contrasted the freedom to experiment and fail in academic settings with the bar of exhaustive reliability in industry settings. One noted that “in academia you can try out an idea and you do not have to make that idea successful. You can try something out, write a paper, done.” By contrast, in industry, another explained, you fixate on “how long does a robot run before something breaks and that needs to be in the hundreds or thousands of hours.” Engineers “focus on reliability in every part of the stack.” Performance, another said, “has to be really robust, really work, and not just be a fancy algorithm.” Respondents emphasized long-term reliability over novel solutions, one saying that commercial success was about “high quality software, high quality process speeds, high quality evaluation.” One respondent used the Technology Readiness Level, or TRL, scale, to explain differences in academic and industry priorities. Originally developed by NASA to describe the flight-readiness of its systems, the TRL scale goes from level 1 (basic research) to level 9 (full scale deployment). In a commercial setting, this respondent said:
“We’re trying to take something that has been shown to work in a lab and take it to a level of polish and reliability that it can be deployed to a customer. The performance and reliability requirements, even for the research group here, are way higher than in academia. If you know what Technology Readiness Level is, it’s like that concept.”
Alongside dependability, respondents said “the most important thing [for commercial robots] is safety.” Safety was connected to, as discussed above, environmental control and complexity. Fully-known environments can be certified as safe with the addition of light curtains or other safety-rated sensors, which would be insufficient for unmapped environments. Safety was key for both autonomous and non-autonomous robots. One researcher noted that a human-machine interface (HMI) might be designed one way for a research setting so that “if you click the wrong button, bad things can happen,” but differently if “you are going to give a tablet to a customer, [where] it is got to be really simple and safe.” Several respondents noted that safety certifications are time challenging, but others raised more structural challenges about making a robot ready for industrial deployment. One asked: “how can we come up with better mechanisms to implement safe systems in the lab that are certifiable?” These standards could guide social robot development in both academic and industry settings, while improving safety standards across the board.
3.3.2 Emergent problems at scale
The second theme in our data about this question was how problems emerged at scale. Respondents outlined how critical industry challenges are often not identifiable or solvable in small lab systems, and require commercial deployment, integration with realistically complex systems, and stress-testing at large scale to emerge. Most engineers emphasized it was not the job of academics to build systems ready for industrial scale deployment–and that academia and industry played complementary roles–others said academics could do more to confront challenges like reliability that tend to separate academic and industrial solutions.
Respondents described how academic research can fall short of solving challenging problems in social navigation not due to a lack of creativity, skill, or effort, but because these problems only emerge when working at the scale and complexity of a real-world system. As one respondent said, “in terms of social navigation, I think scale is probably the big thing that is missing in academic research.” Another added that though academics are understandably focused on small lab contexts, they need to “think about the scale of a warehouse [and] a solution that will work over a football field densely cluttered with stuff.” What may appear to an academic researcher to be a solution, does not necessarily provide “ideas that are rigorous at scale.” And yet, one respondent said, this standard might be “unfair” because in academia “there are problems that you simply cannot observe unless you operate at scale.”
Additionally, some solutions are only feasible using industry-scale infrastructure, such as access to huge computational resources. Many respondents talked about the inverse industry/academia data challenges. While some companies make datasets (stripped of identifying information) public, the amount of data collected by a company operating at scale dwarfs what academic research groups can access. One respondent put it simply, “the amount of data that most of the companies have, it is really a goldmine.” Another observed that “in academia everyone is trying to get more data and our problem [in industry] is that we have so much data that we have to throw 99% of it away because it is not worth the cost to store it.” Another respondent described how once a robot is deployed, it will generate huge amounts of data through its commercial use. Information harvested from this “data flywheel” can be used to refine the product, a course academics do not have the option of pursuing.
Finally, respondents pointed out that emergent problems demand examining an entire system rather than a single issue. While industry engineers deal with real-world complexity, academics “try to simplify the problem to get to a paper” and fail to “look at the problem as a whole and solve it.” Instead, this engineer suggested, academics should “try to formulate research topics that are more realistic for real-world application” by taking on “the clutter and messiness of the real world.” These experiments might tackle “hardware challenges, like sealing sensors properly for navigating in a muddy environment” or move away from “assuming a 2D environment, like most work in mobile navigation, which is really not the case when you go outside and try to go over hills.” Despite agreement that academics were focusing on the performance or complexity conditions required in industry settings, some respondents understood this as an unrealistic bar for lab researchers and that academic researchers and industrial developers held fundamentally different roles; as one respondent reflected, “putting it all into one soup … there’s a lot going on in there. I do not know that I see academia really grappling with that, but I also do not know if that’s actually where academia should be. I have mixed feelings about that overall.”
3.3.3 The role of HRI
Finally, respondents spoke about the perception of the field of HRI in industry, and a ‘wish list’ of research topics academics looking to contribute to the HRI community might address. One respondent shared a view echoed by others that “a lot of people in industry are ignorant about HRI or its importance. I think a lot of them brush it off as being like UX, or the sort of the thing we can just do to add a little polish at the end, and we’ve solved HRI. There’s a very naive approach to it generally.” This first-person account highlights the need for the HRI community to consider how to effectively convey the importance of its work to decision makers in industry that shape how robots out in the world are being built. This respondent noted that some in industry misconstrue HRI as “just part of safety,” even though safety, as discussed above, has a complicated relationship to sociality.
One way to demonstrate the value of HRI research is to solve problems that are meaningful to those in industry. This agenda setting could happen more readily, as one respondent said, “if academia could step into the industrial world before formulating research problems, they could find more valuable problems to work on, or inspiration for better ways to solve problems.” Across our respondents, we identified several broad areas industry experts felt would be desirable for academics to address (the italicized themes below reflect our distillation of the qualitative data).
3.3.4 Incorporating HRI into robot development
Industry respondents highlighted the importance of what we term a hierarchy of deference, or how robots identify types of people, distinguish among them, and define the rules of deference. One said, “I would love to see more detailed research on the hierarchy of deference of social navigation… I think there’s a recognition problem that I’ve never seen anyone try to solve from a social standpoint. Much less how to actually reason about it.” Industry professionals also pointed out that more research should be done into predefined rules for robots, or the social norms robots should adhere to, ignore, or redefine. As one respondent said, it would be good to know “what literature a designer can turn to know how to drop a robot into a setting that has a pre-existing set of rules and responsibilities.”
3.3.5 Addressing industry challenges and standards
Industry experts identified challenges in several areas they felt deserved more attention in academia. Figure 2 highlights those challenges that were mentioned by multiple respondents. Long-term robot deployment, or what methods and designs can prepare a prototype for deployment at scale, came up much more frequently than any others, as illustrated in Figure 2. One respondent said, “a lab of three people that keeps 100 robots running would have done something interesting.” This need and novelty is punctuated by the observation that much recent research in this area has relied on testing in simulation (Eiffert et al., 2020), which removes many important factors from the tests. However, researchers can hardly be blamed for this, as long-term research projects have higher risks with limited additional benefits (Foster et al., 2015). In addition to long-term deployment, experts also focused on “reusability,” or how researchers can develop system architectures that can be repurposed in different contexts. However, as Figure 2 shows that this and subsequent challenges occurred far less frequently than the long-term deployment challenge. Another respondent said researchers should ask themselves: “How does the structure fit together and what parts are reusable, what parts can be redeployed to solve different problems or other areas of the system?” Other respondents said academics take greater account of the compute-constraints of real-time systems and build experiments in frameworks that are realistic about the computational power available for a commercial robot. One respondent said, “It is easy to throw something in a data center and allow infinite compute or many GPUs. This is useful for thinking ‘here’s how we would like to solve a problem,’ I think that’s different from ‘here’s how we can solve the problem.”’ Finally, experts pointed to the value of standardized certification procedures for commercial settings. As one respondent said, “Look at the current certification procedures that are out there and ask how we can improve them and assess them.” Standardized procedures can reduce risk, telling developers exactly what is allowed, and therefore constraining, but also facilitating the research process by removing uncertainty.
The results reveal how engineers approach human detection, integration, and collaboration not as isolated design questions but as interdependent logics within an innovation system. Across themes, three mechanisms appeared repeatedly: (Gartner, 2023) the conceptual starting point of design (safety, sociality, or performance); (Rubio et al., 2019) organizational constraints, such as certification requirements and customer interaction and expectations; (Jahn et al., 2020) pathways toward commercial readiness require balancing these priorities under real-world uncertainty. The following discussion draws these threads together to synthesize how person detection functions as a technical and socio-technical coordination problem, linking laboratory development to commercial deployment.
4 Discussion
4.1 A framework of safety or sociality
Our findings suggest that the conceptual framework or starting point for robotic design is critical for understanding development pathways, particularly for human recognition. This starting framework shapes what problems engineers identify and try to solve. We suggest that industry experts broadly outlined three starting points for robot design: 1. Safety, 2. Sociality, and 3. Technical performance requirements.
Though goals implied by these frameworks may be complementary, we suggest there are also points of tension. For example, beginning with safety as an absolute priority may require robots simply come to a halt around humans and only operate once humans are at a safe distance. This kind of behavior is not social in an interactive or collaborative sense and may, as several respondents said, create awkward or even obstructive behavior, requiring humans to go around, accommodate, or ignore robots in their environments. Safety therefore can come at the expense of social behaviors. However, beginning with sociality as a goal may require engineers to prioritize interaction and continued work or movement in close proximity to humans over absolute safety. Yet, while some industry respondents said robot deployment in human environments was a clear priority in their businesses, many reported that they still did not begin the design process with the idea of social interaction, but started with controlled, often isolated environments designed to optimize safety and technical performance in the forefront, and only then scaled a ladder of complexity (see below).
Finally, the framework of technical performance reflected a commercial bottom line of meeting customer expectations. Robots must perform the tasks they are marketed as being able to do, and do so at a level of high performance and reliability. Technical performance therefore can trump sociality (imagine a delivery robot that people enjoy interacting with on the sidewalk, but never reaches its destination). The tension that emerges reflects efficiency as well as proficiency. Though a team may have the technical acumen to build a socially cognizant robotic system, these additional requirements may slow production or increase costs. Companies therefore weigh safety, performance, costs, and customer feedback in determining what capabilities are necessary for a viable product. Despite these costs, when ‘socially cognizant robots’ are part of the company’s use case, then sociality may be part of technical performance and necessary for a viable product.
The findings can be understood as mapping how engineers position their systems along a safety–sociality continuum In some systems, safety may dominate, with robots designed to simply avoid humans entirely; in other systems, sociality is integral, with robots required to engage or negotiate close proximity in shared space (Mavrogiannis et al., 2023; Bartneck and Moltchanova, 2020). This framing aligns with recent calls in the HRI and innovation-systems literature to treat commercial readiness as a process of socio-technical integration as well as technical improvement (Kuhlmann et al., 2019; Östlund et al., 2023; Eriksson and Music, 2021). Readiness, in this sense, entails reconciling performance reliability with social compatibility (Bernstein et al., 2022). Viewing industry practice through this lens clarifies why many engineers described person detection as “solved” in principle but nevertheless a persistent challenge in deployment since what is technically solved remains socially unready.
4.2 Complexity ladder
Industry experts were divided over whether robotic systems should be optimized for simple environments first or confronted with complexity even at the earliest stages. These decisions have far-reaching implications for the development workflow, including the kinds of expertise needed on a team, as well as hardware and software design choices. This question is closely linked to that of human-detection. While some roboticists did not think human-detection was key to safety, everyone agreed it was an area of active research if not current use. For some roboticists it is already a priority in their day-to-day work and central to their robots’ functionality. These companies tend to begin with rather than build complexity. Finally, some respondents pointed out that while their robots currently do not distinguish humans from other obstacles, they are built on systems with the capacity to incorporate that in the future, or scale the complexity ladder. As one respondent stated, “we are not differentiating between different types of people right now, that is very much an area of research,” and another said, “the answer right now is ‘no,’ but the intent is ‘yes.’ We have machine learning methods to label objects and though we can label ‘people,’ the robots are not doing any particular behavior in response to that.” Or as another put it, “as of now we do not detect people, but it will probably be part of development in the future.”
4.3 Metrics for success
Technology and product design rely on establishing metrics for success, or the measurements used to decide if a project is on track, a system has achieved its goals, or a product is successful. Traditional success metrics include quantified areas like technical performance, budget, and schedule. Common robot technical performance metrics include power consumption, process speed, and mean time between failures. While some HRI-related metrics exist, they tend to focus on human-robot proximity. There is a need for supplemental metrics to guide and assess social robotic system design. Establishing such metrics is challenging given an “absence of a uniformly agreed upon evaluation standard” for robots in social navigation settings (Mavrogiannis et al., 2023). To develop social performance or integration metrics, typical measurements mentioned above can be supplemented by social navigation-focused metrics like collision rate, path irregularity, acceptable buffers or acceleration rate, and others. Any chosen metric may have downsides, such as lack of representativeness or challenges in evaluation, a commitment to include any measure of sociality advances the aim of building more socially acceptable robots. How safety, sociality, and commercial readiness are weighed in performance metrics can determine, therefore, whether human engagement is seen as a feature, a risk, or a cost.
4.4 Communication structures and role-filling
Our research found that some industry professionals draw on academic literature but others find themselves disconnected from cutting-edge research. At the same time, academics often lack an understanding of current industry activity and challenges. There are few outlets for researchers and engineers in industry positions to share technical challenges and evaluation techniques. Some commercial organizations contribute to the research community through sponsor soring competitions or participating in conferences. Sometimes, both industry professionals and academics are represented in working groups, such as those convened by government or international organizations (NIST, ISO, etc.) to discuss industry-wide standards. Industry can communicate research needs through design challenges, such as the Amazon Picking Challenge, which offer companies ways to engage researchers in cutting edge research problems and share some of the resources and data that the company has at its disposal. However, neither conferences, at which companies may only give high-level overviews of their work, nor working groups, which have limited participation, nor design challenges, which often largely benefit the sponsoring company, provide robust avenues of industry/academia exchange. Better and more systematic routes for exchange could inform industry professionals about academic innovations, direct academics to overlooked types of research problems, and offer the public greater information about the state of the field of product development, especially as robots have a greater presence in the human world.
Finally, our research highlights industry challenges and areas of preoccupation that are generally not the focus of academic labs. Yet, industry professionals understood that robotics is a field characterized by an ecosystem of actors. Building robots, particularly robots that will engage with the human world, is a complex challenge that requires a large cast of academic researchers and industry professionals with shared research goals, but different roles to play.
5 Limitations
As with all qualitative research, this study reflects the perspectives of a limited sample and the interpretations of its researchers. In addition to the limited sample size, we recognize that the absence of female representation among our participants is a major limitation to this work. This group could potentially have perspectives that are not captured in the data that was collected. While our team sought to mitigate bias through a cross-disciplinary approach and systematic analysis, our findings should be understood as exploratory. The 16 participants represented a range of professional experiences in mobile robotics, but a larger and more diverse sample could provide additional nuance. Nevertheless, the breadth of views captured here highlights recurring themes and core concerns across the industry, offering an initial typology to inform future HRI research and practice.
6 Conclusion
This study advances HRI by showing how engineers in commercial robotics conceptualize person detection and balance social interaction with safety and technical performance. Our interviews revealed approaches ranging from treating humans as generic obstacles to developing advanced classification models that differentiate among types of people. These strategies reflect not only technical capabilities but also how organizations prioritize between task completion, safety, innovation, and sociality. Prioritizing sociality can facilitate robot integration into human spaces, while extreme sociality, or prioritizing zero-risk HRI, may affect task completion, restrict robot deployment to controlled environments, or slow innovation. Industry professionals understood these trade-offs and articulated nuanced positive and negative effects. While our sample is necessarily limited, the analysis maps key tensions in industry practice: when human identification is treated as critical for safety versus when it is subsumed under performance optimization, and how contextual cues substitute for “acting socially” in constrained environments. By foregrounding these perspectives, we extend prior HRI research that has emphasized prototypes and controlled trials, offering instead a view of the practical considerations shaping commercial deployment.
For HRI researchers, these findings map the boundaries of current industry practice and highlight opportunities to refine evaluation metrics, address data asymmetries between academia and industry, and design robots that are both safe and socially cognizant. For practitioners, the typology we propose clarifies how person detection choices intersect with commercial readiness and deployment constraints. By situating industry perspectives within the broader HRI agenda, this work underscores the importance of bridging controlled experimental research with real-world design challenges. Future research should deepen this dialogue through larger-scale comparative studies and collaborative frameworks that integrate academic and industrial approaches.
Data availability statement
Due to confidentiality agreements, interview transcripts are not publicly available but anonymized excerpts may be provided upon reasonable request.
Ethics statement
The studies involving humans were approved by Rutgers University Institutional Review Board. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.
Author contributions
LM: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Project administration, Resources, Supervision, Validation, Visualization, Writing – original draft, Writing – review and editing. BG: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Project administration, Resources, Supervision, Validation, Visualization, Writing – original draft, Writing – review and editing.
Funding
The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the National Science Foundation (NSF) National Research Traineeship (NRT) grant NSF NRT-FW-HTF: Socially Cognizant Robotics for a Technology Enhanced Society (SOCRATES) No. 2021628.
Acknowledgements
We thank all the participants in this study.
Conflict of interest
The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Banisetty, S. B., Forer, S., Yliniemi, L., Nicolescu, M., and Feil-Seifer, D. (2021). Socially aware navigation: a non-linear multi-objective optimization approach. ACM Trans. Interact. Intell. Syst. 11, 1–26. doi:10.1145/3453445
Bartneck, C., and Moltchanova, E. (2020). Expressing uncertainty in human-robot interaction. PLoS One 15, e0235361. doi:10.1371/journal.pone.0235361
Bernstein, M. J., Nielsen, M. W., Alnor, E., Brasil, A., Birkving, A. L., Chan, T. T., et al. (2022). The societal readiness thinking tool: a practical resource for maturing the societal readiness of research projects. Sci. Eng. Ethics 28, 6. doi:10.1007/s11948-021-00360-3
Bonci, A., Cen Cheng, P. D., Indri, M., Nabissi, G., and Sibona, F. (2021). Human-robot perception in industrial environments: a survey. Sensors 21, 1571. doi:10.3390/s21051571
Brooks, C., Rees, W., and Szafir, D. (2024). Assistance in teleoperation of redundant robots through predictive joint maneuvering. J. Hum. Robot Interact. 13, 1–23. doi:10.1145/3630265
Byeon, S., Choi, J., Zhang, Y., and Hwang, I. (2024). Stochastic-skill-level-based shared control for human training in urban air mobility scenario. J. Hum. Robot Interact. 13, 1–25. doi:10.1145/3603194
Carlsson, B., Jacobsson, S., Holmén, M., and Rickne, A. (2002). Innovation systems: analytical and methodological issues. Res. Policy 31, 233–245. doi:10.1016/S0048-7333(01)00138-X
Cervera, E. (2020). Gpu-accelerated vision for robots: improving system throughput using opencv and cuda. IEEE Robotics Automation Mag. 27, 151–158. doi:10.1109/MRA.2020.2977601
Chen, Y., Lai, S., Cui, J., Wang, B., and Chen, B. M. (2022). Gpu-accelerated incremental euclidean distance transform for online motion planning of mobile robots. IEEE Robotics Automation Lett. 7, 6894–6901. doi:10.1109/LRA.2022.3177852
Cila, N., González González, I., Jacobs, J., and Rozendaal, M. (2024). “Bridging hri theory and practice: design guidelines for robot communication in dairy farming,” in Proceedings of the 2024 ACM/IEEE international conference on human-robot interaction (New York, NY: Association for Computing Machinery). doi:10.1145/3610977.3634991
de Graaf, M. M., and Malle, B. F. (2019). “People’s explanations of robot behavior subtly reveal mental state inferences,” in 2019 14th ACM/IEEE international conference on human-robot interaction (HRI), 239–248. doi:10.1109/HRI.2019.8673308
Dobrosovestnova, A., and Weiss, A. (2024). “What we learn on the streets: situated human-robot interactions from an industry perspective,” in Companion of the 2024 ACM/IEEE international conference on human-robot interaction (New York, NY: Association for Computing Machinery), HRI ’24, 407–411. doi:10.1145/3610978.3640753
Eiffert, S., Wallace, N. D., Kong, H., Pirmarzdashti, N., and Sukkarieh, S. (2020). “A hierarchical framework for long-term and robust deployment of field ground robots in large-scale farming,” in 2020 IEEE 16th international conference on automation science and engineering (CASE), 948–954. doi:10.1109/CASE48305.2020.9216764
Eriksson, A., and Music, A. (2021). Organizational readiness for the implementation of robots in collaborative environments: a case study. Västerås, Sweden: Mälardalen University, School of Innovation, Design and Engineering. Master’s thesis.
Foster, J. G., Rzhetsky, A., and Evans, J. A. (2015). Tradition and innovation in scientists’ research strategies. Am. Sociol. Rev. 80, 875–908. doi:10.1177/0003122415601618
Gartner (2023). Gartner hype cycle shows supply chain adoption of mobile robots will far outpace drones over next three years. Tech. Rep. Available online at:https://www.gartner.com/en/newsroom/press-releases/2023-08-17-gartner-hype-cycle-shows-supply-chain-adoption-of-mobile-robots-will-far-outpace-drones-over-next-three-years#:∼:text=Technology%20Advancements%20Will%20Enable%20Supply,within%20the%20next%20five%20years.
Gil, O., Garrell, A., and Sanfeliu, A. (2021). Social robot navigation tasks: combining machine learning techniques and social force model. Sensors 21, 7087. doi:10.3390/s21217087
Gil, G., Casagrande, D. E., Cortés, L. P., and Verschae, R. (2023). Why the low adoption of robotics in the farms? Challenges for the establishment of commercial agricultural robots. Smart Agric. Technol. 3, 100069. doi:10.1016/j.atech.2022.100069
Hannigan, T. R., Briggs, A. R., Valadao, R., Seidel, M. D. L., and Jennings, P. D. (2022). A new tool for policymakers: mapping cultural possibilities in an emerging AI entrepreneurial ecosystem. Res. Policy 51, 104315. doi:10.1016/j.respol.2021.104315
Hekkert, M., Suurs, R., Negro, S., Kuhlmann, S., and Smits, R. (2007). Functions of innovation systems: a new approach for analysing technological change. Technol. Forecast. Soc. Change 74, 413–432. doi:10.1016/j.techfore.2006.03.002
Jahn, U., Heß, D., Stampa, M., Sutorma, A., Röhrig, C., Schulz, P., et al. (2020). A taxonomy for mobile robots: types, applications, capabilities, implementations, requirements, and challenges. Robotics 9, 109. doi:10.3390/robotics9040109
Joshi, S. (2023). “Community in hri: extending academic and industry collaboration,” in 2023 32nd IEEE international conference on robot and human interactive communication (RO-MAN), 24–29. doi:10.1109/RO-MAN57019.2023.10309498
Kuhlmann, S., Stegmaier, P., and Konrad, K. (2019). The tentative governance of emerging science and technology—a conceptual introduction. Res. Policy 48, 1091–1097. doi:10.1016/j.respol.2019.01.006
Marshall, C., Rossman, G., and Blanco, G. (2021). Designing qualitative research. Thousand Oaks, CA: SAGE Publications.
Matarić, M. (2018). On relevance: balancing theory and practice in hri. J. Hum. Robot Interact. 7, 1–2. doi:10.1145/3209770
Mavrogiannis, C., Baldini, F., Wang, A., Zhao, D., Trautman, P., Steinfeld, A., et al. (2023). Core challenges of social robot navigation: a survey. J. Hum. Robot Interact. 12, 1–39. doi:10.1145/3583741
Mckenna, P. E., Ahmad, M. I., Maisva, T., Nesset, B., Lohan, K., and Hastie, H. (2024). A meta-analysis of vulnerability and trust in human–robot interaction. J. Hum. Robot Interact. 13, 1–25. doi:10.1145/3658897
Mukherjee, D., Gupta, K., Chang, L. H., and Najjaran, H. (2022). A survey of robot learning strategies for human-robot collaboration in industrial settings. Robotics Computer-Integrated Manuf. 73, 102231. doi:10.1016/j.rcim.2021.102231
Navarro, S. E., Mühlbacher-Karrer, S., Alagi, H., Zangl, H., Koyama, K., Hein, B., et al. (2021). Proximity perception in human-centered robotics: a survey on sensing systems and applications. CoRR abs/2108.07206.
Östlund, B., Malvezzi, M., Frennert, S., Funk, M., Gonzalez-Vargas, J., Baur, K., et al. (2023). Interactive robots for health in Europe: technology readiness and adoption potential. Front. Public Health 11, 979225. doi:10.3389/fpubh.2023.979225
Patle, B., Babu, L. G., Pandey, A., Parhi, D., and Jagadeesh, A. (2019). A review: on path planning strategies for navigation of mobile robot. Def. Technol. 15, 582–606. doi:10.1016/j.dt.2019.04.011
Pérez-D’Arpino, C., Khurshid, R. P., and Shah, J. A. (2024). Experimental assessment of human–robot teaming for multi-step remote manipulation with expert operators. J. Hum. Robot Interact. 13, 1–26. doi:10.1145/3618258
Pratt, M. G. (2023). General coding and analysis in qualitative research. Oxford University Press. doi:10.1093/acrefore/9780190236557.013.859
Robey, D., Welke, R., and Turk, D. (2001). Traditional, iterative, and component-based development: a social analysis of software development paradigms. Inf. Technol. Manag. 2, 53–70. doi:10.1023/A:1009982704160
Robinson, N., Tidd, B., Campbell, D., Kulić, D., and Corke, P. (2023). Robotic vision for human-robot interaction and collaboration: a survey and systematic review. J. Hum. Robot Interact. 12, 1–66. doi:10.1145/3570731
Rosenthal-von der Pütten, A., Sirkin, D., Abrams, A., and Platte, L. (2020). “The forgotten in hri: incidental encounters with robots in public spaces,” in Companion of the 2020 ACM/IEEE international conference on human-robot interaction (New York, NY: Association for Computing Machinery). doi:10.1145/3371382.3374852
Rubio, F., Valero, F., and Llopis-Albert, C. (2019). A review of mobile robots: concepts, methods, theoretical framework, and applications. Int. J. Adv. Robotic Syst. 16, 1729881419839596. doi:10.1177/1729881419839596
Salazar, G., and Russi-Vigoya, M. N. (2021). Technology readiness level as the foundation of human readiness level. Ergonomics Des. 29, 25–29. doi:10.1177/10648046211020527
Salvador-Carulla, L., Woods, C., de Miquel, C., and Lukersmith, S. (2024). Adaptation of the technology readiness levels for impact assessment in implementation sciences: the trl-is checklist. Heliyon 10, e29930. doi:10.1016/j.heliyon.2024.e29930
Samsani, S. S., and Muhammad, M. S. (2021). Socially compliant robot navigation in crowded environment by human behavior resemblance using deep reinforcement learning. IEEE Robot. Autom. Lett. 6, 5223–5230. doi:10.1109/LRA.2021.3071954
Serrano, V., and Fischer, T. (2007). Collaborative innovation in ubiquitous systems. J. Intelligent Manuf. 18, 599–615. doi:10.1007/s10845-007-0064-2
Sharma, N., Pandey, J. K., and Mondal, S. (2023). A review of mobile robots: applications and future prospect. Int. J. Precis. Eng. Manuf. 24, 1695–1706. doi:10.1007/s12541-023-00876-7
Sinapov, J., Han, Z., Bagchi, S., Ahmad, M., Leonetti, M., Mead, R., et al. (2024). Introduction to the special issue on artificial intelligence for human–robot interaction (ai-hri). J. Hum. Robot Interact. 13, 1–3. doi:10.1145/3672535
Stam, E. (2015). Entrepreneurial ecosystems and regional policy: a sympathetic critique. Eur. Plan. Stud. 23, 1759–1769. doi:10.1080/09654313.2015.1061484
Thompson, S., Candon, K., and Vázquez, M. (2025). The social context of human–robot interactions. Annu. Rev. Control, Robotics, Aut. Syst. doi:10.1146/annurev-control-030623-015506
Tsoi, N., Romero, J., and Vázquez, M. (2024). “How do robot experts measure the success of social robot navigation?,” in Companion of the 2024 ACM/IEEE international conference on human-robot interaction (New York, NY: Association for Computing Machinery), HRI ’24, 1063–1066. doi:10.1145/3610978.3640636
United Nation (2015). Transforming our world: the 2030 agenda for sustainable development. Tech. rep. New York, NY: United Nations, Department of Economic and Social Affairs - Sustainable Development.
Wang, K. J., Shidujaman, M., Zheng, C. Y., and Thakur, P. (2019). “Hripreneur thinking: strategies towards faster innovation and commercialization of academic hri research,” in 2019 IEEE international conference on advanced robotics and its social impacts (ARSO), 219–226. doi:10.1109/ARSO46408.2019.8948829
Keywords: mobile robots, navigation, industry, interview, qualitative research, person detection
Citation: Moe L and Greenberg B (2025) From complexity to commercial readiness: industry insights on bridging gaps in human-robot interaction and social robot navigation. Front. Robot. AI 12:1711675. doi: 10.3389/frobt.2025.1711675
Received: 23 September 2025; Accepted: 28 November 2025;
Published: 16 December 2025.
Edited by:
Phani Teja Singamaneni, Laboratoire d’analyse et d’architecture Des Systèmes (LAAS), FranceReviewed by:
Muhammad Khan, Atılım University, TürkiyeLotfi Hamzi, NEOMA Business School Reims, France
Copyright © 2025 Moe and Greenberg. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Benjamin Greenberg, YmVuLmdyZWVuYmVyZ0BydXRnZXJzLmVkdQ==