Your new experience awaits. Try the new design now and help us make it even better

OPINION article

Front. Digit. Health, 22 August 2025

Sec. Ethical Digital Health

Volume 7 - 2025 | https://doi.org/10.3389/fdgth.2025.1643238

Designing for dignity: ethics of AI surveillance in older adult care

  • Department of Computer Applications, Marian College Kuttikkanam Autonomous, Kuttikkanam, India

Introduction

Widespread adoption of smart home technologies (SHTs) has changed the face of eldercare. With the help of artificial intelligence, such technologies can enable older individuals to live independently with improved security. Fall sensing, sleeping pattern monitoring, and tracking behaviors have now become features in most care settings (1). Although such technologies aim to optimize care, with them come rooted ethical concerns—one such paramount one being the erosion of privacy and dignity (2, 3). As the author, I argue that technologies designed for eldercare must begin with a moral commitment to human dignity—challenging the assumption that safety and efficiency should override agency, autonomy, and emotional well-being.

Older people, already vulnerable to social isolation and intellectual decline, now face an insidious threat: continuous surveillance under the guise of care. As AI technologies guess at users’ behaviors and thought processes, the risk increases of the reduction of individuals to data points (4). What we are witnessing is not just technological advancement, but an ethical drift—one in which systems designed to care are beginning to control, and the goal of support is replaced by the logic of surveillance. This article argues that in developing AI to enable eldercare, we must move beyond the single-minded prioritization of protection and efficiency. This prompts a deeper question: What does it mean to age with dignity in a digitally mediated home? This question—not just technological, but profoundly human—is the compass for the reflections that follow.

Dignity in the age of ambient surveillance

Dignity, difficult as it is to put into measurement, is the foundation of sound eldercare. It involves autonomy, self-respect, and the liberty to enjoy daily life without unnecessary intrusion. In the realm of AI-enhanced living, all these principles are lost (5). Though smart cameras, motion detectors, and digital aides may sense trouble sooner than human helpers, their ubiquitous presence transforms the older adult's sense of home—to no longer be sanctuaries, but surveillance zones.

It is not just about surveillance. It is about the lack of symmetry in control: data is harvested, manipulated, and frequently acted on without complete knowledge or consent from the user (6). Much older adults either have no knowledge about what data is being harvested or lack the capabilities to adjust their settings as a consequence of cognitive or technical constraints (7). This generates an ethical conflict in which security is valued over agency.

In addition, surveillance is psychologically harmful in the way it is likely to be experienced. Research has revealed that ongoing monitoring-even aimed at protection-actually produces feelings of anxiety, helplessness, or withdrawal from ordinary activities (8). The systems designed to add to quality of life might in fact decrease it in some instances.

Beyond issues of surveillance and autonomy, AI systems in eldercare also pose distinct technical and ethical risks. Algorithmic bias, for instance, can lead to unfair treatment when models are trained on non-representative datasets—more harmful in heterogeneous aging populations. Black-box decision-making makes the issue worse, as the argument in favor of system-driven intervention is difficult to grasp or dispute. In high-risk applications such as fall detection or behavioral research, the lack of explainability will only lower confidence and hasten undesirable results. In addition to substantially helping in eldercare, accuracy and speed must also be transparent, interpretable, and responsible.

Real-world examples: a mixed record

To ground such concerns in daily life, let us turn to two real cases that bring to life the ethical challenges inherent in the use of surveillance in AI eldercare. In a pilot study in California, a fall-detection system was turned off in the home of an older adult woman after frequent false alarms triggered unannounced social service visits (9). Meant for protection, the system made her feel surveilled and powerless. Meant to support, instead, it violated her sense of agency and privacy—inseparable from the need for systems that are not only accurate but also respectful and responsive (10).

In another instance, a dementia care facility deployed emotion-recognition technology to detect signs of agitation (11). While the system occasionally flagged concerning behaviors, it also generated numerous false positives. Over time, staff began to rely more on the system's cues than on their own interactions with patients (12). This shift undermined the relational, human-centered care essential for those with cognitive decline.

These cases illustrate a critical point: when AI tools override human judgment or fail to account for users' comfort and consent, they risk turning care into control. Technologies must enhance, not replace, human connection—and must be designed with adaptability, transparency, and dignity at their core.

A framework for dignity-first AI design

To address these challenges, I propose a “Dignity-First” framework that integrates ethical principles directly into the design and deployment of AI-based eldercare systems.

• Informed and Ongoing Consent: Because current models of consent consistently fail, particularly in the long term with cognitively vulnerable user communities—a consent system must be regarded as an ongoing, dynamic process. In this way, user dignity and autonomy are maintained through regular check-in points and user-friendly settings that enable the user or caregiver to modify permission. With the inclusion of simple, easy-to-visualize dashboards and plain-language explanations of how the data will be collected and used, such systems empower the user rather than overwhelming them.

• Data Minimization and Purpose Limitation: Collecting only the data strictly necessary for specific health objectives not only protects privacy but also reduces cybersecurity vulnerabilities. Systems that avoid storing peripheral data—such as ambient audio or location history—unless clearly justified, demonstrate ethical restraint and reduce the risk of misuse, promoting trust and responsible innovation.

• User-Configurable Privacy Settings: Recognizing that privacy needs vary from person to person, systems should provide users with the ability to set granular preferences. Features like “privacy zones” (e.g., in bathrooms) and the ability to schedule surveillance pauses ensure that users retain control. Even those with cognitive challenges should be offered simplified settings that allow them to maintain a sense of agency and comfort within their living space.

• Transparent Feedback Loops: When users lack insight into how their data is being used, it erodes trust. By integrating feedback tools—such as daily data summaries, alerts for external access, or clear explanations for triggered interventions—systems can enhance transparency. This, in turn, helps users feel more in control and more secure, reinforcing dignity through clarity.

• Design for Dignified Defaults: Since most users do not adjust factory settings, the initial configuration of a system carries significant ethical weight. By implementing privacy-protective defaults—such as limited video surveillance, local data storage, and minimal data collection—developers ensure that dignity is preserved from the start. This approach guards against intrusive overreach and prioritizes the well-being of the user without requiring them to navigate complex settings.

The role of policy and regulation

Design alone cannot bear the burden of dignity; policymakers have a critical role in enforcing ethical standards. Just as healthcare devices are regulated for safety and efficacy, AI systems in eldercare must undergo rigorous audits to ensure ethical compliance (13). Regulatory bodies should mandate periodic privacy and usability audits, establish certification frameworks for systems that meet “dignity-by-design” criteria, and promote incentives for co-designed technologies that actively involve older adults in the development process (14). In parallel, insurance companies and public health systems can contribute meaningfully by subsidizing technologies that uphold both care quality and the fundamental rights of older adult users, ensuring that digital innovations support—not compromise—their autonomy and well-being (15).

Co-design: nothing about us without us

One of the most powerful ways to ensure dignity is to involve older adults directly in the design process. Co-design workshops, usability testing, and participatory prototyping allow for insights that top-down engineering approaches miss. These processes must also consider diversity within aging populations, including differences in culture, language, disability, and digital literacy (16, 17).

Digital inclusion must also account for the wide range of digital literacy levels among older adults. While some may comfortably engage with apps and smart interfaces, many still rely on paper-based routines or struggle with touchscreen navigation and abstract digital icons. To ensure true usability, co-design processes should include older adults with varying levels of technological proficiency. This may involve developing hybrid solutions that combine digital and analog formats, or simplifying interfaces without sacrificing dignity or autonomy. By embracing low-tech-compatible and culturally responsive design options, we can create systems that are not only ethical but genuinely inclusive.

By treating older adults as active collaborators rather than passive recipients, designers can develop technologies that feel like companions rather than watchdogs.

From smart homes to ethical homes

The smart homes of the future need to be turned into ethical homes—homes where care, autonomy, and privacy can coexist. This is not solely a technical challenge, but a moral imperative. Aging is not a problem to be solved, but a natural phase of life that warrants respect. Therefore, technologies should be reflective of the values we have towards the people we care about. That is about creating interfaces that enable rather than overwhelm, policy responses that safeguard without condescending, as well as developing AI systems to gain trust as opposed to commanding it.

Conclusion

The future of eldercare is undeniably digital. But as we invite AI into the homes and lives of older adults, we must ensure that safety does not come at the cost of dignity. The right to age with privacy, agency, and self-respect must be baked into the very code of our systems. Designing for dignity is not a barrier to innovation—it is the standard we must meet to ensure technology serves humanity, not the other way around. As stakeholders—designers, policymakers, healthcare providers, and family members—we have a shared responsibility to ensure that the benefits of AI in eldercare do not overshadow the ethical need to protect what makes us human. Aging with dignity is not a luxury. It is a right.

Author contributions

JJ: Writing – review & editing, Conceptualization, Writing – original draft.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Cao L, Feng S, Chen K, Wu X, Jia Y. An ethical model for smart home-based elder care. Nurs Ethics. (2025) 09697330241312383. doi: 10.1177/09697330241312383

Crossref Full Text | Google Scholar

2. Tian YJ, Duong V, Buhr E, Felber NA, Schwab DR, Wangmo T. Monitored and cared for at home? Privacy concerns when using smart home health technologies to care for older persons. AJOB Empir Bioeth. (2025) 16(2):61–76. doi: 10.1080/23294515.2024.2416121

PubMed Abstract | Crossref Full Text | Google Scholar

3. Percy Campbell J, Buchan J, Chu CH, Bianchi A, Hoey J, Khan SS. User perception of smart home surveillance among adults aged 50 years and older: scoping review. JMIR MHealth UHealth. (2024) 12:e48526. doi: 10.2196/48526

PubMed Abstract | Crossref Full Text | Google Scholar

4. London AJ, Razin YS, Borenstein J, Eslami M, Perkins R, Robinette P. Ethical issues in near-future socially supportive smart assistants for older adults. IEEE Trans Technol Soc. (2023) 4(4):291–301. doi: 10.1109/TTS.2023.3237124

Crossref Full Text | Google Scholar

5. Parihar P, Tiwari S, Srivastava A, Arya ErN. A review on privacy concerns exist in smart home technology. In: 2024 2nd International Conference on Computer, Communication and Control (IC4). Indore, India: IEEE (2024). p. 1–5. https://ieeexplore.ieee.org/document/10486571/ (Accessed June 8, 2025).

Google Scholar

6. Mat Daud M, Mohd Ali F, Wan Zaki WMD. A review on smart home design and technology for elderly. J Inf Syst Technol Manag. (2024) 9(37):77–85. doi: 10.35631/JISTM.937006

Crossref Full Text | Google Scholar

7. Gupta A, Sawhney S, Ahmed S. CareTaker.ai—a smart health-monitoring and caretaker-assistant system for elder healthcare. In: The 1st International Conference on AI Sensors & the 10th International Symposium on Sensor Science. MDPI (2025). p. 7. https://www.mdpi.com/2673-4591/78/1/7 (Accessed June 8, 2025).

Google Scholar

8. Cesta A, Cortellessa G, Rasconi R, Pecora F, Scopelliti M, Tiberio L. Monitoring elderly people with the robocare domestic environment: interaction synthesis and user evaluation. Comput Intell. (2011) 27(1):60–82. doi: 10.1111/j.1467-8640.2010.00372.x

Crossref Full Text | Google Scholar

9. Cola G, Vecchio A, Avvenuti M. Improving the performance of fall detection systems through walk recognition. J Ambient Intell Humaniz Comput. (2014) 5(6):843–55. doi: 10.1007/s12652-014-0235-x

Crossref Full Text | Google Scholar

10. Ferguson EL, Mehta S, Miramontes S, Choi M, Kim YJ, Hill-Jarrett TG, et al. Identifying and addressing housing insecurity in older patients: trends, referrals, and inequities in a California medical system. Gerontologist. (2025) 65(5):gnaf027. doi: 10.1093/geront/gnaf027

PubMed Abstract | Crossref Full Text | Google Scholar

11. Putland E, Chikodzore-Paterson C, Brookes G. Artificial intelligence and visual discourse: a multimodal critical discourse analysis of AI-generated images of “dementia”. Soc Semiot. (2025) 35(2):228–53. doi: 10.1080/10350330.2023.2290555

Crossref Full Text | Google Scholar

12. Al Nahian MJ, Ghosh T, Uddin MN, Islam M, Mahmud M, Kaiser MS. Towards artificial intelligence driven emotion aware fall monitoring framework suitable for elderly people with neurological disorder. Brain Inform. (2020) 12241:275–86. (Lecture Notes in Computer Science). doi: 10.1007/978-3-030-59277-6_25

Crossref Full Text | Google Scholar

13. Fessenko DS, Jasperse A. Ethics at the heart of AI regulation. AI Ethics. (2025) 5(3):3387–98. doi: 10.1007/s43681-024-00562-y

Crossref Full Text | Google Scholar

14. Gebremariam KM, Sadana R. On the ethics of healthy ageing: setting impermissible trade-offs relating to the health and well-being of older adults on the path to universal health coverage. Int J Equity Health. (2019) 18(1):140. doi: 10.1186/s12939-019-0997-z

PubMed Abstract | Crossref Full Text | Google Scholar

15. McNellis C. Reframing aging: through the lens of ethical artificial intelligence. Innov Aging. (2024) 8(Supplement_1):1215. doi: 10.1093/geroni/igae098.3888

Crossref Full Text | Google Scholar

16. Willatt A, Gray SI, Manchester H, Foster T, Cater K. An intersectional lifecourse lens and participatory methods as the foundations for co-designing with and for minoritised older adults. Proc ACM Hum-Comput Interact. (2024) 8(CSCW1):1–29. doi: 10.1145/3637295

PubMed Abstract | Crossref Full Text | Google Scholar

17. Randall N, Šabanović S, Chang W. Engaging older adults with depression as co-designers of assistive in-home robots. In: Proceedings of the 12th EAI International Conference on Pervasive Computing Technologies for Healthcare. New York, NY, USA: ACM (2018). p. 304–9. https://dl.acm.org/doi/10.1145/3240925.3240946 (Accessed June 8, 2025).

Google Scholar

Keywords: artificial intelligence in eldercare, smart home technologies, surveillance ethics, human-centered design, digital privacy

Citation: Joseph J (2025) Designing for dignity: ethics of AI surveillance in older adult care. Front. Digit. Health 7:1643238. doi: 10.3389/fdgth.2025.1643238

Received: 8 June 2025; Accepted: 11 August 2025;
Published: 22 August 2025.

Edited by:

Stephen Gbenga Fashoto, Namibia University of Science and Technology, Namibia

Reviewed by:

Paulo Santos, University of Porto, Portugal

Copyright: © 2025 Joseph. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jeena Joseph, amVlbmFqb3NlcGgwMDVAZ21haWwuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.