ORIGINAL RESEARCH article

Front. Comput. Sci., 21 April 2026

Sec. Human-Media Interaction

Volume 8 - 2026 | https://doi.org/10.3389/fcomp.2026.1772813

Trust rises, attention falls: divergent effects of exposure and education in driving automation

  • 1. Graduate School of Systems and Information Engineering, University of Tsukuba, Tsukuba, Japan

  • 2. Autonomous Driving Research Division, Japan Automobile Research Institute, Tsukuba, Japan

  • 3. Institute of Systems and Information Engineering, University of Tsukuba, Tsukuba, Japan

  • 4. Center for Artificial Intelligence Research, University of Tsukuba, Tsukuba, Japan

  • 5. Tsukuba Institute for Advanced Research, University of Tsukuba, Tsukuba, Japan

Abstract

Introduction:

Drivers supervising Level 2 automation must maintain situation awareness while the system controls steering and speed. Miscalibrated trust can contribute to overreliance and lapses in monitoring, whereas insufficient trust leads to disuse. Prolonged supervision is associated with increased mind-wandering, which can slow reactions to critical events. This study tested whether brief educational interventions affect trust, attention, and takeover readiness during Level 2 driving. Our focus on brief interventions reflects the short, time-constrained onboarding that drivers typically receive when adopting driving automation systems.

Methods:

Fifty-five licensed drivers with no prior hands-on experience of Level 2 automation completed a 15-min automated highway drive. Participants received either minimal instruction (Basic), capability-focused education (Knowledge-based), or limitation-focused education (Rule-based). Trust was measured at four time points; additional measures captured self-reported mind-wandering, gaze behavior, and takeover reaction time.

Results:

Trust increased significantly over time in all groups, and educational framing did not alter this trajectory. Capability-focused education enhanced monitoring of the human-machine interface on two false discovery rate corrected metrics and produced faster takeover reactions than limitation-focused education (no difference vs. Basic). Across participants, greater trust growth correlated with higher mind-wandering, while more structured gaze was associated with lower mind-wandering.

Discussion:

Overall, trust formation appeared to be primarily associated with direct experience with system performance, whereas targeted education refined what drivers monitored and how quickly they responded. Together, these results clarify how experience primarily builds trust while education selectively sharpens attention and response readiness in automated driving. These findings clarify distinct roles of experience and brief education in supervising automation and have implications for driver training, human-machine interface design, and gaze-based monitoring.

1 Introduction

Level 2 driving automation, as defined by the Society of Automotive Engineers (SAE International, 2021), combines longitudinal control (adaptive cruise control) and lateral control (lane-keeping assistance) to relieve drivers of continuous manual steering and speed adjustments. Despite this operational support, Level 2 systems explicitly require continuous driver supervision because the automation may encounter situations beyond its operational design domain or capabilities. Drivers must therefore remain ready to resume manual control at short notice when a takeover request is issued or when traffic conditions exceed system limits. In this context, takeover performance is commonly operationalised as the latency between the takeover request and the driver's first control input that initiates manual control, reflecting readiness to re-engage with the driving task (Merat et al., 2014; Gold et al., 2013; Eriksson and Stanton, 2017).

This supervisory role presents a fundamental human-factors challenge. Maintaining sustained monitoring during prolonged automation is mentally demanding, particularly when the system performs reliably and the task provides limited cognitive stimulation (Warm et al., 2008). Under these conditions, drivers may become overly reliant on automation, showing reduced visual scanning of the road environment and slower readiness to intervene when required (Parasuraman et al., 1993; Parasuraman and Riley, 1997; Inagaki and Itoh, 2013). Japan's road-safety initiatives, including the Ministry of Land, Infrastructure, Transport and Tourism's (MLIT) Advanced Safety Vehicle (ASV) Plan, have identified appropriate human supervision of driver-assistance systems as a policy priority. Similar concerns are reflected in international frameworks such as the European Union's Vision Zero strategy and the United Nations' commitment to reducing global road-traffic casualties by 50 per cent by 2030.

Despite progress in automation design, limited understanding remains of how users learn to supervise such systems effectively. This study addresses that gap by examining how pre-drive educational framings influence trust, attention, and takeover readiness during Level 2 automation. It explores whether targeted instruction can shape drivers' mental models and mitigate the attentional drift observed in prior research.

Long-form driver education is rarely delivered for Level 2 systems. In current practices, onboarding is largely informal and relies on passive materials or self-guided learning: most drivers report learning through trial-and-error whilst driving, brochures, or owner manuals, with only a minority receiving any instruction from dealers (Harms et al., 2020; Oviedo-Trespalacios et al., 2021; DeGuzman and Donmez, 2021). Given that such approaches offer limited time for detailed explanation, it is important to determine whether even minimal pre-drive education can shape how users form trust and supervise automation.

This challenge echoes the classic ironies of automation described by Bainbridge (1983), who observed that the more reliable a system becomes, the more crucial and difficult the human operator's monitoring role becomes. As automation handles routine control, drivers are left with the demanding task of supervising rare but safety-critical situations, often without sufficient opportunity to maintain the necessary monitoring and manual-control skills. These ironies are increasingly evident in modern driving automation, where sustained reliability can erode vigilance and delay intervention.

Trust in automation plays a central role in how drivers allocate attention and effort during supervisory tasks (Lee and See, 2004). In this context, trust refers to the belief that an automated system will support the user's goals when outcomes are uncertain or the situation involves risk. Trust influences how people choose to engage with, depend on, or avoid automation. When trust is well calibrated to system capability, drivers are more likely to maintain appropriate vigilance. Poor calibration can increase safety risks: excessive trust may lead to complacency and reduced monitoring (Parasuraman and Riley, 1997; Parasuraman et al., 1993), whereas insufficient trust can result in disuse, where drivers avoid using automation even when it could assist them (Lee and See, 2004; Parasuraman and Riley, 1997).

Trust calibration poses distinct challenges in driving automation, where incorrect assumptions about system capability can lead to immediate safety risks (Khastgir et al., 2018; Inagaki and Itoh, 2013). As Inagaki and Itoh (2013) note, drivers may overtrust systems beyond their functional limits, resulting in over-reliance and delayed intervention. Unlike automation in controlled industrial domains, driving automation operates within dynamic and unpredictable traffic environments that continually test system boundaries (Campbell et al., 2018). Such variability demands that drivers understand not only what the automation can achieve under ideal conditions but also how its performance degrades as environmental complexity increases (Khastgir et al., 2018).

A foundational model for understanding how drivers maintain effective supervision is Endsley (1995b)'s three-level framework of situation awareness, which distinguishes the perception of environmental elements, their comprehension in relation to goals, and the projection of their future status. Maintaining these levels is essential for safe supervisory control and underpins subsequent theoretical developments, including the situation-awareness discrepancy model adopted in this study.

Theoretical models of trust help explain these mechanisms. The Confidence, Awareness, Understanding, Satisfaction, and Enactment (CAUSE) model (Rowan, 1991) is a structured risk-communication framework showing how sequential messaging can build trust and support behavioral change. Communicators first establish confidence, then raise awareness of risks, foster understanding, promote satisfaction with solutions, and finally encourage enactment. In the context of driver education, this suggests that the sequencing and framing of information, not only the content about system performance, shape how users form and calibrate trust. In contrast, trust-calibration models (Lee and See, 2004; Hoff and Bashir, 2015) describe feedback processes through which users compare observed system behavior with prior expectations and adjust their trust accordingly (Merritt and Ilgen, 2008). These experiential mechanisms operate alongside educational processes that shape initial expectations before interaction begins. The Malleable Attentional Resources Theory (MART) (Young and Stanton, 2002; Warm et al., 2008) complements this view by explaining how consistent automation performance can gradually compress attentional resources. When systems operate without error, monitoring becomes confined to abbreviated scanning routines that suffice in routine conditions but may be insufficient when rare critical events occur (Parasuraman et al., 1993; Parasuraman and Riley, 1997; Warm et al., 2008).

Empirical research supports these theoretical accounts (Hoff and Bashir, 2015; Gold et al., 2015). Studies show that both educational information and direct experience with automation performance can influence trust development, although the relative strength of these effects remains to be determined (Körber et al., 2018; Merritt and Ilgen, 2008). After extended exposure to consistent automation performance, drivers reduce their visual attention to the forward roadway, engage more often in secondary tasks, and respond more slowly to unexpected takeover requests (Louw and Merat, 2017; Merat et al., 2014; Gold et al., 2013). Such patterns appear even when drivers are explicitly instructed to maintain vigilance, indicating that experiential learning exerts a strong influence during ongoing interaction with automation (Gold et al., 2013; Hoff and Bashir, 2015).

Beyond trust dynamics, a related challenge is the natural tendency for attention to drift during monotonous supervision. Mind-wandering describes attention shifting away from the external task toward self-generated thoughts and feelings. In this paper we distinguish mind-wandering from the broader phenomenon of attentional drift. Attentional drift refers to the gradual reduction of active task monitoring during low-demand automation supervision, whereas mind-wandering specifically denotes internally generated thoughts that compete with task-related attention. This state is common in low-stimulation settings and carries measurable costs for ongoing performance (Smallwood and Schooler, 2015). In manual driving studies, mind-wandering has been linked to changes in driver behavior and reduced responsiveness, reflecting a shift of attention from the external driving environment to internally generated thought (He et al., 2011; Baldwin et al., 2017). Evidence from takeover research further shows that drivers respond more slowly when not actively monitoring, underscoring the safety relevance of attentional drift (Eriksson and Stanton, 2017).

Driving automation creates conditions that can encourage attentional drift by reducing immediate task demand. Reduced manual control and predictable system behavior lessen immediate task demands and may draw attention away from the roadway (Merat et al., 2019). Such low-stimulation contexts are known to increase mind-wandering (Smallwood and Schooler, 2015). Empirically, automated driving alters visual-monitoring patterns relative to manual control, indicating shifts in supervisory attention (Louw and Merat, 2017). When attention turns inward, drivers typically need additional time to re-engage with the scene and act, a pattern that aligns with longer and more variable takeover responses when monitoring is reduced (Eriksson and Stanton, 2017).

Visual-attention patterns derived from eye gaze provide behavioral indicators that can be interpreted as proxies for attentional engagement (Underwood et al., 2003). Structured visual scanning, characterized by systematic transitions between the forward roadway, mirrors, and peripheral regions, reflects active monitoring of the traffic environment (Underwood et al., 2003). By contrast, fragmented or unsystematic scanning patterns indicate reduced situational awareness and attentional disengagement (Louw and Merat, 2017; Schnebelen et al., 2020). First-order Markov-chain analysis has recently been applied to quantify the organization of gaze transitions, allowing researchers to describe how drivers structure and maintain monitoring routines under automation (Rabiner and Juang, 1986; Schnebelen et al., 2020; Chouchane et al., 2026). To examine these supervisory behaviors objectively, the study analyzed gaze transitions between predefined areas of interest using eye-tracking metrics that capture the organization of visual scanning during automation supervision.

Given these challenges of trust miscalibration and attentional drift, driver education represents a potentially valuable intervention for supporting both calibrated trust and sustained attention during automation supervision. The concept of informed safety emphasizes that users should receive clear, comprehensive information about the capabilities and limitations of automation to calibrate trust appropriately (Campbell et al., 2018; Khastgir et al., 2018). Effective education therefore extends beyond operational instructions to include understanding of sensing principles, boundary conditions, and expected driver roles. Such knowledge supports realistic mental models and helps prevent both over-trust and disuse when supervising automation.

We introduce two contrasting educational framings for Level 2 supervision: Knowledge-based (capability-focused) and Rule-based (limitation-focused) education. This distinction is grounded in established models of cognitive control and supervisory behavior. Rasmussen's Skills-Rules-Knowledge framework (Rasmussen, 1983) distinguishes capability-focused behavior, which relies on reasoning with an internal model of the system in unfamiliar or variable situations, from limitation-focused behavior, which applies stored procedures to familiar signs and boundary conditions. In driving automation, capability-focused information can help drivers build richer mental models of how the system senses and controls the vehicle, consistent with cognitive load theory's emphasis on instruction that supports schema construction and understanding of system structure (Sweller et al., 2011). Such understanding contributes to the process and purpose information that Lee and See identify as a basis for calibrated trust in automation (Lee and See, 2004). Limitation-focused information, by contrast, highlights contexts in which automated performance may degrade and the actions required from the driver, reinforcing limitation-focused preparedness for boundary conditions and helping to reduce the risk of over-trust (Rasmussen, 1983; Lee and See, 2004). These framings therefore target complementary cognitive pathways-conceptual understanding and conditional responding-that may influence how drivers monitor the system and prepare for intervention during supervision.

Prior researc h on educational framing in automation contexts has produced mixed findings. Some studies suggest that providing clear information about system limitations can reduce inappropriate reliance and support monitoring (DeGuzman et al., 2020; Körber et al., 2018). Other accounts emphasize that trust calibration is strongly shaped by performance feedback during use, which may limit the durability of brief pre-drive instruction (Hoff and Bashir, 2015). Methodologically, many studies assess immediate, short-term effects rather than tracking how trust and monitoring evolve over extended exposure to automation (Gold et al., 2015). Together, these observations highlight the need to examine whether targeted educational interventions can alter trust and attentional processes during sustained automation.

Against this background, the present study examines whether these contrasting educational framings influence trust development, attention management operationalised through gaze behavior, and takeover readiness during Level 2 automation. Building on evidence that exposure to consistent automation performance can degrade monitoring patterns (Louw and Merat, 2017; Merat et al., 2014; Cooper et al., 2023), we investigate whether targeted pre-drive instruction can mitigate these effects by shaping drivers' expectations, mental models, and supervisory strategies.

We adopt the situation-awareness discrepancy model introduced in earlier work (Chouchane et al., 2022), which conceptualizes driver supervision as the difference between the situation awareness required by the driving context (α) and that actually achieved by the driver (α′). Educational interventions are hypothesized to sustain α′ (achieved monitoring) at appropriate levels, thereby reducing the discrepancy δ = α − α′. This model extends the logic of Endsley (1995b)'s three-level situation awareness framework by representing supervisory attention as a measurable balance between normative monitoring demand and descriptive driver behavior. It also aligns with the Monitoring and Response Task (MART) framework (Young and Stanton, 2002; Warm et al., 2008) and with Parasuraman and Riley (1997)'s taxonomy of automation use, misuse, and disuse, linking cognitive state dynamics to observable gaze patterns.

Within this model, we distinguish between trust calibration, which reflects how drivers perceive and internalize the system's reliability and limits, and capability-focused monitoring strategies, which determine how they visually allocate attention and verify the driving scene. These components may respond to educational framing through partially independent pathways: trust primarily develops through accumulated performance feedback and may be less sensitive to brief instruction (Hoff and Bashir, 2015; Merritt and Ilgen, 2008), whereas monitoring strategies may adapt more readily to targeted guidance (Körber et al., 2018).

To test these effects, the experiment employed a between-subjects design with three educational conditions. Basic instruction condition (reference condition) provided minimal information typical of current onboarding, whereas the capability-focused and limitation-focused conditions reflected the capability-focused and limitation-focused framings outlined above. Trust was measured at four time points throughout an 80-min session to capture both immediate post-instruction effects and subsequent evolution during automation exposure (Gold et al., 2015). Additional measures assessed mind-wandering (Smallwood and Schooler, 2015), visual scanning patterns using gaze-transition analysis (Schnebelen et al., 2020; Underwood et al., 2003), and reaction time during a critical takeover event (Gold et al., 2013; Lu et al., 2017).

The central research questions guiding this investigation were: (1) How does the framing of driver education influence trust dynamics over time when supervising Level 2 driving automation? (2) Does instructional framing affect drivers' tendency to mind-wander whilst supervising automation? (3) Do different educational approaches influence visual scanning patterns during supervision? (4) Do different educational approaches influence takeover performance? (5) What relationships exist among trust development, attentional disengagement, and takeover readiness?

Finally, to examine how educational framing shapes supervision in Level 2 driving automation, we proposed a set of hypotheses organized according to the research questions, and derived from established theoretical models and prior empirical evidence.

H1.1 (Trust trajectory). Trust ratings were expected to increase significantly from baseline (T1) to post-drive assessment (T4) across repeated exposure to the automation system (Lee and See, 2004; Hoff and Bashir, 2015).

H1.2 (Educational influence on trust development). Educational framing was expected to influence how trust developed across time (T1 to T4). Limitation-focused instruction highlighting system limitations could foster better-calibrated trust through improved understanding of boundaries (DeGuzman et al., 2020), whereas accumulated reliable experience might override brief instructional effects (Lee and See, 2004; Hoff and Bashir, 2015).

H2 (Educational effects on mind-wandering). Limitation-focused education was expected to reduce mind-wandering relative to capability-focused and Basic instruction by encouraging active monitoring and sustaining achieved monitoring effort (α′), thereby reducing the situation-awareness discrepancy (δ) defined in prior work (Chouchane et al., 2022). This prediction aligns with theories of situation awareness and vigilance suggesting that higher monitoring engagement preserves attentional resources during prolonged automation (Endsley, 1995b; Young and Stanton, 2002).

H3 (Educational effects on gaze behavior). Limitation-focused education was expected to produce more structured gaze patterns than the other conditions, characterized by more frequent recovery transitions to the road center and less peripheral attention (Price et al., 2019; Chouchane et al., 2022).

H4 (Educational effects on takeover performance). Participants receiving limitation-focused education were expected to show shorter reaction times during the takeover task, consistent with evidence that sustained attentional engagement and heightened situation awareness facilitate timely intervention (Endsley, 1995b; Greenlee et al., 2018).

H5.1 (Trust-mind-wandering relationship). Greater trust development (T1 to T4 change) was expected to correlate positively with mind-wandering, as increased confidence in system reliability reduces perceived monitoring demand (Lee and See, 2004; Parasuraman and Riley, 1997; Hergeth et al., 2016; Walker et al., 2019).

H5.2 (Mind-wandering-gaze relationship). Higher mind-wandering scores were expected to be associated with less structured scanning, including fewer mirror-to-road-center recoveries and more peripheral glances (Louw and Merat, 2017; Merat et al., 2014).

Together, these hypotheses address the research questions by linking educational framing, trust calibration, attentional state, gaze behavior, and takeover performance during Level 2 automation.

2 Materials and methods

2.1 Participants

Data were collected from 55 participants recruited through a third-party agency engaged by the Japan Automobile Research Institute (JARI), where the study took place. All participants reported no prior hands-on experience with Level 2 automation systems, and therefore represented an automation-naïve sample at the time of testing.

The study aimed for approximately 20 participants per educational condition to provide a balanced design and reasonable sensitivity for detecting medium-to-large effects. Participants were randomly assigned to one of three groups: Basic Education (n = 20), capability-focused (Method 1, n = 15), and limitation-focused (Method 2, n = 20). The slight imbalance arose from normal variation in attendance and early withdrawals rather than any planned differences in allocation.

Of the 55 participants in the final sample, 16 self-identified as female and 39 as male (Table 1). Sex distribution was broadly comparable across groups: the Basic group comprised 6 females and 14 males, the capability-focused group 5 females and 10 males, and the limitation-focused group 5 females and 15 males. Age and self-reported annual driving mileage by education group are summarized in Table 2.

Table 1

GroupFemaleMaleTotal
B61420
M151015
M251520
Total163955

Contingency table of participants by sex and education condition.

B, Basic education; M1, capability-focused education; M2, limitation-focused education.

Table 2

Age (years)Annual distance (km)
StatisticBM1M2BM1M2
Valid cases201520201520
Missing000000
Mean40.236.743.616,4007,40714,900
Standard deviation11.610.210.913,8804,78214,950
Minimum2324232,0001001,000
Maximum57565950,00015,00070,000
25th percentile302934.57,2502,00010,000
50th percentile39364310,00010,00010,000
75th percentile4942.55224,25010,00015,000

Descriptive statistics for age and annual driving distance (km/year) by education condition.

B, Basic education; M1, capability-focused education; M2, limitation-focused education.

All participants provided informed consent prior to the study. The protocol was reviewed and approved by the JARI Research Ethics Committee. No personally identifying data were retained; all datasets were anonymised using numerical identifiers.

2.2 Experimental design

The study employed a 3 × 4 mixed factorial designwith Educational Condition (Basic, capability-focused, limitation-focused) as a between-subjects factor and Time (T1, T2, T3, T4) as a within-subjects factor corresponding to four measurement points throughout the session. This structure (Figure 1) enabled examination of both immediate and evolving effects of instructional framing across the automation experience, capturing trust calibration and behavioral outcomes over time.

Figure 1

2.3 Educational interventions

Educational briefings were designed to examine how different communication strategies shape drivers' understanding of automation and their expectations of system behavior. Each briefing conveyed the same core informational content: the automation supports driving but still requires active human supervision. The briefings differed in instructional emphasis (capability-focused versus limitation-focused framing). The guiding question was: How should information about system capabilities and limitations be conveyed to adjust driver expectations appropriately?

Educational content was delivered through scripted verbal and written briefings that were standardized across participants within each condition. The three conditions held duration, and delivery format constant, while varying the instructional emphasis (capability-focused versus limitation-focused framing) relative to the reference briefing.

Basic education (reference condition): Participants received only the minimum information typically provided in current vehicle manuals or pre-drive explanations. The briefing stated: “The driving automation system you will use today is a partially automated driving system. When the system is activated, the vehicle can drive automatically without pedal or steering operation. However, the driver must monitor the road and surrounding environment. When an alert occurs, you must take over control of the vehicle.” This condition served as a reference point reflecting minimal onboarding information.

Capability-focused education (method 1): This briefing explained how the automation system functions and perceives its surroundings. It described the system's control of steering and speed on highways up to 100 km/h, and how front and rear cameras detect surrounding vehicles and obstacles. Participants were told that system accuracy depends on image quality, which can deteriorate in poor weather. They were informed that the system issues an alert and disengages when reaching its operational limits, although delayed alerts may occur if sensing is impaired. This education aimed to build a general understanding of how the system works and why its performance changes under certain conditions. Visual aids illustrated camera positions and sensor coverage. The full script is provided in Appendix 4.12.2.

Limitation-focused education (method 2): This briefing conveyed the same core message as Method 1 but focused on operating boundaries. It explained that the system controls steering and pedals, functions in normal traffic and weather, and adjusts speed within its lane up to 100 km/h. Drivers were informed that visibility loss in rain or fog could degrade sensor performance and delay alerts. The message emphasized that when the system reaches its limits, it will issue an alert and disengage. Visual aids showed how adverse weather affects sensing. This framing encouraged drivers to monitor conditions and prepare to take control when necessary. The full script is provided in Appendix 4.12.3.

All briefings were delivered by the same experimenter using standardized scripts to ensure consistency. Each lasted approximately 5 min. Participants were unaware that different educational versions existed or that the study examined trust calibration.

These framings were informed by Rasmussen's Skills-Rules-Knowledge framework (Rasmussen, 1983). The capability-focused condition provided conceptual information about how the system perceives the environment and controls speed and steering. The limitation-focused condition emphasized operating boundaries, environmental constraints, and the driver's responsibility to intervene when sensing is impaired. The Basic condition offered only minimal information typical of current onboarding. These distinctions allowed us to test how different types of pre-drive information shape supervision during Level 2 automation.

2.4 Driving simulator environment and critical scenario

The experiment was conducted using the Japan Automobile Research Institute's omnidirectional driving simulator, a high-fidelity system featuring a complete vehicle cabin mounted on a motion platform with 360-degree visual projection (Figure 2). The simulator provides realistic vehicle dynamics, including motion cueing for acceleration, braking, and lateral movement, which enhances the ecological validity of driver responses during automation supervision and takeover events.

Figure 2

The simulated environment depicted a two-lane, two-direction, Japanese Expressway divided by a barrier, with moderate traffic density. Traffic consisted of computer-controlled vehicles programmed to maintain realistic speeds and following distances. The Level 2 automation system controlled the simulator vehicle's steering and speed throughout a 15-min driving automation segment, maintaining position in the center lane at approximately 80 km/h and adjusting speed to maintain safe following distance from lead vehicles.

The critical takeover scenario (Figure 3) occurred at the conclusion of the 15-min automated segment. A lead vehicle performed an emergency braking maneuver that exceeded the automation's deceleration capability. The system issued a visual and auditory takeover request requiring the driver to resume manual control and apply braking to avoid collision. This scenario was designed to assess drivers' readiness to intervene during safety-critical events following extended automation supervision.

Figure 3

2.5 Measures and data collection

2.5.1 Trust in automation

Trust in the driving automation system was measured using the same single-item question as Abe et al. (2017). Participants responded to the prompt, “At this moment, how much do you trust the driving automation system?,” by marking a value on a continuous 0-100 scale anchored at 0 = “not at all” and 100 = “completely”. The question was administered in Japanese, with both the original Japanese and English wordings provided in Appendix 4.10. Higher scores represent greater subjective trust in the automation.

Ratings were collected at four points during the experiment: baseline (T1), after the education session (T2), after the practice drive (T3), and after the driving automation session (T4). This schedule was designed to capture both the immediate effects of the education and later changes that occurred through direct interaction with the automation. A concise single-item scale was chosen to enable repeated measurement without interrupting the driving task, consistent with prior work showing that simple rating scales can effectively capture moment-to-moment changes in trust toward automation (Lee and Moray, 1992; Muir, 1996).

2.5.2 Mind-wandering

Mind-wandering was measured using a single-item retrospective scale, adapted from Mars et al. (2014) and translated into Japanese. Immediately after the takeover event, participants answered the question: “Compared with manual driving, how much time did you spend thinking about things unrelated to driving during the automated drive (before the alert)?” They marked their response on a continuous horizontal line ranging from 0% (“always thinking about driving as during manual driving”) to 100% (“never thinking about driving”). The complete wording in English and Japanese is provided in Appendix 4.11.

A retrospective self-report method was chosen to preserve the natural flow of the automated driving experience and to avoid intrusive probe techniques (e.g., SAGAT), which can disrupt ongoing cognitive processing and alter task engagement (Endsley, 1995a, 1988). This format allowed participants to reflect on their attention across the entire automated segment rather than at single moments. As noted by Smallwood and Schooler (2015), mind-wandering is an internal, subjective state that is most effectively studied through self-report, especially when researchers aim to capture general fluctuations in attention without disturbing task performance.

Although retrospective reports depend on memory and may involve estimation error, this approach was appropriate here because it maintained an undisturbed and realistic supervision context. It therefore provided an ecologically valid measure of participants' overall attentional disengagement during automation.

2.5.3 Gaze behavior

Eye-tracking data were recorded throughout the driving automation segment using a head-mounted NAC EMR-9 system sampling at 60 Hz. Gaze coordinates were manually annotated frame by frame into nine predefined Areas of Interest (AOIs) following ISO 15007:2020.

The AOI definitions, frame-by-frame annotation protocol, and transition-probability construction procedures follow our previously reported transition-analysis pipeline (Chouchane et al., 2026). These comprised the road center (RC; AOI 1), rear-view mirror (RVM; AOI 2), right side mirror (AOI 3), left side mirror (AOI 4), human-machine interface (HMI; AOI 5), right periphery (RP; AOI 6), left periphery (LP; AOI 7), driver side window (8) and other scene areas (AOI 9) (Figure 4). AOIs were annotated manually frame-by-frame by a single trained annotator using a predefined annotation protocol, with decision rules for ambiguous frames to support consistency despite head-movement artifacts. Additional details of the annotation workflow and quality-control procedures are reported in our transition-analysis pipeline description (Chouchane et al., 2026).

Figure 4

Gaze behavior was analyzed within a 24-s window centered on a side-vehicle cut-in event (beginning 15 s before and ending 5 s after the 4-s maneuver, occurring at 11 min 40 s into the automated drive) (Figure 5). This epoch was selected to capture supervisory attention after extended stable automation and to elicit lateral monitoring followed by recovery to forward scanning; conditions under which vigilance decrements are likely to emerge.

Figure 5

Previous research on spontaneous gaze during driving automation has shown that static gaze allocation captures meaningful aspects of supervision but fails to represent the gaze structure or sequencing of glances (Chouchane et al., 2022, 2026). Building on these insights, the present analysis adopted a dynamic approach using first-order Markov modeling to quantify transition probabilities between AOIs, thereby describing how drivers organized visual scanning. Similar transition-based approaches have demonstrated that gaze sequences between the road center and mirrors better discriminate attentive from out-of-the-loop states than static time-on-AOI measures (Schnebelen et al., 2020). Accordingly, static metrics were treated as complementary and dynamic indicators served as the primary basis for interpretation.

Two composite indices were constructed to summarize supervisory scanning structure. The Recovery to Road Center (RRC) captured how consistently participants redirected attention from mirror locations to the forward roadway, operationalising the reintegration of peripheral information essential to Level 1 and Level 2 situation awareness (Endsley, 1995b). It was computed as:

The Structured Scanning Index (SSI) provided a z-scored, directionally weighted composite integrating recovery frequency, sustained central monitoring, and reduced peripheral dwelling:

where TGTRC is total glance time on the road center and TGTLP, TGTRP are total glance times on the left and right peripheries, respectively. Higher SSI values indicate more structured, forward-anchored scanning routines. Although neither RRC nor SSI directly measure “being in the loop,” both serve as operational proxies for active supervisory control.

In addition, HMI monitoring (the proportion of fixations directed toward instrument-cluster and center-console displays) was analyzed as an exploratory metric to examine how educational framing influenced attention to automation-status information.

Together, these measures capture supervisory gaze at multiple levels: static allocation, dynamic transition structure, and system-information monitoring. Under Malleable Attentional Resources Theory (Young and Stanton, 2002), prolonged low-demand automation can contract attentional investment, which may manifest as simplified or less structured scanning. The present framework quantifies such reorganization empirically through RRC and SSI, offering transferable indicators of supervisory engagement in Level 2 driving.

2.5.4 Takeover performance

Takeover performance was evaluated using the System-recognized reaction time (SRRT) metric, defined as the interval between takeover request (TOR) onset and the first driver input exceeding the automation disengagement thresholds. Automation disengagement was defined as driver input exceeding approximately 10% accelerator stroke, 5% brake stroke, or 5 Nm steering torque, marking effective re-engagement with manual control. These thresholds were applied consistently across all participants to ensure comparability of SRRT values.

2.6 Procedure

Each session followed a standardized 80-min protocol designed to capture behavioral, physiological, and self-report data across all phases of the automation experience while ensuring participant comfort. Figure 6 illustrates the complete timeline and timing of trust assessments (T1 to T4).

Figure 6

Upon arrival at the simulator facility, participants provided written informed consent and completed demographic questionnaires. The session then progressed through the following sequence:

T1 (Baseline trust measurement). Participants first rated their initial level of trust in automation before receiving any detailed system information.

Educational intervention and T2 (Post-education trust).

Participants were randomly assigned to one of three instructional conditions (Basic, capability-focused, or limitation-focused; see Section 2.3). The assigned briefing was delivered by the experimenter while participants remained seated in the briefing room. Immediately after the briefing, participants provided a second trust rating (T2) to assess the immediate effect of the educational content.

Simulator familiarization. Participants then moved to the driving simulator and completed a brief manual driving segment to acclimate to the simulator's steering and pedal dynamics and reduce novelty effects.

Practice drive and T3 (Post-practice trust). Participants then engaged the Level 2 automation during a short, non-critical practice segment on a straight motorway section. This phase provided hands-on experience with system operation and the human-machine interface. A third trust rating (T3) was collected immediately afterward.

Eye-tracking calibration. Before the main drive, a 9-point calibration was performed for each participant using the NAC EMR-9 system, following established protocols to ensure accuracy throughout data collection.

Main automated drive. Participants supervised the Level 2 automation for approximately 15 minutes on a simulated highway with moderate traffic. Several non-critical traffic interactions were embedded to sustain ecological validity, and continuous eye-tracking and vehicle control data were recorded. Participants were instructed to monitor the system and remain ready to resume control at any time.

Critical takeover scenario. At the end of the automated segment, the lead vehicle performed sudden braking, triggering a visual and auditory takeover request prompting the driver to resume manual control. Takeover reaction time was recorded automatically by the simulator.

T4 (Post-drive trust) and mind-wandering assessment. Immediately after the takeover event, participants completed the final trust questionnaire (T4) and a retrospective mind-wandering measure. A structured debriefing followed to collect qualitative feedback about the automation experience.

2.7 Statistical analysis

All analyses were performed using standard parametric and non-parametric methods according to data characteristics. Outliers were screened prior to hypothesis testing, and assumption checks were conducted for normality, sphericity, and homogeneity of variance. When Mauchly's test indicated violation of sphericity, Greenhouse-Geisser correction was applied, with adjusted degrees of freedom and epsilon (ε) values reported. Levene's tests verified that the assumption of equal variances was met for these between-group comparisons.

2.7.1 Trust development

Trust ratings were analyzed using a two-way repeated-measures ANOVA with Time (T1 to T4) as the within-subjects factor and Educational Condition (Basic, capability-focused, limitation-focused) as the between-subjects factor. Greenhouse-Geisser correction was applied where required, and effect sizes are reported as partial eta-squared ().

2.7.2 Mind-wandering

Self-reported mind-wandering scores (MWS) were compared across education groups using one-way ANOVA. Pearson correlations were used to assess associations between mind-wandering, trust change (T4 minus T1), gaze indices (RRC, SSI), and takeover reaction time (SRRT). Effect sizes are reported as Pearson's r.

2.7.3 Gaze behavior

Supervisory gaze metrics were analyzed with Educational Condition as the between-subjects factor. Pre-registered scanning composites (RRC, SSI) were evaluated with Kruskal-Wallis tests due to non-normality; HMI-focused metrics (TGT5, MGD5, MGR5) were analyzed with one-way ANOVA when assumptions were met, otherwise with Kruskal-Wallis. Post-hoc comparisons used Benjamini-Hochberg false discovery rate (FDR) control within metric families (Benjamini and Hochberg, 1995). Both unadjusted p and adjusted qBH values are reported. Pearson (or Spearman, when appropriate) correlations assessed associations between gaze metrics and mind-wandering.

2.7.4 Takeover performance

System-recognized reaction times (SRRT) were slightly right-skewed due to one valid long reaction time (22.9 s) observed in the Basic group. Because this reflected a genuine delayed takeover rather than a measurement error, the value was retained, and non-parametric methods were used. A Kruskal-Wallis test compared SRRT across Educational Conditions. When omnibus results were significant, Dunn pairwise comparisons with Holm correction were conducted. Associations between SRRT, mind-wandering, and gaze metrics were examined using Spearman correlations. Effect sizes are reported as rank-based epsilon-squared () for Kruskal-Wallis tests and Spearman's ρ for correlations.

All tests were two-tailed with an α level of .05. Effect sizes are interpreted as partial for ANOVA, for non-parametric analyses, and r for correlations.

3 Results

3.1 Trust development across time and educational conditions

This section addresses RQ1 (trust dynamics) and evaluates H1.1-H1.2.

Findings from the repeated-measures ANOVA (Table 3) address H1.1 (main effect of Time) and H1.2 (Time × Group), both involving the within-subject factor Time. Outcomes from the between-subjects tests (Table 4) address the between-group component of H1.2 (main effect of Group). Levene's tests verified that the assumption of equal variances was met for these between-group comparisons.

Table 3

CasesSphericity correctionSum of squaresdfMean squareFp
TimeGreenhouse-Geisser7131.0281.6944,209.35621.156<0.0010.092
Time × GroupGreenhouse-Geisser540.7803.388159.6080.8020.5090.007
ResidualsGreenhouse-Geisser16,516.19083.010198.965

Results of repeated-measures ANOVA for within-subjects effects on trust across time.

Degrees of freedom adjusted with Greenhouse-Geisser.

Table 4

EffectSum of squaresdfMean squareFp
Group1158.0122579.0060.6210.5410.015
Residuals45,652.38149931.681

Between-subjects effects for education group on mean trust.

3.1.1 Main effect of time (H1.1)

A repeated-measures ANOVA with Time (T1 to T4) as the within-subjects factor and education group as the between-subjects factor revealed a significant main effect of Time on trust ratings, F(1.694, 83.010) = 21.156, p < 0.001, partialη2 = 0.092 (Greenhouse-Geisser corrected) (Table 3). Descriptive analyses indicated a consistent upward trend in trust scores across all educational groups from the initial time point (T1) to the final assessment (T4) (see Figure 7). Participants in the Basic group showed an increase from M = 58.50 (SD = 19.54) at T1 to M = 68.00 (SD = 16.73) at T4. Those in the Method 1 group improved from M = 50.67 (SD = 19.81) to M = 64.00 (SD = 22.30), while Method 2 participants demonstrated a rise from M = 52.25 (SD = 18.39) to M = 71.75 (SD = 13.70). Descriptive statistics by group and time are in Table 5. These findings support H1.1, which proposed that trust would rise from the baseline (T1) to the post-drive stage (T4) as participants became more familiar with the automation system.

Figure 7

Table 5

Trust T1Trust T2Trust T3Trust T4
StatisticBM1M2BM1M2BM1M2BM1M2
Valid cases201520201520201520201520
Mean58.550.752.360.056.057.568.568.070.368.064.071.8
Standard error4.375.114.113.975.593.554.434.903.453.745.763.06
Standard deviation19.519.818.417.821.715.919.819.015.416.722.313.7
Minimum20205202030202040302050
Maximum9080809080809090909090100

Descriptive statistics for trust scores at each time point (T1 to T4) by education group.

B, Basic; M1, capability-focused; M2, limitation-focused. T1 to T4, baseline, post-education, post-practice, post-drive.

3.1.2 Education group effects (H1.2)

Overall trust did not differ by education group, F(2,49) = 0.621, p = 0.541, (Table 4). The Time × Group interaction was also non-significant, F(3.388,83.010) = 0.802, p = 0.509, η2 = 0.007 (Table 3), suggesting that educational framing did not lead to significant differences in how trust evolved over time. These results do not support H1.2, which predicted that educational framing would influence the development of trust across time (T1 to T4).

3.2 Driver education and mind-wandering

This section addresses RQ2 (mind-wandering) and evaluates H2, followed by RQ5 (relationships) for H5.1.

After identifying the general pattern of trust development, we next explored whether the educational interventions affected mind-wandering during automation supervision.

3.2.1 Education group effects on mind-wandering (H2)

A one-way ANOVA tested for differences in self-reported mind-wandering scores (MWS) at T4. The analysis showed no significant group differences, F(2,52) = 1.571, p =.218, η2 = 0.057 (Table 6). Descriptively, participants in the Method 2 condition reported higher mind-wandering (M = 38.75, SD = 23.39) than those in the Basic (M = 27.50, SD = 17.43) and Method 1 groups (M = 28.67, SD = 24.46; see Table 7 and Figure 8). This pattern contradicts the expectation that limitation-focused education would help reduce mind-wandering. These results do not support H2, which predicted that limitation-focused education would reduce mind-wandering compared with capability-focused and basic instruction.

Table 6

SourceSum of squaresdfMean squareFpω2
Group1482.4622741.2311.5710.2180.0570.020
Residuals24,542.08352471.963

One-way ANOVA results for mind-wandering across education groups.

Table 7

GroupNMeanSDSECoefficient of variation
B2027.5017.433.900.63
M11528.6724.466.320.85
M22038.7523.395.230.60

Descriptive statistics for mind-wandering (MWS) by education group.

Figure 8

3.2.2 Trust development and mind-wandering relationship (H5.1)

A Pearson correlation was conducted to examine the relationship between trust change (T4 minus T1) and mind-wandering scores. The analysis showed a significant positive correlation, r = 0.296, p = 0.028, 95% CI [0.034, 0.521] (Table 8). These results support H5.1. Although the correlation was weak (r = 0.296), a statistically significant positive association was found, indicating that participants who showed larger trust gains also reported more frequent mind-wandering.

Table 8

StatisticTrust change ↔ MWSValue
Pearson's r0.296
p-value0.028
95% CI (lower)0.034
95% CI (upper)0.521
Fisher's z0.305
SE (Fisher's z)0.139

Correlation between trust change and mind-wandering (MWS).

3.3 Gaze behavior and performance outcomes

This section addresses RQ3 (visual scanning) and evaluates H3, then addresses RQ5 for H5.2, and finally addresses RQ4 (takeover performance) for H4.

3.3.1 Education group effects on gaze metrics (pre-registered composites) (H3)

Descriptive statistics for the main gaze measures are shown in Table 9. To test whether educational framing influenced supervisory gaze, omnibus Kruskal-Wallis tests were conducted on the preregistered key dependent variables (KDVs): Recovery to Road Center (RRC) and Structured Scanning Index (SSI). As shown in Table 10, none of these measures differed significantly across groups (RRC: H(2) = 1.37, p =.505; SSI: H(2) = 3.10, p =.212). These results do not support H3.

Table 9

MetricB (mean ± SD)M1M2n
Rear-view mirror → Road center (prob.)0.548 ± 0.3480.357 ± 0.3700.505 ± 0.37642
Recovery to Road center (RRC)2.543 ± 0.9372.560 ± 0.8922.293 ± 1.19447
Structured Scanning Index (SSI, z)0.100 ± 0.411−0.134 ± 0.5220.041 ± 0.69947
Mean glance duration: Mirrors (s)0.563 ± 0.3880.455 ± 0.2470.500 ± 0.29447
Mean glance duration: Periphery (s)0.406 ± 0.1410.459 ± 0.1680.458 ± 0.28647
Mean glance rate: Periphery (Hz)0.082 ± 0.0420.135 ± 0.0750.094 ± 0.07247
Total glance time: Periphery (s)1.865 ± 0.9413.209 ± 1.7041.963 ± 1.47047
Total glance time: Road center (s)11.716 ± 4.95111.178 ± 4.80111.776 ± 5.87447

Descriptive statistics for primary gaze outcomes and key derived static indicators.

B, Basic; M1, capability-focused; M2, limitation-focused.

Table 10

MetricTestHp
Recovery to Road center (RRC)Kruskal-Wallis1.3670.505
Structured Scanning Index (SSI)Kruskal-Wallis3.1020.212

Omnibus Kruskal-Wallis tests for pre-registered scanning dependent variables.

Within the static AOI measures (Table 11), two HMI-related metrics showed significant group differences after Benjamini–Hochberg correction: HMI total glance time (TGT5), H(2) = 11.02, p =.004, qBH =.036, and HMI mean glance duration (MGD5), H(2) = 11.28, p =.0036, qBH =.032. Post hoc Dunn tests indicated higher values for Method 1 compared with both Basic and Method 2 groups (pairwise q's = 0.006–0.015). HMI glance rate (MGR5) followed the same pattern but did not remain significant after FDR adjustment (p =.006, qBH = 0.057).

Table 11

MetricHpqBHB vs M1 qM1 vs M2 qB vs M2 q
TGT5 (HMI total time)11.020.0040.0360.0060.0150.718
MGD5 (HMI mean glance duration)11.280.0040.0320.0080.0080.985
MGR5 (HMI glance rate)10.140.0060.0570.0080.0230.621

HMI-centered static gaze metrics: omnibus Kruskal-Wallis tests and Benjamini-Hochberg-adjusted Dunn post hocs.

Exploratory transition analyses (Table 12) revealed several nominal trends (e.g., Right-periphery → HMI p = 0.051), but none survived FDR correction within the transition family. Overall, preregistered scanning metrics showed no educational effects, although exploratory HMI measures suggested greater display monitoring for the Method 1 group in the scenario.

Table 12

TransitionnHp (uncorr.)pFDR
Right periphery → HMI455.9650.0511.000
Left periphery → HMI315.1980.0741.000
Rear-view mirror → HMI424.9930.0821.000
Road center → HMI454.9230.0850.896
Right periphery → Left periphery454.7360.0940.787

Exploratory transition probabilities with the smallest omnibus p-values.

3.3.2 Mind-wandering and gaze behavior relationship (H5.2)

More structured scanning was associated with lower mind-wandering. Participants with higher Recovery to Road Center (RRC) scores reported less mind-wandering (Table 13; Pearson r = −0.329, p =.024, 95% CI [−0.563, −0.046]; Spearman ρ = −0.367, p = 0.011). The Structured Scanning Index (SSI) showed a similar pattern (Pearson r = −0.364, p = 0.012, 95% CI [−0.590, −0.086]; Spearman ρ = −0.367, p = 0.011). Although these correlations were weak to modest in magnitude, both measures showed a consistent negative association with self-reported mind-wandering. Together, the results indicate that structured scanning accounted for roughly 11–13 per cent of the variance in mind-wandering (r2), supporting H5.2.

Table 13

xynPearson rp95% CISpearman ρp
RRCMWS47−0.3290.024[−0.563, −0.046]−0.3670.011
RRCSRRT (s)470.0900.545[−0.200, 0.370]−0.1600.287
SSIMWS47−0.3640.012[−0.590, −0.086]−0.3670.011
SSISRRT (s)470.1100.480[−0.190, 0.380]−0.1100.473

Correlations between structured scanning composites and mind-wandering (MWS) and system-recognized reaction time (SRRT).

3.3.3 Takeover performance (H4)

System-recognized reaction times (SRRT) differed significantly across educational groups, H(2) = 7.898, p =.019, , 95% CI [0.024, 0.381] (Table 14). Median SRRTs were shortest for the Method 1 group (Mdn = 2.142 s), followed by Method 2 (Mdn = 2.770 s) and Basic groups (Mdn = 2.715 s; see Table 15 and Figure 9). Post hoc Dunn tests with Holm correction confirmed a significant difference between Method 1 and Method 2 (pHolm =.015), with no other contrasts reaching significance (Table 16). These results do not support H4, which predicted shorter SRRTs for participants in the limitation-focused education condition.

Table 14

EffectHdfpRank ε2 (95% CI)
EstimateLowerUpper
Group7.89820.0190.1520.0240.381

Kruskal-Wallis omnibus test for SRRT across education groups.

Table 15

GroupnMeanSDMedianQ1Q3MinMax
B183.6514.8372.7152.1072.8931.55022.90
M1142.1270.5572.1451.7232.4601.2303.00
M2212.8610.7292.7702.4303.1301.7705.02

Descriptive statistics for SRRT by education group.

B, Basic; M1, capability-focused; M2, limitation-focused. One high outlier (22.9 s) was retained.

Figure 9

Table 16

Comparisonzrrbp (raw)pHolm
B vs M11.8040.3730.0710.143
B vs M2−1.0090.1880.3130.313
M1 vs M2−2.8020.5650.0050.015

Dunn post hoc pairwise comparisons for SRRT with Holm correction.

4 Discussion

This study investigated whether educational interventions could calibrate trust and sustain supervisory attention during Level 2 driving automation. The results suggested an asymmetry: trust formation appeared to be primarily associated with direct exposure to automation performance, whereas educational framing was associated with differences in HMI monitoring and takeover readiness, while broader scanning structure remained unaffected. Exposure was the dominant factor shaping trust trajectories (H1.1–H1.2), while capability-focused education was associated with differences in HMI monitoring and takeover readiness (H4). Trust growth was also associated with increased mind-wandering (H5.1), and mind-wandering correlated with less structured scanning (H5.2), linking subjective and behavioral markers of supervisory disengagement. Collectively, these findings help to clarify both the limitations and the potential, targeted benefits of pre-drive education for supporting safe human-automation interaction (Table 17).

Table 17

HypothesisPredictionResultSupport
H1.1Trust increases across T1 to T4 during Level 2 automation.Significant main effect of Time; trust rose consistently across groups.
H1.2Educational framing influences trust trajectory across time.No Group or Time by Group effects; trajectories were comparable.×
H2Limitation-focused education reduces mind-wandering compared with other groups.No significant group differences in mind-wandering.×
H3Educational framing affects structured scanning (RRC, SSI).No group differences in preregistered scanning composites.×
H4Limitation-focused education produces the fastest takeover reactions.Group differences observed, but Method 1 (capability-focused) was faster than Method 2 (limitation-focused); prediction not met.×
H5.1Greater trust increase is associated with higher mind-wandering.Weak but significant positive correlation (r = 0.296, p = 0.028).
H5.2More structured scanning is linked to lower mind-wandering.Weak to modest but significant negative correlations for RRC and SSI (r = −0.329, p = 0.024; r = −0.364, p = 0.012).

Summary of hypothesis testing outcomes.

RRC, Recovery to Road Center; SSI, Structured Scanning Index.

□ = Supported; × = Not supported.

4.1 Trust development is primarily exposure driven

Trust increased across all educational conditions as participants gained exposure to Level 2 automation, supporting H1.1. This pattern is consistent with earlier evidence that direct interaction with automation gradually aligns trust with actual system performance, even when users begin with different initial information (Beggiato and Krems, 2013; Lee and See, 2004; Hoff and Bashir, 2015). Neither capability-focused nor limitation-focused framing appeared to alter this trajectory (H1.2), suggesting that direct exposure to the system's behavior outweighed prior instruction.

This pattern aligns with the CAUSE model (Rowan, 1991), which describes effective communication as a sequential process: establishing credibility, raising awareness, promoting understanding, achieving satisfaction, and supporting enactment. The brief instructional framings in this study were unlikely to complete this process; instead, participants appeared to calibrate trust through their direct observation of system reliability. In this context, performance feedback rather than message framing likely served as the main influence on trust development.

The positive correlation between trust growth and mind-wandering (H3) indicated a potential tension between confidence in the system and supervisory engagement. Participants who showed greater increases in trust also showed lower values on monitoring-related indicators, consistent with a trust-associated reduction in supervisory engagement. This pattern is consistent with Malleable Attentional Resources Theory (MART) (Young and Stanton, 2002; Warm et al., 2008), which proposes that attentional investment contracts under low task demand. The relationship was observed across educational conditions, suggesting that reduced monitoring may emerge generally as drivers gain experience with stable automation.

4.2 Capability-focused education and monitoring priorities

Although capability-focused education did not substantially alter trust trajectories, it appeared to yield two modest benefits (H8): greater attention to the HMI and shorter system-recognized reaction times during takeover. Participants who received knowledge-based (capability-focused) instruction monitored automation status displays more frequently and responded more promptly when manual control was required. These findings suggest that a conceptual understanding of system operation may have refined monitoring selectivity and readiness for authority transfer.

Within Endsley's situation-awareness framework (Endsley, 1995b), these improvements may reflect an increased ability to interpret what automation feedback signals about system state. Capability-focused instruction might have rendered HMI cues more meaningful and diagnostic, enabling participants to interpret them more efficiently without compromising roadway monitoring. The faster SRRT further suggests somewhat smoother coordination during manual resumption.

Hence, brief conceptual instruction may provide targeted benefits by sharpening monitoring priorities and readiness, even though it does not seem to mitigate trust-related vigilance loss. These effects illustrate how education and exposure can influence distinct aspects of supervision: exposure shaping trust and perceived monitoring demand, and education fine-tuning particular monitoring behaviors.

4.3 Limitation-focused education and its limited impact

Rule-based (limitation-oriented) education showed little measurable influence on trust, attention, or takeover performance. Several factors may explain this outcome. One possible interpretation is that emphasizing system limits did not translate into actionable monitoring strategies during a stable scenario, which may have limited observable behavioral effects. The instruction listed boundary conditions such as poor weather or degraded lane markings but offered little procedural guidance on how to detect or respond to them. Without actionable cues, participants had limited opportunity to apply these rules during the stable highway scenario.

The motivational tone may also have contributed. Emphasizing system fallibility might have led drivers to expect that failures would occur only under specific conditions, such as poor weather or unclear lane markings. Because the simulated drive presented no such cues, participants may have focused less on continuous monitoring and more on waiting for visible signs of malfunction. This conditional form of supervision differs from the general vigilance loss associated with growing trust, yet it can be viewed as another expression of trust miscalibration, where expectations about when the system might fail are narrower than its actual range of vulnerabilities (Parasuraman and Riley, 1997). Moreover, the drive remained within the automation's operational domain, so the limitation messages were not especially relevant. Under these conditions, participants may have regarded the Limitation-focused briefing as less meaningful, which could have limited its behavioral impact.

4.4 Educational framing and control modes

The Skills-Rules-Knowledge (SRK) framework (Rasmussen, 1983) provides a useful lens for understanding these differences. Limitation-focused instruction depends on stored “if-then” procedures, which apply only when matching cues arise. In this study, such cues were absent, and the learned rules were therefore unlikely to be triggered. Capability-focused education, by contrast, fosters conceptual models that apply across diverse conditions. Understanding how the automation senses and acts may have improved the perceived relevance of HMI information and facilitated faster recognition of control transitions. In Rasmussen's framework, capability-focused processing is required when operators face unfamiliar or unexpected situations for which no stored rules apply. Such conditions are analogous to automation takeovers, where drivers must interpret system behavior and re-establish manual control.

4.5 Integration with theoretical frameworks

These findings can be interpreted within the situation-awareness discrepancy model, δ = α−α′, introduced in earlier work (Chouchane et al., 2022). Within this model, trust is likely to shape perceived monitoring demand (α). As trust increases, drivers may judge that less active supervision is necessary. This change reflects heightened confidence in automation reliability rather than an actual reduction in the monitoring required. The observed association between higher trust and mind-wandering supports this interpretation, indicating that lower perceived demand can promote attentional drift and weaker supervisory engagement.

Educational interventions, by contrast, may affect achieved monitoring (α′) through partially independent pathways. Capability-focused education seemed to enhance certain components of monitoring, such as HMI attention and takeover readiness. These effects represent localized gains in α′, without offsetting the overall decline in perceived demand. Consequently, brief instruction may reinforce specific monitoring habits but is unlikely to counter the wider vigilance contraction associated with rising trust.

Malleable Attentional Resources Theory (MART) (Young and Stanton, 2002; Warm et al., 2008) provides a complementary explanation for this contraction. Under stable, low-demand automation, attentional resources tend to narrow, leading to less structured and less frequent scanning between key visual areas. The observed correlation between mind-wandering and reduced scanning organization (lower SSI and RRC) aligns with this mechanism. The CAUSE model (Rowan, 1991) further clarifies why instructional messages had limited effect: effective risk communication depends on credible and sequenced engagement rather than isolated briefings. Finally, performance-based trust models (Lee and See, 2004; Hoff and Bashir, 2015) help explain why direct exposure to automation performance dominated trust development.

Taken together, these frameworks suggest that trust most plausibly influences perceived monitoring demand, while education can refine selective monitoring behaviors. The combination of rising trust and stable system performance therefore creates conditions in which drivers monitor less actively even though objective supervisory requirements remain unchanged.

4.6 Practical implications

The results suggest that brief pre-drive education has limited influence on trust trajectories, which develop mainly through direct exposure to automation performance. Verbal or written instructions, though valuable for setting initial expectations, appear insufficient to shape how trust evolves during actual use. Nonetheless, such education can enhance comprehension and promote more focused monitoring.

The modest benefits of capability-focused instruction indicate that more extensive, scenario-based training could produce broader effects. Programmes combining conceptual explanation with guided practice might help drivers encounter automation in varied conditions and near its operational limits, allowing better calibration of both trust and monitoring strategies. Real-time feedback from instructors could reinforce how and when to intervene, supporting balanced supervision.

The improvement in HMI monitoring following capability-focused instruction also points to the importance of interface transparency. Displays that provide meaningful cues, such as sensor coverage or automation confidence, can enhance monitoring, provided the information is interpretable and not overwhelming. Education and interface design may therefore work best when developed in tandem.

Finally, the link between scanning structure and mind-wandering highlights opportunities for driver-monitoring systems. Beyond tracking gaze-on-road duration, systems might assess how organized scanning patterns remain over time. Identifying reduced mirror-to-road transitions or increased randomness could enable early alerts before performance declines.

4.7 Limitations

Several factors limit interpretation and generalisability. First, between-group sample sizes were modest (n = 15–20 per educational condition), which limits statistical power to detect small effects and interactions; null findings for group differences should therefore be interpreted cautiously. Second, the educational interventions were brief and delivered once; repeated or interactive training might yield stronger and more durable effects. Third, the simulated drive involved only one takeover event after 15 minutes of automation in stable conditions. Real driving includes greater variability in road, traffic, and weather conditions, which may influence both trust and attention. Fourth, participants were inexperienced with hands-on Level 2 automation use, and the findings therefore reflect early exposure rather than longer-term adaptation; future studies should examine whether educational benefits or trust-attention dynamics persist over extended use. Fifth, mind-wandering was measured retrospectively, which could have underestimated momentary fluctuations; incorporating concurrent or physiological measures would strengthen future assessments. Finally, the instructional materials represented one implementation of capability-focused and limitation-focused framing. Although the briefings differed in emphasis, we did not include a direct manipulation check to verify whether participants internalized the framings as intended. Given the shared informational content across the briefings, it is possible that overlap attenuated between-group differences in trust and gaze metrics, and effects should therefore be interpreted cautiously. Variations in content, delivery medium, or realism, as well as brief post-briefing checks of capability versus boundary understanding, would strengthen future validation of framing-based interventions.

4.8 Future directions

Future research should examine how educational and exposure factors interact over time and across contexts. Longitudinal studies could track trust and monitoring behavior over extended periods of real-world driving. Scenario-based training could identify how conceptual understanding and hands-on exposure combine to calibrate trust and attention effectively. Research in diverse driving environments, including urban and adverse-weather conditions, would clarify when limitation-focused education becomes most relevant. Finally, combining educational interventions with adaptive HMIs and monitoring systems could help sustain driver engagement.

4.9 Conclusions

Brief pre-drive education modulated specific monitoring behaviors but did not influence the trajectory of trust formation during Level 2 automation. Trust increased over time across all groups and was not affected by instructional framing, consistent with experience-driven adjustment to observed system performance. Capability-focused education was associated with greater attention to automation feedback and faster takeover responses compared with limitation-focused education, indicating that conceptual information may refine how drivers monitor system status without altering overall trust development.

Across participants, increases in trust were accompanied by patterns consistent with reduced structured monitoring, suggesting a potential trade-off between growing confidence and sustained supervisory engagement during routine automation. More structured gaze organization was associated with lower self-reported mind-wandering, supporting the interpretation of gaze metrics as behavioral proxies for attentional allocation rather than direct measures of attention.

These findings suggest the potential value of integrated approaches combining extended experiential training, transparent interface design, and gaze-based monitoring systems to support sustained supervisory engagement in partially automated driving.

Statements

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Japan Automobile Research Institute, Research Ethics Committee. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

HC: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing. YS: Methodology, Software, Writing – review & editing. KS: Funding acquisition, Project administration, Software, Writing – review & editing. GA: Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Writing – review & editing. MI: Conceptualization, Funding acquisition, Investigation, Methodology, Resources, Supervision, Writing – review & editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This work was supported by JSPS KAKENHI Grant Number JP24H00361.

Acknowledgments

The authors thank JARI Autonomous Driving Research Division members for the technical support and valuable discussions throughout the project development.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fcomp.2026.1772813/full#supplementary-material

References

  • 1

    AbeG.SatoK.ItohM. (2017). Driver trust in automated driving systems: The case of overtaking and passing. IEEE Trans. Human-Mach. Syst. 48, 8594. doi: 10.1109/THMS.2017.2781619

  • 2

    BainbridgeL. (1983). “Ironies of automation,” in Analysis, Design and Evaluation of Man-Machine Systems (London: Elsevier), 129135.

  • 3

    BaldwinC. L.RobertsD. M.BarraganD.LeeJ. D.LernerN.HigginsJ. S. (2017). Detecting and quantifying mind wandering during simulated driving. Front. Hum. Neurosci. 11:406. doi: 10.3389/fnhum.2017.00406

  • 4

    BeggiatoM.KremsJ. F. (2013). The evolution of mental model, trust and acceptance of adaptive cruise control in relation to initial information. Transport. Res. Part F: Traffic Psychol. Behav. 18, 4757. doi: 10.1016/j.trf.2012.12.006

  • 5

    BenjaminiY.HochbergY. (1995). Controlling the false discovery rate: a practical and powerful approach to multiple testing. J. Royal Statist. Soc.: Series B. 57, 289300. doi: 10.1111/j.2517-6161.1995.tb02031.x

  • 6

    CampbellJ. L.BrownJ. L.GravingJ. S.RichardC. M.LichtyM. G.BaconL. P.et al. (2018). “Human factors design guidance for level 2 and level 3 automated driving concepts,” in Technical Report.

  • 7

    ChouchaneH.LeeJ.SakamuraY.NakamuraH.AbeG.ItohM. (2026). Supervisory gaze behaviour under different automation durations in level 2 driving: a first-order transition analysis. Appl. Sci. 16:1401. doi: 10.3390/app16031401

  • 8

    ChouchaneH.NakamuraH.SatoK.Antona-MakoshiJ.AbeG.ItohM. (2022). “Identifying the out of the loop phenomenon during driving automation using spontaneous gaze behavior,” in Proceedings of the Human Factors and Ergonomics Society Annual Meeting (Los Angeles, CA: SAGE Publications), 305309.

  • 9

    CooperJ. M.CrabtreeK. W.McDonnellA. S.MayD.StrayerS. C.TsogtbaatarT.et al. (2023). Driver behavior while using level 2 vehicle automation: a hybrid naturalistic study. Cognit. Res.: Princ. Implicat. 8:71. doi: 10.1186/s41235-023-00527-5

  • 10

    DeGuzmanC. A.DonmezB. (2021). Drivers still have limited knowledge about adaptive cruise control even when they own the system. Transp. Res. Rec. 2675, 328339. doi: 10.1177/03611981211011482

  • 11

    DeGuzmanC. A.HopkinsS. A.DonmezB. (2020). Driver takeover performance and monitoring behavior with driving automation at system-limit versus system-malfunction failures. Transp. Res. Rec. 2674, 140151. doi: 10.1177/0361198120912228

  • 12

    EndsleyM. R. (1988). “Design and evaluation for situation awareness enhancement,” in Proc. Human Fact. Soc. Annual Meet. 32, 97101. doi: 10.1177/154193128803200221

  • 13

    EndsleyM. R. (1995a). Measurement of situation awareness in dynamic systems. Hum. Factors37, 6584. doi: 10.1518/001872095779049499

  • 14

    EndsleyM. R. (1995b). Toward a theory of situation awareness in dynamic systems. Hum. Factors37, 3264. doi: 10.1518/001872095779049543

  • 15

    ErikssonA.StantonN. A. (2017). Takeover time in highly automated vehicles: noncritical transitions to and from manual control. Hum. Factors59, 689705. doi: 10.1177/0018720816685832

  • 16

    GoldC.DamböckD.LorenzL.BenglerK. (2013). ““take over!” how long does it take to get the driver back into the loop?,” in Proceedings of the Human Factors and Ergonomics Society Annual Meeting (Los Angeles, CA: Sage Publications), 19381942.

  • 17

    GoldC.KörberM.HohenbergerC.LechnerD.BenglerK. (2015). Trust in automation-before and after the experience of take-over scenarios in a highly automated vehicle. Procedia Manufact. 3, 30253032. doi: 10.1016/j.promfg.2015.07.847

  • 18

    GreenleeE. T.DeLuciaP. R.NewtonD. C. (2018). Driver vigilance in automated vehicles: Hazard detection failures are a matter of time. Hum. Factors60, 465476. doi: 10.1177/0018720818761711

  • 19

    HarmsI. M.BingenL.SteffensJ. (2020). Addressing the awareness gap: a combined survey and vehicle registration analysis to assess car owners' usage of adas in fleets. Transport. Res. Part A: Policy Pract. 134, 6577. doi: 10.1016/j.tra.2020.01.018

  • 20

    HeJ.BecicE.LeeY.-C.McCarleyJ. S. (2011). Mind wandering behind the wheel: Performance and oculomotor correlates. Hum. Factors53, 1321. doi: 10.1177/0018720810391530

  • 21

    HergethS.LorenzL.VilimekR.KremsJ. F. (2016). Keep your scanners peeled: Gaze behavior as a measure of automation trust during highly automated driving. Hum. Factors58, 509519. doi: 10.1177/0018720815625744

  • 22

    HoffK. A.BashirM. (2015). Trust in automation: Integrating empirical evidence on factors that influence trust. Hum. Factors57, 407434. doi: 10.1177/0018720814547570

  • 23

    InagakiT.ItohM. (2013). Human's overtrust in and overreliance on advanced driver assistance systems: a theoretical framework. Int. J. Vehicular Technol. 2013:951762. doi: 10.1155/2013/951762

  • 24

    KhastgirS.BirrellS.DhadyallaG.JenningsP. (2018). Calibrating trust through knowledge: Introducing the concept of informed safety for automation in vehicles. Transport. Res. Part C: Emerg. Technol. 96, 290303. doi: 10.1016/j.trc.2018.07.001

  • 25

    KörberM.BaselerE.BenglerK. (2018). Introduction matters: manipulating trust in automation and reliance in automated driving. Appl. Ergon. 66, 1831. doi: 10.1016/j.apergo.2017.07.006

  • 26

    LeeJ.MorayN. (1992). Trust, control strategies and allocation of function in human-machine systems. Ergonomics35, 12431270. doi: 10.1080/00140139208967392

  • 27

    LeeJ. D.SeeK. A. (2004). Trust in automation: Designing for appropriate reliance. Hum. Factors46, 5080. doi: 10.1518/hfes.46.1.50.30392

  • 28

    LouwT.MeratN. (2017). Are you in the loop? Using gaze dispersion to understand driver visual attention during vehicle automation. Transport. Res. Part C: Emerg. Technol. 76, 3550. doi: 10.1016/j.trc.2017.01.001

  • 29

    LuZ.CosterX.de WinterJ. (2017). How much time do drivers need to obtain situation awareness? A laboratory-based study of automated driving. Appl. Ergon. 60, 293304. doi: 10.1016/j.apergo.2016.12.003

  • 30

    MarsF.DerooM.CharronC. (2014). Driver adaptation to haptic shared control of the steering wheel. In 2014 ieee international conference on systems, man, and cybernetics (SMC) (San Diego, CA: IEEE), 15051509.

  • 31

    MeratN.JamsonA. H.LaiF. C.DalyM.CarstenO. M. (2014). Transition to manual: Driver behaviour when resuming control from a highly automated vehicle. Transport. Res. Part F: Traffic Psychol. Behav. 27, 274282. doi: 10.1016/j.trf.2014.09.005

  • 32

    MeratN.SeppeltB.LouwT.EngströmJ.LeeJ. D.JohanssonE.et al. (2019). The “out-of-the-loop” concept in automated driving: proposed definition, measures and implications. Cognit. Technol. Work21, 8798. doi: 10.1007/s10111-018-0525-8

  • 33

    MerrittS. M.IlgenD. R. (2008). Not all trust is created equal: dispositional and history-based trust in human-automation interactions. Hum. Factors50, 194210. doi: 10.1518/001872008X288574

  • 34

    MuirV. (1996). Trust in automation: part II. Theoretical issues in the study of trust and human intervention in automation systems. Ergonomics39, 429460. doi: 10.1080/00140139608964474

  • 35

    Oviedo-TrespalaciosO.TichonJ.BriantO. (2021). Is a flick-through enough? A content analysis of advanced driver assistance systems (adas) user manuals. PLoS ONE16:e0252688. doi: 10.1371/journal.pone.0252688

  • 36

    ParasuramanR.MolloyR.SinghI. L. (1993). Performance consequences of automation-induced 'complacency'. Int. J. Aviat. Psychol. 3, 123. doi: 10.1207/s15327108ijap0301_1

  • 37

    ParasuramanR.RileyV. (1997). Humans and automation: use, misuse, disuse, abuse. Hum. Factors39, 230253. doi: 10.1518/001872097778543886

  • 38

    PriceM.LeeJ. D.DinparastdjadidA.ToyodaH.DomeyerJ. (2019). Effect of automation instructions and vehicle control algorithms on eye behavior in highly automated vehicles. Int. J. Automot. Eng. 10, 7379. doi: 10.20485/jsaeijae.10.1_73

  • 39

    RabinerL. R.JuangB. (1986). A tutorial on hidden markov models. IEEE ASSp Magazine3, 416. doi: 10.1109/MASSP.1986.1165342

  • 40

    RasmussenJ. (1983). Skills, rules, and knowledge; signals, signs and symbols, and other distinction in human performance models. IEEE Trans. Syst. Man Cybernet. 13, 6157. doi: 10.1109/TSMC.1983.6313160

  • 41

    RowanK. E. (1991). Goals, obstacles, and strategies in risk communication: a problem-solving approach to improving communication about risks. J. Appl. Commun. Res. 19, 300329. doi: 10.1080/00909889109365311

  • 42

    SAE International (2021). “Taxonomy and definitions for terms related to driving automation systems for on-road motor vehicles,” in Technical Report J3016_202104. Warrendale, PA: SAE International.

  • 43

    SchnebelenD.CharronC.MarsF. (2020). Estimating the out-of-the-loop phenomenon from visual strategies during highly automated driving. Accid. Analy. Prevent. 148:105776. doi: 10.1016/j.aap.2020.105776

  • 44

    SmallwoodJ.SchoolerJ. W. (2015). The science of mind wandering: Empirically navigating the stream of consciousness. Annu. Rev. Psychol. 66, 487518. doi: 10.1146/annurev-psych-010814-015331

  • 45

    SwellerJ.AyresP.KalyugaS. (2011). Cognitive Load Theory. Cham: Springer.

  • 46

    UnderwoodG.ChapmanP.BrocklehurstN.UnderwoodJ.CrundallD. (2003). Visual attention while driving: sequences of eye fixations made by experienced and novice drivers. Ergonomics46, 629646. doi: 10.1080/0014013031000090116

  • 47

    WalkerF.WangJ.MartensM. H.VerweyW. B. (2019). Gaze behaviour and electrodermal activity: Objective measures of drivers' trust in automated vehicles. Transport. Res. Part F: Traffic Psychol. Behav. 64, 401412. doi: 10.1016/j.trf.2019.05.021

  • 48

    WarmJ. S.ParasuramanR.MatthewsG. (2008). Vigilance requires hard mental work and is stressful. Hum. Factors50, 433441. doi: 10.1518/001872008X312152

  • 49

    YoungM. S.StantonN. A. (2002). Malleable attentional resources theory: a new explanation for the effects of mental underload on performance. Hum. Factors44, 365375. doi: 10.1518/0018720024497709

Summary

Keywords

driver education, gaze behavior, Level 2 automation, mind-wandering, situation awareness, supervisory attention, take over performance, trust in automation

Citation

Chouchane H, Sakamura Y, Sato K, Abe G and Itoh M (2026) Trust rises, attention falls: divergent effects of exposure and education in driving automation. Front. Comput. Sci. 8:1772813. doi: 10.3389/fcomp.2026.1772813

Received

21 December 2025

Revised

10 March 2026

Accepted

26 March 2026

Published

21 April 2026

Volume

8 - 2026

Edited by

Andrej Košir, University of Ljubljana, Slovenia

Reviewed by

Liza Dixon, University of Ulm, Germany

Jinzhen Dou, Southeast University, China

Updates

Copyright

*Correspondence: Hanna Chouchane,

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics