Abstract
Introduction:
Drivers supervising Level 2 automation must maintain situation awareness while the system controls steering and speed. Miscalibrated trust can contribute to overreliance and lapses in monitoring, whereas insufficient trust leads to disuse. Prolonged supervision is associated with increased mind-wandering, which can slow reactions to critical events. This study tested whether brief educational interventions affect trust, attention, and takeover readiness during Level 2 driving. Our focus on brief interventions reflects the short, time-constrained onboarding that drivers typically receive when adopting driving automation systems.
Methods:
Fifty-five licensed drivers with no prior hands-on experience of Level 2 automation completed a 15-min automated highway drive. Participants received either minimal instruction (Basic), capability-focused education (Knowledge-based), or limitation-focused education (Rule-based). Trust was measured at four time points; additional measures captured self-reported mind-wandering, gaze behavior, and takeover reaction time.
Results:
Trust increased significantly over time in all groups, and educational framing did not alter this trajectory. Capability-focused education enhanced monitoring of the human-machine interface on two false discovery rate corrected metrics and produced faster takeover reactions than limitation-focused education (no difference vs. Basic). Across participants, greater trust growth correlated with higher mind-wandering, while more structured gaze was associated with lower mind-wandering.
Discussion:
Overall, trust formation appeared to be primarily associated with direct experience with system performance, whereas targeted education refined what drivers monitored and how quickly they responded. Together, these results clarify how experience primarily builds trust while education selectively sharpens attention and response readiness in automated driving. These findings clarify distinct roles of experience and brief education in supervising automation and have implications for driver training, human-machine interface design, and gaze-based monitoring.
1 Introduction
Level 2 driving automation, as defined by the Society of Automotive Engineers (SAE International, 2021), combines longitudinal control (adaptive cruise control) and lateral control (lane-keeping assistance) to relieve drivers of continuous manual steering and speed adjustments. Despite this operational support, Level 2 systems explicitly require continuous driver supervision because the automation may encounter situations beyond its operational design domain or capabilities. Drivers must therefore remain ready to resume manual control at short notice when a takeover request is issued or when traffic conditions exceed system limits. In this context, takeover performance is commonly operationalised as the latency between the takeover request and the driver's first control input that initiates manual control, reflecting readiness to re-engage with the driving task (Merat et al., 2014; Gold et al., 2013; Eriksson and Stanton, 2017).
This supervisory role presents a fundamental human-factors challenge. Maintaining sustained monitoring during prolonged automation is mentally demanding, particularly when the system performs reliably and the task provides limited cognitive stimulation (Warm et al., 2008). Under these conditions, drivers may become overly reliant on automation, showing reduced visual scanning of the road environment and slower readiness to intervene when required (Parasuraman et al., 1993; Parasuraman and Riley, 1997; Inagaki and Itoh, 2013). Japan's road-safety initiatives, including the Ministry of Land, Infrastructure, Transport and Tourism's (MLIT) Advanced Safety Vehicle (ASV) Plan, have identified appropriate human supervision of driver-assistance systems as a policy priority. Similar concerns are reflected in international frameworks such as the European Union's Vision Zero strategy and the United Nations' commitment to reducing global road-traffic casualties by 50 per cent by 2030.
Despite progress in automation design, limited understanding remains of how users learn to supervise such systems effectively. This study addresses that gap by examining how pre-drive educational framings influence trust, attention, and takeover readiness during Level 2 automation. It explores whether targeted instruction can shape drivers' mental models and mitigate the attentional drift observed in prior research.
Long-form driver education is rarely delivered for Level 2 systems. In current practices, onboarding is largely informal and relies on passive materials or self-guided learning: most drivers report learning through trial-and-error whilst driving, brochures, or owner manuals, with only a minority receiving any instruction from dealers (Harms et al., 2020; Oviedo-Trespalacios et al., 2021; DeGuzman and Donmez, 2021). Given that such approaches offer limited time for detailed explanation, it is important to determine whether even minimal pre-drive education can shape how users form trust and supervise automation.
This challenge echoes the classic ironies of automation described by Bainbridge (1983), who observed that the more reliable a system becomes, the more crucial and difficult the human operator's monitoring role becomes. As automation handles routine control, drivers are left with the demanding task of supervising rare but safety-critical situations, often without sufficient opportunity to maintain the necessary monitoring and manual-control skills. These ironies are increasingly evident in modern driving automation, where sustained reliability can erode vigilance and delay intervention.
Trust in automation plays a central role in how drivers allocate attention and effort during supervisory tasks (Lee and See, 2004). In this context, trust refers to the belief that an automated system will support the user's goals when outcomes are uncertain or the situation involves risk. Trust influences how people choose to engage with, depend on, or avoid automation. When trust is well calibrated to system capability, drivers are more likely to maintain appropriate vigilance. Poor calibration can increase safety risks: excessive trust may lead to complacency and reduced monitoring (Parasuraman and Riley, 1997; Parasuraman et al., 1993), whereas insufficient trust can result in disuse, where drivers avoid using automation even when it could assist them (Lee and See, 2004; Parasuraman and Riley, 1997).
Trust calibration poses distinct challenges in driving automation, where incorrect assumptions about system capability can lead to immediate safety risks (Khastgir et al., 2018; Inagaki and Itoh, 2013). As Inagaki and Itoh (2013) note, drivers may overtrust systems beyond their functional limits, resulting in over-reliance and delayed intervention. Unlike automation in controlled industrial domains, driving automation operates within dynamic and unpredictable traffic environments that continually test system boundaries (Campbell et al., 2018). Such variability demands that drivers understand not only what the automation can achieve under ideal conditions but also how its performance degrades as environmental complexity increases (Khastgir et al., 2018).
A foundational model for understanding how drivers maintain effective supervision is Endsley (1995b)'s three-level framework of situation awareness, which distinguishes the perception of environmental elements, their comprehension in relation to goals, and the projection of their future status. Maintaining these levels is essential for safe supervisory control and underpins subsequent theoretical developments, including the situation-awareness discrepancy model adopted in this study.
Theoretical models of trust help explain these mechanisms. The Confidence, Awareness, Understanding, Satisfaction, and Enactment (CAUSE) model (Rowan, 1991) is a structured risk-communication framework showing how sequential messaging can build trust and support behavioral change. Communicators first establish confidence, then raise awareness of risks, foster understanding, promote satisfaction with solutions, and finally encourage enactment. In the context of driver education, this suggests that the sequencing and framing of information, not only the content about system performance, shape how users form and calibrate trust. In contrast, trust-calibration models (Lee and See, 2004; Hoff and Bashir, 2015) describe feedback processes through which users compare observed system behavior with prior expectations and adjust their trust accordingly (Merritt and Ilgen, 2008). These experiential mechanisms operate alongside educational processes that shape initial expectations before interaction begins. The Malleable Attentional Resources Theory (MART) (Young and Stanton, 2002; Warm et al., 2008) complements this view by explaining how consistent automation performance can gradually compress attentional resources. When systems operate without error, monitoring becomes confined to abbreviated scanning routines that suffice in routine conditions but may be insufficient when rare critical events occur (Parasuraman et al., 1993; Parasuraman and Riley, 1997; Warm et al., 2008).
Empirical research supports these theoretical accounts (Hoff and Bashir, 2015; Gold et al., 2015). Studies show that both educational information and direct experience with automation performance can influence trust development, although the relative strength of these effects remains to be determined (Körber et al., 2018; Merritt and Ilgen, 2008). After extended exposure to consistent automation performance, drivers reduce their visual attention to the forward roadway, engage more often in secondary tasks, and respond more slowly to unexpected takeover requests (Louw and Merat, 2017; Merat et al., 2014; Gold et al., 2013). Such patterns appear even when drivers are explicitly instructed to maintain vigilance, indicating that experiential learning exerts a strong influence during ongoing interaction with automation (Gold et al., 2013; Hoff and Bashir, 2015).
Beyond trust dynamics, a related challenge is the natural tendency for attention to drift during monotonous supervision. Mind-wandering describes attention shifting away from the external task toward self-generated thoughts and feelings. In this paper we distinguish mind-wandering from the broader phenomenon of attentional drift. Attentional drift refers to the gradual reduction of active task monitoring during low-demand automation supervision, whereas mind-wandering specifically denotes internally generated thoughts that compete with task-related attention. This state is common in low-stimulation settings and carries measurable costs for ongoing performance (Smallwood and Schooler, 2015). In manual driving studies, mind-wandering has been linked to changes in driver behavior and reduced responsiveness, reflecting a shift of attention from the external driving environment to internally generated thought (He et al., 2011; Baldwin et al., 2017). Evidence from takeover research further shows that drivers respond more slowly when not actively monitoring, underscoring the safety relevance of attentional drift (Eriksson and Stanton, 2017).
Driving automation creates conditions that can encourage attentional drift by reducing immediate task demand. Reduced manual control and predictable system behavior lessen immediate task demands and may draw attention away from the roadway (Merat et al., 2019). Such low-stimulation contexts are known to increase mind-wandering (Smallwood and Schooler, 2015). Empirically, automated driving alters visual-monitoring patterns relative to manual control, indicating shifts in supervisory attention (Louw and Merat, 2017). When attention turns inward, drivers typically need additional time to re-engage with the scene and act, a pattern that aligns with longer and more variable takeover responses when monitoring is reduced (Eriksson and Stanton, 2017).
Visual-attention patterns derived from eye gaze provide behavioral indicators that can be interpreted as proxies for attentional engagement (Underwood et al., 2003). Structured visual scanning, characterized by systematic transitions between the forward roadway, mirrors, and peripheral regions, reflects active monitoring of the traffic environment (Underwood et al., 2003). By contrast, fragmented or unsystematic scanning patterns indicate reduced situational awareness and attentional disengagement (Louw and Merat, 2017; Schnebelen et al., 2020). First-order Markov-chain analysis has recently been applied to quantify the organization of gaze transitions, allowing researchers to describe how drivers structure and maintain monitoring routines under automation (Rabiner and Juang, 1986; Schnebelen et al., 2020; Chouchane et al., 2026). To examine these supervisory behaviors objectively, the study analyzed gaze transitions between predefined areas of interest using eye-tracking metrics that capture the organization of visual scanning during automation supervision.
Given these challenges of trust miscalibration and attentional drift, driver education represents a potentially valuable intervention for supporting both calibrated trust and sustained attention during automation supervision. The concept of informed safety emphasizes that users should receive clear, comprehensive information about the capabilities and limitations of automation to calibrate trust appropriately (Campbell et al., 2018; Khastgir et al., 2018). Effective education therefore extends beyond operational instructions to include understanding of sensing principles, boundary conditions, and expected driver roles. Such knowledge supports realistic mental models and helps prevent both over-trust and disuse when supervising automation.
We introduce two contrasting educational framings for Level 2 supervision: Knowledge-based (capability-focused) and Rule-based (limitation-focused) education. This distinction is grounded in established models of cognitive control and supervisory behavior. Rasmussen's Skills-Rules-Knowledge framework (Rasmussen, 1983) distinguishes capability-focused behavior, which relies on reasoning with an internal model of the system in unfamiliar or variable situations, from limitation-focused behavior, which applies stored procedures to familiar signs and boundary conditions. In driving automation, capability-focused information can help drivers build richer mental models of how the system senses and controls the vehicle, consistent with cognitive load theory's emphasis on instruction that supports schema construction and understanding of system structure (Sweller et al., 2011). Such understanding contributes to the process and purpose information that Lee and See identify as a basis for calibrated trust in automation (Lee and See, 2004). Limitation-focused information, by contrast, highlights contexts in which automated performance may degrade and the actions required from the driver, reinforcing limitation-focused preparedness for boundary conditions and helping to reduce the risk of over-trust (Rasmussen, 1983; Lee and See, 2004). These framings therefore target complementary cognitive pathways-conceptual understanding and conditional responding-that may influence how drivers monitor the system and prepare for intervention during supervision.
Prior researc h on educational framing in automation contexts has produced mixed findings. Some studies suggest that providing clear information about system limitations can reduce inappropriate reliance and support monitoring (DeGuzman et al., 2020; Körber et al., 2018). Other accounts emphasize that trust calibration is strongly shaped by performance feedback during use, which may limit the durability of brief pre-drive instruction (Hoff and Bashir, 2015). Methodologically, many studies assess immediate, short-term effects rather than tracking how trust and monitoring evolve over extended exposure to automation (Gold et al., 2015). Together, these observations highlight the need to examine whether targeted educational interventions can alter trust and attentional processes during sustained automation.
Against this background, the present study examines whether these contrasting educational framings influence trust development, attention management operationalised through gaze behavior, and takeover readiness during Level 2 automation. Building on evidence that exposure to consistent automation performance can degrade monitoring patterns (Louw and Merat, 2017; Merat et al., 2014; Cooper et al., 2023), we investigate whether targeted pre-drive instruction can mitigate these effects by shaping drivers' expectations, mental models, and supervisory strategies.
We adopt the situation-awareness discrepancy model introduced in earlier work (Chouchane et al., 2022), which conceptualizes driver supervision as the difference between the situation awareness required by the driving context (α) and that actually achieved by the driver (α′). Educational interventions are hypothesized to sustain α′ (achieved monitoring) at appropriate levels, thereby reducing the discrepancy δ = α − α′. This model extends the logic of Endsley (1995b)'s three-level situation awareness framework by representing supervisory attention as a measurable balance between normative monitoring demand and descriptive driver behavior. It also aligns with the Monitoring and Response Task (MART) framework (Young and Stanton, 2002; Warm et al., 2008) and with Parasuraman and Riley (1997)'s taxonomy of automation use, misuse, and disuse, linking cognitive state dynamics to observable gaze patterns.
Within this model, we distinguish between trust calibration, which reflects how drivers perceive and internalize the system's reliability and limits, and capability-focused monitoring strategies, which determine how they visually allocate attention and verify the driving scene. These components may respond to educational framing through partially independent pathways: trust primarily develops through accumulated performance feedback and may be less sensitive to brief instruction (Hoff and Bashir, 2015; Merritt and Ilgen, 2008), whereas monitoring strategies may adapt more readily to targeted guidance (Körber et al., 2018).
To test these effects, the experiment employed a between-subjects design with three educational conditions. Basic instruction condition (reference condition) provided minimal information typical of current onboarding, whereas the capability-focused and limitation-focused conditions reflected the capability-focused and limitation-focused framings outlined above. Trust was measured at four time points throughout an 80-min session to capture both immediate post-instruction effects and subsequent evolution during automation exposure (Gold et al., 2015). Additional measures assessed mind-wandering (Smallwood and Schooler, 2015), visual scanning patterns using gaze-transition analysis (Schnebelen et al., 2020; Underwood et al., 2003), and reaction time during a critical takeover event (Gold et al., 2013; Lu et al., 2017).
The central research questions guiding this investigation were: (1) How does the framing of driver education influence trust dynamics over time when supervising Level 2 driving automation? (2) Does instructional framing affect drivers' tendency to mind-wander whilst supervising automation? (3) Do different educational approaches influence visual scanning patterns during supervision? (4) Do different educational approaches influence takeover performance? (5) What relationships exist among trust development, attentional disengagement, and takeover readiness?
Finally, to examine how educational framing shapes supervision in Level 2 driving automation, we proposed a set of hypotheses organized according to the research questions, and derived from established theoretical models and prior empirical evidence.
H1.1 (Trust trajectory). Trust ratings were expected to increase significantly from baseline (T1) to post-drive assessment (T4) across repeated exposure to the automation system (Lee and See, 2004; Hoff and Bashir, 2015).
H1.2 (Educational influence on trust development). Educational framing was expected to influence how trust developed across time (T1 to T4). Limitation-focused instruction highlighting system limitations could foster better-calibrated trust through improved understanding of boundaries (DeGuzman et al., 2020), whereas accumulated reliable experience might override brief instructional effects (Lee and See, 2004; Hoff and Bashir, 2015).
H2 (Educational effects on mind-wandering). Limitation-focused education was expected to reduce mind-wandering relative to capability-focused and Basic instruction by encouraging active monitoring and sustaining achieved monitoring effort (α′), thereby reducing the situation-awareness discrepancy (δ) defined in prior work (Chouchane et al., 2022). This prediction aligns with theories of situation awareness and vigilance suggesting that higher monitoring engagement preserves attentional resources during prolonged automation (Endsley, 1995b; Young and Stanton, 2002).
H3 (Educational effects on gaze behavior). Limitation-focused education was expected to produce more structured gaze patterns than the other conditions, characterized by more frequent recovery transitions to the road center and less peripheral attention (Price et al., 2019; Chouchane et al., 2022).
H4 (Educational effects on takeover performance). Participants receiving limitation-focused education were expected to show shorter reaction times during the takeover task, consistent with evidence that sustained attentional engagement and heightened situation awareness facilitate timely intervention (Endsley, 1995b; Greenlee et al., 2018).
H5.1 (Trust-mind-wandering relationship). Greater trust development (T1 to T4 change) was expected to correlate positively with mind-wandering, as increased confidence in system reliability reduces perceived monitoring demand (Lee and See, 2004; Parasuraman and Riley, 1997; Hergeth et al., 2016; Walker et al., 2019).
H5.2 (Mind-wandering-gaze relationship). Higher mind-wandering scores were expected to be associated with less structured scanning, including fewer mirror-to-road-center recoveries and more peripheral glances (Louw and Merat, 2017; Merat et al., 2014).
Together, these hypotheses address the research questions by linking educational framing, trust calibration, attentional state, gaze behavior, and takeover performance during Level 2 automation.
2 Materials and methods
2.1 Participants
Data were collected from 55 participants recruited through a third-party agency engaged by the Japan Automobile Research Institute (JARI), where the study took place. All participants reported no prior hands-on experience with Level 2 automation systems, and therefore represented an automation-naïve sample at the time of testing.
The study aimed for approximately 20 participants per educational condition to provide a balanced design and reasonable sensitivity for detecting medium-to-large effects. Participants were randomly assigned to one of three groups: Basic Education (n = 20), capability-focused (Method 1, n = 15), and limitation-focused (Method 2, n = 20). The slight imbalance arose from normal variation in attendance and early withdrawals rather than any planned differences in allocation.
Of the 55 participants in the final sample, 16 self-identified as female and 39 as male (Table 1). Sex distribution was broadly comparable across groups: the Basic group comprised 6 females and 14 males, the capability-focused group 5 females and 10 males, and the limitation-focused group 5 females and 15 males. Age and self-reported annual driving mileage by education group are summarized in Table 2.
Table 1
| Group | Female | Male | Total |
|---|---|---|---|
| B | 6 | 14 | 20 |
| M1 | 5 | 10 | 15 |
| M2 | 5 | 15 | 20 |
| Total | 16 | 39 | 55 |
Contingency table of participants by sex and education condition.
B, Basic education; M1, capability-focused education; M2, limitation-focused education.
Table 2
| Age (years) | Annual distance (km) | |||||
|---|---|---|---|---|---|---|
| Statistic | B | M1 | M2 | B | M1 | M2 |
| Valid cases | 20 | 15 | 20 | 20 | 15 | 20 |
| Missing | 0 | 0 | 0 | 0 | 0 | 0 |
| Mean | 40.2 | 36.7 | 43.6 | 16,400 | 7,407 | 14,900 |
| Standard deviation | 11.6 | 10.2 | 10.9 | 13,880 | 4,782 | 14,950 |
| Minimum | 23 | 24 | 23 | 2,000 | 100 | 1,000 |
| Maximum | 57 | 56 | 59 | 50,000 | 15,000 | 70,000 |
| 25th percentile | 30 | 29 | 34.5 | 7,250 | 2,000 | 10,000 |
| 50th percentile | 39 | 36 | 43 | 10,000 | 10,000 | 10,000 |
| 75th percentile | 49 | 42.5 | 52 | 24,250 | 10,000 | 15,000 |
Descriptive statistics for age and annual driving distance (km/year) by education condition.
B, Basic education; M1, capability-focused education; M2, limitation-focused education.
All participants provided informed consent prior to the study. The protocol was reviewed and approved by the JARI Research Ethics Committee. No personally identifying data were retained; all datasets were anonymised using numerical identifiers.
2.2 Experimental design
The study employed a 3 × 4 mixed factorial designwith Educational Condition (Basic, capability-focused, limitation-focused) as a between-subjects factor and Time (T1, T2, T3, T4) as a within-subjects factor corresponding to four measurement points throughout the session. This structure (Figure 1) enabled examination of both immediate and evolving effects of instructional framing across the automation experience, capturing trust calibration and behavioral outcomes over time.
Figure 1
2.3 Educational interventions
Educational briefings were designed to examine how different communication strategies shape drivers' understanding of automation and their expectations of system behavior. Each briefing conveyed the same core informational content: the automation supports driving but still requires active human supervision. The briefings differed in instructional emphasis (capability-focused versus limitation-focused framing). The guiding question was: How should information about system capabilities and limitations be conveyed to adjust driver expectations appropriately?
Educational content was delivered through scripted verbal and written briefings that were standardized across participants within each condition. The three conditions held duration, and delivery format constant, while varying the instructional emphasis (capability-focused versus limitation-focused framing) relative to the reference briefing.
Basic education (reference condition): Participants received only the minimum information typically provided in current vehicle manuals or pre-drive explanations. The briefing stated: “The driving automation system you will use today is a partially automated driving system. When the system is activated, the vehicle can drive automatically without pedal or steering operation. However, the driver must monitor the road and surrounding environment. When an alert occurs, you must take over control of the vehicle.” This condition served as a reference point reflecting minimal onboarding information.
Capability-focused education (method 1): This briefing explained how the automation system functions and perceives its surroundings. It described the system's control of steering and speed on highways up to 100 km/h, and how front and rear cameras detect surrounding vehicles and obstacles. Participants were told that system accuracy depends on image quality, which can deteriorate in poor weather. They were informed that the system issues an alert and disengages when reaching its operational limits, although delayed alerts may occur if sensing is impaired. This education aimed to build a general understanding of how the system works and why its performance changes under certain conditions. Visual aids illustrated camera positions and sensor coverage. The full script is provided in Appendix 4.12.2.
Limitation-focused education (method 2): This briefing conveyed the same core message as Method 1 but focused on operating boundaries. It explained that the system controls steering and pedals, functions in normal traffic and weather, and adjusts speed within its lane up to 100 km/h. Drivers were informed that visibility loss in rain or fog could degrade sensor performance and delay alerts. The message emphasized that when the system reaches its limits, it will issue an alert and disengage. Visual aids showed how adverse weather affects sensing. This framing encouraged drivers to monitor conditions and prepare to take control when necessary. The full script is provided in Appendix 4.12.3.
All briefings were delivered by the same experimenter using standardized scripts to ensure consistency. Each lasted approximately 5 min. Participants were unaware that different educational versions existed or that the study examined trust calibration.
These framings were informed by Rasmussen's Skills-Rules-Knowledge framework (Rasmussen, 1983). The capability-focused condition provided conceptual information about how the system perceives the environment and controls speed and steering. The limitation-focused condition emphasized operating boundaries, environmental constraints, and the driver's responsibility to intervene when sensing is impaired. The Basic condition offered only minimal information typical of current onboarding. These distinctions allowed us to test how different types of pre-drive information shape supervision during Level 2 automation.
2.4 Driving simulator environment and critical scenario
The experiment was conducted using the Japan Automobile Research Institute's omnidirectional driving simulator, a high-fidelity system featuring a complete vehicle cabin mounted on a motion platform with 360-degree visual projection (Figure 2). The simulator provides realistic vehicle dynamics, including motion cueing for acceleration, braking, and lateral movement, which enhances the ecological validity of driver responses during automation supervision and takeover events.
Figure 2
The simulated environment depicted a two-lane, two-direction, Japanese Expressway divided by a barrier, with moderate traffic density. Traffic consisted of computer-controlled vehicles programmed to maintain realistic speeds and following distances. The Level 2 automation system controlled the simulator vehicle's steering and speed throughout a 15-min driving automation segment, maintaining position in the center lane at approximately 80 km/h and adjusting speed to maintain safe following distance from lead vehicles.
The critical takeover scenario (Figure 3) occurred at the conclusion of the 15-min automated segment. A lead vehicle performed an emergency braking maneuver that exceeded the automation's deceleration capability. The system issued a visual and auditory takeover request requiring the driver to resume manual control and apply braking to avoid collision. This scenario was designed to assess drivers' readiness to intervene during safety-critical events following extended automation supervision.
Figure 3
2.5 Measures and data collection
2.5.1 Trust in automation
Trust in the driving automation system was measured using the same single-item question as Abe et al. (2017). Participants responded to the prompt, “At this moment, how much do you trust the driving automation system?,” by marking a value on a continuous 0-100 scale anchored at 0 = “not at all” and 100 = “completely”. The question was administered in Japanese, with both the original Japanese and English wordings provided in Appendix 4.10. Higher scores represent greater subjective trust in the automation.
Ratings were collected at four points during the experiment: baseline (T1), after the education session (T2), after the practice drive (T3), and after the driving automation session (T4). This schedule was designed to capture both the immediate effects of the education and later changes that occurred through direct interaction with the automation. A concise single-item scale was chosen to enable repeated measurement without interrupting the driving task, consistent with prior work showing that simple rating scales can effectively capture moment-to-moment changes in trust toward automation (Lee and Moray, 1992; Muir, 1996).
2.5.2 Mind-wandering
Mind-wandering was measured using a single-item retrospective scale, adapted from Mars et al. (2014) and translated into Japanese. Immediately after the takeover event, participants answered the question: “Compared with manual driving, how much time did you spend thinking about things unrelated to driving during the automated drive (before the alert)?” They marked their response on a continuous horizontal line ranging from 0% (“always thinking about driving as during manual driving”) to 100% (“never thinking about driving”). The complete wording in English and Japanese is provided in Appendix 4.11.
A retrospective self-report method was chosen to preserve the natural flow of the automated driving experience and to avoid intrusive probe techniques (e.g., SAGAT), which can disrupt ongoing cognitive processing and alter task engagement (Endsley, 1995a, 1988). This format allowed participants to reflect on their attention across the entire automated segment rather than at single moments. As noted by Smallwood and Schooler (2015), mind-wandering is an internal, subjective state that is most effectively studied through self-report, especially when researchers aim to capture general fluctuations in attention without disturbing task performance.
Although retrospective reports depend on memory and may involve estimation error, this approach was appropriate here because it maintained an undisturbed and realistic supervision context. It therefore provided an ecologically valid measure of participants' overall attentional disengagement during automation.
2.5.3 Gaze behavior
Eye-tracking data were recorded throughout the driving automation segment using a head-mounted NAC EMR-9 system sampling at 60 Hz. Gaze coordinates were manually annotated frame by frame into nine predefined Areas of Interest (AOIs) following ISO 15007:2020.
The AOI definitions, frame-by-frame annotation protocol, and transition-probability construction procedures follow our previously reported transition-analysis pipeline (Chouchane et al., 2026). These comprised the road center (RC; AOI 1), rear-view mirror (RVM; AOI 2), right side mirror (AOI 3), left side mirror (AOI 4), human-machine interface (HMI; AOI 5), right periphery (RP; AOI 6), left periphery (LP; AOI 7), driver side window (8) and other scene areas (AOI 9) (Figure 4). AOIs were annotated manually frame-by-frame by a single trained annotator using a predefined annotation protocol, with decision rules for ambiguous frames to support consistency despite head-movement artifacts. Additional details of the annotation workflow and quality-control procedures are reported in our transition-analysis pipeline description (Chouchane et al., 2026).
Figure 4
Gaze behavior was analyzed within a 24-s window centered on a side-vehicle cut-in event (beginning 15 s before and ending 5 s after the 4-s maneuver, occurring at 11 min 40 s into the automated drive) (Figure 5). This epoch was selected to capture supervisory attention after extended stable automation and to elicit lateral monitoring followed by recovery to forward scanning; conditions under which vigilance decrements are likely to emerge.
Figure 5
Previous research on spontaneous gaze during driving automation has shown that static gaze allocation captures meaningful aspects of supervision but fails to represent the gaze structure or sequencing of glances (Chouchane et al., 2022, 2026). Building on these insights, the present analysis adopted a dynamic approach using first-order Markov modeling to quantify transition probabilities between AOIs, thereby describing how drivers organized visual scanning. Similar transition-based approaches have demonstrated that gaze sequences between the road center and mirrors better discriminate attentive from out-of-the-loop states than static time-on-AOI measures (Schnebelen et al., 2020). Accordingly, static metrics were treated as complementary and dynamic indicators served as the primary basis for interpretation.
Two composite indices were constructed to summarize supervisory scanning structure. The Recovery to Road Center (RRC) captured how consistently participants redirected attention from mirror locations to the forward roadway, operationalising the reintegration of peripheral information essential to Level 1 and Level 2 situation awareness (Endsley, 1995b). It was computed as:
The Structured Scanning Index (SSI) provided a z-scored, directionally weighted composite integrating recovery frequency, sustained central monitoring, and reduced peripheral dwelling:
where TGTRC is total glance time on the road center and TGTLP, TGTRP are total glance times on the left and right peripheries, respectively. Higher SSI values indicate more structured, forward-anchored scanning routines. Although neither RRC nor SSI directly measure “being in the loop,” both serve as operational proxies for active supervisory control.
In addition, HMI monitoring (the proportion of fixations directed toward instrument-cluster and center-console displays) was analyzed as an exploratory metric to examine how educational framing influenced attention to automation-status information.
Together, these measures capture supervisory gaze at multiple levels: static allocation, dynamic transition structure, and system-information monitoring. Under Malleable Attentional Resources Theory (Young and Stanton, 2002), prolonged low-demand automation can contract attentional investment, which may manifest as simplified or less structured scanning. The present framework quantifies such reorganization empirically through RRC and SSI, offering transferable indicators of supervisory engagement in Level 2 driving.
2.5.4 Takeover performance
Takeover performance was evaluated using the System-recognized reaction time (SRRT) metric, defined as the interval between takeover request (TOR) onset and the first driver input exceeding the automation disengagement thresholds. Automation disengagement was defined as driver input exceeding approximately 10% accelerator stroke, 5% brake stroke, or 5 Nm steering torque, marking effective re-engagement with manual control. These thresholds were applied consistently across all participants to ensure comparability of SRRT values.
2.6 Procedure
Each session followed a standardized 80-min protocol designed to capture behavioral, physiological, and self-report data across all phases of the automation experience while ensuring participant comfort. Figure 6 illustrates the complete timeline and timing of trust assessments (T1 to T4).
Figure 6
Upon arrival at the simulator facility, participants provided written informed consent and completed demographic questionnaires. The session then progressed through the following sequence:
T1 (Baseline trust measurement). Participants first rated their initial level of trust in automation before receiving any detailed system information.
Educational intervention and T2 (Post-education trust).
Participants were randomly assigned to one of three instructional conditions (Basic, capability-focused, or limitation-focused; see Section 2.3). The assigned briefing was delivered by the experimenter while participants remained seated in the briefing room. Immediately after the briefing, participants provided a second trust rating (T2) to assess the immediate effect of the educational content.
Simulator familiarization. Participants then moved to the driving simulator and completed a brief manual driving segment to acclimate to the simulator's steering and pedal dynamics and reduce novelty effects.
Practice drive and T3 (Post-practice trust). Participants then engaged the Level 2 automation during a short, non-critical practice segment on a straight motorway section. This phase provided hands-on experience with system operation and the human-machine interface. A third trust rating (T3) was collected immediately afterward.
Eye-tracking calibration. Before the main drive, a 9-point calibration was performed for each participant using the NAC EMR-9 system, following established protocols to ensure accuracy throughout data collection.
Main automated drive. Participants supervised the Level 2 automation for approximately 15 minutes on a simulated highway with moderate traffic. Several non-critical traffic interactions were embedded to sustain ecological validity, and continuous eye-tracking and vehicle control data were recorded. Participants were instructed to monitor the system and remain ready to resume control at any time.
Critical takeover scenario. At the end of the automated segment, the lead vehicle performed sudden braking, triggering a visual and auditory takeover request prompting the driver to resume manual control. Takeover reaction time was recorded automatically by the simulator.
T4 (Post-drive trust) and mind-wandering assessment. Immediately after the takeover event, participants completed the final trust questionnaire (T4) and a retrospective mind-wandering measure. A structured debriefing followed to collect qualitative feedback about the automation experience.
2.7 Statistical analysis
All analyses were performed using standard parametric and non-parametric methods according to data characteristics. Outliers were screened prior to hypothesis testing, and assumption checks were conducted for normality, sphericity, and homogeneity of variance. When Mauchly's test indicated violation of sphericity, Greenhouse-Geisser correction was applied, with adjusted degrees of freedom and epsilon (ε) values reported. Levene's tests verified that the assumption of equal variances was met for these between-group comparisons.
2.7.1 Trust development
Trust ratings were analyzed using a two-way repeated-measures ANOVA with Time (T1 to T4) as the within-subjects factor and Educational Condition (Basic, capability-focused, limitation-focused) as the between-subjects factor. Greenhouse-Geisser correction was applied where required, and effect sizes are reported as partial eta-squared ().
2.7.2 Mind-wandering
Self-reported mind-wandering scores (MWS) were compared across education groups using one-way ANOVA. Pearson correlations were used to assess associations between mind-wandering, trust change (T4 minus T1), gaze indices (RRC, SSI), and takeover reaction time (SRRT). Effect sizes are reported as Pearson's r.
2.7.3 Gaze behavior
Supervisory gaze metrics were analyzed with Educational Condition as the between-subjects factor. Pre-registered scanning composites (RRC, SSI) were evaluated with Kruskal-Wallis tests due to non-normality; HMI-focused metrics (TGT5, MGD5, MGR5) were analyzed with one-way ANOVA when assumptions were met, otherwise with Kruskal-Wallis. Post-hoc comparisons used Benjamini-Hochberg false discovery rate (FDR) control within metric families (Benjamini and Hochberg, 1995). Both unadjusted p and adjusted qBH values are reported. Pearson (or Spearman, when appropriate) correlations assessed associations between gaze metrics and mind-wandering.
2.7.4 Takeover performance
System-recognized reaction times (SRRT) were slightly right-skewed due to one valid long reaction time (22.9 s) observed in the Basic group. Because this reflected a genuine delayed takeover rather than a measurement error, the value was retained, and non-parametric methods were used. A Kruskal-Wallis test compared SRRT across Educational Conditions. When omnibus results were significant, Dunn pairwise comparisons with Holm correction were conducted. Associations between SRRT, mind-wandering, and gaze metrics were examined using Spearman correlations. Effect sizes are reported as rank-based epsilon-squared () for Kruskal-Wallis tests and Spearman's ρ for correlations.
All tests were two-tailed with an α level of .05. Effect sizes are interpreted as partial for ANOVA, for non-parametric analyses, and r for correlations.
3 Results
3.1 Trust development across time and educational conditions
This section addresses RQ1 (trust dynamics) and evaluates H1.1-H1.2.
Findings from the repeated-measures ANOVA (Table 3) address H1.1 (main effect of Time) and H1.2 (Time × Group), both involving the within-subject factor Time. Outcomes from the between-subjects tests (Table 4) address the between-group component of H1.2 (main effect of Group). Levene's tests verified that the assumption of equal variances was met for these between-group comparisons.
Table 3
| Cases | Sphericity correction | Sum of squares | df | Mean square | F | p | |
|---|---|---|---|---|---|---|---|
| Time | Greenhouse-Geisser | 7131.028 | 1.694 | 4,209.356 | 21.156 | <0.001 | 0.092 |
| Time × Group | Greenhouse-Geisser | 540.780 | 3.388 | 159.608 | 0.802 | 0.509 | 0.007 |
| Residuals | Greenhouse-Geisser | 16,516.190 | 83.010 | 198.965 | – | – | – |
Results of repeated-measures ANOVA for within-subjects effects on trust across time.
Degrees of freedom adjusted with Greenhouse-Geisser.
Table 4
| Effect | Sum of squares | df | Mean square | F | p | |
|---|---|---|---|---|---|---|
| Group | 1158.012 | 2 | 579.006 | 0.621 | 0.541 | 0.015 |
| Residuals | 45,652.381 | 49 | 931.681 | – | – | – |
Between-subjects effects for education group on mean trust.
3.1.1 Main effect of time (H1.1)
A repeated-measures ANOVA with Time (T1 to T4) as the within-subjects factor and education group as the between-subjects factor revealed a significant main effect of Time on trust ratings, F(1.694, 83.010) = 21.156, p < 0.001, partialη2 = 0.092 (Greenhouse-Geisser corrected) (Table 3). Descriptive analyses indicated a consistent upward trend in trust scores across all educational groups from the initial time point (T1) to the final assessment (T4) (see Figure 7). Participants in the Basic group showed an increase from M = 58.50 (SD = 19.54) at T1 to M = 68.00 (SD = 16.73) at T4. Those in the Method 1 group improved from M = 50.67 (SD = 19.81) to M = 64.00 (SD = 22.30), while Method 2 participants demonstrated a rise from M = 52.25 (SD = 18.39) to M = 71.75 (SD = 13.70). Descriptive statistics by group and time are in Table 5. These findings support H1.1, which proposed that trust would rise from the baseline (T1) to the post-drive stage (T4) as participants became more familiar with the automation system.
Figure 7
Table 5
| Trust T1 | Trust T2 | Trust T3 | Trust T4 | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Statistic | B | M1 | M2 | B | M1 | M2 | B | M1 | M2 | B | M1 | M2 |
| Valid cases | 20 | 15 | 20 | 20 | 15 | 20 | 20 | 15 | 20 | 20 | 15 | 20 |
| Mean | 58.5 | 50.7 | 52.3 | 60.0 | 56.0 | 57.5 | 68.5 | 68.0 | 70.3 | 68.0 | 64.0 | 71.8 |
| Standard error | 4.37 | 5.11 | 4.11 | 3.97 | 5.59 | 3.55 | 4.43 | 4.90 | 3.45 | 3.74 | 5.76 | 3.06 |
| Standard deviation | 19.5 | 19.8 | 18.4 | 17.8 | 21.7 | 15.9 | 19.8 | 19.0 | 15.4 | 16.7 | 22.3 | 13.7 |
| Minimum | 20 | 20 | 5 | 20 | 20 | 30 | 20 | 20 | 40 | 30 | 20 | 50 |
| Maximum | 90 | 80 | 80 | 90 | 80 | 80 | 90 | 90 | 90 | 90 | 90 | 100 |
Descriptive statistics for trust scores at each time point (T1 to T4) by education group.
B, Basic; M1, capability-focused; M2, limitation-focused. T1 to T4, baseline, post-education, post-practice, post-drive.
3.1.2 Education group effects (H1.2)
Overall trust did not differ by education group, F(2,49) = 0.621, p = 0.541, (Table 4). The Time × Group interaction was also non-significant, F(3.388,83.010) = 0.802, p = 0.509, η2 = 0.007 (Table 3), suggesting that educational framing did not lead to significant differences in how trust evolved over time. These results do not support H1.2, which predicted that educational framing would influence the development of trust across time (T1 to T4).
3.2 Driver education and mind-wandering
This section addresses RQ2 (mind-wandering) and evaluates H2, followed by RQ5 (relationships) for H5.1.
After identifying the general pattern of trust development, we next explored whether the educational interventions affected mind-wandering during automation supervision.
3.2.1 Education group effects on mind-wandering (H2)
A one-way ANOVA tested for differences in self-reported mind-wandering scores (MWS) at T4. The analysis showed no significant group differences, F(2,52) = 1.571, p =.218, η2 = 0.057 (Table 6). Descriptively, participants in the Method 2 condition reported higher mind-wandering (M = 38.75, SD = 23.39) than those in the Basic (M = 27.50, SD = 17.43) and Method 1 groups (M = 28.67, SD = 24.46; see Table 7 and Figure 8). This pattern contradicts the expectation that limitation-focused education would help reduce mind-wandering. These results do not support H2, which predicted that limitation-focused education would reduce mind-wandering compared with capability-focused and basic instruction.
Table 6
| Source | Sum of squares | df | Mean square | F | p | ω2 | |
|---|---|---|---|---|---|---|---|
| Group | 1482.462 | 2 | 741.231 | 1.571 | 0.218 | 0.057 | 0.020 |
| Residuals | 24,542.083 | 52 | 471.963 | – | – | – | – |
One-way ANOVA results for mind-wandering across education groups.
Table 7
| Group | N | Mean | SD | SE | Coefficient of variation |
|---|---|---|---|---|---|
| B | 20 | 27.50 | 17.43 | 3.90 | 0.63 |
| M1 | 15 | 28.67 | 24.46 | 6.32 | 0.85 |
| M2 | 20 | 38.75 | 23.39 | 5.23 | 0.60 |
Descriptive statistics for mind-wandering (MWS) by education group.
Figure 8
3.2.2 Trust development and mind-wandering relationship (H5.1)
A Pearson correlation was conducted to examine the relationship between trust change (T4 minus T1) and mind-wandering scores. The analysis showed a significant positive correlation, r = 0.296, p = 0.028, 95% CI [0.034, 0.521] (Table 8). These results support H5.1. Although the correlation was weak (r = 0.296), a statistically significant positive association was found, indicating that participants who showed larger trust gains also reported more frequent mind-wandering.
Table 8
| Statistic | Trust change ↔ MWS | Value |
|---|---|---|
| Pearson's r | 0.296 | |
| p-value | 0.028 | |
| 95% CI (lower) | 0.034 | |
| 95% CI (upper) | 0.521 | |
| Fisher's z | 0.305 | |
| SE (Fisher's z) | 0.139 |
Correlation between trust change and mind-wandering (MWS).
3.3 Gaze behavior and performance outcomes
This section addresses RQ3 (visual scanning) and evaluates H3, then addresses RQ5 for H5.2, and finally addresses RQ4 (takeover performance) for H4.
3.3.1 Education group effects on gaze metrics (pre-registered composites) (H3)
Descriptive statistics for the main gaze measures are shown in Table 9. To test whether educational framing influenced supervisory gaze, omnibus Kruskal-Wallis tests were conducted on the preregistered key dependent variables (KDVs): Recovery to Road Center (RRC) and Structured Scanning Index (SSI). As shown in Table 10, none of these measures differed significantly across groups (RRC: H(2) = 1.37, p =.505; SSI: H(2) = 3.10, p =.212). These results do not support H3.
Table 9
| Metric | B (mean ± SD) | M1 | M2 | n |
|---|---|---|---|---|
| Rear-view mirror → Road center (prob.) | 0.548 ± 0.348 | 0.357 ± 0.370 | 0.505 ± 0.376 | 42 |
| Recovery to Road center (RRC) | 2.543 ± 0.937 | 2.560 ± 0.892 | 2.293 ± 1.194 | 47 |
| Structured Scanning Index (SSI, z) | 0.100 ± 0.411 | −0.134 ± 0.522 | 0.041 ± 0.699 | 47 |
| Mean glance duration: Mirrors (s) | 0.563 ± 0.388 | 0.455 ± 0.247 | 0.500 ± 0.294 | 47 |
| Mean glance duration: Periphery (s) | 0.406 ± 0.141 | 0.459 ± 0.168 | 0.458 ± 0.286 | 47 |
| Mean glance rate: Periphery (Hz) | 0.082 ± 0.042 | 0.135 ± 0.075 | 0.094 ± 0.072 | 47 |
| Total glance time: Periphery (s) | 1.865 ± 0.941 | 3.209 ± 1.704 | 1.963 ± 1.470 | 47 |
| Total glance time: Road center (s) | 11.716 ± 4.951 | 11.178 ± 4.801 | 11.776 ± 5.874 | 47 |
Descriptive statistics for primary gaze outcomes and key derived static indicators.
B, Basic; M1, capability-focused; M2, limitation-focused.
Table 10
| Metric | Test | H | p |
|---|---|---|---|
| Recovery to Road center (RRC) | Kruskal-Wallis | 1.367 | 0.505 |
| Structured Scanning Index (SSI) | Kruskal-Wallis | 3.102 | 0.212 |
Omnibus Kruskal-Wallis tests for pre-registered scanning dependent variables.
Within the static AOI measures (Table 11), two HMI-related metrics showed significant group differences after Benjamini–Hochberg correction: HMI total glance time (TGT5), H(2) = 11.02, p =.004, qBH =.036, and HMI mean glance duration (MGD5), H(2) = 11.28, p =.0036, qBH =.032. Post hoc Dunn tests indicated higher values for Method 1 compared with both Basic and Method 2 groups (pairwise q's = 0.006–0.015). HMI glance rate (MGR5) followed the same pattern but did not remain significant after FDR adjustment (p =.006, qBH = 0.057).
Table 11
| Metric | H | p | qBH | B vs M1 q | M1 vs M2 q | B vs M2 q |
|---|---|---|---|---|---|---|
| TGT5 (HMI total time) | 11.02 | 0.004 | 0.036 | 0.006 | 0.015 | 0.718 |
| MGD5 (HMI mean glance duration) | 11.28 | 0.004 | 0.032 | 0.008 | 0.008 | 0.985 |
| MGR5 (HMI glance rate) | 10.14 | 0.006 | 0.057 | 0.008 | 0.023 | 0.621 |
HMI-centered static gaze metrics: omnibus Kruskal-Wallis tests and Benjamini-Hochberg-adjusted Dunn post hocs.
Exploratory transition analyses (Table 12) revealed several nominal trends (e.g., Right-periphery → HMI p = 0.051), but none survived FDR correction within the transition family. Overall, preregistered scanning metrics showed no educational effects, although exploratory HMI measures suggested greater display monitoring for the Method 1 group in the scenario.
Table 12
| Transition | n | H | p (uncorr.) | pFDR |
|---|---|---|---|---|
| Right periphery → HMI | 45 | 5.965 | 0.051 | 1.000 |
| Left periphery → HMI | 31 | 5.198 | 0.074 | 1.000 |
| Rear-view mirror → HMI | 42 | 4.993 | 0.082 | 1.000 |
| Road center → HMI | 45 | 4.923 | 0.085 | 0.896 |
| Right periphery → Left periphery | 45 | 4.736 | 0.094 | 0.787 |
Exploratory transition probabilities with the smallest omnibus p-values.
3.3.2 Mind-wandering and gaze behavior relationship (H5.2)
More structured scanning was associated with lower mind-wandering. Participants with higher Recovery to Road Center (RRC) scores reported less mind-wandering (Table 13; Pearson r = −0.329, p =.024, 95% CI [−0.563, −0.046]; Spearman ρ = −0.367, p = 0.011). The Structured Scanning Index (SSI) showed a similar pattern (Pearson r = −0.364, p = 0.012, 95% CI [−0.590, −0.086]; Spearman ρ = −0.367, p = 0.011). Although these correlations were weak to modest in magnitude, both measures showed a consistent negative association with self-reported mind-wandering. Together, the results indicate that structured scanning accounted for roughly 11–13 per cent of the variance in mind-wandering (r2), supporting H5.2.
Table 13
| x | y | n | Pearson r | p | 95% CI | Spearman ρ | p |
|---|---|---|---|---|---|---|---|
| RRC | MWS | 47 | −0.329 | 0.024 | [−0.563, −0.046] | −0.367 | 0.011 |
| RRC | SRRT (s) | 47 | 0.090 | 0.545 | [−0.200, 0.370] | −0.160 | 0.287 |
| SSI | MWS | 47 | −0.364 | 0.012 | [−0.590, −0.086] | −0.367 | 0.011 |
| SSI | SRRT (s) | 47 | 0.110 | 0.480 | [−0.190, 0.380] | −0.110 | 0.473 |
Correlations between structured scanning composites and mind-wandering (MWS) and system-recognized reaction time (SRRT).
3.3.3 Takeover performance (H4)
System-recognized reaction times (SRRT) differed significantly across educational groups, H(2) = 7.898, p =.019, , 95% CI [0.024, 0.381] (Table 14). Median SRRTs were shortest for the Method 1 group (Mdn = 2.142 s), followed by Method 2 (Mdn = 2.770 s) and Basic groups (Mdn = 2.715 s; see Table 15 and Figure 9). Post hoc Dunn tests with Holm correction confirmed a significant difference between Method 1 and Method 2 (pHolm =.015), with no other contrasts reaching significance (Table 16). These results do not support H4, which predicted shorter SRRTs for participants in the limitation-focused education condition.
Table 14
| Effect | H | df | p | Rank ε2 (95% CI) | ||
|---|---|---|---|---|---|---|
| Estimate | Lower | Upper | ||||
| Group | 7.898 | 2 | 0.019 | 0.152 | 0.024 | 0.381 |
Kruskal-Wallis omnibus test for SRRT across education groups.
Table 15
| Group | n | Mean | SD | Median | Q1 | Q3 | Min | Max |
|---|---|---|---|---|---|---|---|---|
| B | 18 | 3.651 | 4.837 | 2.715 | 2.107 | 2.893 | 1.550 | 22.90 |
| M1 | 14 | 2.127 | 0.557 | 2.145 | 1.723 | 2.460 | 1.230 | 3.00 |
| M2 | 21 | 2.861 | 0.729 | 2.770 | 2.430 | 3.130 | 1.770 | 5.02 |
Descriptive statistics for SRRT by education group.
B, Basic; M1, capability-focused; M2, limitation-focused. One high outlier (22.9 s) was retained.
Figure 9
Table 16
| Comparison | z | rrb | p (raw) | pHolm |
|---|---|---|---|---|
| B vs M1 | 1.804 | 0.373 | 0.071 | 0.143 |
| B vs M2 | −1.009 | 0.188 | 0.313 | 0.313 |
| M1 vs M2 | −2.802 | 0.565 | 0.005 | 0.015 |
Dunn post hoc pairwise comparisons for SRRT with Holm correction.
4 Discussion
This study investigated whether educational interventions could calibrate trust and sustain supervisory attention during Level 2 driving automation. The results suggested an asymmetry: trust formation appeared to be primarily associated with direct exposure to automation performance, whereas educational framing was associated with differences in HMI monitoring and takeover readiness, while broader scanning structure remained unaffected. Exposure was the dominant factor shaping trust trajectories (H1.1–H1.2), while capability-focused education was associated with differences in HMI monitoring and takeover readiness (H4). Trust growth was also associated with increased mind-wandering (H5.1), and mind-wandering correlated with less structured scanning (H5.2), linking subjective and behavioral markers of supervisory disengagement. Collectively, these findings help to clarify both the limitations and the potential, targeted benefits of pre-drive education for supporting safe human-automation interaction (Table 17).
Table 17
| Hypothesis | Prediction | Result | Support |
|---|---|---|---|
| H1.1 | Trust increases across T1 to T4 during Level 2 automation. | Significant main effect of Time; trust rose consistently across groups. | □ |
| H1.2 | Educational framing influences trust trajectory across time. | No Group or Time by Group effects; trajectories were comparable. | × |
| H2 | Limitation-focused education reduces mind-wandering compared with other groups. | No significant group differences in mind-wandering. | × |
| H3 | Educational framing affects structured scanning (RRC, SSI). | No group differences in preregistered scanning composites. | × |
| H4 | Limitation-focused education produces the fastest takeover reactions. | Group differences observed, but Method 1 (capability-focused) was faster than Method 2 (limitation-focused); prediction not met. | × |
| H5.1 | Greater trust increase is associated with higher mind-wandering. | Weak but significant positive correlation (r = 0.296, p = 0.028). | □ |
| H5.2 | More structured scanning is linked to lower mind-wandering. | Weak to modest but significant negative correlations for RRC and SSI (r = −0.329, p = 0.024; r = −0.364, p = 0.012). | □ |
Summary of hypothesis testing outcomes.
RRC, Recovery to Road Center; SSI, Structured Scanning Index.
□ = Supported; × = Not supported.
4.1 Trust development is primarily exposure driven
Trust increased across all educational conditions as participants gained exposure to Level 2 automation, supporting H1.1. This pattern is consistent with earlier evidence that direct interaction with automation gradually aligns trust with actual system performance, even when users begin with different initial information (Beggiato and Krems, 2013; Lee and See, 2004; Hoff and Bashir, 2015). Neither capability-focused nor limitation-focused framing appeared to alter this trajectory (H1.2), suggesting that direct exposure to the system's behavior outweighed prior instruction.
This pattern aligns with the CAUSE model (Rowan, 1991), which describes effective communication as a sequential process: establishing credibility, raising awareness, promoting understanding, achieving satisfaction, and supporting enactment. The brief instructional framings in this study were unlikely to complete this process; instead, participants appeared to calibrate trust through their direct observation of system reliability. In this context, performance feedback rather than message framing likely served as the main influence on trust development.
The positive correlation between trust growth and mind-wandering (H3) indicated a potential tension between confidence in the system and supervisory engagement. Participants who showed greater increases in trust also showed lower values on monitoring-related indicators, consistent with a trust-associated reduction in supervisory engagement. This pattern is consistent with Malleable Attentional Resources Theory (MART) (Young and Stanton, 2002; Warm et al., 2008), which proposes that attentional investment contracts under low task demand. The relationship was observed across educational conditions, suggesting that reduced monitoring may emerge generally as drivers gain experience with stable automation.
4.2 Capability-focused education and monitoring priorities
Although capability-focused education did not substantially alter trust trajectories, it appeared to yield two modest benefits (H8): greater attention to the HMI and shorter system-recognized reaction times during takeover. Participants who received knowledge-based (capability-focused) instruction monitored automation status displays more frequently and responded more promptly when manual control was required. These findings suggest that a conceptual understanding of system operation may have refined monitoring selectivity and readiness for authority transfer.
Within Endsley's situation-awareness framework (Endsley, 1995b), these improvements may reflect an increased ability to interpret what automation feedback signals about system state. Capability-focused instruction might have rendered HMI cues more meaningful and diagnostic, enabling participants to interpret them more efficiently without compromising roadway monitoring. The faster SRRT further suggests somewhat smoother coordination during manual resumption.
Hence, brief conceptual instruction may provide targeted benefits by sharpening monitoring priorities and readiness, even though it does not seem to mitigate trust-related vigilance loss. These effects illustrate how education and exposure can influence distinct aspects of supervision: exposure shaping trust and perceived monitoring demand, and education fine-tuning particular monitoring behaviors.
4.3 Limitation-focused education and its limited impact
Rule-based (limitation-oriented) education showed little measurable influence on trust, attention, or takeover performance. Several factors may explain this outcome. One possible interpretation is that emphasizing system limits did not translate into actionable monitoring strategies during a stable scenario, which may have limited observable behavioral effects. The instruction listed boundary conditions such as poor weather or degraded lane markings but offered little procedural guidance on how to detect or respond to them. Without actionable cues, participants had limited opportunity to apply these rules during the stable highway scenario.
The motivational tone may also have contributed. Emphasizing system fallibility might have led drivers to expect that failures would occur only under specific conditions, such as poor weather or unclear lane markings. Because the simulated drive presented no such cues, participants may have focused less on continuous monitoring and more on waiting for visible signs of malfunction. This conditional form of supervision differs from the general vigilance loss associated with growing trust, yet it can be viewed as another expression of trust miscalibration, where expectations about when the system might fail are narrower than its actual range of vulnerabilities (Parasuraman and Riley, 1997). Moreover, the drive remained within the automation's operational domain, so the limitation messages were not especially relevant. Under these conditions, participants may have regarded the Limitation-focused briefing as less meaningful, which could have limited its behavioral impact.
4.4 Educational framing and control modes
The Skills-Rules-Knowledge (SRK) framework (Rasmussen, 1983) provides a useful lens for understanding these differences. Limitation-focused instruction depends on stored “if-then” procedures, which apply only when matching cues arise. In this study, such cues were absent, and the learned rules were therefore unlikely to be triggered. Capability-focused education, by contrast, fosters conceptual models that apply across diverse conditions. Understanding how the automation senses and acts may have improved the perceived relevance of HMI information and facilitated faster recognition of control transitions. In Rasmussen's framework, capability-focused processing is required when operators face unfamiliar or unexpected situations for which no stored rules apply. Such conditions are analogous to automation takeovers, where drivers must interpret system behavior and re-establish manual control.
4.5 Integration with theoretical frameworks
These findings can be interpreted within the situation-awareness discrepancy model, δ = α−α′, introduced in earlier work (Chouchane et al., 2022). Within this model, trust is likely to shape perceived monitoring demand (α). As trust increases, drivers may judge that less active supervision is necessary. This change reflects heightened confidence in automation reliability rather than an actual reduction in the monitoring required. The observed association between higher trust and mind-wandering supports this interpretation, indicating that lower perceived demand can promote attentional drift and weaker supervisory engagement.
Educational interventions, by contrast, may affect achieved monitoring (α′) through partially independent pathways. Capability-focused education seemed to enhance certain components of monitoring, such as HMI attention and takeover readiness. These effects represent localized gains in α′, without offsetting the overall decline in perceived demand. Consequently, brief instruction may reinforce specific monitoring habits but is unlikely to counter the wider vigilance contraction associated with rising trust.
Malleable Attentional Resources Theory (MART) (Young and Stanton, 2002; Warm et al., 2008) provides a complementary explanation for this contraction. Under stable, low-demand automation, attentional resources tend to narrow, leading to less structured and less frequent scanning between key visual areas. The observed correlation between mind-wandering and reduced scanning organization (lower SSI and RRC) aligns with this mechanism. The CAUSE model (Rowan, 1991) further clarifies why instructional messages had limited effect: effective risk communication depends on credible and sequenced engagement rather than isolated briefings. Finally, performance-based trust models (Lee and See, 2004; Hoff and Bashir, 2015) help explain why direct exposure to automation performance dominated trust development.
Taken together, these frameworks suggest that trust most plausibly influences perceived monitoring demand, while education can refine selective monitoring behaviors. The combination of rising trust and stable system performance therefore creates conditions in which drivers monitor less actively even though objective supervisory requirements remain unchanged.
4.6 Practical implications
The results suggest that brief pre-drive education has limited influence on trust trajectories, which develop mainly through direct exposure to automation performance. Verbal or written instructions, though valuable for setting initial expectations, appear insufficient to shape how trust evolves during actual use. Nonetheless, such education can enhance comprehension and promote more focused monitoring.
The modest benefits of capability-focused instruction indicate that more extensive, scenario-based training could produce broader effects. Programmes combining conceptual explanation with guided practice might help drivers encounter automation in varied conditions and near its operational limits, allowing better calibration of both trust and monitoring strategies. Real-time feedback from instructors could reinforce how and when to intervene, supporting balanced supervision.
The improvement in HMI monitoring following capability-focused instruction also points to the importance of interface transparency. Displays that provide meaningful cues, such as sensor coverage or automation confidence, can enhance monitoring, provided the information is interpretable and not overwhelming. Education and interface design may therefore work best when developed in tandem.
Finally, the link between scanning structure and mind-wandering highlights opportunities for driver-monitoring systems. Beyond tracking gaze-on-road duration, systems might assess how organized scanning patterns remain over time. Identifying reduced mirror-to-road transitions or increased randomness could enable early alerts before performance declines.
4.7 Limitations
Several factors limit interpretation and generalisability. First, between-group sample sizes were modest (n = 15–20 per educational condition), which limits statistical power to detect small effects and interactions; null findings for group differences should therefore be interpreted cautiously. Second, the educational interventions were brief and delivered once; repeated or interactive training might yield stronger and more durable effects. Third, the simulated drive involved only one takeover event after 15 minutes of automation in stable conditions. Real driving includes greater variability in road, traffic, and weather conditions, which may influence both trust and attention. Fourth, participants were inexperienced with hands-on Level 2 automation use, and the findings therefore reflect early exposure rather than longer-term adaptation; future studies should examine whether educational benefits or trust-attention dynamics persist over extended use. Fifth, mind-wandering was measured retrospectively, which could have underestimated momentary fluctuations; incorporating concurrent or physiological measures would strengthen future assessments. Finally, the instructional materials represented one implementation of capability-focused and limitation-focused framing. Although the briefings differed in emphasis, we did not include a direct manipulation check to verify whether participants internalized the framings as intended. Given the shared informational content across the briefings, it is possible that overlap attenuated between-group differences in trust and gaze metrics, and effects should therefore be interpreted cautiously. Variations in content, delivery medium, or realism, as well as brief post-briefing checks of capability versus boundary understanding, would strengthen future validation of framing-based interventions.
4.8 Future directions
Future research should examine how educational and exposure factors interact over time and across contexts. Longitudinal studies could track trust and monitoring behavior over extended periods of real-world driving. Scenario-based training could identify how conceptual understanding and hands-on exposure combine to calibrate trust and attention effectively. Research in diverse driving environments, including urban and adverse-weather conditions, would clarify when limitation-focused education becomes most relevant. Finally, combining educational interventions with adaptive HMIs and monitoring systems could help sustain driver engagement.
4.9 Conclusions
Brief pre-drive education modulated specific monitoring behaviors but did not influence the trajectory of trust formation during Level 2 automation. Trust increased over time across all groups and was not affected by instructional framing, consistent with experience-driven adjustment to observed system performance. Capability-focused education was associated with greater attention to automation feedback and faster takeover responses compared with limitation-focused education, indicating that conceptual information may refine how drivers monitor system status without altering overall trust development.
Across participants, increases in trust were accompanied by patterns consistent with reduced structured monitoring, suggesting a potential trade-off between growing confidence and sustained supervisory engagement during routine automation. More structured gaze organization was associated with lower self-reported mind-wandering, supporting the interpretation of gaze metrics as behavioral proxies for attentional allocation rather than direct measures of attention.
These findings suggest the potential value of integrated approaches combining extended experiential training, transparent interface design, and gaze-based monitoring systems to support sustained supervisory engagement in partially automated driving.
Statements
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving humans were approved by Japan Automobile Research Institute, Research Ethics Committee. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
HC: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing. YS: Methodology, Software, Writing – review & editing. KS: Funding acquisition, Project administration, Software, Writing – review & editing. GA: Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Writing – review & editing. MI: Conceptualization, Funding acquisition, Investigation, Methodology, Resources, Supervision, Writing – review & editing.
Funding
The author(s) declared that financial support was received for this work and/or its publication. This work was supported by JSPS KAKENHI Grant Number JP24H00361.
Acknowledgments
The authors thank JARI Autonomous Driving Research Division members for the technical support and valuable discussions throughout the project development.
Conflict of interest
The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fcomp.2026.1772813/full#supplementary-material
References
1
AbeG.SatoK.ItohM. (2017). Driver trust in automated driving systems: The case of overtaking and passing. IEEE Trans. Human-Mach. Syst. 48, 85–94. doi: 10.1109/THMS.2017.2781619
2
BainbridgeL. (1983). “Ironies of automation,” in Analysis, Design and Evaluation of Man-Machine Systems (London: Elsevier), 129–135.
3
BaldwinC. L.RobertsD. M.BarraganD.LeeJ. D.LernerN.HigginsJ. S. (2017). Detecting and quantifying mind wandering during simulated driving. Front. Hum. Neurosci. 11:406. doi: 10.3389/fnhum.2017.00406
4
BeggiatoM.KremsJ. F. (2013). The evolution of mental model, trust and acceptance of adaptive cruise control in relation to initial information. Transport. Res. Part F: Traffic Psychol. Behav. 18, 47–57. doi: 10.1016/j.trf.2012.12.006
5
BenjaminiY.HochbergY. (1995). Controlling the false discovery rate: a practical and powerful approach to multiple testing. J. Royal Statist. Soc.: Series B. 57, 289–300. doi: 10.1111/j.2517-6161.1995.tb02031.x
6
CampbellJ. L.BrownJ. L.GravingJ. S.RichardC. M.LichtyM. G.BaconL. P.et al. (2018). “Human factors design guidance for level 2 and level 3 automated driving concepts,” in Technical Report.
7
ChouchaneH.LeeJ.SakamuraY.NakamuraH.AbeG.ItohM. (2026). Supervisory gaze behaviour under different automation durations in level 2 driving: a first-order transition analysis. Appl. Sci. 16:1401. doi: 10.3390/app16031401
8
ChouchaneH.NakamuraH.SatoK.Antona-MakoshiJ.AbeG.ItohM. (2022). “Identifying the out of the loop phenomenon during driving automation using spontaneous gaze behavior,” in Proceedings of the Human Factors and Ergonomics Society Annual Meeting (Los Angeles, CA: SAGE Publications), 305–309.
9
CooperJ. M.CrabtreeK. W.McDonnellA. S.MayD.StrayerS. C.TsogtbaatarT.et al. (2023). Driver behavior while using level 2 vehicle automation: a hybrid naturalistic study. Cognit. Res.: Princ. Implicat. 8:71. doi: 10.1186/s41235-023-00527-5
10
DeGuzmanC. A.DonmezB. (2021). Drivers still have limited knowledge about adaptive cruise control even when they own the system. Transp. Res. Rec. 2675, 328–339. doi: 10.1177/03611981211011482
11
DeGuzmanC. A.HopkinsS. A.DonmezB. (2020). Driver takeover performance and monitoring behavior with driving automation at system-limit versus system-malfunction failures. Transp. Res. Rec. 2674, 140–151. doi: 10.1177/0361198120912228
12
EndsleyM. R. (1988). “Design and evaluation for situation awareness enhancement,” in Proc. Human Fact. Soc. Annual Meet. 32, 97–101. doi: 10.1177/154193128803200221
13
EndsleyM. R. (1995a). Measurement of situation awareness in dynamic systems. Hum. Factors37, 65–84. doi: 10.1518/001872095779049499
14
EndsleyM. R. (1995b). Toward a theory of situation awareness in dynamic systems. Hum. Factors37, 32–64. doi: 10.1518/001872095779049543
15
ErikssonA.StantonN. A. (2017). Takeover time in highly automated vehicles: noncritical transitions to and from manual control. Hum. Factors59, 689–705. doi: 10.1177/0018720816685832
16
GoldC.DamböckD.LorenzL.BenglerK. (2013). ““take over!” how long does it take to get the driver back into the loop?,” in Proceedings of the Human Factors and Ergonomics Society Annual Meeting (Los Angeles, CA: Sage Publications), 1938–1942.
17
GoldC.KörberM.HohenbergerC.LechnerD.BenglerK. (2015). Trust in automation-before and after the experience of take-over scenarios in a highly automated vehicle. Procedia Manufact. 3, 3025–3032. doi: 10.1016/j.promfg.2015.07.847
18
GreenleeE. T.DeLuciaP. R.NewtonD. C. (2018). Driver vigilance in automated vehicles: Hazard detection failures are a matter of time. Hum. Factors60, 465–476. doi: 10.1177/0018720818761711
19
HarmsI. M.BingenL.SteffensJ. (2020). Addressing the awareness gap: a combined survey and vehicle registration analysis to assess car owners' usage of adas in fleets. Transport. Res. Part A: Policy Pract. 134, 65–77. doi: 10.1016/j.tra.2020.01.018
20
HeJ.BecicE.LeeY.-C.McCarleyJ. S. (2011). Mind wandering behind the wheel: Performance and oculomotor correlates. Hum. Factors53, 13–21. doi: 10.1177/0018720810391530
21
HergethS.LorenzL.VilimekR.KremsJ. F. (2016). Keep your scanners peeled: Gaze behavior as a measure of automation trust during highly automated driving. Hum. Factors58, 509–519. doi: 10.1177/0018720815625744
22
HoffK. A.BashirM. (2015). Trust in automation: Integrating empirical evidence on factors that influence trust. Hum. Factors57, 407–434. doi: 10.1177/0018720814547570
23
InagakiT.ItohM. (2013). Human's overtrust in and overreliance on advanced driver assistance systems: a theoretical framework. Int. J. Vehicular Technol. 2013:951762. doi: 10.1155/2013/951762
24
KhastgirS.BirrellS.DhadyallaG.JenningsP. (2018). Calibrating trust through knowledge: Introducing the concept of informed safety for automation in vehicles. Transport. Res. Part C: Emerg. Technol. 96, 290–303. doi: 10.1016/j.trc.2018.07.001
25
KörberM.BaselerE.BenglerK. (2018). Introduction matters: manipulating trust in automation and reliance in automated driving. Appl. Ergon. 66, 18–31. doi: 10.1016/j.apergo.2017.07.006
26
LeeJ.MorayN. (1992). Trust, control strategies and allocation of function in human-machine systems. Ergonomics35, 1243–1270. doi: 10.1080/00140139208967392
27
LeeJ. D.SeeK. A. (2004). Trust in automation: Designing for appropriate reliance. Hum. Factors46, 50–80. doi: 10.1518/hfes.46.1.50.30392
28
LouwT.MeratN. (2017). Are you in the loop? Using gaze dispersion to understand driver visual attention during vehicle automation. Transport. Res. Part C: Emerg. Technol. 76, 35–50. doi: 10.1016/j.trc.2017.01.001
29
LuZ.CosterX.de WinterJ. (2017). How much time do drivers need to obtain situation awareness? A laboratory-based study of automated driving. Appl. Ergon. 60, 293–304. doi: 10.1016/j.apergo.2016.12.003
30
MarsF.DerooM.CharronC. (2014). Driver adaptation to haptic shared control of the steering wheel. In 2014 ieee international conference on systems, man, and cybernetics (SMC) (San Diego, CA: IEEE), 1505–1509.
31
MeratN.JamsonA. H.LaiF. C.DalyM.CarstenO. M. (2014). Transition to manual: Driver behaviour when resuming control from a highly automated vehicle. Transport. Res. Part F: Traffic Psychol. Behav. 27, 274–282. doi: 10.1016/j.trf.2014.09.005
32
MeratN.SeppeltB.LouwT.EngströmJ.LeeJ. D.JohanssonE.et al. (2019). The “out-of-the-loop” concept in automated driving: proposed definition, measures and implications. Cognit. Technol. Work21, 87–98. doi: 10.1007/s10111-018-0525-8
33
MerrittS. M.IlgenD. R. (2008). Not all trust is created equal: dispositional and history-based trust in human-automation interactions. Hum. Factors50, 194–210. doi: 10.1518/001872008X288574
34
MuirV. (1996). Trust in automation: part II. Theoretical issues in the study of trust and human intervention in automation systems. Ergonomics39, 429–460. doi: 10.1080/00140139608964474
35
Oviedo-TrespalaciosO.TichonJ.BriantO. (2021). Is a flick-through enough? A content analysis of advanced driver assistance systems (adas) user manuals. PLoS ONE16:e0252688. doi: 10.1371/journal.pone.0252688
36
ParasuramanR.MolloyR.SinghI. L. (1993). Performance consequences of automation-induced 'complacency'. Int. J. Aviat. Psychol. 3, 1–23. doi: 10.1207/s15327108ijap0301_1
37
ParasuramanR.RileyV. (1997). Humans and automation: use, misuse, disuse, abuse. Hum. Factors39, 230–253. doi: 10.1518/001872097778543886
38
PriceM.LeeJ. D.DinparastdjadidA.ToyodaH.DomeyerJ. (2019). Effect of automation instructions and vehicle control algorithms on eye behavior in highly automated vehicles. Int. J. Automot. Eng. 10, 73–79. doi: 10.20485/jsaeijae.10.1_73
39
RabinerL. R.JuangB. (1986). A tutorial on hidden markov models. IEEE ASSp Magazine3, 4–16. doi: 10.1109/MASSP.1986.1165342
40
RasmussenJ. (1983). Skills, rules, and knowledge; signals, signs and symbols, and other distinction in human performance models. IEEE Trans. Syst. Man Cybernet. 13, 61–57. doi: 10.1109/TSMC.1983.6313160
41
RowanK. E. (1991). Goals, obstacles, and strategies in risk communication: a problem-solving approach to improving communication about risks. J. Appl. Commun. Res. 19, 300–329. doi: 10.1080/00909889109365311
42
SAE International (2021). “Taxonomy and definitions for terms related to driving automation systems for on-road motor vehicles,” in Technical Report J3016_202104. Warrendale, PA: SAE International.
43
SchnebelenD.CharronC.MarsF. (2020). Estimating the out-of-the-loop phenomenon from visual strategies during highly automated driving. Accid. Analy. Prevent. 148:105776. doi: 10.1016/j.aap.2020.105776
44
SmallwoodJ.SchoolerJ. W. (2015). The science of mind wandering: Empirically navigating the stream of consciousness. Annu. Rev. Psychol. 66, 487–518. doi: 10.1146/annurev-psych-010814-015331
45
SwellerJ.AyresP.KalyugaS. (2011). Cognitive Load Theory. Cham: Springer.
46
UnderwoodG.ChapmanP.BrocklehurstN.UnderwoodJ.CrundallD. (2003). Visual attention while driving: sequences of eye fixations made by experienced and novice drivers. Ergonomics46, 629–646. doi: 10.1080/0014013031000090116
47
WalkerF.WangJ.MartensM. H.VerweyW. B. (2019). Gaze behaviour and electrodermal activity: Objective measures of drivers' trust in automated vehicles. Transport. Res. Part F: Traffic Psychol. Behav. 64, 401–412. doi: 10.1016/j.trf.2019.05.021
48
WarmJ. S.ParasuramanR.MatthewsG. (2008). Vigilance requires hard mental work and is stressful. Hum. Factors50, 433–441. doi: 10.1518/001872008X312152
49
YoungM. S.StantonN. A. (2002). Malleable attentional resources theory: a new explanation for the effects of mental underload on performance. Hum. Factors44, 365–375. doi: 10.1518/0018720024497709
Summary
Keywords
driver education, gaze behavior, Level 2 automation, mind-wandering, situation awareness, supervisory attention, take over performance, trust in automation
Citation
Chouchane H, Sakamura Y, Sato K, Abe G and Itoh M (2026) Trust rises, attention falls: divergent effects of exposure and education in driving automation. Front. Comput. Sci. 8:1772813. doi: 10.3389/fcomp.2026.1772813
Received
21 December 2025
Revised
10 March 2026
Accepted
26 March 2026
Published
21 April 2026
Volume
8 - 2026
Edited by
Andrej Košir, University of Ljubljana, Slovenia
Reviewed by
Liza Dixon, University of Ulm, Germany
Jinzhen Dou, Southeast University, China
Updates
Copyright
© 2026 Chouchane, Sakamura, Sato, Abe and Itoh.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Hanna Chouchane, hanna@chouchane.com
Disclaimer
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.