Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 28 November 2022
Sec. Brain Imaging Methods
This article is part of the Research Topic Brain Imaging Relations Through Simultaneous Recordings View all 12 articles

Reliability of MUSE 2 and Tobii Pro Nano at capturing mobile application users' real-time cognitive workload changes

\nLimin Zhang
Limin Zhang1*Hong CuiHong Cui2
  • 1China School of Fine Arts, Huaiyin Normal University, Huaian, China
  • 2USA School of Information, University of Arizona, Tucson, AZ, United States

Introduction: Despite the importance of cognitive workload in examining the usability of smartphone applications and the popularity of smartphone usage globally, cognitive workload as one attribute of usability tends to be overlooked in Human-Computer Interaction (HCI) studies. Moreover, limited studies that have examined the cognitive workload aspect often measured some summative workloads using subjective measures (e.g., questionnaires). A significant limitation of subjective measures is that they can only assess the overall, subject-perceived cognitive workload after the procedures/tasks have been completed. Such measurements do not reflect the real-time workload fluctuation during the procedures. The reliability of some devices on a smartphone setting has not been thoroughly evaluated.

Methods: This study used mixed methods to empirically study the reliability of an eye-tracking device (i.e., Tobii Pro Nano) and a low-cost electroencephalogram (EEG) device (i.e., MUSE 2) for detecting real-time cognitive workload changes during N-back tasks.

Results: Results suggest that the EEG measurements collected by MUSE 2 are not very useful as indicators of cognitive workload changes in our setting, eye movement measurements collected by Tobii Pro Nano with mobile testing accessory are useful for monitoring cognitive workload fluctuations and tracking down interface design issues in a smartphone setting, and more specifically, the maximum pupil diameter is the preeminent indicator of cognitive workload surges.

Discussion: In conclusion, the pupil diameter measure combined with other subjective ratings would provide a comprehensive user experience assessment of mobile applications. They can also be used to verify the successfulness of a user interface design solution in improving user experience.

Introduction

Portable media devices, such as smartphones, have become an increasingly pervasive part of our lives. In 2020, the number of smartphone users in the United States was estimated to reach 294.15 million and will reach 311.53 million by 2025 (O'Dea, 2021). American adults spent around 3 h and 30 min per day using mobile phones in 2019, with an increase of about 20 min from 2018, according to Zenith (Molla, 2020). Correspondingly, the number of applications in the App Store has soared from the initial 500 in 2008 to roughly 2.22 million available applications in 2021 (Ceci, 2022). As a result, mobile phone applications receive greater attention from the Human–Computer Interaction (HCI) field, resulting in a surge in the number of publications. We input a query “usability AND phone AND application” with custom time ranges: 1991–2000, 2001–2010, and 2011–2020 in Google Scholar, and get 8,500, 55,500, and 68,200 results.

Researchers in Human–Computer Interaction (HCI) fields have long recognized usability as the core of product design, including the application design of smartphones (Shneiderman, 1986; Nielsen, 1993; Brooke, 1996; Dumas et al., 1999). Previous research has manifested that cognitive workload is an essential aspect of product usability (Harrison et al., 2013; Davids et al., 2015).

Measuring cognitive workload has been recognized as one challenge when taking objectivity and causality into consideration (Brunken et al., 2003; Brünken et al., 2010). Instruments such as the NASA questionnaire (Hart and Staveland, 1988) help solicit perceived cognitive workload from users after a task is completed. Results obtained through such instruments are tinted with a level of subjectivity and put the causality between stimuli and reported cognitive workload in question.

On the other hand, electroencephalogram (EEG) devices can objectively monitor and record the brain's electrical activities and researchers have successfully identified signals from EEG to measure cognitive workload (Gevins and Smith, 2003; Antonenko et al., 2010a; Makransky et al., 2019). And eye movement data have been collected and analyzed to guide and advise various aspects of product design: navigation, page layout, user interface (UI) visualization style with design elements, advertisement, user viewing behaviors, and user cognitive workload (Goldberg and Wichansky, 2003; Nielsen and Pernice, 2010).

However, most of the studies were not executed in a smartphone setting and they cannot provide direct evidence for the reliability of EEG and eye-tracking devices to measure cognitive workload in a smartphone setting, due to several variabilities between desktop/laptop computer settings and smartphone settings. The screen sizes of desktop/laptop computers and smartphones are different: large vs. small. Users' interactions with these devices are distinct: cursors vs. gestures. The content compositions are not the same either: columns vs. scrolling. Physically, the users interact with their smartphones in different manners, such as: (1) one-handed, (2) two-handed, and (3) cradled, (4) no-handed; and in three body postures: walking, standing, and sitting/lying (Hoober, 2013).

Based on a thorough review of the related literature, we have identified three gaps as follows:

(1) Despite the significance of cognitive workload, it tends to be overlooked in the HCI field (Zhang and Adipat, 2005; Coursaris and Kim, 2006; Harrison et al., 2013).

(2) The majority of studies we reviewed only examine the overall cognitive workload during tasks and fail to study the instantaneous or peak cognitive workload during tasks and its relationship with product interface design and usability.

(3) There is little direct evidence to suggest that EEG and eye-tracking devices are reliable in measuring cognitive workload in a smartphone setting.

To address these gaps, we need to answer the two questions first:

(1) Are EEG data collected by MUSE 2 and eye movement data recorded by Tobii Pro Nano valid, reliable, and feasible as assessment tools for the real-time cognitive workload?

(2) Are measures collected by the two devices (averages of Event-related (de-)synchronization (ERD) of Alpha, Beta, and Event-related synchronization (ERS) of Theta for TP9, TP10, AF7, and AF8; pupil dilation, saccade duration and saccade number, fixation duration, and fixation number) sensitive to the cognitive workload of different N-back tasks in real time when the tasks are completed on a smartphone?

To answer the questions asked above, we employed a low-cost and portable electroencephalogram (EEG) device (MUSE 2, https://choosemuse.com/muse-2-guided-bundle/) and a user-friendly eye-tracking device (Tobii Pro Nano, https://www.tobiipro.com/product-listing/nano/) to detect real-time cognitive workload changes during N-back tasks on a smartphone. Our hypotheses were simple—we predict that the EEG device, MUSE 2, and the eye tracker device, Tobii Pro Nano with smartphone adopters, are reliably quantifying the cognitive workload of users performing tasks on a smartphone by these measures listed above.

Background

Cognitive workload in usability

According to the latest ISO 9241-11 (2018), usability is “the extent to which a system, product, service can be used by specified users to achieve specified goals with effectiveness, efficiency, and satisfaction in a specified context of use.” Various standards and models list a range of attributes for usability. Among these attributes, the cognitive workload is defined by Bevan and MacLeod (1994) as the mental effort required to perform tasks and is particularly important in safety-critical applications. It refers to the user's cognitive processing amount to using the application (Harrison et al., 2013).

Cognitive workload measurements in HCI

The cognitive workload measurements can be roughly grouped into three broad categories: subjective self-assessment rating scales, performance measures, and psychophysiological measures (Wilson and Eggemeier, 1991; Cain, 2007; Evans and Fendley, 2017). Here, we only introduce two measures adopted in this research: electroencephalogram (EEG) and eye movement in psychophysiological measures.

Measurement of cognitive workload using electroencephalogram

Electroencephalogram (EEG) is an electrophysiological method of monitoring and recording the brain's electrical activity. Most of the time, an EEG device that comprises non-invasive electrodes is placed along a subject's scalp. These electrodes capture voltage fluctuations resulting from ionic currents within the brain's neurons.

In recent years, researchers have been evaluating the potential of the EEG as a measure of cognitive workload in different task conditions: arithmetic tasks (Anderson et al., 2011; Cirett Galán and Beal, 2012; Kumar and Kumar, 2016; Borys et al., 2017; Chin et al., 2018); cognitive tasks (Trammell et al., 2017); reading tasks (Dimigen et al., 2011; Knoll et al., 2011; Gwizdka et al., 2017); listening to music tasks (Asif et al., 2019); visual search task (Winslow et al., 2013; Hild et al., 2014); learning tasks (Dan and Reiner, 2017; Mazher et al., 2017; Notaro and Diamond, 2018); and vehicle driving task (Cernea et al., 2012). These studies confirm the fact that EEG provides reliable signals for studying cognitive workload in their respective settings.

Event-related (de-)synchronization (ERD/ERS) with Alpha, Theta, and Beta bands is one of the three most popular analysis techniques (Cabañero et al., 2019). Event-related (de)synchronization (ERD) is a recognized rate-of-change metric for oscillatory EEG dynamics, which was originally developed to quantify changes in the Alpha band (Pfurtscheller and Aranibar, 1977). Synchronization is a process where neurons are getting in line (synchronized) to enter an idling state. Desynchronization is a process where individual neurons get ready to perform their parts in a task. The steps of performing a task are: neurons desynchronize (wake up), perform tasks, and neurons synchronize (rest).

To obtain percentage values for ERD/ERS, the power within the frequency band of interest in the period after the event is given by A, whereas that of the preceding baseline or reference period is given by R. The percentage decrease (or increase) from the reference interval (R) to the activation interval (A) (before responding) was defined as

ERD/ERS%=[(R--A)/R]*100%    (1)

(Pfurtscheller and Aranibar, 1977; Pfurtscheller and Lopes da Silva, 1999; Pfurtscheller, 2001).

Negative values computed by Equation 1 indicate power increase and desynchronization (ERD), and positive values indicate power decrease and synchronization (ERS).

Pfurtscheller and Lopes da Silva (1999) recommended that the term ERD is meaningful only if the baseline measured some seconds before the event represents rhythmicity seen as a clear peak in the power spectrum. Similarly, the term ERS only has a meaning if the event results in the appearance of a rhythmic component and therefore in a spectral peak that was initially not detectable (Pfurtscheller and Lopes da Silva, 1999).

The quantification of ERD/ERS was divided into four steps, first, the bandpass filtering was carried out for all Event-related trials; second, the amplitude samples were squared to obtain the power samples; third, the power samples of all trials were averaged; and fourth, the time samples were averaged to make the data smooth and reduce (Pfurtscheller and Lopes da Silva, 1999).

The review articles (Klimesch, 1999; Antonenko and Niederhauser, 2010b) concluded that with increasing task demands Theta synchronizes (decreases), whereas Alpha and Beta desynchronize (increase) (Pfurtscheller and Berghold, 1989; Neubauer and Fink, 2003; Stipacek et al., 2003; Klimesch et al., 2005; Neubauer et al., 2006; Scharinger et al., 2016; Saitis et al., 2018).

Measuring cognitive workload using eye movement data

Multiple kinds of eye movement data related to cognitive workload can be reliably collected using a high-quality eye-tracking device.

Pupil dilation is an involuntary response, in which the pupil diameter changes to protect the retina or to respond to a shift in fixation between objects at different distances. Previous research has shown that users' pupils dilate when the difficulty of the task increases and more cognitive effort has been allocated to solve the task (Granholm et al., 1996; Pomplun and Sunkara, 2003; Klingner et al., 2008; Chen et al., 2011; Porta et al., 2012; Rafiqi et al., 2015; Gavas et al., 2017; Ehlers, 2020). Accounting for individual and environmental differences, it is necessary to measure pupil diameters while referencing an adaptive baseline (Lallé et al., 2016).

According to Purves et al. (2001), saccades are rapid and ballistic movements of eyes that change the fixations abruptly. Previous research has found that growth in saccade velocity indicates a greater task difficulty (Barrios et al., 2004; Chen et al., 2011; Lallé et al., 2016; Zagermann et al., 2018).

Eye fixation refers to a focused state when eyes dwell voluntarily over some time and is the most common type of eye-tracking event (Zagermann et al., 2016). Previous research has proven that the correlation between the duration of fixation and the cognitive processing level is positive (Rudmann et al., 2003; Goldberg and Helfman, 2010; Chen et al., 2011; Wang et al., 2014; Zagermann et al., 2018).

Devices in measuring cognitive workload

As previously reviewed literature shows, the measures of computing from EEG and eye movement data have been proven to be effective for detecting cognitive workload changes. However, most of the studies were conducted in smartphone settings, and the devices adopted in these studies are not suitable for use in smartphone usability testing environment.

Grateful to technology development, there are a wide range of choices in the selection of devices to capture the EEG data and eye-tracking data, respectively. Some examples of the EEG devices, ordered at prices, from low to high include: MUSE 2 headband, Emotiv Insight, OpenBCI, ANT Neuro, BioSemi, etc. (Farnsworth, 2019). A ranking of the top eye-tracking companies, ordered by the number of publications found through Google Scholar, is Tobii, SMI, EyeLink, Smart Eye, LC Technologies, Gazepoint, The Eye Tribe, etc. (Farnsworth, 2020).

Among listed choices, the MUSE 2 headband (Figure 1A, $250) is an easy-to-use, affordable, and portable EEG recording system from InteraXon Inc. It is a four-channel headband with dry electrodes at positions AF7, AF8, TP9, and TP10 (Figure 1B). The headband is connected to the app on phone via Bluetooth, which makes it a great tool for detecting cognitive workload while the user is performing the task on smartphones, especially in some field experiments, of course, after its reliability is verified.

FIGURE 1
www.frontiersin.org

Figure 1. (A) MUSE 2 EEG headband; (B) EEG electrode positions in the 10-10 system using modified combinatorial nomenclature, along with the fiducials and associated lobes of the brain. Adopted from https://en.wikipedia.org/wiki/10%E2%80%9320_system_%28EEG%29#/media/File:EEG_10-10_system_with_additional_information.svg.

Despite the small number of sensors and the mismatch in the locations of the sensors to the standard 10–20 electrode positioning system, several studies have shown that the MUSE headband has the potential to provide good quality EEG data. Two studies (Arsalan et al., 2019; Asif et al., 2019) adopted MUSE 2 to capture EEG data and adopted classifiers to classify stress levels. Another study (Papakostas et al., 2017) also adopted MUSE EEG to predict the user task performance, and they achieved a maximum accuracy rate of 74%. Krigolson et al. (2017) collected data by MUSE EEG system, and the results showed quantifiable N200 and P300 Event-related potential (ERP) components in the visual oddball task and the reward positivity. However, these studies cannot provide direct evidence on the EEG data captured by the MUSE EEG system is reliable for cognitive workload changes.

Some studies have pointed out MUSE's limitations. Ratti et al. (2017) compared two medical grade (B-Alert, Enobio) and two consumer (MUSE, Mindware) EEG systems in five healthy subjects. Results showed that EEG data can be successfully collected from four devices, yet MUSE showed a broadband increase in power spectra and the highest relative variation across test–retest acquisitions. Another study has also shown that the data collected by MUSE headband were of poor quality under noisy conditions, such as at a public lecture (Przegalinska et al., 2018). To explore MUSE 2's potential as a great tool in smartphone usability testing, we still need direct empirical evidence on the reliability of MUSE 2 in capturing EEG data for measuring cognitive workload.

Having picked an EEG device with its usefulness still under investigation, we selected a well-established eye-tracking device for this study to control the risk. We chose Tobii Pro Nano because it is one of the top eye-tracking companies and has been used in 20.5 k publications. It is also an accessible and efficient approach to capturing eye movement (Figure 2) and is used by many HCI researchers (Sugaya, 2019; Ehlers, 2020; Lee and Chenkin, 2020). Ehlers (2020) adopted Tobii Pro Nano to capture the pupil diameter and confirmed that it is a valid indicator of cognitive workload. Lee and Chenkin (2020) evaluated Tobii Pro Nano's potential to differentiate between experts and novices in the interpretation of POCUS clips in medical fields. Sugaya (2019) used Tobii Pro Nano to test an assumption about the meaning-making process of adjective expression formation.

FIGURE 2
www.frontiersin.org

Figure 2. Tobii Pro Nano with mobile testing accessory.

Tobii Pro Nano can be mounted on a mobile testing accessory, also manufactured by Tobii (Figure 2). It has a screen capture device (Figure 2) connecting directly to smartphones. The screen capture device records a high-definition (HD) video of the mobile device's screen at 60 frames per second with a latency of only 10 milliseconds (Mobile Testing Accessory | Perfect for Usability Tests., 2020). Yet, the mobile testing accessory is just in the market, with no research done on it. The other great device for smartphone experiments is Tobii Pro Glass with a much higher price. If we can provide a piece of evidence on the reliability of Tobii Pro Nano with a mobile testing accessory, it could be a high-performance cost ratio choice for researchers.

Methods

The goal of the experiment was to examine the reliability of the MUSE 2 headband and Tobii Pro Nano with a mobile testing accessory for detecting cognitive workload changes during a smartphone task and to select the best measure(s) computed from data collected by the two devices. The measures adopted in the experiment are (1) ERD percentage for Alpha, Beta, and ERS Theta rhythms extracted from EEG data; (2) multiple eye movements: pupil dilation, saccade duration, saccade number in second, fixation duration, and fixation number in second; and (3) user performance data: reaction time and accuracy rate.

Participants

This study was approved by the UA IRB office (Protocol Number: 2101428836) and obtained permission from Qinghe High School, Jiangsu, China.

We recruited 5 students as pilots and 30 students as participants from Qinghe High School, Jiangsu, China. The inclusion criteria were normal vision or correct to normal vision, normal cognitive function, and proficiency in smartphones.

Students who participated either as pilots or as participants were compensated 50 in Chinese currency after they complete the task successfully.

Apparatus

We used the MUSE 2 headband and Tobii Pro Nano as the devices to collect cognitive workload-related measures.

The environment's brightness variations produce changes in the pupil size (Pfleging et al., 2016; Zagermann et al., 2016). Therefore, the experiment was conducted in a room with lightproof curtains down to avoid natural lighting conditions, and electric lights on the room ceiling created a consistent lamination for the experiment. Environment, such as noise, also impacts cognitive load (Örün and Akbulut, 2019). We made sure the experiment room was free of all noise during the experimental sessions. All devices were sanitized before the next participant came.

When using EEG devices, one has to fulfill several other requirements. These include a clean scalp, clean electrodes, minimum participant activities, including head movements, since a small movement could generate muscle-based signals known as artifacts (Pratama et al., 2020). We instructed all participants to stay as still as possible and not to wear any makeup during the experiment. Also, as the electrodes need to be attached to the back of the ears, we encouraged the participants to wear contacts instead of glasses. We also provided a disposable wet cloth for participants to moisturize their foreheads and back of the ears to get a better connection of the EEG headset.

Task: N-back task

N-back tasks are continuous-recognition measures that present stimulus sequences, such as letters or pictures. A sequence of stimuli is presented to the participants one by one. The participants are required to make a decision as to whether the current stimulus is the same as the one presented in N trials ago (Coulacoglou and Saklofske, 2017). The N can be 0, 1, 2, 3, etc. There is an increase in difficulty in tasks while N increases. An N-back task is a useful tool for experimental research on working memory (Jaeggi et al., 2010), and it has been adopted to manipulate cognitive workload (Reimer et al., 2009; Ayaz et al., 2010; Yokota and Naruse, 2015).

In this study, all participants completed an N-back task. When employed in a computer setting, the participants of the experiment can press individual keys on keyboards as “YES” or “NO”. To cope with the touch screen of a smartphone, we placed “ × ” on the left bottom corner, and “√” on the right bottom corner of the smartphone screen (Figure 3).

FIGURE 3
www.frontiersin.org

Figure 3. A screenshot of the N-back task.

In this study, we employed a 1 back task and a 2 back task to create a low cognitive workload condition and a high cognitive workload condition. The rationale of only including 1 and 2 levels is to simulate cognitive workload levels that smartphone users would experience in the real world.

The key features of the N-back task implementation were:

• Four sets of letters were created and arranged in two groups for a training block and an experiment block.

° Training Block:

* Five trails of one back task (EEIPP) as a training session,

* Six trails of two back task (OSOMLI) as a training session;

° Experiment Block:

* low cognitive workload block: 20 trials of 1 back task (DAABEEDRRODHHRDSSELDD);

* high cognitive workload block: 21 trails of 2 back (BAEAAEASHSAELEOBBBOSHS).

• These two sets of letters in the experiment block were designed to have an identical “YES” or “NO” response sequential: YNNNYNNNYNNNYNNYNNYN.

• Each stimulus was presented for maximally 3,000 milliseconds.

The low and high cognitive workload blocks were randomly and evenly assigned to participants. More specifically, 15 participants assigned an odd ID completed the task in Order 1: low cognitive workload block, high cognitive workload block; and another 15 participants assigned an even ID completed the task in Order 2: high cognitive workload block, low cognitive workload block.

Procedure

All participants entered the experiment room and performed the experiment once at a time.

First, the participants watched an instructional video of the instruments and experimental procedure (https://youtu.be/_d24CRSwhuQ). They were free to ask any questions after viewing the video.

The experiment started with the participant filling in a demographic questionnaire (Appendix A). This questionnaire covered subjects' age, gender, strong hand, experience with smartphones, and current smartphone usage situation. Then, they wore the MUSE 2 headband and adjusted themselves to a comfortable sitting position. After that, the participants completed an eye-tracking calibration with Tobii Pro Nano followed by 10 s with an eye-open relaxed position and another 10 s with an eye-closed comfortable position.

After the preparation step, the participants completed the training session. They can ask any questions about the N-back task during or after the training section. The training sessions were excluded from data analysis.

Then the participants completed the experiment session of the N-back task at the experiment station, wearing the MUSE 2 headband. They first completed 20 trails of 1-back/2-back stimuli, followed by 20 trails of 2-back/1-back stimuli, with intervals of approximately 1–2 s in between each 1-back/2back stimuli (the time between a response and the display of the next stimuli) and a rest period of 5 s in between 1-back and 2-back blocks. The participants were instructed to respond to tasks as accurately and rapidly as possible. The variation of intervals between each 1-back/2back stimuli caused by the internet loading time varied.

The MUSE 2 headband collected raw EEG data of TP9, AF7, AF8, and TP10 through an application called Mind Monitor (iOS Version 2.2.0) (Figure 4A). And the Tobii Pro Nano recorded multiple types of eye movement data: pupil dilation, saccade length, saccade velocity, fixation duration, and fixation number (Figure 4B). The self-developed website for the N-back task collected the reaction time and accuracy rate during the experiment (Figure 4C) (N-back task website: http://n-back.artkey.xin/). It is designed for an experiment on a smartphone, and it works best on a smartphone.

FIGURE 4
www.frontiersin.org

Figure 4. (A) The MUSE 2 headband was transmitted over Bluetooth to a phone via an application called Mind Monitor for EEG data collection, and the EEG data were uploaded to the research's selected cloud drive after each participant completed the task. (B) The N back task was recorded on a website designed by researchers, and the user performance data were collected via the server. (C) The Tobii Pro Nano eye tracker was connected to a laptop with Tobii Pro Lab software installed, and the eye movement data were collected by the Tobii Pro Lab software and stored in the hard drive of the laptop.

Besides the procedure described above, the procedure described in Appendix B when conducting the experiments as a precaution against COVID-19 was followed as well.

Measures

Data collected from the experiment allow us to examine and compare the following measures:

(1) Event-Related Synchronization percentage (ERS) of Theta, Event-Related Desynchronization percentage (ERD) of Alpha, and Event-Related Desynchronization percentage (ERD) (Equation 1).

(2) Multiple eye movements: pupil dilation, saccade duration, saccade number in second, fixation duration, and fixation number in second.

(3) User performance data: reaction time (RT) and accuracy rate (AR). Reaction time (RT) is the period between the onset of a letter and the response made by a participant. The accuracy rate (AR) is the ratio of the number of correct inputs and the total number of inputs.

Results, analysis, and discussion

The experiment has investigated the feasibility of using data acquired wirelessly from an EEG headband (MUSE 2) and an eye-tracking device (Tobii Pro Nano) to assess cognitive workload in a well-controlled N-back task in a smartphone setting.

Demographic data

A total of 30 high school students from Qinghe High School in Huaian, Jiangsu, China completed the experiment, including the demographic questionnaire (Appendix A). Eleven of the 30 participants were female and 19 were male, and the female and male ratio is 11/19. Their average age was 16.34 years old (SD = 0.61). All participants were from the first year in high school. The right hand was the dominant in 29 participants, and the other participant was ambidextrous. The answers to Questions 4–8 of the demographic questionnaire are presented in Table 1. To summarize the data in Table 1, all participants were frequent and proficient smartphone users.

TABLE 1
www.frontiersin.org

Table 1. Participants' answers for Questions 4–8 in the demographic questionnaire.

User performance data

Only 29 participants' user performance data were processed, analyzed, and discussed here. Participant ID 126's experimental data were not recorded due to Internet connection issues.

We analyzed user performance data (reaction time and accuracy rate) in order to confirm that participants perceived the various N-Back conditions as different. User performance data processing, user performance data results, and user performance data analysis are included in this section.

User performance data processing

The N-Back website recorded participants' ID, N-back order, current letter, participants' choices, accuracy (yes/no/null), reaction time, start timestamp in Unix time, and end timestamp in Unix time. Unix time is a system for describing a point in time and it is the number of seconds that have elapsed since the Unix epoch, excluding leap seconds (Ritchie and Thompson, 1978).

I conducted a Shapiro–Wilk test for the reaction time (RT) of all participants for 1 back and 2 back tests to check its normality. The result is significant (p < 0.001), which indicates that the data are not normally distributed. Therefore, I conducted a Mann–Whitney–Wilcoxon Test for the reaction time (RT) between 1 and 2 back.

User performance data results

The descriptive statistics and the Mann–Whitney–Wilcoxon Test results for the reaction time (RT) and accuracy rate (AR) between conditions are presented in Table 2.

TABLE 2
www.frontiersin.org

Table 2. The descriptive statistics and Mann–Whitney–Wilcoxon Test results for reaction time (RT) (unit: second) and accuracy rate (AR) between 1 and 2 back.

User performance data analysis

Reaction time

The Mann–Whitney U test was conducted to examine whether the reaction time (RT) had statistically significant differences between 1 and 2 back for all participants, for participants with odd IDs, and for participants with even IDs. The p-values (< 0.001) indicate the answer is yes, as expected (Table 2).

Accuracy rate

The Mann–Whitney–Wilcoxon test was conducted to examine whether the accuracy rate (AR) had significant differences between 1 back and 2 back for the 29 participants. The p-value (<0.001) indicates the answer is yes, as expected. The median of 1 back accuracy rate is 1, which is higher than the median of 2 back accuracy rate (0.9) (Table 2).

User performance data discussion

As expected, the reaction time (RT)increased and accuracy rate (AR) decreased with the leveled-up differential of N-Back tasks, and 1 and 2 back tasks did create low and high cognitive workload conditions for the experiment's participants.

EEG data

EEG data processing and analysis

Only 29 participants' EEG data were processed, analyzed, and discussed here. Participant ID 126's EEG was not recorded due to Internet connection issues.

During the experiment, participants wore a MUSE 2 headband connected to the Mind Monitor. The Mind Monitor collected their EEG data. According to the Technical Manual from the Mind Monitor website, bandpass filtering was carried out on the raw data with power noise at 50 Hz or 60 Hz. Then, a fast Fourier transform (FFT) calculation (Heckbert, 1995) was applied to the raw data to get Theta, Alpha, and Beta.

The recorded EEG signals were processed using Excel and R to get two baselines:

• baseline_near: Based on the timestamps recorded by the N-back task website, we sectioned the intervals starting from−200 ms to the onset of each letter as the baseline interval for each letter (Xiang et al., 2021).

• baseline_away: According to the timestamps recorded by the N-back task website, we segmented the first 3,000 ms of the 10 s relaxing eyes open relaxing as a baseline.

Computed ERD of Alpha {TP9, AF7, AF8, TP10}, ERD of Beta{TP9, AF7, AF8, TP10}, and ERS of Theta {TP9, AF7, AF8, TP10} with baseline near for each letter interval; and ERD of Alpha AF7, Alpha AF8, Beta AF8, and Beta TP9 with baseline away for each letter interval. The interval starts from the onset of each letter to the time point that a choice is being made by participants, which is definitely ≤ 3 s.

A non-parametric test, the Mann–Whitney–Wilcoxon test, was selected for non-normal data. The Mann–Whitney–Wilcoxon test was conducted for 12 measures (Alpha {TP9, AF7, AF8, TP10}, ERD of Beta{TP9, AF7, AF8, TP10}, and ERS of Theta {TP9, AF7, AF8, TP10} with baseline_near between 1 back and 2 back; and ERD of Alpha AF7, Alpha AF8, Beta AF8, and Beta TP9 with baseline away 1 back and 2 back. See Appendix C for the details.

EEG data results

The average workload is the average value of instantaneous loads within a task duration (Xie and Salvendy, 2000). In this study, the average cognitive workload is represented by the average ERD of Alpha, Beta, and ERS of Theta for TP9, TP10, AF7, and AF8 of the intervals of each letter with baseline_near.

To examine whether there is a difference between 1 and 2 back for averages of ERD of Alpha, Beta, and ERS of Theta for TP9, TP10, AF7, and AF8 with baseline near, we conducted Mann–Whitney–Wilcoxon tests between 1 and 2 back of all participants. The results are included in Table 3. Due to computational problems, the numbers of subjects (N) vary between the analyses.

TABLE 3
www.frontiersin.org

Table 3. The descriptive statistics and Mann–Whitney–Wilcoxon test results for averages of ERD of Alpha, Beta, and ERS of Theta for TP9, TP10, AF7, and AF8 with baseline near between 1 and 2 back.

It is not feasible to have a baseline for each stimulus in scenarios of users experiencing an application on smartphones. Hence, to explore the feasibility of adopting a single baseline away with stimulus, we conducted Mann–Whitney–Wilcoxon tests for averages of ERD of Alpha AF7 and Beta TP9 with baseline away between 1 and 2 back for all participants, participants with odd IDs, and participants with even IDs, respectively. We find no significant results (Table 4).

TABLE 4
www.frontiersin.org

Table 4. The descriptive statistics and Mann–Whitney–Wilcoxon test results for averages of ERD of Alpha_AF7 and Beta_TP9 with baseline away between 1 back and 2 back.

Analysis of EEG results

Different cognitive workloads evoked associated human brain oscillatory responses (Krause et al., 2000; Pesonen et al., 2007) that made it possible to measure the corresponding cognitive workload levels.

The results in Table 3 show that only Alpha ERD AF7 and Beta ERD TP9 among the 12 measures (TP9, AF7, AF8, TP10 * Alpha, Beta, and Theta) are sensitive to the different workloads of between 1 and 2 back conditions for all participants (p < 0.05).

The magnitudes of Alpha ERD AF7 and Beta ERD TP9 are significantly greater for the 2 back than for the 1 back (Figure 5), indicating that Alpha and Beta increase as tasks demand more cognitive workload. This is in line with previous studies that found with inclining task demands, Alpha and Beta desynchronize (increase) (Klimesch, 1999; Stipacek et al., 2003; Klimesch et al., 2005; Neubauer et al., 2006; Antonenko et al., 2010a; Antonenko and Niederhauser, 2010b; Xiang et al., 2021).

FIGURE 5
www.frontiersin.org

Figure 5. Box plots with medians between 1 back and 2 back for Alpha ERD AF7 (A) and Beta ERD TP9 (B).

The insensitivity of the six measures (AF8 and TP10 * Alpha, Beta, and Theta) can be explained by the functions of the cerebral cortex.

The TP10 electrode is positioned behind the right ear, which is the right temporal lobe, and the AF8 electrode is on the right forehead, which is the right frontal lobe (Figure 1). The TP9 electrode is positioned behind the left ear, which is the left temporal lobe, and the AF7 electrode is on the left forehead, which is the left frontal lobe (Figure 1).

The left temporal lobe is associated with understanding language, learning, memorizing, forming speech, and remembering verbal information (Guy-Evans, 2021b). We used English letters as a stimulus in the N-back task and participants' primary language is Chinese, thus it makes sense that Beta ERD TP9 are found to have significant differences between 1 and 2 back.

The AF7 electrode is on the left side of the frontal lobe. The frontal lobe is located behind the forehead, at the front of the brain. Each lobe controls the operations on opposite sides of the body: the left hemisphere controls the right side of the body and vice versa (Guy-Evans, 2021a). It is believed that the left frontal lobe works predominantly with language, logical thinking, and analytical reasoning. The right frontal lobe, on the other hand, is mostly associated with non-verbal abilities, creativity, imagination, musical, and art skills (Guy-Evans, 2021a).

The dominant hands of 29 participants were the right hand, and the remaining one was both hands. We observed all participants only used their right hands to make the choices. This explains that the Alpha ERD AF7 (left forehead) but not AF8 (right forehead) was found to have significant differences between 1 and 2 back. However, the other four measures of TP9 and AF7 shall be sensitive to the difference in cognitive workload, as previous studies prove.

Discussion on EEG data

In summary, MUSE 2 outputs good signals, but these signals may not be readily useful in the studies on the usability of smartphone applications for an entire and consecutive user experience as a result of the difficulty in selecting a sensible baseline due to two rationales.

The first rationale behind the difficulty in the selection of a sensible baseline lies in the fact that only Alpha ERD AF7 and Beta ERD TP9 show sensitivity to the difference in cognitive workload between 1 and 2 back. It is not consistent with that given in previous studies (Klimesch, 1999; Stipacek et al., 2003; Klimesch et al., 2005; Neubauer et al., 2006; Antonenko and Niederhauser, 2010b; Xiang et al., 2021).

Second, according to Pfurtscheller and Lopes da Silva (1999), ERD/ERS is required to have a baseline captured some seconds right before the events. Yet, it is not feasible to have a baseline for each stimulus in the scenarios of users experiencing any applications on smartphones. We had carried out an exploration of adopting a single baseline away with stimulus, but unfortunately it did not show any statistically significant differences for averages of ERD of Alpha AF7 and Beta TP9 between 1 and 2 back.

All in one sentence, although MUSE 2 is of consumer grade, comfortable to wear, and wireless connected, it is a reliable device for researchers to capure stable EEG data for measuring cognitive workload. It does show some promise for detecting cognitive workload elicited by isolated/independent elements in user interface (UI) design, and selective signals may be combined with eye-tracking data to detect UI issues that invoke user errors.

Eye movement data

Eye movement data processing and analysis

During the experiment, participants completed N-back tasks on a smartphone attached to the mobile testing accessory with the Tobii Pro Nano mounted on the top (Figure 2). Eye movement data were collected by the Tobii Pro Nano via the Tobii Pro Lab software (version 1.162.32461).

Similar to the EEG data, the recorded eye movement data were processed using Excel and R through several steps to get averaged pupil dilation left, averaged pupil dilation right, averaged fixation duration, averaged fixation number in second, averaged saccade duration, averaged saccade number in second, maximums of pupil dilation left, maximums of pupil dilation right, maximums of fixation duration, and maximums of saccade duration.

A non-parametric test was selected for non-normal data. See Appendix C for data processing and analysis detailed steps.

Eye movement data results

The Tobii Pro Nano was extremely sensitive to angle changes between participant's eyes and the device. Based on the experience we had gathered from pilots, we had a higher chair for the participants to improve the capture rate and adjusted the Tobii Pro Nano angle according to each participant. Despite the higher chair we had employed and the active adjustments made to the mobile testing accessory, the capture rates varied across participants. Eighteen participants' data with relatively higher capture rates were processed and analyzed here.

Averaged cognitive workload is quantified by the averaged pupil dilation left and the averaged pupil dilation right, the averaged fixation duration and the averaged fixation number, and the averaged saccade duration and the averaged saccade number.

To examine whether there were statistically significant differences for averaged pupil dilation changes between low and high cognitive workload conditions, we conducted a Mann–Whitney–Wilcoxon test for averaged pupil dilation left, averaged pupil dilation right, between 1 and 2 back. The results are presented in Table 5.

TABLE 5
www.frontiersin.org

Table 5. The descriptive statistics and Mann–Whitney–Wilcoxon test results for the averaged pupil dilation left, the averaged pupil dilation right, the averaged fixation duration, the averaged fixation number, the averaged saccade duration, and the averaged saccade number in second between 1 and 2 back.

In this research, the averaged fixation duration and the fixation number in second were adopted as the representative of the averaged cognitive workload during the intervals, starting from the appearance of each letter to the time point that choices were made.

To examine whether there were statistically significant differences for averaged fixation duration and fixation number in second between low and high cognitive workload conditions were observed, we conducted Mann–Whitney–Wilcoxon tests between 1 and 2 back. The results are presented in Table 5.

In this research, the averaged saccade duration and the averaged saccade number in second were adopted as the representative of the averaged cognitive workload during the intervals, starting from the appearance of each letter to the time point that choices were made.

To examine whether there were statistically significant differences for the averaged saccade duration and the averaged saccade number in second between low and high cognitive workload conditions, we conducted Mann–Whitney–Wilcoxon tests between 1 and 2 back. The results are presented in Table 5.

The maximum pupil dilation left and the maximum pupil dilation right were adopted as the representative of the peak cognitive workload during the intervals, starting from the appearance of each letter to the time point that choices were made.

To examine whether statistically significant differences for maximums of pupil dilation changes between low and high cognitive workload condition, we conducted Mann–Whitney–Wilcoxon tests for maximums of pupil dilation left and of pupil dilation right between 1 and 2 back. The results are presented in Table 6.

TABLE 6
www.frontiersin.org

Table 6. The descriptive statistics and Mann–Whitney–Wilcoxon test results for the maximum pupil dilation left and the maximum pupil dilation right between 1 and 2 back.

In this research, the maximums of fixation duration were adopted as the representative of the peak cognitive workload during the intervals, starting from the appearance of each letter to the time point that choices were made.

To examine whether statistically significant differences exist for the maximums of fixation duration between low and high cognitive workload conditions, we conducted Mann–Whitney–Wilcoxon tests between 1 and 2 back. We find no significant results (Table 7).

TABLE 7
www.frontiersin.org

Table 7. The descriptive statistics and Mann–Whitney–Wilcoxon test results for the maximums of fixation duration between 1 and 2 back for all selected participants.

In this research, the maximum saccade duration was adopted as the representative of the peak cognitive workload during the intervals, starting from the appearance of each letter to the time point that choices were made.

To examine whether statistically significant differences exist for the maximums of saccade duration between low and high cognitive workload conditions, we conducted Mann–Whitney–Wilcoxon tests for it between 1 and 2 back. We find no significant results (Table 7).

Eye movement results' analysis

Overall, the eye movement data collected by the Tobii Pro Nano are valid and reliable. Some measures (pupil dilation, saccade number in second, fixation number in second) are sensitive to the difference of average cognitive workload and peak cognitive workload introduced by the 1 or 2 back tasks.

The averages of pupil dilations of both eyes have been proven to be reactive to the differences of average cognitive workload between 1 and 2 back tasks consistently. As Table 5 reveals, there are statistically significant differences between 1 and 2 back for the averages of pupil dilation of both eyes (p < 0.05). The medians of the averages of pupil dilations of both eyes are larger in 2 back than in 1 back. The medians of the averages of pupil dilations of both eyes remain greater in the 2 back and in the 1 back, so that the bigger average of pupil dilations means a higher averaged cognitive workload. This finding is in line with earlier studies (Granholm et al., 1996; Pomplun and Sunkara, 2003; Klingner et al., 2008; Chen et al., 2011; Porta et al., 2012; Rafiqi et al., 2015).

The same pattern is discovered in the maximums of pupil dilation. There are statistically significant differences between 1 back and 2 back for maximums of pupil dilation of both eyes (p < 0.05 or < 0.001) (Table 6). The medians of the maximum pupil dilation of both eyes are larger in 2 back than in 1 back for both eyes, which indicates the larger maximum of pupil dilations means a higher peak cognitive workload.

As for fixation and saccade, statistically significant differences were observed between 1 and 2 back in the fixation number in second and saccade number in second (Table 6). Irreconcilable with previous findings is that the correlation between the number of fixation and cognitive workload is negative. Previous studies have concluded that an upswing number of fixations correlate with an increased cognitive load level (Goldberg and Helfman, 2010; Chen et al., 2011; Wang et al., 2014; Zagermann et al., 2018). And the higher number of saccades in second (saccade velocity) is also related to lower cognitive workload, opposing the previous research (Barrios et al., 2004; Chen et al., 2011; Lallé et al., 2016; Zagermann et al., 2018).

Discussion on eye movement data

In this study, we found that eye-tracking device, Tobii Pro Nano with mobile testing accessory, appears to be a valid instrument for monitoring the cognitive workload difference in a smartphone setting. This finding along with previous studies (Sugaya, 2019; Ehlers, 2020; Lee and Chenkin, 2020) can provide an initial empirical evidence on the reliability of Tobii Pro Nano with mobile testing accessory. Moreover, the average pupil dilation and the maximum pupil dilation have been ratified as the effective measures of cognitive workload difference in a smartphone setting, and they enlarge along with the difficulty levels of N-back task rising.

One incongruent finding is that the fixation velocity and saccade velocity decline with the increment of cognitive workload, while the previous studies found an upswing number of fixations correlate with an increased cognitive load level (Barrios et al., 2004; Goldberg and Helfman, 2010; Chen et al., 2011; Wang et al., 2014; Lallé et al., 2016; Alonso Dos Santos and Calabuig Moreno, 2018; Zagermann et al., 2018).

One possible justification for this reverse is the different task design. The N-back task only required participants to look at one spot on the screen, while the previous studies required participants to observe, scan, and search during tasks and the gazes were not fixed in one spot (Table 8).

TABLE 8
www.frontiersin.org

Table 8. The tasks in the papers.

Another obvious concern about the Tobii Pro Nano is the unstable capture rate. Only nearly half of the participants' data was captured enough to be adopted.

One pilot participant's capture rate was 0% and he mentioned that he had a high degree of astigmatism, around 500–600 in both eyes. Astigmatism is an imperfection in the curvature of your eye's cornea or lens (Boyd, 2021). It may be helpful to think of the normal eye as being shaped like a basketball. With astigmatism, it is shaped more like an American football. The Tobii Pro Nano may not effectively recognize the eyes of people with astigmatism. This suggests that the low capture rate for some participants may be caused by astigmatism. Therefore, information about astigmatism was obtained from the participants.

For the astigmatism degree, we averaged two eye degrees. The capture rate was recorded in the Tobii Pro Nano. We adopted Spearman's rho statistic to assess the correlation between capture rate and astigmatism degree, and the correlation coefficients and p values are given in Figure 6. The result shows that there is a statistically significant negative correlation between capture rate and astigmatism degree (p < 0.05 or < 0.001). The correlation coefficient is −0.55. The negative correlation between astigmatism and capture rate may have resulted from the changes in the shape of eyeballs.

FIGURE 6
www.frontiersin.org

Figure 6. The correlation between capture rate and astigmatism degree.

Overall, the objective of Experiment 1 was to assess the feasibility of using wirelessly acquired EEG (MUSE 2) and eye-tracking device (Tobii Pro Nano) to assess cognitive workload in a well-controlled N-back task in a smartphone setting. And the eye-tracking device, Tobii Pro Nano, can be adopted as a device to collect eye movement data to monitor cognitive workload fluctuations in a smartphone setting with a screen for high astigmatism, and pupil dilation can be measured for cognitive workload differences.

Conclusion and future directions

This study aimed to verify the feasibility of using eye-tracking (i.e., Tobii Pro Nano) and low-cost electroencephalogram (EEG, i.e., MUSE 2) devices to measure real-time cognitive workload changes during mobile application use, and which measures are sensitive to cognitive workload differences. Results from the experiment manifest that the eye-tracking device (Tobii Pro Nano) can be adopted as a device to collect eye movement data to monitor cognitive workload fluctuations in a smartphone setting, and pupil dilations can be used to measure the cognitive workload differences, with a screening test to filter out people with a high astigmatism degree.

There are three main directions for future research. The first one is to adopt pupil dilations as an effective measure to assess users' cognitive workloads while experiencing a smartphone application and to improve the UI design of the application based on the assessment.

The second one is to expand the age range to cover middle-aged and older adults. Only younger users who were born with smartphones were included in the study and we found that they have a high level of endurance for design issues and proficient capability to resolve issues by themselves. While it has been found that cognitive performance declines with age (Deary et al., 2009), it is reasonable to expand research to include middle-aged and older adults to verify the findings in the different age groups to investigate whether experience with smartphones overcomes cognitive ability's recession.

The third one is to test findings in other settings, e.g., virtual reality (VR), wearable devices, etc. Some ubiquitous screens (e.g., smart watches, etc.) and certain brand immersive experience technologies (e.g., VR glasses, etc.) have been winning consumers' heart with acceptable prices and great user experiences. It is necessary to test our findings in these settings as well.

In the far future, one major direction we desire to explore is to establish a multidimensional assessment tool for product usability, including subjective ratings, psychophysiological measures, and performance measures. We understand this objective is extensive and requires considerable time and human resources to complete.

Another direction is to expand from application-focused studies to include cognitive-focused studies. Instead of studying how to improve the usability of specific kinds of software applications, we aim to study cognitive processes, such as how to help people focus or refocus in different settings.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Ethics statement

The studies involving human participants were reviewed and approved by University of Arizona IRB Office (Protocol Number: 2101428836). Written informed consent to participate in this study was provided by the participants' legal guardian/next of kin.

Author contributions

Study conception and design, data collection, and draft manuscript preparation: LZ. Analysis and interpretation of results: LZ and HC. All authors reviewed the results and approved the final version of the manuscript.

Funding

NSF award # 1661485 and University of Arizona SBS summer dissertation fellowship.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnins.2022.1011475/full#supplementary-material

References

Alonso Dos Santos, M., and Calabuig Moreno, F. (2018). Assessing the effectiveness of sponsorship messaging: measuring the impact of congruence through electroencephalogram. Int. J. Sports Mark. Spons. 19, 25–40. doi: 10.1108/IJSMS-09-2016-0067

CrossRef Full Text | Google Scholar

Anderson, E. W., Potter, K. C., Matzen, L. E., Shepherd, J. F., Preston, G. A., and Silva, C. T. (2011). A user study of visualization effectiveness using EEG and cognitive load. Computer Graphics Forum. 30, 791–800. doi: 10.1111/j.1467-8659.2011.01928.x

CrossRef Full Text | Google Scholar

Antonenko, P., Paas, F., Grabner, R., and van Gog, T. (2010a). Using Electroencephalography to measure cognitive load. Educ. Psychol. Rev. 22, 425–438. doi: 10.1007/s10648-010-9130-y

CrossRef Full Text | Google Scholar

Antonenko, P. D., and Niederhauser, D. S. (2010b). The influence of leads on cognitive load and learning in a hypertext environment. Comput. Human Behav. 26, 140–150. doi: 10.1016/j.chb.2009.10.014

CrossRef Full Text | Google Scholar

Arsalan, A., Majid, M., Butt, A. R., and Anwar, S. M. (2019). Classification of perceived mental stress using a commercially available EEG headband. IEEE J. Biomed. Health Info. 23, 2257–2264. doi: 10.1109/JBHI.2019.2926407

PubMed Abstract | CrossRef Full Text | Google Scholar

Asif, A., Majid, M., and Anwar, S. M. (2019). Human stress classification using EEG signals in response to music tracks. Comput. Biol. Med. 107, 182–196. doi: 10.1016/j.compbiomed.2019.02.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Ayaz, H., Willems, B., Bunce, S., Shewokis, P. A., Hah, S., Deshmukh, A. R., et al. (2010). “Cognitive workload assessment of air traffic controllers using optical brain imaging sensors,” in Advances in Understanding Human Performance: Neuroergonomics, Human Factors Design, and Special Populations. 21–32. doi: 10.1201/EBK1439835012-c3

CrossRef Full Text | Google Scholar

Bevan, N., and MacLeod, M. (1994). Usability measurement in context. Behav. Infm.Technol. 13, 132–145. doi: 10.1080/01449299408914592

CrossRef Full Text | Google Scholar

Borys, M., Plechawska-Wójcik, M., Wawrzyk, M., and Wesołowska, K. (2017). “Classifying Cognitive Workload Using Eye Activity and EEG Features in Arithmetic Tasks,” in Damaševičius, R., and Mikašyte, V. Information and Software Technologies. Midtown Manhattan, New York City: Springer International Publishing. p. 90–105. doi: 10.1007/978-3-319-67642-5_8

CrossRef Full Text | Google Scholar

Boyd, K. (2021). What Is Astigmatism? - American Academy of Ophthalmology. Available online at: https://www.aao.org/eye-health/diseases/what-is-astigmatism

Google Scholar

Brooke, J. (1996). SUS: A “Quick and Dirty” Usability Scale. In Usability Evaluation In Industry. CRC Press. p. 7.

Google Scholar

Brunken, R., Plass, J. L., and Leutner, D. (2003). Direct measurement of cognitive load in multimedia learning. Educ. Psychol. 38, 53–61. doi: 10.1207/S15326985EP3801_7

CrossRef Full Text | Google Scholar

Brünken, R., Seufert, T., and Paas, F. (2010). Measuring cognitive load. doi: 10.1017/CBO9780511844744.011

CrossRef Full Text | Google Scholar

Cabañero, L., Hervás, R., González, I., Fontecha, J., Mondéjar, T., and Bravo, J. (2019). Analysis of cognitive load using EEG when interacting with mobile devices. Proceedings. 31, 70. doi: 10.3390/proceedings2019031070

CrossRef Full Text | Google Scholar

Cain, B. (2007). “A Review of the Mental Workload Literature,” Defence Research And Development Toronto (Canada). Available online at: https://apps.dtic.mil/docs/citations/ADA474193

PubMed Abstract | Google Scholar

Ceci, L. (2022). “Biggest app stores in the world 2021,” in Statista. Available online at: https://www.statista.com/statistics/276623/number-of-apps-available-in-leading-app-stores/

Cernea, D., Olech, P.-S., Ebert, A., and Kerren, A. (2012). “Controlling In-Vehicle Systems with a Commercial EEG Headset: Performance and Cognitive Load,” in Garth, C., Middel, V., and Hagen, H. (Eds.). Visualization of Large and Unstructured Data Sets: Applications in Geospatial Planning, Modeling and Engineering—Proceedings of IRTG 1131 Workshop 2011. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik. p. 113–122

Google Scholar

Chen, S., Epps, J., Ruiz, N., and Chen, F. (2011). “Eye activity as a measure of human mental effort in HCI,” Proceedings of the 16th International Conference on Intelligent User Interfaces. p. 315–318. doi: 10.1145/1943403.1943454

CrossRef Full Text | Google Scholar

Chin, Z. Y., Zhang, X., Wang, C., and Ang, K. K. (2018). “EEG-based discrimination of different cognitive workload levels from mental arithmetic,” in 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). p. 1984–1987. doi: 10.1109/EMBC.2018.8512675

PubMed Abstract | CrossRef Full Text | Google Scholar

Cirett Galán, F., and Beal, C. R. (2012). “EEG estimates of engagement and cognitive workload predict math problem solving outcomes,” in Masthoff, J., Mobasher, B., Desmarais, M. C., and Nkambou, R. (Eds.). User Modeling, Adaptation, and Personalization. Springer. p. 51–62. doi: 10.1007/978-3-642-31454-4_5

CrossRef Full Text | Google Scholar

Coulacoglou, C., and Saklofske, D. H. (2017). “Executive Function, Theory of Mind, and Adaptive Behavior,” in Psychometrics and Psychological Assessment. Elsevier. p. 91–130. doi: 10.1016/B978-0-12-802219-1.00005-5

CrossRef Full Text | Google Scholar

Coursaris, C., and Kim, D. (2006). “A qualitative review of empirical mobile usability studies,” Conference: Connecting the Americas. Acapulco, México: 12th Americas Conference on Information Systems, AMCIS 2006. p. 352.

Google Scholar

Dan, A., and Reiner, M. (2017). Real time EEG based measurements of cognitive load indicates mental states during learning. JEDM. 9, 31–44. doi: 10.5281/zenodo.3554719

CrossRef Full Text | Google Scholar

Davids, M. R., Halperin, M. L., and Chikte, U. M. E. (2015). Review: optimising cognitive load and usability to improve the impact of e-learning in medical education. Afr. J. Health Prof. Educ. 7, 147–152. doi: 10.7196/AJHPE.569

CrossRef Full Text | Google Scholar

Deary, I. J., Corley, J., Gow, A. J., Harris, S. E., Houlihan, L. M., Marioni, R. E., et al. (2009). Age-associated cognitive decline. Br. Med. Bull. 92, 135–152. doi: 10.1093/bmb/ldp033

PubMed Abstract | CrossRef Full Text | Google Scholar

Dimigen, O., Sommer, W., Hohlfeld, A., Jacobs, A. M., and Kliegl, R. (2011). Coregistration of eye movements and EEG in natural reading: analyses and review. J. Exp. Psychol.: General. 140, 552–572. doi: 10.1037/a0023885

PubMed Abstract | CrossRef Full Text | Google Scholar

Dumas, J. S., Dumas, J. S., and Redish, J. (1999). A Practical Guide to Usability Testing. Intellect Books.

Google Scholar

Ehlers, J. (2020). “Exploring the effect of transient cognitive load on bodily arousal and secondary task performance,” Proceedings of the Conference on Mensch Und Computer. p. 7–10. doi: 10.1145/3404983.3410017

CrossRef Full Text | Google Scholar

Evans, D. C., and Fendley, M. (2017). A multi-measure approach for connecting cognitive workload and automation. Int. J. Hum. Comput. Stud. 97, 182–189. doi: 10.1016/j.ijhcs.2016.05.008

CrossRef Full Text | Google Scholar

Farnsworth, B. (2019). “EEG Headset Prices – An Overview of 15+ EEG Devices,” Imotions. Available online at: https://imotions.com/blog/eeg-headset-prices/

Farnsworth, B. (2020). “Top 12 Eye Tracking Hardware Companies (Ranked),” Imotions. Available online at: https://imotions.com/blog/top-eye-tracking-hardware-companies/

Google Scholar

Gavas, R., Chatterjee, D., and Sinha, A. (2017). “Estimation of cognitive load based on the pupil size dilation,” in 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC). p. 1499–1504. doi: 10.1109/SMC.2017.8122826

PubMed Abstract | CrossRef Full Text | Google Scholar

Gevins, A., and Smith, M. E. (2003). Neurophysiological measures of cognitive workload during human-computer interaction. Theor. Issues Ergon. Sci. 4, 113–131. doi: 10.1080/14639220210159717

CrossRef Full Text | Google Scholar

Goldberg, J., and Helfman, J. (2010). “Comparing information graphics: a critical look at eye tracking,” Conference on Human Factors in Computing Systems - Proceedings. doi: 10.1145/2110192.2110203

CrossRef Full Text | Google Scholar

Goldberg, J. H., and Wichansky, A. M. (2003). “Chapter 23 - Eye Tracking in Usability Evaluation: A Practitioner's Guide,” in Hyönä, J., Radach, R., and Deubel, H. (Eds.). The Mind's Eye. North-Holland. p. 493–516. doi: 10.1016/B978-044451020-4/50027-X

CrossRef Full Text | Google Scholar

Granholm, E., Asarnow, R. F., Sarkin, A. J., and Dykes, K. L. (1996). Pupillary responses index cognitive resource limitations. Psychophysiology. 33, 457–461. doi: 10.1111/j.1469-8986.1996.tb01071.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Guy-Evans, O. (2021a). Frontal Lobe Function, Location in Brain, Damage, More | Simply Psychology. Available online at: https://www.simplypsychology.org/frontal-lobe.html

Guy-Evans, O. (2021b). Temporal Lobe: Definition, Functions, and Location | Simply Psychology. Available online at: https://www.simplypsychology.org/temporal-lobe.html

Google Scholar

Gwizdka, J., Hosseini, R., Cole, M., and Wang, S. (2017). Temporal dynamics of eye-tracking and EEG during reading and relevance decisions. J. Assoc. Inf. Sci. Technol. 68, 2299–2312. doi: 10.1002/asi.23904

CrossRef Full Text | Google Scholar

Harrison, R., Flood, D., and Duce, D. (2013). Usability of mobile applications: literature review and rationale for a new usability model. J. Interact. Sci. 1, 1. doi: 10.1186/2194-0827-1-1

CrossRef Full Text | Google Scholar

Hart, S. G., and Staveland, L. E. (1988). “Development of NASA-TLX (Task Load Index): results of empirical and theoretical research,” in Hancock, P. A., and Meshkati, N. (Eds.). Advances in Psychology. North-Holland. p. 139–183. doi: 10.1016/S0166-4115(08)62386-9

CrossRef Full Text | Google Scholar

Heckbert, P. (1995). Fourier Transforms and the Fast Fourier Transform (FFT) Algorithm. p. 13.

Google Scholar

Hild, J., Putze, F., Kaufman, D., Kühnle, C., Schultz, T., and Beyerer, J. (2014). Spatio-Temporal Event Selection in Basic Surveillance Tasks using Eye Tracking and EEG. p. 6.

Google Scholar

Hoober, S. (2013). How Do Users Really Hold Mobile Devices? UXmatters. Available online at: https://www.uxmatters.com/mt/archives/2013/02/how-do-users-really-hold-mobile-devices.php?

Google Scholar

ISO 9241-11. (2018). Ergonomics of human-system interaction—Part 11: Usability: Definitions and concepts. Available online at: https://www.iso.org/obp/ui/#iso:std:iso:9241:-11:ed-2:v1:en

Google Scholar

Jaeggi, S. M., Buschkuehl, M., Perrig, W. J., and Meier, B. (2010). The concurrent validity of the N-back task as a working memory measure. Memory. 18, 394–412. doi: 10.1080/09658211003702171

PubMed Abstract | CrossRef Full Text | Google Scholar

Klimesch, W. (1999). EEG alpha and theta oscillations reflect cognitive and memory performance: a review and analysis. Brain Res. Rev. 29, 169–195. doi: 10.1016/S0165-0173(98)00056-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Klimesch, W., Schack, B., and Sauseng, P. (2005). The functional significance of theta and upper alpha oscillations. Exp. Psychol. 52, 99–108. doi: 10.1027/1618-3169.52.2.99

PubMed Abstract | CrossRef Full Text | Google Scholar

Klingner, J., Kumar, R., and Hanrahan, P. (2008). “Measuring the task-evoked pupillary response with a remote eye tracker,” in Proceedings of the 2008 Symposium on Eye Tracking Research and Applications. p. 69–72. doi: 10.1145/1344471.1344489

CrossRef Full Text | Google Scholar

Knoll, A., Wang, Y., Chen, F., Xu, J., Ruiz, N., Epps, J., et al. (2011). “Measuring Cognitive Workload with Low-Cost Electroencephalograph,” in Campos, P., Graham, N., Jorge, J., Nunes, N., Palanque, P., and Winckler, M. (Eds.). Human-Computer Interaction – INTERACT 2011. Springer. p. 568–571. doi: 10.1007/978-3-642-23768-3_84

CrossRef Full Text | Google Scholar

Krause, C. M., Sillanmäki, L., Koivisto, M., Saarela, C., Häggqvist, A., Laine, M., et al. (2000). The effects of memory load on event-related EEG desynchronization and synchronization. Clini. Neurophysiol. 111, 2071–2078. doi: 10.1016/S1388-2457(00)00429-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Krigolson, O. E., Williams, C. C., Norton, A., Hassall, C. D., and Colino, F. L. (2017). Choosing MUSE: validation of a low-cost, portable EEG system for ERP research. Front. Neurosci. 11. doi: 10.3389/fnins.2017.00109

PubMed Abstract | CrossRef Full Text | Google Scholar

Kumar, N., and Kumar, J. (2016). Measurement of cognitive load in HCI systems using EEG power spectrum: an experimental study. Procedia Comp. Sci. 84, 70–78. doi: 10.1016/j.procs.2016.04.068

CrossRef Full Text | Google Scholar

Lallé, S., Conati, C., and Carenini, G. (2016). Prediction of individual learning curves across information visualizations. User Model. User-adapt. Interact. 26, 307–345. doi: 10.1007/s11257-016-9179-5

CrossRef Full Text | Google Scholar

Lee, W. F., and Chenkin, J. (2020). “Exploring Eye-tracking Technology as an Assessment Tool for Point-of-care Ultrasound Training,” in AEM Education and Training. 5:e10508. doi: 10.1002/aet2.10508

PubMed Abstract | CrossRef Full Text | Google Scholar

Makransky, G., Terkildsen, T. S., and Mayer, R. E. (2019). Adding immersive virtual reality to a science lab simulation causes more presence but less learning. Learn. Instruct. 60, 225–236. doi: 10.1016/j.learninstruc.2017.12.007

CrossRef Full Text | Google Scholar

Manuel, V., Barrios, G., Gütl, C., Preis, A. M., Andrews, K., Pivec, M., et al. (2004). “AdELE: A framework for adaptive e-learning through eye tracking,” in Proceedings of I-Know'04. p. 609–616.

Google Scholar

Mazher, M., Abd Aziz, A., Malik, A. S., and Ullah Amin, H. (2017). An EEG-based cognitive load assessment in multimedia learning using feature extraction and partial directed coherence. IEEE Access. 5, 14819–14829. doi: 10.1109/ACCESS.2017.2731784

CrossRef Full Text | Google Scholar

Mobile Testing Accessory | Perfect for Usability Tests. (2020). Available online at: https://www.tobiipro.com/product-listing/mobile-testing-accessory/

Google Scholar

Molla, R. (2020). “Americans spent about 3.5 hours per day on their phones last year—A number that keeps going up despite the “time well spent” movement,” in Vox. Available online at: https://www.vox.com/recode/2020/1/6/21048116/tech-companies-time-well-spent-mobile-phone-usage-data

Neubauer, A. C., and Fink, A. (2003). Fluid intelligence and neural efficiency: effects of task complexity and sex. Pers. Individ. Dif. 35, 811–827. doi: 10.1016/S0191-8869(02)00285-4

CrossRef Full Text | Google Scholar

Neubauer, A. C., Fink, A., and Grabner, R. H. (2006). “Sensitivity of alpha band ERD to individual differences in cognition,” in Neuper, C., and Klimesch, W. (Eds.). Progress in Brain Research. Elsevier. p. 167–178. doi: 10.1016/S0079-6123(06)59011-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Nielsen, J. (1993). Iterative user-interface design. Computer. 26, 32–41. doi: 10.1109/2.241424

CrossRef Full Text | Google Scholar

Nielsen, J., and Pernice, K. (2010). Eyetracking Web Usability. New Riders.

Google Scholar

Notaro, G. M., and Diamond, S. G. (2018). Simultaneous EEG, eye-tracking, behavioral, and screen-capture data during online German language learning. Data in Brief. 21, 1937–1943. doi: 10.1016/j.dib.2018.11.044

PubMed Abstract | CrossRef Full Text | Google Scholar

O'Dea, S. (2021). Number of smartphone users in the U.S. 2025. Statista. Available online at: https://www.statista.com/statistics/201182/forecast-of-smartphone-users-in-the-us/

Google Scholar

Örün, Ö., and Akbulut, Y. (2019). Effect of multitasking, physical environment and electroencephalography use on cognitive load and retention. Comput. Human Behav. 92, 216–229. doi: 10.1016/j.chb.2018.11.027

CrossRef Full Text | Google Scholar

Papakostas, M., Tsiakas, K., Giannakopoulos, T., and Makedon, F. (2017). “Towards predicting task performance from EEG signals,” in 2017 IEEE International Conference on Big Data (Big Data). p. 4423–4425. doi: 10.1109/BigData.2017.8258478

PubMed Abstract | CrossRef Full Text | Google Scholar

Pesonen, M., Hämäläinen, H., and Krause, C. M. (2007). Brain oscillatory 4–30 Hz responses during a visual n-back memory task with varying memory load. Brain Res. 1138, 171–177. doi: 10.1016/j.brainres.2006.12.076

PubMed Abstract | CrossRef Full Text | Google Scholar

Pfleging, B., Fekety, D. K., Schmidt, A., and Kun, A. L. (2016). “A model relating pupil diameter to mental workload and lighting conditions,” in Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems. p. 5776–5788. doi: 10.1145/2858036.2858117

CrossRef Full Text | Google Scholar

Pfurtscheller, G. (2001). Functional brain imaging based on ERD/ERS. Vision Res. 41, 1257–1260. doi: 10.1016/S0042-6989(00)00235-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Pfurtscheller, G., and Aranibar, A. (1977). Event-related cortical desynchronization detected by power measurements of scalp EEG. Electroencephalogr. Clin. Neurophysiol. 42, 817–826. doi: 10.1016/0013-4694(77)90235-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Pfurtscheller, G., and Berghold, A. (1989). Patterns of cortical activation during planning of voluntary movement. Electroencephalogr. Clin. Neurophysiol. 72, 250–258. doi: 10.1016/0013-4694(89)90250-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Pfurtscheller, G., and Lopes da Silva, F. H. (1999). Event-related EEG/MEG synchronization and desynchronization: basic principles. Clini. Neurophysiol. 110, 1842–1857. doi: 10.1016/S1388-2457(99)00141-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Pomplun, M., and Sunkara, S. (2003). “Pupil dilation as an indicator of cognitive workload in human-computer interaction,” in Human-Centered Computing, (CRC Press).

Google Scholar

Porta, M., Ricotti, S., and Perez, C. J. (2012). “Emotional e-learning through eye tracking,” in Proceedings of the 2012 IEEE Global Engineering Education Conference (EDUCON). p. 1–6. doi: 10.1109/EDUCON.2012.6201145

PubMed Abstract | CrossRef Full Text | Google Scholar

Pratama, S. H., Rahmadhani, A., Bramana, A., Oktivasari, P., Handayani, N., Haryanto, F., et al. (2020). Signal comparison of developed EEG device and emotiv insight based on brainwave characteristics analysis. J. Physics: Conference Series. 1505, 012071. doi: 10.1088/1742-6596/1505/1/012071

CrossRef Full Text | Google Scholar

Przegalinska, A., Ciechanowski, L., Magnuski, M., and Gloor, P. (2018). “Muse Headband: Measuring Tool or a Collaborative Gadget?” in Grippa, F., Leitão, J., Gluesing, J., Riopelle, K., and Gloor, P. (Eds.). Collaborative Innovation Networks: Building Adaptive and Resilient Organizations. Midtown Manhattan, New York City: Springer International Publishing. p. 93–101. doi: 10.1007/978-3-319-74295-3_8

PubMed Abstract | CrossRef Full Text | Google Scholar

Purves, D., Augustine, G. J., Fitzpatrick, D., Katz, L. C., LaMantia, A.-S., McNamara, J. O., et al. (2001). Types of Eye Movements and Their Functions. Neuroscience. 2nd Edition. Available online at: https://www.ncbi.nlm.nih.gov/books/NBK10991/

Rafiqi, S., Wangwiwattana, C., Kim, J., Fernandez, E., Nair, S., and Larson, E. C. (2015). “PupilWare: towards pervasive cognitive load measurement using commodity devices,” in Proceedings of the 8th ACM International Conference on PErvasive Technologies Related to Assistive Environments. p. 1–8. doi: 10.1145/2769493.2769506

CrossRef Full Text | Google Scholar

Ratti, E., Waninger, S., Berka, C., Ruffini, G., and Verma, A. (2017). Comparison of medical and consumer wireless EEG systems for use in clinical trials. Front. Hum. Neurosci. 11, 398. doi: 10.3389/fnhum.2017.00398

PubMed Abstract | CrossRef Full Text | Google Scholar

Reimer, B., Mehler, B., Coughlin, J. F., Godfrey, K. M., and Tan, C. (2009). “An on-road assessment of the impact of cognitive workload on physiological arousal in young adult drivers,” in Proceedings of the 1st International Conference on Automotive User Interfaces and Interactive Vehicular Applications. p. 115–118. doi: 10.1145/1620509.1620531

CrossRef Full Text | Google Scholar

Ritchie, D. M., and Thompson, K. (1978). The UNIX time-sharing system†. Bell System Tech. J. 57, 1905–1929. doi: 10.1002/j.1538-7305.1978.tb02136.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Rudmann, D. S., McConkie, G. W., and Zheng, X. S. (2003). “Eyetracking in cognitive state detection for HCI,” in Proceedings of the 5th International Conference on Multimodal Interfaces. p. 159–163. doi: 10.1145/958432.958464

CrossRef Full Text | Google Scholar

Saitis, C., Parvez, M. Z., and Kalimeri, K. (2018). Cognitive Load Assessment from EEG and Peripheral Biosignals for the Design of Visually Impaired Mobility Aids [Research Article]. Hindawi: Wireless Communications and Mobile Computing. doi: 10.1155/2018/8971206

CrossRef Full Text | Google Scholar

Scharinger, C., Kammerer, Y., and Gerjets, P. (2016). “Fixation-Related EEG Frequency Band Power Analysis: A Promising Neuro-Cognitive Methodology to Evaluate the Matching-Quality of Web Search Results?” in Stephanidis, C. (Ed.). HCI International 2016 – Posters' Extended Abstracts. Midtown Manhattan, New York City: Springer International Publishing. p. 245–250. doi: 10.1007/978-3-319-40548-3_41

CrossRef Full Text | Google Scholar

Shneiderman, B. (1986). Designing menu selection systems. J. Am. Soc. Inf. Sci. 37, 57–70. doi: 10.1002/(SICI)1097-4571(198603)37:2<57::AID-ASI2>3.0.CO;2-S

CrossRef Full Text | Google Scholar

Stipacek, A., Grabner, R. H., Neuper, C., Fink, A., and Neubauer, A. C. (2003). Sensitivity of human EEG alpha band desynchronization to different working memory components and increasing levels of memory load. Neurosci. Lett. 353, 193–196. doi: 10.1016/j.neulet.2003.09.044

PubMed Abstract | CrossRef Full Text | Google Scholar

Sugaya, Y. (2019). Distributed Pragmatic Processing for Adjective Expression: An Experimental Study. Available online at: https://www.jcss.gr.jp/meetings/jcss2019/proceedings/area14.html

Google Scholar

Trammell, J. P., MacRae, P. G., Davis, G., Bergstedt, D., and Anderson, A. E. (2017). The relationship of cognitive performance and the theta-alpha power ratio is age-dependent: an EEG study of short term memory and reasoning during task and resting-state in healthy young and old adults. Front. Aging Neurosci. 9. doi: 10.3389/fnagi.2017.00364

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, Q., Yang, S., Liu, M., Cao, Z., and Ma, Q. (2014). An eye-tracking study of website complexity from cognitive load perspective. Decis. Support Syst. 62, 1–10. doi: 10.1016/j.dss.2014.02.007

CrossRef Full Text | Google Scholar

Wilson, G. F., and Eggemeier, F. T. (1991). “Psychophysiological assessment of workload in multi-task environments,” in Multiple-task performance. CRC Press. doi: 10.1201/9781003069447-15

CrossRef Full Text | Google Scholar

Winslow, B., Carpenter, A., Flint, J., Wang, X., Tomasetti, D., Johnston, M., et al. (2013). Combining EEG and eye tracking: using fixation-locked potentials in visual search. J. Eye Mov. Res. 6, 4. doi: 10.16910/jemr.6.4.5

CrossRef Full Text | Google Scholar

Xiang, Z., Huang, Y., Luo, G., Ma, H., and Zhang, D. (2021). Decreased event-related desynchronization of mental rotation tasks in young tibetan immigrants. Front. Hum. Neurosci. 15, 664039. doi: 10.3389/fnhum.2021.664039

PubMed Abstract | CrossRef Full Text | Google Scholar

Xie, B., and Salvendy, G. (2000). Review and reappraisal of modelling and predicting mental workload in single- and multi-task environments. Work and Stress. 1, 74–99. doi: 10.1080/026783700417249

CrossRef Full Text | Google Scholar

Yokota, Y., and Naruse, Y. (2015). Phase coherence of auditory steady-state response reflects the amount of cognitive workload in a modified N-back task. Neurosci. Res. 100, 39–45. doi: 10.1016/j.neures.2015.06.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Zagermann, J., Pfeil, U., and Reiterer, H. (2016). “Measuring Cognitive Load using Eye Tracking Technology in Visual Computing,” Proceedings of the Beyond Time and Errors on Novel Evaluation Methods for Visualization – BELIV. p. 78–85. doi: 10.1145/2993901.2993908

CrossRef Full Text | Google Scholar

Zagermann, J., Pfeil, U., and Reiterer, H. (2018). “Studying Eye Movements as a Basis for Measuring Cognitive Load,” in Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems. p. 1–6. doi: 10.1145/3170427.3188628

CrossRef Full Text | Google Scholar

Zhang, D., and Adipat, B. (2005). Challenges, methodologies, and issues in the usability testing of mobile applications. Int. J. Human–Computer Interact 18, 293–308. doi: 10.1207/s15327590ijhc1803_3

CrossRef Full Text | Google Scholar

Keywords: cognitive workload, EEG, eye tracking, eye movement, GUI

Citation: Zhang L and Cui H (2022) Reliability of MUSE 2 and Tobii Pro Nano at capturing mobile application users' real-time cognitive workload changes. Front. Neurosci. 16:1011475. doi: 10.3389/fnins.2022.1011475

Received: 04 August 2022; Accepted: 18 October 2022;
Published: 28 November 2022.

Edited by:

Waldemar Karwowski, University of Central Florida, United States

Reviewed by:

Debatri Chatterjee, Tata Consultancy Services, India
Yanlu Wang, Karolinska Institutet (KI), Sweden
Yinglei Song, Jiangsu University of Science and Technology, China

Copyright © 2022 Zhang and Cui. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Limin Zhang, lmzhang@hytc.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.