ORIGINAL RESEARCH article

Front. Artif. Intell., 27 June 2025

Sec. AI for Human Learning and Behavior Change

Volume 8 - 2025 | https://doi.org/10.3389/frai.2025.1516295

Data stream-pairwise bottleneck transformer for engagement estimation from video conversation


Keita Suzuki
Keita Suzuki*Nobukatsu HojoNobukatsu HojoKazutoshi ShinodaKazutoshi ShinodaSaki MizunoSaki MizunoRyo MasumuraRyo Masumura
  • NTT Human Informatics Laboratories, NTT Corporation, Yokosuka, Japan

This study aims to assess participant engagement in multiparty conversations using video and audio data. For this task, the interaction among numerous data streams, such as video and audio from multiple participants, should be modeled effectively, considering the redundancy of video and audio across frames. To efficiently model participant interactions while accounting for such redundancy, a previous study proposed inputting participant feature sequences into global token-based transformers, which constrain attention across feature sequences to pass through only a small set of internal units, allowing the model to focus on key information. However, this approach still faces the challenge of redundancy in participant-feature estimation based on standard cross-attention transformers, which can connect all frames across different modalities. To address this, we propose a joint model for interactions among all data streams using global token-based transformers, without distinguishing between cross-modal and cross-participant interactions. Experiments on the RoomReader corpus confirm that the proposed model outperforms previous models, achieving accuracy ranging from 0.720 to 0.763, weighted F1 scores from 0.733 to 0.771, and macro F1 scores from 0.236 to 0.277.

1 Introduction

Online meetings are essential tools for today's work environment, and the adoption of remote work has advanced due to the COVID-19 pandemic. Many companies continue to use hybrid work even after the pandemic. Therefore, the demand for online meetings has not decreased. However, unlike face-to-face meetings, online meetings present the challenge of enabling everyone to stay focused due to a diminished sense of participation. In such online environments, it can be difficult to capture nonverbal cues such as gaze, facial expressions, and tone of voice, leading to an increased risk of declines in engagement (Sukumaran and Manoharan, 2024). Therefore, it has become important to continually estimate the engagement of online meeting participants and provide appropriate feedback.

Engagement estimation has evolved from analyzing individual behaviors to modeling complex group interactions across various modalities. Early work focused on single-person unimodal signals, such as facial expressions or gaze (Savchenko et al., 2022; Singh et al., 2023), followed by multimodal approaches that combine audio and video to improve robustness (Pan et al., 2023; Kumar et al., 2024). Dyadic settings introduced the importance of interpersonal cues (Dermouche and Pelachaud, 2019; Chen et al., 2022). Recent work has expanded to multiparty scenarios using only visual data (Lee et al., 2023). However, as shown in recent studies (Kim et al., 2023; Suzuki et al., 2024), modeling both audio and visual signals across participants improves engagement prediction. We therefore address multiparty multimodal engagement estimation, defined as estimating a target participant's engagement from the audiovisual data of all participants, which remains a challenging and underexplored area.

Figure 1 illustrates the existing and proposed multiparty multimodal engagement estimation models. A central challenge in multiparty multimodal engagement estimation lies in modeling interactions among multiple data streams—namely, different modalities (e.g., video and audio) from multiple participants. To make accurate predictions for a target participant, the model used must effectively capture both cross-modal interactions within each participant and cross-person interactions across participants. Previous studies have addressed this by introducing cross-person transformers (CPTs) (Lee et al., 2023; Kim et al., 2023), which rely on hierarchical combinations of cross-attention layers to separately model these two types of interaction. More recently, global token-based architectures have been proposed as a more efficient alternative for representing interactions between high-dimensional streams, avoiding the combinatorial explosion of direct attention (Sun et al., 2023; Nagrani et al., 2021). In the context of engagement estimation, participant-pairwise global tokens have shown promise in modeling cross-person interactions more effectively (Suzuki et al., 2024). However, the previous model still depends on cross-attention mechanisms for cross-modal fusion, leaving challenges related to redundancy and scalability unresolved. In this work, we address this limitation by extending global token-based modeling to unify both cross-modal and cross-person interactions within a single architecture. Our proposed model introduces a data stream-pairwise structure that enables efficient and accurate engagement estimation across all modalities and participants in multiparty conversations.

Figure 1
Three diagrams labeled as CPT, PPBT, and DSPBT. Each diagram shows a sequence of layers processing video and speech inputs, with CAL and GTL blocks. Outputs are combined with a ‘̀C ” component leading to ‘̀eng. ” CPT and PPBT have separate video and speech pathways, while PPBT and DSPBT share GTL sections.

Figure 1. Overview of previous and proposed models. “CAL” and “GTL” respectively denote cross-attention and global-token layers in Figure 2. “c” denotes classification layer. Each figure illustrates case in which number of participants N is 3 and target participant ID n* is 2. (a) CPT (Lee et al., 2023; Kim et al., 2023). (b) PPBT (Suzuki et al., 2024). (c) DSPBT (Proposed).

To thoroughly investigate this question in a realistic multiparty environment, we require a dataset that meets four key criteria: it should (1) include engagement annotations, (2) contain both video and audio data, (3) capture multiparty interactions, and (4) ideally be publicly available. To the best of our knowledge, only the RoomReader corpus (Reverdy et al., 2022) satisfies all these requirements simultaneously (see Section 2.4 and Table 1). RoomReader provides over 8 h of online multiparty conversations with synchronized video and audio, along with annotations of each participant's engagement level, making it uniquely suited to this research.

Table 1
www.frontiersin.org

Table 1. Representative engagement estimation datasets: comparison of participants, modalities, and availability.

While RoomReader provides behavior-based engagement annotations suitable for real-time modeling, it is important to contrast this with more traditional definitions and measurement approaches. Engagement is often defined as a multi-dimensional construct encompassing behavioral, emotional, cognitive, and agentic components and is typically measured via self-reports (Fredricks et al., 2004, 2016; Sinatra et al., 2015). However, such labels lack temporal granularity and may diverge from observable behavior due to annotation bias. These limitations make them ill-suited for the application assumed in this study, which requires not only real-time inference and feedback during conversations but also engagement labels that align consistently with audiovisual input–since misalignment can cause models to learn spurious or non-generalizable associations. We therefore adopt the RoomReader corpus (Reverdy et al., 2022), which provides temporally dense, behavior-based annotations that are well-suited for multimodal modeling.

The contributions of this research are as follows.

Introduction of global tokens that handle interactions among data streams, i.e., multiple modalities from multiple participants in conversations: To estimate engagement in multiparty conversations, we introduce global tokens to manage the interactions among multiple input data streams, enabling efficient modeling.

Proof of effect: By introducing global tokens that manage interactions between data streams, we demonstrate that higher accuracy can be achieved through engagement-estimation experiments on the publicly available RoomReader corpus compared with previous methods.

These contributions provide new directions for estimating multi-participant engagement.

This article is structured as follows. Section 2 reviews the previous research on engagement estimation, small-group interaction modeling, transformer-based multimodal approaches, and relevant corpora. Section 3 details the baseline and our proposed method. Section 4 explains the experimental setup and datasets, and Section 5 reports and discusses the experimental results. Finally, Section 6 concludes the paper and suggests possible future directions.

2 Related works

Research on automatic engagement recognition has evolved from early work on individual behaviors to more complex modeling of social interaction. In this section, we briefly review approaches that estimate engagement of individual participants before shifting to methods that incorporate interpersonal group-level dynamics (see Table 2).

Table 2
www.frontiersin.org

Table 2. Representative engagement estimation methods: comparison of study, dataset, participants, modality and model.

2.1 Engagement estimation

Early research on automatic engagement recognition spans diverse contexts, including education, social robotics, and conversational interfaces. Rich et al. (2010) pioneered engagement recognition in human–robot interaction by using backchannel cues. Others leveraged nonverbal signals: for instance, Bednarik et al. (2012) used gaze patterns to recognize engagement in group conversations, and Sanghvi et al. (2011) analyzed body posture features to estimate child engagement with a robot tutor. In educational settings, detecting student engagement from face and body cues has been a major focus (Grafsgaard et al., 2014).

Previous studies estimated engagement using the convolutional neural network-long short-term memory (CNN-LSTM) and CNN-transformer on the basis of the video and speech data of the target participant (Li and Hung, 2019; Xiong et al., 2023). Engagement estimation has advanced with the introduction of models such as the bootstrap model ensemble (BOOT) and the ensemble model (ENS-MODEL), which have facilitated the use of bootstrapping and ensembling methods (Wang et al., 2019; Huynh et al., 2019). Wang et al. (2019) introduced a model ensemble with a rank-based loss function for engagement intensity regression. By aggregating multiple models, their approach achieved a top rank in the EmotiW 2019 engagement challenge (mean squared error of 0.0626 on the test set). Similarly, Huynh et al. (2019) developed an ensemble-based regression using facial behavior features such as action units and head motion, demonstrating high-level performance in the same challenge. These ensemble models mitigate individual model biases and variance, leading to more robust engagement estimation.

Additionally, hierarchical temporal multi-instance learning (HTMIL) uses a bidirectional long short-term memory (Bi-LSTM) with multi-scale attention to achieve both clip-level and video-level objectives, effectively capturing short- and long-term patterns (e.g., momentary distraction vs. sustained attention) by splitting a video into temporal segments, as proposed in Ma et al. (2021).

These methods primarily focus on estimating the engagement of an individual participant based on unimodal or multimodal signals, without modeling interactions among participants. In contrast, our work considers engagement as a phenomenon that emerges through social interaction, particularly in multiparty settings. We therefore turn next to models that explicitly handle small-group interactions.

2.2 Small-group interaction models

Graphical models have been instrumental in analyzing interactions, notably for group-performance prediction, behavior recognition, social-field modeling, and interaction recognition (Lin and Lee, 2020; Yang et al., 2020; Zhou et al., 2019; Li et al., 2020). However, in the context of larger, multi-participant settings, the effectiveness of graphical models suffers. The complication arises from the scalability of these models and the added complexity of interactions that come with larger group sizes. The web of intricate interactions in such settings introduces a level of complexity that conventional graphical models cannot handle; the nuances of complex group dynamics prove difficult to accurately capture. It seems rather that more sophisticated approaches are needed to handle the diverse and dynamic nature of interactions in multi-participant environments. Recently, transformer-based methods that employ attention mechanisms among participants have been proposed, offering new possibilities for modeling interactions more effectively in large-scale social settings (Lee et al., 2023; Kim et al., 2023; Suzuki et al., 2024).

2.3 Transformer modeling by using global tokens

There are studies that take a transformer-based approach to representing interactions between modalities using global tokens for multimodal sentiment analysis (Sun et al., 2023; Nagrani et al., 2021). These studies embrace the novel approach of using interactions based on global tokens to help make the process much more efficient. Global tokens are abstract representations of information that are shared across the modalities, so they can exchange info, such as video and audio, without resorting to the attention mechanism. They incur lower computational costs and information redundancy than the attention mechanism. It has also been reported that the use of global tokens facilitates the integration of information across different modalities, resulting in overall models with higher accuracy. Sentiment analysis generally aims to identify or categorize emotional states (e.g., positive or negative feelings) based on audiovisual data. In contrast, engagement analysis focuses on the level of participation of meeting attendees, which differs from emotional analysis. Nevertheless, the global-token approach remains relevant for both sentiment and engagement tasks because it offers an efficient way to capture cross-modal relationships by acting as an attention bottleneck. In this work, we build on previous research involving global tokens but shift our emphasis from sentiment analysis to engagement estimation, centering on how actively participants are involved in the conversation rather than on their emotional valence. There are studies that focused on multi-participant meetings in engagement estimation, where global tokens are used to represent interactions in part of the model: their effectiveness has been demonstrated (Suzuki et al., 2024).

2.4 Corpus for engagement estimation

In Table 1, we compare representative datasets commonly used for engagement estimation. Key attributes are summarized, including the participant setup (dyadic or multiparty), recorded modalities, and data availability. Notably, only the RoomReader dataset provides publicly available audiovisual data of multiparty interactions, which is why our work utilizes RoomReader exclusively (Reverdy et al., 2022). Other datasets either focus on dyadic interactions or are not publicly released or lack combined video and audio modalities. The primary uses for engagement estimation in human–computer-interaction research have been the Remote Collaborative and Affective Interaction (RECOLA) and Nonverbal Interaction in Expert–Novice Interaction (NoXi) corpora (Ringeval et al., 2013; Cafaro et al., 2017). These corpora have played an important role in the field, recording dyadic interactions through detailed recordings of conversational exchanges between two participants. In particular, as shown in Table 1, existing engagement datasets vary widely, but early works often focused on dyadic interactions: for example, the NoXi corpus captures two-person video chats—specifically remote expert–novice conversations (Cafaro et al., 2017). Meanwhile, the RECOLA corpus contains remote dyadic collaborations featuring rich modalities (audio, video, ECG, and EDA) that include engagement (Ringeval et al., 2013). Consequently, the RECOLA and NoXi corpora serve as essential research resources for researchers seeking to understand the subtle variations in engagement between two individuals.

Some datasets focus on parent–child scenarios: Dyadic Affect in Multimodal Interaction—Parent to Child (DAMI-P2C) records in-lab story-reading sessions (dyads), annotated for child engagement and parent–child relationship measures (e.g., attachment, relational frustration, and parenting stress), among others (Chen et al., 2022). There are also “in-the-wild” single-participant datasets: the Emotion Recognition in the Wild 2018 and 2019 challenges (EmotiW 2018 and 2019) introduced webcam videos of students watching MOOC lectures, annotated for engagement (Dhall et al., 2018, 2019). Meanwhile, the Dataset for Affective States in E-Environments (DAiSEE) captures user engagement in online learning contexts (Gupta et al., 2016), alongside other affective states such as boredom, confusion, and frustration. However, these datasets are not multiparty (involving only one learner) and were only released for competition use.

Our research goes beyond the scope of dyadic engagement estimation by introducing the RoomReader corpus to understand engagement in multiparty conversational settings, a method also used in a previous study because fewer datasets feature multiparty (group) interactions (Lee et al., 2023). Therefore, to enable research on engaged group conversations with reproducible results, we focus on the RoomReader corpus in our study. RoomReader uniquely offers a public, multimodal, group-interaction dataset containing over 8 h of online multiparty conversations with synchronized video and audio, along with student engagement annotations. This combination of group interaction, audio-visual modality, and public accessibility is exclusive to RoomReader among current datasets, making it essential for our multiparty engagement research.

3 Method

3.1 Cross-attention layer and global token layer for data stream interaction modeling

To model interactions among multiple data streams–that is, different modalities (e.g., audio and video) across multiple participants–previous work has proposed various mechanisms, notably the Cross-Attention Layer (CAL) and the Global Token Layer (GTL) (Figure 2). While CAL has been widely adopted in recent approaches to handle pairwise stream interactions, our proposed architecture adopts a unified modeling approach based on GTL, which enables more efficient and scalable fusion across modalities and participants. When more than two input streams are involved, GTL can be implemented in two forms: the common GTL, which aggregates all streams into shared tokens, and the pairwise GTL, which processes each pair of streams individually through a shared bottleneck (Suzuki et al., 2024). In this section, we present the formal definitions of CAL, GTL (common), and GTL (pairwise) to clarify their functional differences and modeling characteristics. We begin by introducing the notation for input data streams and then describe the mathematical formulation of each layer.

Figure 2
Comparison of (a) Cross-attention layer (CAL), where each input attends to the other's features using cross-attention, and (b) Global token layer (GTL), where a shared global token interacts with all inputs through self-attention and is updated via aggregation across layers.

Figure 2. Illustration of interaction models between two data streams. “Cross-attention” and “self-attention” denote cross-attention and self-attention transformer-encoder layers, respectively. (a) Cross-attention layer (CAL). (b) Global token layer (GTL).

3.1.1 Cross-attention layer

The Cross-Attentional Layer (CAL) (Figure 2a) takes two input data streams, X1 and X2, and models their interaction to output two corresponding streams, Y1 and Y2. The inputs and outputs are matrices of shape D × T1 for X1, Y1 and D × T2 for X2, Y2, where D is the feature dimension and T1 and T2 are the time lengths.

CAL is implemented by repeating a cross-attention transformer layer (Vaswani et al., 2017) L times. Let H10=X1 and H20=X2 denote the initial inputs. Then, for each layer = 1, …, L, the intermediate representations are computed as follows:

H1=CrossAttention(H10,H2-1,H2-1;θ1),    (1)
H2=CrossAttention(H20,H1-1,H1-1;θ2).    (2)

Here, CrossAttention(Q, K, V; θ) denotes a cross-attention transformer layer, where Q, K, and V represent the query, key, and value matrices, respectively. θ1 and θ2 are the parameters of each respective transformer at layer .

The final outputs of CAL are defined as follows:

Y1=H1L,Y2=H2L.    (3)

In this study, we define CAL layers as a function mapping from X1 and X2 to Y1 and Y2 using shared notation:

Y1,Y2=CAL(X1,X2;θCAL).    (4)

Here, θCAL={θ1,θ2} represents the set of trainable parameters used in the CAL layers.

3.1.2 Global token layer (common)

By contrast, the common GTL (Figure 3a) forces all data streams to exchange information through a set of trainable tokens, G, greatly reducing potential redundancy across frames. GTL takes N(≥ 2) input data streams X1, …, XN and models their interactions via G. It outputs the corresponding data streams and updated global tokens for use in subsequent layers. Each stream Xn and Yn is a matrix in D×Tn, and the global tokens are represented as a matrix G ∈ ℝD×B, where B represents the dimension and length of the bottleneck tokens.

Figure 3
Comparison of (a) GTL with a common global token shared across all inputs and updated collectively, and (b) GTL with pairwise tokens, where each input interacts with others via distinct pairwise global tokens that are updated and integrated separately into each input's representation.

Figure 3. Illustration of common and pairwise GTL when there are three input data streams. (a) GTL (common). (b) GTL (pairwise).

GTL is implemented by repeating a transformer encoder with self-attention for L layers. Let Hn0=Xn and G0 denote the initial inputs. At each layer = 1, …, L, the intermediate representations are computed as follows:

[HnGn]=TransformerEnc([Hn-1G-1];θn),    (5)

where [·||·] denotes concatenation along the temporal (sequence) dimension. The global token is updated by aggregating across all n as follows:

G=nGn.    (6)

The final outputs of the GTL are defined as follows:

Yn=HnL.    (7)

In this study, we define the GTL module as follows:

Y1,,YN=GTLcommon(X1,,XN;θGTLc).    (8)

Here, TransformerEnc(·;θ) denotes a self-attention-based Transformer Encoder layer, and θGTLc={{θn},n,G0} represents the parameters used in the GTL module.

The advantage of GTL is that interactions between different modalities or participants are channeled through a compact set of tokens, potentially alleviating computational bottlenecks and improving the learning of cross-modal relationships. On the other hand, CAL can suffer from higher complexity when dealing with longer input sequences, particularly those exhibiting high redundancy across frames, such as audio or video.

3.1.3 Global token layer (pairwise)

To accurately model the interaction between each pair of data streams, the previous study proposed defining multiple global token sequences, each corresponding to a pair of data streams (Suzuki et al., 2024). The pairwise Global Token Layer (GTL) (Figure 3b) also takes N (N ≥ 2) input data streams X1, …, XN and models their interactions, producing the corresponding output streams Y1, …, YN, in a manner similar to the common GTL.

However, it differs from the common GTL in that it defines and utilizes multiple global token sequences Gmn to capture pairwise interactions among the input streams, m and n. Each token sequence GmnD×B is dedicated to modeling the interaction between Xm and Xn.

Let Hn0=Xn and Gmn0=G0 denote the initial input and pairwise tokens. Then, at each layer = 1, …, L, we compute:

[HmnJ\mGmn]=TransformerEnc([Hm-1nJ\mGmn-1];θm-1),    (9)

where ⊕ represents vector concatenation (⊕n = 1, 2 An = [A1||A2]), and θm-1 denotes the parameters of the l-th layer of the Transformer-encoder block for the m-th data stream.

Instead of Equation 6 in common GTL, each global-token sequence is updated by summing the variables that represent the dependencies of participants in both directions as follows:

Gmn=Gmn+Gnm    (10)
Gnm=Gnm+Gmn.    (11)

The final outputs of the pairwise GTL are defined as follows:

Yn=HnL,    (12)

We define the pairwise GTL function as follows:

Y1,,YN=GTLpairwise(X1,,XN;θGTLp).    (13)

Here, θGTLp={{θn}n,,G0} represents the parameters of the pairwise GTL module.

3.2 Task

We adopt the four engagement classes for training and evaluation, as in Multipar-T, as a baseline method (Lee et al., 2023). This four-label design is consistent with other engagement corpora (e.g., DAiSEE) and has effectively captured varying degrees of participant attentiveness and dis-engagement. Figure 4 provides an outline of the engagement-estimation task. Let N represent the total number of participants involved in a conversation. Our goal is to estimate the engagement of a specific participant n* ∈ {1, ⋯ , N} at a certain time t based on conversation data. More specifically, we estimate engagement over the time interval [tD, t], meaning that a video and audio segment with a duration of D s serves as the context. In the remainder of this section, we omit the explicit mention of time t and refer to the video and audio clips of participant n as Xnv and Xna, respectively. We also denote the engagement label of the target participant as ln*. We framed the task as a four-class classification problem, where ln ∈ {1 (High Dis-Engagement), 2 (Low Dis-Engagement), 3 (Low Engagement), 4 (High Engagement)}. The input data X are represented as follows:

X={X1v,X1a,,XNv,XNa},    (14)

Thus, the engagement-estimation task can be formulated as follows:

l^n*=f(X,n*;Θ),    (15)

where f(·) is the classification function provided by the model, and Θ denotes the parameters of that model. This setup follows the approach in a previous study (Lee et al., 2023), with the addition of audio data streams.

Figure 4
Diagram showing a multi-modal engagement/disengagement classification model. Video and audio inputs pass through encoders and feature extractors to produce outputs (Zv) and (Za). These outputs undergo processing across (K) layers with GTL (Graph Transformer Layers), producing (UK) outputs. A combination of these outputs goes through AP (Average Pooling), FC (Fully Connected layer), and Softmax, resulting in a classification labeled as high/low engagement or disengagement.

Figure 4. Schematic diagram of engagement-estimation task, DSPBT. “AP” and “FC” denote attention pooling and fully connected layer, respectively.

3.3 Multi-person bottleneck transformers

3.3.1 Baseline multi-person bottleneck transformer

The baseline multi-person bottleneck transformer (MPBT) model (Suzuki et al., 2024) uses CALs for the cross-modal interaction model and GTLs for the cross-person interaction model. It initially uses pre-trained encoders to extract audio and video features from the input data for each participant n ∈ {1, ⋯ , N} through speech and video encoders as follows:

Zna=SpeechEncoder(Xna;θa),    (16)
Znvenc=VideoEncoder(Xnv;θvenc),    (17)

where VideoEncoder(·) and SpeechEncoder(·) are functions that project data into feature vectors for video and speech, respectively. The parameters of the encoders are denoted as θvenc and θa, and ZnmDmodel×Tm represents the feature vectors for the modality m ∈ {venc, a}, where Dmodel is the feature dimension and Tm is the time length. Similar to Multipar-T (Lee et al., 2023), we also extract features, such as head movements, from the video using OpenFace (Baltrušaitis et al., 2016), which are then added to the output of the video encoder as follows:

Znvfeat=VideoFeatureExtractor(Xnv;θvfeat),    (18)
Znv=Znvenc+FC(Znvfeat;θFC1),    (19)

where VideoFeatureExtractor(·) represents the projection function that maps data to feature vectors, θvfeat denotes its corresponding parameters, ZnvfeatDvfeat×Tvenc represents the feature vectors, where Dvfeat denotes the feature dimension, FC(·) refers to a fully connected layer, and θFC1 represents its parameters. The details of these extracted features are described in the “Encoder Configurations” paragraph in Section 4.2.

Znv and Zna are used as inputs to the CAL with L layers,

Sna,Snv=CAL(Zna,Znv;θCAL).    (20)

Subsequently, the interaction among participants was modeled using a global-token sequence. The outputs of the CALs are concatenated to create a participant feature vector SnDmodel×(Tv+Ta) for each participant. To manage the quadratic complexity of attention, we introduce GTL with K layers;

Sn=[Sna,LSnv,L],    (21)
U1,,UN=GTLcommon(S1,,SN;θGTLc).    (22)

Finally, pooling is applied to the output of the final layer of the transformer encoder to obtain the posterior probabilities of the labels as follows:

P(ln*|X,n*,Θ)=softmax(FC(AP(Un*;θAP);θFC2)),    (23)

where softmax(·), AP(·), θAP and θFC2 represent the softmax-function, attention-pooling-layer, attention-pooling-layer parameters, and fully connected layer parameters, respectively.

The model parameters Θ are optimized by minimizing the cross-entropy loss on the training data D as follows:

Θ={θFC1,θCAL,θGTLc,θAP,θFC2}    (24)
L=-X,n,lnDlogP(ln|X,n,Θ).    (25)

Notably, the encoder parameters θvenc, θa, and θvfeat are pre-trained and remain frozen during training.

3.3.2 Baseline participant-pairwise bottleneck transformer

To accurately model the interaction between each pair of participants, the baseline PPBT model (Suzuki et al., 2024) defines multiple global token sequences, with each token corresponding to a pair of participants. Specifically, instead of using Equation 22 in the baseline MPBT, PPBT utilizes pairwise GTL to model cross-person interaction.

U1,,UN=GTLpairwise(S1,,SN;θGTLp).    (26)

The calculation of posterior probabilities from the output of the final layer follows the same process as the baseline MPBT.

3.4 Multi-data stream bottleneck transformers

3.4.1 Baseline multi-data stream bottleneck transformer

We introduce the baseline multi-data stream bottleneck transformer (MDSBT). The baseline MDSBT was designed to model interactions across input data streams, i.e., multiple modalities from multiple participants by common global tokens. Instead of modeling cross-modal and cross-person interactions by using Equations 2022 of a hierarchical MPBT model, the MDSBT models the interactions among all input data streams by GTLs with K layers,

U(1,a),U(1,v),,U(N,a),U(N,v)=GTLcommon(Z1a,Z1v,,ZNa,ZNv;θGTLc).    (27)

Finally, the output variables associated with the target participant are concatenated to form Un*K as follows:

Un*=[U(n*,a)U(n*,v)].    (28)

The posterior probabilities of the labels are calculated in the same manner as Equation 23.

3.4.2 Proposed data stream-pairwise bottleneck transformer

The proposed DSPBT introduces pairwise global tokens to the baseline MDSBT. Instead of using Equation 27 of the baseline MDSBT, DSPBT uses pairwise GTL as follows:

U(1,a),U(1,v),,U(N,a),U(N,v)=GTLpairwise(Z1a,Z1v,,ZNa,ZNv;θGTLp).    (29)

Following the process of the baseline MDSBT, the output variables associated with the target participant are concatenated and fed to the classification layer, in the same manner as Equation 23.

4 Experiment

4.1 Experimental dataset

We used the RoomReader corpus (Reverdy et al., 2022), which comprises multimodal, multiparty conversational interactions where participants engaged in a collaborative online student-tutor scenario designed to elicit spontaneous speech. This dataset was chosen because it reflects realistic online multiparty interactions (4—5 participants), aligning with our aim to capture cross-participant dynamics in engagement. Notably, it has also been utilized in Multipar-T, allowing us to compare results under consistent conditions (Lee et al., 2023). The corpus was processed to separate the audio and video for each participant and synchronize them. The video resolution is 2,560 × 1,440, with a frame rate of 60 fps, and the audio is sampled at 32 kHz with 16-bit quantization. In the experiment, the frame rate was reduced to 8 fps. Although higher frame rates would capture more detailed facial expressions, they would significantly increase GPU memory usage and limit the computing environments in which the model could be trained. The corpus also includes continuous annotations for engagement. The data are labeled every second, and the label at the last second of each clip was used as the target. The labels range from –2 to 2. Instead of a regression task, we defined the task as a four-class classification, where labels in the range [–2, –1] indicate high disengagement (l = 1), (–1, 0] indicate low disengagement (l = 2), (0, 1] indicate low engagement (l = 3), and (1, 2] indicate high engagement (l = 4). We trained models on data from 24 groups and tested them on data from 6 groups. Each group consisted of five participants. We split the training and test sets so that the test set did not include any participant in the training set. The number of video clips used for training was 53,192 and 12,756 for testing (a total of 65,948). This was smaller than a previous study (Lee et al., 2023), where 152,614 clips were extracted from the RoomReader corpus, of which 121,305 were allocated for training and 31,309 for testing. This difference was due to our exclusion of clips with errors in face detection and OpenFace-based feature extraction. Specifically, if face region detection or feature extraction failed for even a single frame in a clip, that clip was excluded from the experiments.

4.2 Setups

4.2.1 Pre-processing

For video inputs, we detected face regions in each frame using YOLOv3 (Redmon and Farhadi, 2018), which was trained on the Wider Face dataset (Yang et al., 2016). In our experiments, we resized the detected facial regions to a resolution of 128 × 128. While ResNet-50 (He et al., 2016) typically employs inputs with a resolution of 224 × 224, we opted for this lower resolution to reduce GPU memory consumption. This decision was made due to the large volume of data we used, which included both video and audio recordings from five participants. Previous studies have demonstrated that a resolution of 128 × 128 can still achieve sufficiently high accuracy (Touvron et al., 2019), making it a viable choice for our setup. Table 3 shows the frequency of each label in the 8-second video clips from the training and test sets. It is evident that there is a significant class imbalance. To mitigate the effects of class imbalance, we oversampled the infrequent classes to balance the frequency distribution in the training set. We did not use Focal Loss (Lin et al., 2017) because it was not effective under our experimental conditions.

Table 3
www.frontiersin.org

Table 3. Frequency of each label in the 8-second video clips from the training and test data.

4.2.2 Encoder configurations

We used ResNet-50 (He et al., 2016) as the video encoder. We used the normalized eye-gaze direction, head location, 3D landmark positions, and facial-action units extracted through OpenFace (Baltrušaitis et al., 2016) as the video features. The xlsr-53 features from the final layer were used as the audio features (Conneau et al., 2021).1 The feature dimensions were Dvenc = 2,048, Da = 1,024, Dvfeat = 709. The length of each feature was Tvenc, Tvfeat = 64, Ta = 799. No further feature selection or dimensionality reduction was applied to any of these extracted features, as we opted to retain all available information for engagement estimation.

4.2.3 Methods

We evaluated the baseline CPT (Lee et al., 2023; Kim et al., 2023), MPBT (Section 3.3.1), PPBT (Section 3.3.2), MDSBT (Section 3.4.1), and the proposed DSPBT (Section 3.4.2). To examine the effect of not using a hierarchical structure (“joint modeling”) in MDSBT and DSPBT, we also evaluated MDSBT and DSPBT using a hierarchical structure, i.e., models in which CALs were replaced with GTLs in Figure 1b. We denote these hierarchical models as MDSBT1 and DSPBT1 and the originals as MDSBT2 and DSPBT2. In order to clarify the effect of using multimodal information in each model, not only the multimodal conditions (using video and audio features) but also the single-modal conditions (using only video features) were evaluated. The single-modal condition using only audio features was not evaluated because of the need to estimate the engagement of participants who were silent. Note that in the single-modal condition, MPBT is attributed to the same model structure as MDSBT, and similarly, PPBT is attributed to the same model structure as DSPBT, as there is no cross-modal interaction model.

We describe the setup that was common to all ten models. We set the number of people N to 5. For all models, the total number of transformer encoder blocks was unified; L and K were set to 2 for the hierarchical models (CPT, MPBT, PPBT, MDSBT1, and DSPBT1), while K was set to 4 for the joint models (MPBT and PPBT using single modality, MDSBT2, and DSPBT2). We used 8 s of video and audio context information, i.e., D = 8. The number of multi-head attention heads was set to 4. We used the rectified linear unit activation function; we trained all models using three seeds to calculate an average score for each experimental condition.

For the conventional CPT, instead of Equation 27 from the baseline model, we used a CAL, where the target participant's features were used as keys and values, and another participant's features were used as queries, following a previous study (Lee et al., 2023). The outputs of the five CPTs were combined, and the model dimension Dmodel was set to 256. The batch size was 4, the learning rate was 0.00001, the optimizer was Radam, and early stopping was applied (Liu et al., 2020). To ensure that the results are not due to coincidence along the feature dimension and are used as Un* in Equation 23. For all models, the length of the global tokens B was set to 4.

5 Results and discussion

Table 4 shows the results (the number of participants N = 5). We first compared models that use CALs with those that use GTLs for cross-modal interactions. CPT showed lower accuracy, weighted F1, and macro F1 than all those using all GTL cross-modal interaction models. In the models using common global tokens (MPBT, MDSBT1, and MDSBT2), MDSBT1 and MDSBT2 showed higher accuracy, weighted F1, and macro F1 than MPBT. Similarly, in models using pairwise global tokens (PPBT, DSPBT1, and DSPBT2), DSPBT1 and DSPBT2 showed higher accuracy, weighted F1, and macro F1 than PPBT. On the basis of these results, using GTL for cross-modal interaction is effective.

Table 4
www.frontiersin.org

Table 4. Evaluation results (the number of participants N = 5; modalities: V, video features; A, audio features).

Next, the GTL interaction models were compared from the viewpoint of the hierarchical structure. Using common global tokens, we compared MDSBT1 with MDSBT2. MDSBT2 showed higher accuracy, weighted F1, and macro F1. By using pairwise global tokens and GTLs for all interactions, we compared DSPBT1 with DSPBT2, and DSPBT2 showed higher accuracy, weighted F1, and macro F1. On the basis of these results, using a joint model for cross-modal and cross-person interaction modeling is effective.

Additionally, models were compared from the viewpoint of the modality. As in previous studies (Kim et al., 2023), multimodal conditions showed higher accuracy, weighted F1, and macro F1 than the single-modal conditions for each model. Given these results, by incorporating auditory cues–such as the presence or absence of speech, vocal intonation, and speech rhythm–alongside visual information (e.g., facial expression), We can improve the accuracy of engagement estimation for each participant. On the basis of these results, multimodal modeling is effective for the proposed DSPBT as well as the conventional models.

Finally, we examined the effect of adopting the pairwise global tokens for the proposed DSPBT2. We compared MDSBT2 (using common global tokens) and DSPBT2 (using pairwise global tokens), and DSPBT2 showed higher accuracy, weighted F1 and macro F1. The model that uses pairwise global tokens showed the highest accuracy, weighted F1, and macro F1 among all models. On the basis of these results, using pairwise global tokens for the proposed method is effective.

Figures 5, 6 show the accuracy and macro F1 for different numbers of participants. On the basis of the results of the models using common global tokens (MPBT/MDSBT1) and using pairwise global tokens (PPBT/DSPBT1), using GTLs for cross-modal interaction yielded higher accuracy and macro F1. On the basis of the results of the models using common global tokens (MDSBT1/MDSBT2) and using pairwise global tokens (DSPBT1/DSPBT2), not using a hierarchical structure showed higher accuracy and macro F1. The tendency of DSPBT2 remained consistent regardless of the number of participants. In a two-participant scenario, the difference between “common” and “pairwise” tokens effectively disappears as there is only one pair. As the number of participants decreases, the total number of interactions likewise diminishes, which tends to reduce the model's accuracy.

Figure 5
Bar chart showing accuracy versus number of participants (3, 4, 5) for eight models. Proposed models DSPBT1 and DSPBT2 consistently achieve the highest accuracy, followed by baselines MDSBT2, MDSBT1, and PPBT. Conventional CPT and MPBT perform the worst. Accuracy increases with more participants across all models.

Figure 5. Accuracy across different numbers of participants.

Figure 6
Bar chart showing Macro F1 scores for models across different participant counts (3, 4, 5). DSPBT2 (Proposed) consistently achieves the highest F1 scores, followed by DSPBT1 and baseline models. Macro F1 generally increases with more participants, with CPT (Conventional) and MPBT (Baseline) showing the lowest performance.

Figure 6. Macro F1 across different numbers of participants.

Regarding the per-class F1-scores, the results for the high disengagement label were not improved. This is likely due to the label-imbalance issue, as shown in Table 3. Addressing this issue requires either using a corpus with balanced labels or further investigating the details of the oversampling techniques.

Figures 7, 8 show the confusion matrix for the proposed DSPBT2 and the baseline PPBT. DSPBT2 showed a higher accuracy than DSPBT1, and the baseline PPBT showed the highest accuracy among the baseline models. DSPBT decreased misclassifications in classifying high disengagement as high engagement. This means that even if it couldn't classify high disengagement, it classified it as low disengagement, which improved the classification performance. DSPBT was more accurate in classifying low disengagement and decreased the misclassification of low disengagement as high engagement. This means that even if it couldn't classify low disengagement, it classified it as low engagement, which improved the classification performance. Low engagement and high engagement can be classified more accurately. In summary, our findings demonstrate that leveraging pairwise global tokens for non-hierarchical cross-modal and cross-person interaction modeling can enhance engagement classification performance. This approach holds promise for multiparty, multimodal tasks in fields such as human–computer interaction, group communication analysis, or social robotics.

Figure 7
Confusion matrix showing classification results for four engagement levels. Most predictions are biased toward “High Engagement,” with over 80% of each true class being classified as such. Misclassifications are primarily to adjacent engagement categories, with very low percentages assigned to “High Dis-Engagement.”

Figure 7. Confusion matrix of baseline PPBT.

Figure 8
Confusion matrix showing predicted versus true labels for four engagement levels. “High Engagement” is most accurately classified (86.1%), while “Low Dis-Engagement” shows significant confusion, with 33.8% predicted as “Low Engagement” and 50.9% as “High Engagement.” Overall, misclassifications are frequent between adjacent engagement categories.

Figure 8. Confusion matrix of DSPBT2.

The model performs best when each participant is recorded with a front-facing webcam that keeps the face largely inside the frame, mirroring the setup of the RoomReader corpus. Consequently, gallery view video conference recordings or any scenario in which every participant has a dedicated webcam tile are the most suitable inputs. Recordings where faces are small, heavily occluded, or only intermittently visible (e.g., speaker view or wide-angle room cameras) may require additional front facing cues or model retraining to maintain accuracy. By integrating both auditory (e.g., speech presence, vocal intonation, and rhythm) and visual (e.g., facial expression) cues, we achieve more accuracy. Future work will focus on extending these methods to larger and more diverse datasets, as well as exploring strategies to mitigate class imbalance and further improve the classification of minority classes.

6 Conclusion

We proposed the data stream-pairwise bottleneck transformer (DSPBT), which uses pairwise global tokens while simultaneously handling both cross-modal and cross-person interactions. Compared with the baseline PPBT, DSPBT not using a hierarchical structure showed better accuracy, weighted F1, and macro F1. These findings confirm our main concept—that unifying cross-modal and cross-person interactions through global token-based transformer effectively reduces redundancy and facilitates more accurate engagement estimation in multiparty settings. Additionally, the method is effective in multimodal conditions. We also showed that this tendency remains consistent regardless of the number of participants from the viewpoint of the multimodal condition.

Building on these findings, our future work will explore more diverse data scenarios, such as varying conversation lengths and many participants, while also addressing class-imbalance challenges through improved data augmentation or tailored loss functions. Ultimately, we aim to make the DSPBT framework more robust, scalable, and adaptable to a broader range of real-world multiparty interactions.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.

Author contributions

KSu: Writing – original draft, Writing – review & editing. NH: Writing – original draft, Writing – review & editing. KSh: Writing – review & editing. SM: Writing – review & editing. RM: Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was funded by NTT Corporation's internal research and development budget.

Acknowledgments

We would like to thank our colleagues who participated in the discussion.

Conflict of interest

KSu, NH, KSh, SM, and RM were employed by NTT Corporation and conducted the research as part of their employment. NTT Corporation had the following involvement in the study: the study design and the decision to submit it for publication.

Generative AI statement

The author(s) declare that Gen AI was used in the creation of this manuscript. We used GPT-o1 and GPT-o1-pro for proofreading.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

References

Baltrušaitis, T., Robinson, P., and Morency, L.-P. (2016). “OpenFace: an open source facial behavior analysis toolkit,” in Proceedings of the IEEE Winter Conference on Applications of Computer Vision, 1–10. doi: 10.1109/WACV.2016.7477553

Crossref Full Text | Google Scholar

Bednarik, R., Eivazi, S., and Hradis, M. (2012). “Gaze and conversational engagement in multiparty video conversation: an annotation scheme and classification of high and low levels of engagement,” in Proceedings of the 4th Workshop on Eye Gaze in Intelligent Human Machine Interaction. doi: 10.1145/2401836.2401846

Crossref Full Text | Google Scholar

Cafaro, A., Wagner, J., Baur, T., Dermouche, S., Torres Torres, M., Pelachaud, C., et al. (2017). “The NoXi database: multimodal recordings of mediated novice-expert interactions,” in Proceedings of the International Conference on Multimodal Interaction, 350–359. doi: 10.1145/3136755.3136780

Crossref Full Text | Google Scholar

Chen, H., Alghowinem, S. M., Jang, S. J., Breazeal, C., and Park, H. W. (2022). Dyadic affect in parent-child multi-modal interaction: introducing the DAMI-P2C dataset and its preliminary analysis. IEEE Trans. Affect. Comput. 14, 3345–3361. doi: 10.1109/TAFFC.2022.3178689

Crossref Full Text | Google Scholar

Conneau, A., Baevski, A., Collobert, R., Mohamed, A., and Auli, M. (2021). “Unsupervised cross-lingual representation learning for speech recognition,” in Proceedings of the Annual Conference of the International Speech Communication Association, 2426–2430. doi: 10.21437/Interspeech.2021-329

Crossref Full Text | Google Scholar

Dermouche, S., and Pelachaud, C. (2019). “Engagement modeling in dyadic interaction,” in Proceedings of the International Conference on Multimodal Interaction, 440–445. doi: 10.1145/3340555.3353765

Crossref Full Text | Google Scholar

Dhall, A., Ghosh, S., Goecke, R., and Gedeon, T. (2019). “EmotiW 2019: automatic emotion, engagement and cohesion prediction tasks,” in Proceedings of the International Conference on Multimodal Interaction, 546–550. doi: 10.1145/3340555.3355710

Crossref Full Text | Google Scholar

Dhall, A., Kaur, A., Goecke, R., and Gedeon, T. (2018). “EmotiW 2018: audio-video, student engagement and group-level affect prediction tasks,” in Proceedings of the International Conference on Multimodal Interaction, 653–656. doi: 10.1145/3242969.3264993

Crossref Full Text | Google Scholar

Fredricks, J. A., Blumenfeld, P. C., and Paris, A. H. (2004). School engagement: potential of the concept, state of the evidence. Rev. Educ. Res. 74, 59–109. doi: 10.3102/00346543074001059

PubMed Abstract | Crossref Full Text | Google Scholar

Fredricks, J. A., Filsecker, M., and Lawson, M. A. (2016). Student engagement, context, and adjustment: addressing definitional, measurement, and methodological issues. Learn. Instr. 43, 1–4. doi: 10.1016/j.learninstruc.2016.02.002

PubMed Abstract | Crossref Full Text | Google Scholar

Grafsgaard, J. F., Wiggins, J. B., Vail, A. K., Boyer, K. E., Wiebe, E. N., and Lester, J. C. (2014). “The additive value of multimodal features for predicting engagement, frustration, and learning during tutoring,” in Proceedings of the International Conference on Multimodal Interaction, 42–49. doi: 10.1145/2663204.2663264

Crossref Full Text | Google Scholar

Gupta, A., D'Cunha, A., Awasthi, K., and Balasubramanian, V. (2016). DAiSEE: towards user engagement recognition in the wild. arXiv preprint arXiv:1609.01885.

Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 770–778. doi: 10.1109/CVPR.2016.90

Crossref Full Text | Google Scholar

Huynh, V. T., Kim, S.-H., Lee, G.-S., and Yang, H.-J. (2019). “Engagement intensity prediction with facial behavior features,” in Proceedings of the International Conference on Multimodal Interaction, 567–571. doi: 10.1145/3340555.3355714

Crossref Full Text | Google Scholar

Kim, Y., Lee, D. W., Liang, P. P., Alghowinem, S., Breazeal, C., and Park, H. W. (2023). “HIINT: historical, intra-and inter-personal dynamics modeling with cross-person memory transformer,” in Proceedings of the International Conference on Multimodal Interaction, 314–325. doi: 10.1145/3577190.3614122

Crossref Full Text | Google Scholar

Kumar, D., Madan, S., Singh, P., Dhall, A., and Raman, B. (2024). “Towards engagement prediction: a cross-modality dual-pipeline approach using visual and audio features,” in Proceedings of the International Conference on Multimedia, 11383–11389. doi: 10.1145/3664647.3688986

Crossref Full Text | Google Scholar

Lee, D. W., Kim, Y., Picard, R. W., Breazeal, C., and Park, H. W. (2023). “Multipar-T: multiparty-transformer for capturing contingent behaviors in group conversations,” in Proceedings of the International Joint Conference on Artificial Intelligence, 3893–3901. doi: 10.24963/ijcai.2023/433

Crossref Full Text | Google Scholar

Li, W., Duan, Y., Lu, J., Feng, J., and Zhou, J. (2020). “Graph-based social relation reasoning,” in Proceedings of the European Conference on Computer Vision, 18–34. doi: 10.1007/978-3-030-58555-6_2

Crossref Full Text | Google Scholar

Li, Y.-Y., and Hung, Y.-P. (2019). “Feature fusion of face and body for engagement intensity detection,” in Proceeding of the IEEE International Conference on Image Processing, 3312–3316. doi: 10.1109/ICIP.2019.8803488

Crossref Full Text | Google Scholar

Lin, T.-Y., Goyal, P., Girshick, R., He, K., and Dollár, P. (2017). “Focal loss for dense object detection,” in Proceedings of the IEEE International Conference on Computer Vision, 2980–2988. doi: 10.1109/ICCV.2017.324

Crossref Full Text | Google Scholar

Lin, Y.-S., and Lee, C.-C. (2020). “Predicting performance outcome with a conversational graph convolutional network for small group interactions,” in Proceedings of the International Conference on Acoustics, Speech and Signal Processing, 8044–8048. doi: 10.1109/ICASSP40776.2020.9053308

Crossref Full Text | Google Scholar

Liu, L., Jiang, H., He, P., Chen, W., Liu, X., Gao, J., et al. (2020). “On the variance of the adaptive learning rate and beyond,” in Proceedings of the International Conference on Learning Representations.

PubMed Abstract | Google Scholar

Ma, J., Jiang, X., Xu, S., and Qin, X. (2021). “Hierarchical temporal multi-instance learning for video-based student learning engagement assessment,” in Proceedings of the International Joint Conference on Artificial Intelligence, 2782–2789. doi: 10.24963/ijcai.2021/383

Crossref Full Text | Google Scholar

Nagrani, A., Yang, S., Arnab, A., Jansen, A., Schmid, C., and Sun, C. (2021). “Attention bottlenecks for multimodal fusion,” in Proceedings of the International Conference on Neural Information Processing Systems, 14200–14213.

Google Scholar

Pan, S., Xu, G. J., Guo, K., Park, S. H., and Ding, H. (2023). Video-based engagement estimation of game streamers: an interpretable multimodal neural network approach. IEEE Trans. Games 16, 746–757. doi: 10.1109/TG.2023.3348230

Crossref Full Text | Google Scholar

Redmon, J., and Farhadi, A. (2018). YOLOv3: an incremental improvement. arXiv preprint arXiv:1804.02767.

Google Scholar

Reverdy, J., O'Connor Russell, S., Duquenne, L., Garaialde, D., Cowan, B. R., and Harte, N. (2022). “RoomReader: a multimodal corpus of online multiparty conversational interactions,” in Proceedings of the International Conference on Language Resources and Evaluation Conference, 2517–2527.

Google Scholar

Rich, C., Ponsler, B., Holroyd, A., and Sidner, C. L. (2010). “Recognizing engagement in human-robot interaction,” in Proceedings of the ACM/IEEE International Conference on Human-Robot Interaction. doi: 10.1145/1734454.1734580

Crossref Full Text | Google Scholar

Ringeval, F., Sonderegger, A., Sauer, J., and Lalanne, D. (2013). “Introducing the RECOLA multimodal corpus of remote collaborative and affective interactions,” in Proceedings of the IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, 1–8. doi: 10.1109/FG.2013.6553805

Crossref Full Text | Google Scholar

Sanghvi, J., Castellano, G., Leite, I., Pereira, A., McOwan, P. W., and Paiva, A. (2011). “Automatic analysis of affective postures and body motion to detect engagement with a game companion,” in Proceedings of the ACM/IEEE International Conference on Human-Robot Interaction, 305–312. doi: 10.1145/1957656.1957781

Crossref Full Text | Google Scholar

Savchenko, A. V., Savchenko, L. V., and Makarov, I. (2022). “Classifying emotions and engagement in online learning based on a single facial expression recognition neural network. IEEE Trans. Affect. Comput. 13, 2132–2143. doi: 10.1109/TAFFC.2022.3188390

Crossref Full Text | Google Scholar

Sinatra, G. M., Heddy, B. C., and Lombardi, D. (2015). The challenges of defining and measuring student engagement in science. Educ. Psychol. 50, 1–13. doi: 10.1080/00461520.2014.1002924

Crossref Full Text | Google Scholar

Singh, M., Hoque, X., Zeng, D., Wang, Y., Ikeda, K., and Dhall, A. (2023). “Do I have your attention: a large scale engagement prediction dataset and baselines,” in Proceedings of the International Conference on Multimodal Interaction, 174–182. doi: 10.1145/3577190.3614164

Crossref Full Text | Google Scholar

Sukumaran, A., and Manoharan, A. (2024). Multimodal engagement recognition from image traits using deep learning techniques. IEEE Access 12, 25228–25244. doi: 10.1109/ACCESS.2024.3353053

Crossref Full Text | Google Scholar

Sun, L., Lian, Z., Liu, B., and Tao, J. (2023). Efficient multimodal transformer with dual-level feature restoration for robust multimodal sentiment analysis. IEEE Trans. Affect. Comput. 15, 309–325. doi: 10.1109/TAFFC.2023.3274829

Crossref Full Text | Google Scholar

Suzuki, K., Hojo, N., Shinoda, K., Mizuno, S., and Masumura, R. (2024). “Participant-pair-wise bottleneck transformer for engagement estimation from video conversation,” in Proceedings of the Annual Conference of the International Speech Communication Association, 4079–4083. doi: 10.21437/Interspeech.2024-1329

Crossref Full Text | Google Scholar

Touvron, H., Vedaldi, A., Douze, M., and Jegou, H. (2019). “Fixing the train-test resolution discrepancy,” in Advances in Neural Information Processing Systems, eds. H. Wallach, H. Larochelle, A. Beygelzimer, F. d' Alché-Buc, E. Fox, and R. Garnett (London: Curran Associates, Inc.).

Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need,” in Proceedings of the International Conference on Neural Information Processing Systems, 6000–6010.

Google Scholar

Wang, K., Yang, J., Guo, D., Zhang, K., Peng, X., and Qiao, Y. (2019). “Bootstrap model ensemble and rank loss for engagement intensity regression,” in Proceeding of the International Conference on Multimodal Interaction, 551–556. doi: 10.1145/3340555.3355711

Crossref Full Text | Google Scholar

Xiong, Y., Xinya, G., and Xu, J. (2023). CNN-Transformer: a deep learning method for automatically identifying learning engagement. Educ. Inf. Technol. 15, 9989–10008. doi: 10.1007/s10639-023-12058-z

Crossref Full Text | Google Scholar

Yang, F., Yin, W., Inamura, T., Björkman, M., and Peters, C. (2020). “Group behavior recognition using attention- and graph-based neural networks,” in Proceedings of the European Conference on Artificial Intelligence, 1626–1633.

Google Scholar

Yang, S., Luo, P., Loy, C.-C., and Tang, X. (2016). “WIDER FACE: a face detection benchmark,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 5525–5533. doi: 10.1109/CVPR.2016.596

Crossref Full Text | Google Scholar

Zhou, C., Han, M., Liang, Q., Hu, Y.-F., and Kuai, S.-G. (2019). A social interaction field model accurately identifies static and dynamic social groupings. Nat. Hum. Behav. 3, 847–855. doi: 10.1038/s41562-019-0618-2

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: transformer, engagement, multiparty conversation, multimodal, classification, global token

Citation: Suzuki K, Hojo N, Shinoda K, Mizuno S and Masumura R (2025) Data stream-pairwise bottleneck transformer for engagement estimation from video conversation. Front. Artif. Intell. 8:1516295. doi: 10.3389/frai.2025.1516295

Received: 24 October 2024; Accepted: 05 May 2025;
Published: 27 June 2025.

Edited by:

Amelia Zafra, University of Cordoba, Spain

Reviewed by:

Arghir-Nicolae Moldovan, National College of Ireland, Ireland
Valentina Saccone, University of Florence, Italy

Copyright © 2025 Suzuki, Hojo, Shinoda, Mizuno and Masumura. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Keita Suzuki, a2VpdGF4cy5zdXp1a2lAbnR0LmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.