Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Neurorobot., 06 February 2026

Volume 20 - 2026 | https://doi.org/10.3389/fnbot.2026.1760494

This article is part of the Research TopicMultimodal human action recognition in real or virtual environmentsView all 5 articles

Multimodal sequence dynamics and convergence optimization in dual-stream LSTM networks for complex physiological state estimation

  • Department of Public Physical Education, China Academy of Art, Hangzhou, Zhejiang, China

Introduction: The integration of virtual simulation with intelligent modeling is crucial for advancing the scientization and personalization of volleyball physical training. This study aims to overcome the convergence instability and feature misalignment in modeling multimodal kinematic and physiological sequences.

Methods: A dynamical framework based on a Dual-Stream Long Short-Term Memory network integrated with a temporal attention mechanism is proposed. The framework decouples heterogeneous feature learning and optimizes temporal weight distribution.

Results: Experimental validation on complex motion state estimation demonstrates that the proposed model reduces load modeling error to 3.8% and achieves a motion classification accuracy of 93.1%. The velocity trajectory fitting coefficient of determination is 0.91 with a peak deviation of 0.05 m/s.

Discussion: These results confirm the effectiveness of the attention-based DS-LSTM in optimizing multimodal sequence modeling for training state estimation and feedback.

1 Introduction

As a sport that combines both competitive and entertaining elements, volleyball places high demands on athletes’ physical fitness (Albaladejo-Saura et al., 2023; Lin et al., 2024; Pawlik and Mroczek, 2023). Physical training not only determines an athlete’s explosive power, endurance, and agility during competition but also plays a crucial role in injury prevention and extending their athletic career (Esposito et al., 2024; Rebelo et al., 2022; Tan et al., 2023). In recent years, the development of virtual simulation technology has brought new opportunities to sports training. Virtual reality and augmented reality platforms, combined with sensor acquisition and intelligent analysis, can provide athletes with immersive training environments and data-driven feedback (Li et al., 2024; Sousa et al., 2023). Research on how to combine virtual simulation with intelligent modeling for volleyball physical training not only promotes innovation in training models but also has practical value in improving the scientific nature and personalization of training (Duan, 2021; Yunchao et al., 2023).

Current research on volleyball physical training faces three key challenges. First, motion capture and muscle load monitoring data often exhibit high dimensionality and temporal dependencies, making it difficult for traditional statistical methods to accurately characterize their dynamic patterns (Amin et al., 2023; Bae et al., 2024; Suo et al., 2024). Second, the data types collected by multimodal sensors vary widely, and asynchrony and noise exist between posture, muscle load, and velocity curves, impacting data fusion and modeling accuracy (Hafer et al., 2023; Tsanousa et al., 2022). Third, existing training feedback mechanisms largely rely on empirical judgment and lack real-time prediction and interpretable feedback based on neural network models, making personalized training optimization difficult to implement (Hribernik et al., 2022; Vec et al., 2024). These difficulties have hindered the practical application of virtual simulation technology in volleyball physical training.

To address these challenges, various technical approaches have been proposed (Putranto et al., 2023; Richlan et al., 2023). Convolutional neural networks have been used for motion recognition, but their structure is biased towards static feature extraction, making it difficult to capture the dynamic evolution of long time series (Huang and Cai, 2023; Le et al., 2022). Recurrent neural networks have demonstrated effectiveness in movement classification tasks (Amin and Noh, 2024), and possess certain advantages in time series modeling. However, they are susceptible to vanishing gradients in multimodal data fusion, resulting in insufficient convergence speed and stability (Su et al., 2025; Waqas and Humphries, 2024). Multi-sensor fusion methods have attempted to use weighted averaging and Kalman filtering, but their ability to fit complex nonlinear relationships is limited, making it difficult to accurately reflect the coupling effects between different training features (Tang et al., 2023; Urrea and Agramonte, 2021). Attention mechanisms have been introduced in some studies to highlight key features, but they are mostly used in single-modal scenarios and lack cross-modal interaction (Lu et al., 2023; Wang and Liang, 2023). Overall, these methods still have shortcomings in accurately modeling multimodal time series data and optimizing personalized training, and have not yet addressed the need for scientific and real-time feedback in volleyball physical training (Salim et al., 2024; Wang et al., 2022).

To address the shortcomings of traditional methods in multimodal time-series data modeling and training feedback, this paper proposes a dynamic modeling approach based on DS-LSTM and a temporal attention mechanism. This framework uniquely applies temporal attention after separate kinematic and physiological temporal encoding, rather than before fusion. This structured decoupling and reweighting of modalities enhances convergence stability and mitigates temporal misalignment. This approach first synchronizes and suppresses the temporal noise of posture, electromyography, and acceleration data. Then, in a dual-stream architecture, kinematic and physiological features are fed into recursive units to capture their internal dynamic patterns. Interactive feature fusion is then used to integrate cross-modal information. Furthermore, a temporal attention mechanism is introduced to automatically identify key performance-affecting periods during training and weight these features, ultimately outputting training status predictions and optimization recommendations. This approach enables dynamic modeling and intelligent feedback for volleyball physical training in a virtual simulation environment, enhancing the interpretability and convergence stability of training data and providing athletes with a personalized, scientific training path.

2 Methods

2.1 Data collection and preprocessing

In virtual simulation experiments of volleyball physical training, the acquisition of multimodal time-series data is a prerequisite for dynamic modeling. This paper constructs a dataset based on motion capture sensors, myoelectric sensors, and accelerometers. Motion capture sensors output the athlete’s joint positions and angles in three-dimensional space, reflecting the temporal changes in posture and movement. Myoelectric sensors record the electrophysiological signals of major muscle groups during training, revealing the muscle load under high-intensity exercise. Accelerometers measure the velocity and acceleration curves of training movements to characterize the explosive power and rhythmic characteristics of the movement. Due to differences in sampling rate, data dimension, and duration settings among the three sensor types, the raw data must be standardized. The data acquisition parameters are shown in Table 1.

Table 1
www.frontiersin.org

Table 1. Data acquisition parameters.

The raw data will be affected by noise and drift during the acquisition process, and directly inputting the model will cause convergence difficulties. To this end, targeted preprocessing methods are used for different signals: the electromyographic signal is filtered by a Butterworth bandpass filter to remove low-frequency drift and high-frequency interference, the motion capture and acceleration signals are filtered by sliding average to eliminate jitter, and the mechanical data are filtered by low-pass filtering to reduce high-frequency noise. All signals are then normalized to eliminate dimensional differences so that data from different modes are comparable within the same range. Normalization uses the standard scaler method, as shown in Equation 1 (Sinsomboonthong, 2022).

x = ( x μ ) σ     (1)

In Equation 1, x is the original data value, μ represents the mean of the feature dimension, σ represents its standard deviation, and x is the normalized output. This process centers the data and scales it based on variance, which is more robust for handling potential out-of-range data during model application.

Because the sampling rates of various sensors vary, timing alignment is necessary on a unified time axis. This paper uses the electromyographic signal with the highest sampling rate as the time reference and employs linear interpolation to interpolate the lower-sampling-rate signals, synchronizing all data at the same time step. To verify the effectiveness of preprocessing, a comparative analysis of signal quality metrics before and after filtering and alignment was performed, including signal-to-noise ratio, timing alignment error, and stability index. The results are shown in Table 2.

Table 2
www.frontiersin.org

Table 2. Comparison of data preprocessing effects.

2.2 DS-LSTM structure modeling

In the dynamic modeling of multimodal time series data, kinematic and physiological signals differ significantly. Previous dual-stream LSTM models commonly perform early or mid-level fusion without explicit temporal reweighting, which limits their ability to suppress redundant segments in long training sequences. Attention-based recurrent architectures reported in related studies typically operate on concatenated multimodal representations, leading to interference between heterogeneous dynamics during attention computation. By contrast, the proposed DS-LSTM with temporal attention performs attention weighting only after independent temporal encoding, ensuring that the attention distribution reflects modality-consistent temporal patterns rather than mixed feature responses. Posture capture sequences are characterized by high dimensionality, strong temporal dependencies, and dynamic continuity, while electromyographic and acceleration signals reflect the changing patterns of muscle load and exercise intensity. To avoid the coupling interference of single-stream networks when processing cross-modal features, this paper employs a DS-LSTM architecture to perform temporal modeling of kinematic and physiological features in independent branches, and then implements interactive fusion of cross-modal features at a high level.

The kinematic flow takes the coordinate sequence of the posture key points as input and recursively updates the hidden state through the LSTM unit X k = { x 1 k , x 2 k x T k } expanded in time steps to capture the dynamic dependency between spatial displacement and joint motion. Let the input sequence be, where x T k represents t the key point feature vector at the moment, and the recursive process of LSTM can be expressed as Lees et al. (2021):

h t k , c t k = LSTM ( x t k , h t 1 k , c t 1 k )     (2)

In Equation 2, h t k represents the hidden state of c t k the kinematic flow at time t, t is the unit state, and the recursive function includes nonlinear updates of the input gate, forget gate, and output gate. The output sequence of this branch { h 1 k , h 2 k h T k } describes the dynamic structure of the action sequence.

The physiological flow takes the fusion vector of myoelectric and acceleration signals as input and focuses on the activation intensity of muscle groups and the load changes during exercise. Assuming the input sequence is X p = { x 1 p , x 2 p x T p } , its recursive process is consistent with Equation 2, but the hidden state and unit state are recorded as and, respectively, h t p . c t p This branch is formed through feature modeling of time series expansion { h 1 p , h 2 p h T p } and is used to characterize the dynamic changes at the physiological level of exercise.

In the dual-stream architecture, the kinematic and physiological streams maintain the integrity of temporal feature representation within independent recursive units, thus avoiding feature aliasing in early stages. High-level feature fusion will subsequently incorporate attention mechanisms and fully connected mapping to achieve cross-modal interaction modeling and prediction optimization. The overall framework of the dual-stream architecture is shown in Figure 1.

Figure 1
Flowchart diagram illustrating a multi-stream data processing pipeline for training state prediction and personalized feedback. Pose and physiological data are acquired, preprocessed, modeled using LSTM networks, fused via temporal attention weighting, then used for prediction and feedback generation.

Figure 1. The proposed dual-stream LSTM (DS-LSTM) modeling framework. The architecture decouples kinematic (posture) and physiological (EMG/acceleration) data processing into independent streams. This ensures feature integrity before high-level fusion, allowing the model to capture distinct dynamic patterns for each modality.

2.3 Temporal attention mechanism fusion

After the DS-LSTM architecture completes modeling of both kinematic and physiological features, a temporal attention mechanism is introduced to achieve weighted screening of key training segments. Multimodal time series data exhibits significant phase-by-phase differences during training. Features from some time periods contribute significantly to overall training state prediction, while others represent noise or redundant information. To address this issue, the hidden state sequences output by the DS-LSTM are set to be H k = { h 1 k , h 2 k h T k } and H p = { h 1 p , h 2 p , , h T p } , where h t k and represent h t p the kinematic and physiological flow feature vectors T at time t, respectively, t and is the total length of the sequence.

The calculation of attention weight is achieved through the interaction between the learnable parameter vector and the feature latent representation, and the weight distribution is defined as Hernández and Amigó (2021):

α t = exp ( u tanh ( W h t + b ) ) j = 1 T exp ( u tanh ( W h j + b ) )     (3)

In Equation 3, h t represents the candidate feature vector after stacking and fusion of the dual-stream output, W and b is the trainable linear transformation parameter, and u is the global context vector. In Equation 3, the numerator is used to measure the moment t . The denominator ensures that the weights are normalized at all times to satisfy the constraints of the probability distribution.

The weights obtained α t based on Equation 3 are used to weight the feature sequence to obtain the context representation after time series fusion:

c = t = 1 T α t h t     (4)

In Equation 4, c the temporal context vector integrates the dynamic contributions of key segments across modalities, eliminating interference from redundant features. In this process, training phases corresponding to moments with larger weights have a higher proportion in the final representation, ensuring that the model prioritizes high-value segments during prediction and feedback.

To further improve the effectiveness of cross-modal interaction, this paper introduces a feature fusion operation after the attention layer. The context vectors of kinematic and physiological features filtered by attention are spliced and nonlinearly mapped in the feature space:

z = σ ( W f [ c k ; c p ] + b f )     (5)

Where, c k and are c p the attention-weighted contextual representations of the kinematic and physiological flows, respectively, represents [ c k ; c p ] the vector concatenation operation, W f and are b f the fusion layer parameters, σ ( · ) and is a nonlinear activation function. Equation 5 implements cross-modal interactive expression, establishing an association between kinematic and physiological features in a shared space, ensuring that the final training state prediction not only relies on information from a single modality, but also integrates the interactive features of multimodal key time series segments.

While ensuring information selectivity, this structure also provides a more compact and discriminative input for the subsequent fully connected prediction layer, effectively improving the accuracy and stability of training optimization feedback. The overall process is shown in Figure 2.

Figure 2
Flowchart diagram showing a deep learning model architecture with four main layers: Input Layer (Kinematic and Physiological LSTM output sequences), Temporal Attention Mechanism Layer, Fusion Layer, and Output Layer generating fused interaction feature representation.

Figure 2. Schematic diagram of the temporal attention mechanism fusion structure. This layer calculates weight distributions for different time segments to filter key motion phases. It effectively integrates kinematic and physiological context vectors to generate a robust fused feature representation for state prediction.

2.4 Training optimization and feedback generation

The fused time series feature set Z = { z t } t = 1 T Through a multi-layer feedforward network and task-specific output heads, the training objectives are mapped into posture prediction y ̂ pose , electromyography prediction y ̂ emg , velocity prediction y ̂ vel , and attention weight sequences A = { a h , t } . The training objectives are uniformly expressed as a multi-task joint loss function, which is expressed as shown in Equation 6:

L total = w p L pose + w e L emg + w v L vel + λ att L att + λ smoot h L smoot h + λ reg θ 2 2     (6)

Among them, L pose is the posture prediction error, L emg is the electromyography regression error, L vel is the velocity curve fitting error, L att is the attention regularization term, L smoot h is the time series smoothing term, θ 2 2 is the L2 regularization term of the model parameters; w p , w e , w v is the task weight, λ att , λ smoot h , λ reg is the regularization weight, and θ is the model parameter vector.

The posture loss is in the form of mean square error, which is defined as Equation 7.

L pose = 1 N j T t = 1 T j = 1 N j y ̂ t , j pose y t , j pose 2 2     (7)

Equation 7, N j is the number of joints, T is the number of time steps, y ̂ t , j pose is the model’s t Step j . The prediction of each joint y t , j pose is the corresponding annotation.

The EMG regression loss is defined as Equation 8.

L emg = 1 N m T t = 1 T m = 1 N m y ̂ t , m emg y t , m emg 2 2     (8)

Equation 8, N m is the number of myoelectric channels, y ̂ t , m emg is the t Step m . The prediction of the channel y t , m emg is the labeled value.

The velocity fitting loss is defined as Equation 9:

L vel = 1 DT t = 1 T y ̂ t vel y t vel 2 2     (9)

Equation 9, D is the velocity vector dimension, y ̂ t vel is the model for the t . The prediction of y t vel the step speed is the true speed vector.

The attention regularization uses entropy terms to constrain the sparsity and interpretability of attention distribution, which is defined as follows:

L att = 1 H h = 1 H t = 1 T a h , t log ( a h , t + ε )     (10)

In H Equation 10, is the number of attention heads or channels, a h , t is the normalized attention weight (satisfies for each head t a h , t = 1 ), ε and is a numerical stabilization term to prevent logarithmic divergence.

In order to ensure the smoothness of the forecast time series, the first-order difference smoothing term is introduced, which is defined as follows:

L smoot h = 1 T 1 t = 2 T y ̂ t y ̂ t 1 2 2     (11)

Equation 11 where y ̂ t is the joint prediction vector obtained by splicing on each task, which is used to measure the continuity of predictions at adjacent moments.

In the mini-batch training framework, the batch loss is the average of samples within the batch, expressed as Equation 12:

L batc h = 1 B i = 1 B L total ( x ( i ) )     (12)

Equation 12, B is the batch size, x ( i ) is the multimodal sequence and label of the i -th sample in the batch.

Backpropagation uses truncated backpropagation (BPTT) to expand to length in the time dimension τ . To suppress gradient explosion, the norm of the gradient is clipped. The clipping rule is as shown in Equation 13:

g ˜ = g · min ( 1 , C g 2 )     (13)

Equation 13, g is the unclipped gradient vector, g 2 is its two-norm, C is the clipping threshold, and g ˜ is the clipped gradient.

The parameter update uses the Adam optimizer with bias correction, and the update process is based on Equation 14:

m t = β 1 m t 1 + ( 1 β 1 ) g t , v t = β 2 v t 1 + ( 1 β 2 ) g t g t , m ̂ t = m t 1 β 1 t , v ̂ t = v t 1 β 2 t , θ t + 1 = θ t η m ̂ t v ̂ t + ε .     (14)

Equation 14, g t is the clipped gradient, m t and v t are the first-order and second-order moment estimates respectively, β 1 , β 2 is the momentum coefficient, m ̂ t , v ̂ t is the bias-corrected estimate, η is the learning rate, ε is the numerical stability term.

The cosine annealing learning rate scheduling is used, and t the learning rate during the training step is calculated according to Equation 15:

η t = η min + 1 2 ( η 0 η min ) ( 1 + cos ( πt / T max ) )     (15)

In Equation 15, η 0 is the initial learning rate, η min is the minimum learning rate, T max and is the maximum number of training steps; to be consistent with the convergence observation, take T max = 500 .

When the validation set loss does not decrease in consecutive p = 20 epochs, the training is terminated and the model parameters are rolled back to the lowest point of the validation set to retain the optimal performance.

The fused features are fed into the fully connected layer to generate personalized training feedback. Feedback generation forms a closed loop with four steps: feature mapping, deviation measurement, intensity mapping, and output mapping. The feedback feature vector is given by Equation 16:

f t = σ ( W f z t + b f )     (16)

In Equation 16, W f is the size R f × D z . The weight matrix is b f , the bias vector is D z , the fusion feature dimension is σ , the component-by-component Sigmoid activation function is used to limit the output to ( 0 , 1 ) the interval, and the feedback feature vector is f t .

The deviation metric uses the Euclidean deviation from the target baseline and is normalized within the window, defined as follows:

Δ t = y t ref y ̂ t , s t = Δ t 2 max t ' W Δ t ' 2 + δ     (17)

In Equation 17, y t ref is the target baseline vector, Δ t is the deviation vector, 2 . is the Euclidean norm, and the window W . Used to calculate the normalized reference value, δ a small constant s t to prevent division by zero, and a normalized deviation scalar.

The deviation scalar is nonlinearly mapped to obtain the feedback strength I t , and the mapping is performed according to Equation 18.

I t = sigmoid ( γ s t )     (18)

In Equation 18, γ is the gain coefficient used to adjust the sensitivity of the deviation to intensity, sigmoid ( · ) is the standard S-shaped mapping, I t and is the scalar intensity.

The continuous control vector actually sent to the virtual simulation/teaching interface is obtained by linear mapping and clipping after splicing the feedback characteristics and intensity, and is expressed as follows:

u t = clip ( W u [ f t [ 2 pt ] I t ] + b u , u min , u max )     (19)

In Equation 19, W u is the output mapping matrix, b u is the bias, clip ( · , umin , umax ) clip ( · , u min , u max ) represents the component— [ u min , u max ] by-component clipping of the vector components by interval, u t and is the continuous feedback signal finally sent to the virtual simulation engine or the coaching end.

In order to make the generated feedback show a positive effect in the posterior evaluation, a closed-loop evaluation of the feedback utility is introduced and the evaluation signal is converted into an auxiliary loss for joint training, which can be rewritten as follows:

R t t + K = 1 K i = 1 K ( m t + i m t )     (20)

In Equation 20, m t is the scalar performance metric used to evaluate the training effect, K is the evaluation window length, R t t + K and is the average improvement within a given window.

L fb = R t t + K , L ' total = L total + λ fb L fb     (21)

In Equation 21, λ fb is the feedback utility weight, L ' total is the joint loss with feedback utility constraint, and is used to guide the model to generate feedback signals that reflect positive improvements in the posterior evaluation.

The training hyperparameters in this implementation are: batch size B = 32 , truncation length τ = 100 , gradient clipping threshold C = 5.0 , Adam momentum coefficient β 1 = 0.9 , β 2 = 0.999 , numerical stability term ε = 10 8 , initial learning rate η 0 = 10 3 , minimum learning rate η min = 10 6 , maximum number of training steps T max = 500 , early stopping window p = 20 , attention regularization weight λ att = 0.01 , time series smoothing weight λ smoot h = 0.1 , feedback utility weight λ fb = 0.5 , task weight is set to w p = 1.0 , w e = 0.8 , w v = 0.8 , feedback gain γ = 10.0 , output clipping interval is taken u min = 1 , u max = 1 . During the training process, batch normalization and layer normalization are used to maintain numerical stability of the hidden layer after each parameter update, and the training is rolled back at the lowest point of the validation set to retain the optimal weights.

The above training optimization and feedback generation process transforms cross-modal fusion time series predictions into executable personalized feedback signals through multi-task loss construction, constraints based on entropy and time series smoothing, a stable numerical update mechanism, and closed-loop feedback utility constraints. It also uses closed-loop evaluation to drive the adaptive fine-tuning of the model during training.

3 Experimental design and implementation

3.1 Experimental dataset construction

The study recruited 45 participants with varying levels of volleyball proficiency, including young professionals, university athletes, and amateur enthusiasts (see Table 3). The participants were aged between 15 and 28 years (Mean = 21.4, SD = 3.2), with heights ranging from 165 cm to 195 cm. Prior to the experiment, all participants (and legal guardians for those under 18) provided written informed consent. The experimental procedures regarding human subjects were strictly adhered to the Declaration of Helsinki and were approved by the institutional review board. During the experiment, the research team constructed a multi-layered dataset targeting individuals at different training levels to ensure sufficient motion diversity and sample coverage for the multimodal time series modeling process. The experimental subjects included young professionals, university teams, amateur clubs, middle school interest classes, and students from physical education colleges. Gesture capture, electromyography sensors, and accelerometers were used throughout the data collection process to continuously sample core volleyball training movements such as serving, spiking, blocking, passing, padding, and run-ups, and the cumulative training duration was recorded. To ensure the reliability of the subsequent model in motion recognition and load modeling, this section summarizes the experimental subjects, motion categories, and acquisition duration. The results are shown in Table 3.

Table 3
www.frontiersin.org

Table 3. Distribution of experimental subjects and actions.

All participants were within the age range of 15 to 28 years and their height ranged from 165 to 195 cm. Surface EMG sensors were placed on the skin overlying the bellies of the following major muscle groups involved in volleyball actions: biceps brachii, triceps brachii, anterior deltoid, pectoralis major, rectus femoris, and gastrocnemius. The specific combination of muscles monitored was adjusted per the focus of the training movement for each experimental group, ensuring coverage of primary movers for spiking, blocking, serving, and jumping actions. The distribution of experimental subjects showed that the professional youth group and physical education college students had the highest proportion of action collection time, reaching 370 and 340 min, respectively. This was primarily due to the stability and continuity of these groups’ movements during training, which reduced the number of invalid segments removed, effectively extending data retention. The university volleyball team’s training time was 320 min, indicating that their training intensity and frequency approached professional standards, but the movement continuity was slightly lower, resulting in incomplete data retention. The amateur club group’s collection time was 260 min, lower than that of the university and professional groups. This was due to insufficient standardization of movement execution, resulting in the rejection of some sequences due to substandard quality. The middle school interest class’s total time was 210 min, the lowest of all groups. This was primarily due to their lack of sports experience, resulting in frequent interruptions and unstable segments between movements, which reduced the proportion of valid data. Overall, differences in training proficiency and technical stability across different groups directly influenced the distribution of data collection time. For each subject and action sequence, the multimodal time-series data were sorted chronologically and split into training, validation, and test sets with a ratio of 70%/15%/15%. The earliest segment was used for training, the subsequent segment for validation, and the final segment for testing, ensuring that all test samples occurred strictly after the training data.

3.2 Experimental platform and implementation environment

During the experiment, the research team established a complete training environment based on a virtual simulation platform. They combined posture capture, electromyography, and accelerometers to collect multimodal data, ensuring the simultaneous acquisition of kinematic and physiological signals on the same platform. After data collection, a unified hardware and software environment was used for storage, modeling, and training, achieving a closed-loop support from virtual training scenarios to algorithm implementation. To clarify the experimental conditions and environmental configuration, the relevant parameters are shown in Table 4.

Table 4
www.frontiersin.org

Table 4. Experimental environment parameters.

This table summarizes the construction conditions of the virtual simulation platform, the sensor acquisition accuracy and number of channels, and the core hardware and software parameters, ensuring that experimental data is processed in an efficient and stable environment. The GPU and CPU configurations meet the training requirements of deep learning models, while the matching memory and operating system support the storage and access of large-scale time series data. The deep learning framework and programming environment lay the foundation for the implementation of DS-LSTM and temporal attention mechanisms.

3.3 Ablation study of model architecture components

To examine the roles of two-stream structures and temporal attention mechanisms in multimodal temporal modeling, this study conducted comparative modeling of different structural configurations while maintaining consistency in data partitioning, preprocessing procedures, network depth, hidden dimensions, and training strategies. The single-stream LSTM and the proposed DS-LSTM were implemented and trained under identical experimental conditions, including the same input features, data splits, network depth, hidden dimensions, optimizer, learning rate schedule, batch size, and stopping criteria, so that the comparison directly reflects architectural differences. Experimental setups included single-stream temporal modeling, modeling with only two-stream structures, modeling with only temporal attention, and a complete model employing both two-stream structures and temporal attention mechanisms. All models converged within the same number of training epochs, and action classification accuracy, electromyography load modeling error, and velocity curve fitting performance were evaluated on the same test set to observe the impact of structural changes on multimodal modeling behavior. The relevant results are summarized in Table 5.

Table 5
www.frontiersin.org

Table 5. Ablation results of model components.

The results in Table 5 show that the single-stream LSTM exhibits high error levels in action recognition and EMG load modeling, with an action classification accuracy of 86.4 and an EMG error of 8.7. This indicates that heterogeneous modalities generate significant information interference when modeled in a unified temporal space. When a two-stream structure is introduced without using an attention mechanism, the action accuracy improves to 90.2, the EMG error decreases to 6.0, and the velocity fitting coefficient increases to 0.87, indicating that temporal decoupling of kinematic and physiological signals makes the model’s representation of their respective dynamic modes more stable. The single-stream model using only temporal attention achieves an action accuracy of 88.1 and an EMG error of 7.4, showing that temporal weight allocation has a suppressive effect on redundant segments, but it is still limited by the modeling inconsistencies caused by feature mixing. When the dual-stream structure and time attention are introduced simultaneously, the motion accuracy is improved to 93.1, the electromyographic error is reduced to 3.8, and the velocity fitting coefficient reaches 0.91. This indicates that performing time weighting after completing intramodal temporal modeling effectively coordinates cross-modal dynamic information, thereby forming a stable and consistent modeling result.

4 Results analysis

4.1 Motion capture accuracy analysis

In the motion capture accuracy experiment, by constructing prediction results under different action categories, the performance of the method proposed in this paper is compared with other typical methods in the posture recognition task. The stability of the model in different actions is tested by combining multiple repeated experiments. At the same time, the statistical distribution analysis of the key point coordinate prediction error is carried out to fully reflect the dynamic performance of the model and the differences in data characteristics. The results are shown in Figure 3.

Figure 3
Three charts are displayed: (a) Line graph comparing action-wise accuracy for DS-LSTM+Attention, Single-stream LSTM, CNN-LSTM, and Traditional methods across spike, block, serve, dig, and average categories. (b) Line graph depicting prediction stability for accuracy with mean and standard deviation for spike, block, serve, and dig. (c) Bar graph illustrating prediction error distribution in centimeters, with frequencies decreasing from 0-1 cm to 4-5 cm.

Figure 3. Motion capture accuracy and stability analysis comparing the proposed DS-LSTM against baseline methods. (a) Action-wise accuracy comparison showing the proposed method (blue) achieving higher accuracy in spike and block actions. (b) Prediction stability error bars indicating consistent performance. (c) Histogram of prediction error distribution, highlighting that 77% of errors fall within the minimal 0–2 cm range.

The data in the figure show that our method achieves 93.1% accuracy for the spike action, while traditional methods achieve only 78.6% for the same action. This difference stems from the dual-stream architecture’s superior ability to capture temporal dependencies in high-speed movements, thereby reducing transient loss. For the block action, the accuracy is 91.8%, compared to 85.2% for the single-stream LSTM. This higher accuracy is primarily due to the interaction of cross-modal features, which enables more complete temporal modeling of upper and lower limb coordination. In the stability analysis, the standard deviation of the serve action is only 1.1, less than the 1.4 for the block action. This is because the serve action pattern is relatively fixed and has high temporal repeatability, which makes the model converge more stably during learning. Error distribution statistics show that 77 cases, representing 77% of the total data, have prediction errors less than 2 cm. This concentration is related to the attention mechanism’s ability to filter critical time periods, effectively reducing the impact of non-critical movements on prediction accuracy. Overall, this method demonstrates consistent advantages in both accuracy and stability (Figure 3).

4.2 Muscle load modeling error analysis

In the EMG load modeling experiment, multimodal time-series data from volleyball-specific physical training was collected simultaneously using posture capture, EMG, and accelerometers. A dual-stream architecture was used to model kinematic and physiological characteristics separately, and a temporal attention mechanism was incorporated to achieve cross-modal fusion. During training, 500 iterations were set to monitor model error convergence. Comparative experiments were further conducted to verify the differences between the DS-LSTM and single-stream LSTM models. The error distributions of five typical volleyball movements under repeated experiments were statistically analyzed to assess movement-level stability and volatility. The results obtained based on this experimental process are shown in Figure 4.

Figure 4
Three-panel graph showing: (a) Convergence of Muscle Load Error, with a decreasing trend in modeling error percentage over iterations; (b) Model Comparison, displaying error trends for DS-LSTM and Single LSTM models, both declining over iterations; (c) Error Distribution Across Actions, using box plots for spike, block, serve, dig, and set actions, indicating varying error percentages.

Figure 4. Analysis of EMG load modeling error convergence and motion-specific distribution. (a) The loss curve demonstrates rapid convergence to a 3.8% error rate after 500 iterations. (b) Model comparison showing the DS-LSTM (blue) reduces error more effectively than the single-stream LSTM (orange). (c) Boxplot of error distribution across varying volleyball actions, reflecting higher variability in dynamic movements like Digging.

Experimental results show that the overall convergence curve error dropped to 3.8 percent after 500 iterations, indicating that the model gradually overcomes high-frequency noise interference with continued training. This is because time alignment and filtering preprocessing suppress redundant signals, enabling the LSTM memory units to effectively capture the long-term dependencies of the EMG signals. In a model comparison, the DS-LSTM achieved an error of 6.0 percent after 200 iterations, while the single-stream LSTM achieved an error of 8.7 percent during the same period. This difference is primarily due to the dual-stream architecture’s separation of kinematic and physiological features, avoiding gradient convergence instability caused by feature aliasing. A boxplot of the action category errors shows that the median errors for Block and Set are 3.6 percent and 3.5 percent, respectively, while the median error for Dig reaches 4.3 percent, indicating that high-speed directional changes introduce EMG fluctuations, making accurate predictions in this category more difficult for the model. Overall, the multimodal fusion strategy offers advantages in reducing error levels and improving training data stability. The observed EMG load error patterns align with established EMG feature analysis practices, where lower errors correspond to consistent muscle activation timing and amplitude across trials, reflecting stable neuromuscular coordination during volleyball-specific movements.

4.3 Speed curve fitting analysis

In this experiment, we used multimodal sensors to collect posture and velocity sequences during volleyball physical training. We then used the DS-LSTM model to fit and model the movement velocity to examine the discrepancies between the predicted results and the actual measurements. After normalization and time-series alignment, the raw data was input into the model to produce the predicted output. We then systematically evaluated the model’s fitting performance and error distribution characteristics by comparing the actual velocity curves with the predicted curves, analyzing the residual distribution, and performing regression scatter point fitting. The results are shown in Figure 5.

Figure 5
Graph set showing speed analysis. Panel (a) displays a line graph comparing real speed and predicted speed over time, with closely matched curves. Panel (b) presents a residual curve indicating error variation over time, peaking at around 0.08 meters per second. Panel (c) features a scatter plot of real versus predicted speed, with a regression line (R² = 0.99) closely following the ideal y=x line.

Figure 5. Velocity curve fitting analysis verifying model precision. (a) Comparison of Real (black) vs. Predicted (red) speed curves showing tight alignment with a peak deviation of only 0.05 m/s. (b) Residual curve highlighting minor deviations primarily during peak velocity phases. (c) Regression scatter plot (R2 = 0.91) confirming a strong linear correlation between predicted and actual velocities.

The results in the figure show that the true and predicted velocity curves closely match each other for most periods of time, reaching a peak between 1.0 and 1.5 s. The true maximum value is 2.85 meters per second, while the predicted value is 2.80 meters per second, a difference of 0.05 meters per second. This minor error is primarily due to transient fluctuations in the accelerometer during high-speed motion. The residual curve shows a positive deviation of 0.05 to 0.08 meters per second between 0.4 and 1.0 s, which is related to phase differences caused by synchronization errors in the electromyographic signals. In the scatter regression plot, the data points are distributed close to the diagonal, and the coefficient of determination calculated by the regression is 0.91, indicating that the model has strong fitting stability across different velocity ranges. In summary, this method can effectively characterize the dynamic characteristics of velocity during volleyball physical training and maintain high prediction accuracy.

4.4 Explanatory analysis of attention weight distribution

In a virtual simulation experiment of volleyball physical training, to explore the model’s dynamic nature and cross-modal explanatory power in temporal feature selection, we conducted comparative modeling of the attention distribution of kinematic and physiological streams across different training time segments. This was further extended to a two-dimensional interactive analysis of movement categories and time segments. Based on multimodal time series data collected by posture, electromyography, and accelerometer sensors, the experiment extracted weight information for key time segments after two-stream LSTM encoding and attention weighting. The weight differences were then compared across kinematic and physiological features. Finally, multi-view diagrams of the model’s attention patterns across temporal segments and movement components were generated, including histograms, broken line trends, and heat matrix diagrams, as shown in Figure 6.

Figure 6
Three-part data visualization showing attention distribution over time. (a) Bar chart compares motion and physiological flow attention weights across four time segments. (b) Line chart shows attention trends over time for both flows. (c) Heatmap displays attention allocation for different actions, with colors indicating weight intensity.

Figure 6. Spatiotemporal visualization of attention weight distribution. (a) Bar chart contrasting kinematic vs. physiological attention weights across time segments. (b) Line chart illustrating the dynamic shift in attention intensity during movement execution. (c) Heatmap showing how the model prioritizes specific time windows (e.g., 4–6 s) for different action types such as blocking and spiking.

The results show that the kinematic stream’s attention weight for the 4–6 s period is 0.28, higher than the physiological stream’s 0.22. This phenomenon stems from the dominance of postural changes caused by the rapid jump during this period over the motion characteristics. The physiological stream’s weight for the 2–4 s period reaches 0.25, while the kinematic stream’s is only 0.20. This is primarily due to the increased muscle load early in training, which makes the EMG signal intensity more pronounced in driving the model weights. In the movement comparison, the Block’s weight for the 4–6 s period is 0.30, higher than the Spike’s 0.28 and the Jump’s 0.25. This result is due to the aggregation of physiological and kinematic signal characteristics generated by the combined upper limb force and core stability during the blocking process. Overall, the differences in weights over time and movement type indicate that the model’s attention to kinematic and physiological characteristics switches periodically during training.

4.5 Reliability analysis across trials and across subjects

To systematically evaluate the predictive consistency of the model across different trials and subjects, this experiment selected three key indicators: motion recognition accuracy, electromyographic load modeling error, and velocity fitting error. Based on data from all 45 subjects in multiple repeated trials, the intraclass correlation coefficient (ICC) and coefficient of variation (CV%) were calculated to quantify the model’s repeatability and stability. The relevant reliability statistics are shown in Table 6.

Table 6
www.frontiersin.org

Table 6. Statistics of model prediction reliability indicators.

The model exhibits high consistency in action recognition tasks, with ICC values of 0.94 and 0.92 for spiking and blocking actions, respectively, and CV% values below 4.0%, indicating that the dual-stream structure effectively extracts common temporal features of motion across different subjects. These results indicate that the proposed model maintains consistent prediction behavior at the individual subject level across repeated trials, demonstrating subject-specific stability rather than reliance on averaged population performance. In electromyography (EMG) load modeling, the professional group achieved an ICC of 0.91 and a CV% of 4.1% for serving, superior to the amateur group’s 0.88 and 5.3% for passing. This reflects the impact of technical standardization on EMG signal stability, with greater variability in action execution among amateur subjects leading to increased predictive volatility. Velocity fitting showed an ICC of 0.93 and a CV% of 4.5% for the approach jump, indicating strong generalization ability of the model to dynamic velocity curves. The consistency ICC of attention weights across trials was 0.89, demonstrating that the temporal attention mechanism stably captures key motion phases. Overall, the model demonstrates good cross-trial and cross-subject reliability across multiple tasks, making it suitable for training and analysis scenarios with diverse populations.

To facilitate a direct quantitative comparison between data-driven temporal modeling and handcrafted signal analysis, reliability metrics reported in this study were contrasted with representative sEMG and IMU feature-based results reported in prior musculoskeletal monitoring research. The comparison focuses on intraclass correlation coefficients and coefficients of variation, which were used in both studies to assess reproducibility and precision across repeated trials. The handcrafted feature results were drawn from a validated wearable-based rehabilitation study that evaluated time-domain and frequency-domain sEMG features together with IMU acceleration characteristics under controlled exercise protocols, while the DS-LSTM results were obtained from the multimodal volleyball training dataset analyzed in this work. The comparison is summarized in Table 7.

Table 7
www.frontiersin.org

Table 7. Reliability comparison between DS-LSTM outputs and handcrafted sEMG–IMU features.

The reliability metrics indicate that the DS-LSTM outputs exhibit a narrower variability range and higher consistency across trials compared with handcrafted sEMG and IMU features. The handcrafted feature-based analysis shows ICC values spanning from 0.81 to 0.98, accompanied by coefficients of variation extending up to 14.4 percent, reflecting sensitivity to signal amplitude fluctuations and feature extraction stability across repetitions. In contrast, the DS-LSTM results demonstrate ICC values concentrated between 0.88 and 0.94, with coefficients of variation remaining below 6.1 percent, indicating reduced dispersion in repeated measurements. This difference arises from the end-to-end temporal modeling of multimodal sequences, which suppresses trial-level signal variability through recurrent state propagation rather than relying on isolated signal descriptors. These results confirm that DS-LSTM-based modeling provides more stable reliability characteristics when evaluated using the same statistical validation criteria.

4.6 Analysis of training optimization feedback effects

In this experiment, using a virtual simulation platform as support, we conducted a multi-dimensional dynamic comparison of traditional training and personalized feedback training. We focused on collecting four core metrics: endurance, explosive power, agility, and recovery efficiency. We tracked the temporal changes in athletic performance during training and, combined with the weight distribution of the attention mechanism output, revealed differences in the model’s focus on different features and training phases. This approach not only compared the final performance differences between the two training models, but also provided in-depth analysis of the data across both the temporal and feature dimensions, ultimately yielding the results shown in Figure 7.

Figure 7
Four charts compare training methods in performance metrics. (a) Bar chart shows higher scores for

Figure 7. Evaluation of training optimization feedback effects. (a) Comparison of physical metrics (Endurance, Explosiveness, Agility, Recovery) between traditional and feedback-based training. (b) Endurance progress curve over 6 weeks showing steady improvement. (c) Heatmap of feature-specific attention weights. (d) Percentage improvement across all metrics, demonstrating the efficacy of the personalized feedback loop.

As shown in the results, endurance reached 38.7 min with personalized feedback training, compared to 32.5 min with traditional training. This improvement is primarily attributed to the dynamic adjustment of load distribution during training, which reduces the rate of lactate accumulation and delays the onset of fatigue. Explosive power reached 52.8 cm with feedback training, a significant improvement compared to 46.2 cm with traditional training. This result is related to modifications in joint angle control and muscle activation patterns. Sensitivity decreased to 10.9 s with feedback training, compared to 12.4 s with traditional training. This is due to the model’s increased focus on acceleration features during high-intensity transitions, which improves the latency of movement responses. Recovery efficiency increased to 82.6% with feedback training, compared to 71.3% with traditional training. This is due to more effective control of heart rate fluctuations, which enhances the recovery capacity of the autonomic nervous system. Overall, the personalized feedback mechanism, through optimized load distribution and feature focus strategies, leads to comprehensive improvements in athletic performance.

5 Conclusion

This paper proposes a dynamic modeling method for volleyball physical training based on the DS-LSTM and temporal attention mechanism. By collecting multimodal time series data using posture capture, electromyography, and accelerometers, and performing time alignment and feature fusion, this method achieves intelligent prediction of training status and personalized feedback. Experimental results show that this method achieves 93.1% accuracy in spike motion recognition, significantly exceeding the 78.6% of traditional methods, demonstrating the advantages of cross-modal interaction in high-speed motion modeling. In muscle load prediction, the error drops to 3.8% after 500 iterations, surpassing the 8.7% of a single-stream LSTM. This demonstrates the significant effectiveness of the dual-stream architecture in stabilizing convergence and reducing noise. In velocity curve fitting experiments, the predicted peak value of 2.80 meters per second differs only 0.05 meters per second from the true value of 2.85 meters per second, with a coefficient of determination of 0.91, demonstrating the model’s high fitting accuracy during dynamic motion. Research shows that combined with the feedback mechanism of the virtual simulation platform, the trainees have achieved significant improvements in endurance, explosive power and sensitivity. This shows that this method not only breaks through the limitations of traditional training that relies on experience and judgment, but also shows strong interpretability and scientificity in multimodal feature interaction and key period screening, providing reliable support for the personalized optimization of volleyball physical training and the promotion and application of virtual simulation teaching.

Despite the promising results, this study has certain limitations that outline directions for future research. First, the current dataset focuses primarily on specific age groups and skill levels; future work will expand the sample diversity to include a wider range of demographics to enhance model generalization. Second, while the current system operates effectively in a simulation environment, future iterations will focus on optimizing the algorithm for edge computing devices to minimize latency in real-world, on-court training scenarios. Finally, we plan to extend this dual-stream attention framework to other complex team sports, such as basketball and football, to validate its cross-disciplinary applicability.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

The studies involving humans were approved by the Ethics Committee of China Academy of Art (Approval Number: CAA-REC-2023-11-058). The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants’ legal guardians/next of kin where applicable.

Author contributions

XC: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Albaladejo-Saura, M., Vaquero-Cristóbal, R., García-Roca, J. A., and Esparza-Ros, F. (2023). What variables allow the differentiation between more and less successful adolescent volleyball players? J. Hum. Kinet. 88:229. doi: 10.2478/hukin-2023-0022

Crossref Full Text | Google Scholar

Amin, A. B., Asabre, E., Razaghi, S., and Noh, Y. (2025). Quantitative musculoskeletal monitoring and analysis in aquatic rehabilitation. Front. Electron. 6:1566899. doi: 10.3389/felec.2025.1566899

Crossref Full Text | Google Scholar

Amin, A. B., Asabre, E., Sahay, A., Razaghi, S., and Noh, Y. (2023). "Feasibility testing of wearable device for musculoskeletal monitoring during aquatic therapy and rehabilitation," 2023 45th annual international conference of the IEEE engineering in Medicine & Biology Society (EMBC), Sydney, Australia. pp. 1–4.

Google Scholar

Amin, A. B., and Noh, Y., (2024). "Classification of upper limb movements based on a LSTM model in aquatic rehabilitation," 2024 46th annual international conference of the IEEE engineering in medicine and biology society (EMBC), Orlando, FL, USA. pp. 1–4.

Google Scholar

Bae, K., Lee, S., Bak, S. Y., Kim, H. S., Ha, Y., and You, J. H. (2024). Concurrent validity and test reliability of the deep learning markerless motion capture system during the overhead squat. Sci. Rep. 14:29462. doi: 10.1038/s41598-024-79707-2,

PubMed Abstract | Crossref Full Text | Google Scholar

Duan, C. (2021). Design of online volleyball remote teaching system based on AR technology. Alex. Eng. J. 60, 4299–4306. doi: 10.1016/j.aej.2021.03.006

Crossref Full Text | Google Scholar

Esposito, G., Altavilla, G., Giardullo, G., Ceruso, R., and D’Isanto, T. (2024). The effects of the use of plyometric exercises with and without the ball in the development of explosive strength in volleyball. J. Funct. Morphol. Kinesiol. 9:126. doi: 10.3390/jfmk9030126

Crossref Full Text | Google Scholar

Hafer, J. F., Vitali, R., Gurchiek, R., Curtze, C., Shull, P., and Cain, S. M. (2023). Challenges and advances in the use of wearable sensors for lower extremity biomechanics. J. Biomech. 157:111714. doi: 10.1016/j.jbiomech.2023.111714,

PubMed Abstract | Crossref Full Text | Google Scholar

Hernández, A., and Amigó, J. M. (2021). Attention mechanisms and their applications to complex systems. Entropy 23:283. doi: 10.3390/e23030283,

PubMed Abstract | Crossref Full Text | Google Scholar

Hribernik, M., Umek, A., Tomažič, S., and Kos, A. (2022). Review of real-time biomechanical feedback systems in sport and rehabilitation. Sensors 22:3006. doi: 10.3390/s22083006,

PubMed Abstract | Crossref Full Text | Google Scholar

Huang, X., and Cai, Z. (2023). A review of video action recognition based on 3D convolution. Comput. Electr. Eng. 108:108713. doi: 10.1016/j.compeleceng.2023.108713

Crossref Full Text | Google Scholar

Le, V. T., Tran-Trung, K., and Hoang, V. T. (2022). A comprehensive review of recent deep learning techniques for human activity recognition. Comput. Intell. Neurosci. 2022:8323962. doi: 10.1155/2022/8323962

Crossref Full Text | Google Scholar

Lees, T., Buechel, M., Anderson, B., Slater, L., Reece, S., Coxon, G., et al. (2021). Benchmarking data-driven rainfall–runoff models in Great Britain: a comparison of long short-term memory (LSTM)-based models with four lumped conceptual models. Hydrol. Earth Syst. Sci. 25, 5517–5534. doi: 10.5194/hess-25-5517-2021

Crossref Full Text | Google Scholar

Li, X., Fan, D., Deng, Y., Lei, Y., and Omalley, O. (2024). Sensor fusion-based virtual reality for enhanced physical training. Robot. Intellig. Autom. 44, 48–67. doi: 10.1108/ria-08-2023-0103

Crossref Full Text | Google Scholar

Lin, H. S., Wu, H. J., Wu, C. C., Chen, J. Y., and Chang, C. K. (2024). Quantifying internal and external training loads in collegiate male volleyball players during a competitive season. BMC Sports Sci. Med. Rehabil. 16:168. doi: 10.1186/s13102-024-00958-7,

PubMed Abstract | Crossref Full Text | Google Scholar

Lu, S., Liu, M., Yin, L., Yin, Z., Liu, X., and Zheng, W. (2023). The multi-modal fusion in visual question answering: a review of attention mechanisms. PeerJ Comput. Sci. 9:e1400. doi: 10.7717/peerj-cs.1400,

PubMed Abstract | Crossref Full Text | Google Scholar

Pawlik, D., and Mroczek, D. (2023). Influence of jump height on the game efficiency in elite volleyball players. Sci. Rep. 13:8931. doi: 10.1038/s41598-023-35729-w,

PubMed Abstract | Crossref Full Text | Google Scholar

Putranto, J. S., Heriyanto, J., Achmad, S., and Kurniawan, A. (2023). Implementation of virtual reality technology for sports education and training: systematic literature review. Proc. Comp. Sci. 216, 293–300. doi: 10.1016/j.procs.2022.12.138

Crossref Full Text | Google Scholar

Rebelo, A., Valente-dos-Santos, J., Pires, I. G., Arrais, I., Pereira, J. R., and Turner, A. N. (2022). Strength and conditioning for volleyball: a review. Strength Cond. J. 44, 10–1519. doi: 10.1519/SSC.0000000000000674

Crossref Full Text | Google Scholar

Richlan, F., Weiß, M., Kastner, P., and Braid, J. (2023). Virtual training, real effects: a narrative review on sports performance enhancement through interventions in virtual reality. Front. Psychol. 14:1240790. doi: 10.3389/fpsyg.2023.1240790,

PubMed Abstract | Crossref Full Text | Google Scholar

Salim, F. A., Postma, D. B., Haider, F., Luz, S., Beijnum, B. J. F. V., and Reidsma, D. (2024). Enhancing volleyball training: empowering athletes and coaches through advanced sensing and analysis. Front. Sports Act. Living 6:1326807. doi: 10.3389/fspor.2024.1326807

Crossref Full Text | Google Scholar

Sinsomboonthong, S. (2022). Performance comparison of new adjusted min-max with decimal scaling and statistical column normalization methods for artificial neural network classification. Int. J. Math. Math. Sci. 2022:3584406. doi: 10.1155/2022/3584406

Crossref Full Text | Google Scholar

Sousa, A. C., Marques, D. L., Marinho, D. A., Neiva, H. P., and Marques, M. C. (2023). Assessing and monitoring physical performance using wearable technologies in volleyball players: a systematic review. Appl. Sci. 13:4102. doi: 10.3390/app13074102

Crossref Full Text | Google Scholar

Su, L., Zuo, X., Li, R., Wang, X., Zhao, H., and Huang, B. (2025). A systematic review for transformer-based long-term series forecasting. Artif. Intell. Rev. 58:80. doi: 10.1007/s10462-024-11044-2

Crossref Full Text | Google Scholar

Suo, X., Tang, W., and Li, Z. (2024). Motion capture technology in sports scenarios: a survey. Sensors 24:2947. doi: 10.3390/s24092947,

PubMed Abstract | Crossref Full Text | Google Scholar

Tan, B., Tian, S., Wang, E., Xiao, L., Cao, K., Zheng, B., et al. (2023). Research on the development and testing methods of physical education and agility training equipment in universities. Front. Psychol. 14:1155490. doi: 10.3389/fpsyg.2023.1155490,

PubMed Abstract | Crossref Full Text | Google Scholar

Tang, Q., Liang, J., and Zhu, F. (2023). A comparative review on multi-modal sensors fusion based on deep learning. Signal Process. 213:109165. doi: 10.1016/j.sigpro.2023.109165

Crossref Full Text | Google Scholar

Tsanousa, A., Bektsis, E., Kyriakopoulos, C., González, A. G., Leturiondo, U., Gialampoukidis, I., et al. (2022). A review of multisensor data fusion solutions in smart manufacturing: systems and trends. Sensors 22:1734. doi: 10.3390/s22051734,

PubMed Abstract | Crossref Full Text | Google Scholar

Urrea, C., and Agramonte, R. (2021). Kalman filter: historical overview and review of its use in robotics 60 years after its creation. J. Sens. 2021:9674015. doi: 10.1155/2021/9674015

Crossref Full Text | Google Scholar

Vec, V., Tomažič, S., Kos, A., and Umek, A. (2024). Trends in real-time artificial intelligence methods in sports: a systematic review. J. Big Data 11:148. doi: 10.1186/s40537-024-01026-0

Crossref Full Text | Google Scholar

Wang, M., and Liang, Z. (2023). Cross-modal self-attention mechanism for controlling robot volleyball motion. Front. Neurorobot. 17:1288463. doi: 10.3389/fnbot.2023.1288463,

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, Y., Shan, G., Li, H., and Wang, L. (2022). A wearable-sensor system with AI technology for real-time biomechanical feedback training in hammer throw. Sensors 23:425. doi: 10.3390/s23010425,

PubMed Abstract | Crossref Full Text | Google Scholar

Waqas, M., and Humphries, U. W. (2024). A critical review of RNN and LSTM variants in hydrological time series predictions. MethodsX 13:102946. doi: 10.1016/j.mex.2024.102946,

PubMed Abstract | Crossref Full Text | Google Scholar

Yunchao, M., Mengyao, R., and Xingman, L. (2023). Application of virtual simulation technology in sports decision training: a systematic review. Front. Psychol. 14:1164117. doi: 10.3389/fpsyg.2023.1164117,

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: attention mechanism, convergence analysis, multimodal dynamics, recurrent neural networks, sequence modeling, state estimation

Citation: Cao X (2026) Multimodal sequence dynamics and convergence optimization in dual-stream LSTM networks for complex physiological state estimation. Front. Neurorobot. 20:1760494. doi: 10.3389/fnbot.2026.1760494

Received: 04 December 2025; Revised: 07 January 2026; Accepted: 08 January 2026;
Published: 06 February 2026.

Edited by:

Michalis Vrigkas, University of Western Macedonia, Greece

Reviewed by:

Sawan Kumar, L&T Technology Services Limited, India
Abu Bony Amin, University of Massachusetts Amherst, United States

Copyright © 2026 Cao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xiaoxiao Cao, Y2FveGlhb3hpYW8xN0BvdXRsb29rLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.