Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Earth Sci., 22 January 2026

Sec. Geochemistry

Volume 14 - 2026 | https://doi.org/10.3389/feart.2026.1736569

This article is part of the Research TopicBig Data Mining & Artificial Intelligence in Earth ScienceView all 9 articles

Atmospheric CO2 concentration prediction based on bidirectional long short-term memory

Yina QiaoYina Qiao1Hui Yang,
Hui Yang1,2*Liu CuiLiu Cui1Yuan ZhangYuan Zhang1Gefei FengGefei Feng3Huaiwei FanHuaiwei Fan1Qingzhou LvQingzhou Lv1Yuejing YaoYuejing Yao1
  • 1School of Resources and Geosciences, China University of Mining and Technology, Xuzhou, China
  • 2Key Laboratory of Coalbed Methane Resources and Reservoir Formation Process of the Ministry of Education, China University of Mining and Technology, Xuzhou, China
  • 3Collaborative Innovation Center for Language Competence, Jiangsu Normal University, Xuzhou, China

Introduction: Accurate prediction of atmospheric CO2 concentration is essential for evaluating local emission dynamics, supporting regional carbon management, and promoting carbon neutrality goals. However, CO2 variations are highly influenced by complex interactions between meteorological conditions and anthropogenic activities, leading to highly nonlinear and time-dependent behavior that challenges conventional prediction methods.

Methods: To address this issue, a time series prediction framework based on Bidirectional Long Short-Term Memory (BILSTM) was developed for ground-based CO2 concentration forecasting. Observations from four stations of the European Integrated Carbon Observation System (ICOS) were used. Correlation analysis was first conducted between CO2 concentration and multiple meteorological variables to identify dominant driving factors. Two prediction schemes were then designed: one incorporating all meteorological variables and another using only highly correlated features. The proposed BILSTM model was evaluated against LSTM, GRU, and Support Vector Regression (SVR) models using RMSE, MAE, and R2 as performance metrics.

Results: The results indicate that correlation-based feature selection effectively improves prediction accuracy and model stability. Among all models, BILSTM consistently achieves superior performance in both short-term and long-term CO2 prediction, exhibiting greater stability and robustness than LSTM, GRU, and SVR.

Discussion: The superior performance of BILSTM is attributed to its bidirectional recurrent structure, which enables effective learning of both past and future temporal dependencies in CO2 time series. Differences in prediction accuracy among stations reflect varying environmental and anthropogenic influences.

1 Introduction

Carbon dioxide (CO2) is the most critical greenhouse gas, accounting for more than three-quarters of global greenhouse forcing. Since the Industrial Revolution, rapid industrialization and fossil fuel consumption have led to a continuous rise in atmospheric CO2 levels, intensifying the greenhouse effect and triggering frequent extreme weather events (Das et al., 2023; Diffenbaugh et al., 2017). To address these challenges, the Paris Agreement in 2015 (UNFCCC, 2016) marked the international community’s collective commitment to combat climate change, aiming to limit the global average temperature rise to below 2 °C compared to pre-industrial levels and make efforts to further reduce the increase to below 1.5 °C (Ayugi et al., 2023; Jin et al., 2024; Zhang et al., 2023; Zhang and Chen, 2022). Currently, over 190 parties have legally committed to reducing their national greenhouse gas emissions (Lelandais et al., 2022). Achieving this target relies not only on emission reduction commitments but also on accurate monitoring and prediction of atmospheric CO2 concentration at regional and local scales, which provide scientific evidence for climate policy formulation and emission regulation.

At present, satellite observation and surface measurement are the main methods to obtain the CO2 concentration (Li Y. et al., 2025; Yang et al., 2020). Satellite observation, due to its extensive spatial coverage, has become an important tool for analyzing the spatio-temporal variations of global CO2 concentration (Xiang et al., 2022; Yang et al., 2023). However, the narrow swaths and revisit cycles (Jin et al., 2022; Zhang et al., 2022) make the satellite observation data suffer from temporal discontinuity and accuracy limitations (Heymann et al., 2017; Wang W. et al., 2022). As a result, it is difficult to accurately reflect the temporal continuous variation of CO2 concentration. To make up for the deficiency, ground-based monitoring stations provide high-precision and temporal continuous CO2 observation data (Mingwei et al., 2017), which can more finely depict the real-time change of CO2 concentration (Wu C. et al., 2023). As early as 1958, the Mauna Loa Observatory in Hawaii, USA, began long-term observations of CO2 concentration (Keeling et al., 1976). Subsequently, the World Meteorological Organization (WMO) established the Global Atmosphere Watch Programme (GAW) in 1991 (Ferrarese et al., 2015; Li et al., 2022) with the goal of developing a comprehensive network observation to facilitate in-depth research into global CO2 concentration changes. However, most of these monitoring stations are located in high-altitude and remote areas far from CO2 sources and sinks (Curcoll et al., 2018). They are primarily dedicated to studying long-term CO2 concentration changes on global trend rather than regional and local variations (Conil et al., 2019). Therefore, in order to accurately monitor regional and local CO2 concentration, the Integrated Carbon Observation System (ICOS) has been established in Europe since 2012 (Yver-Kwok et al., 2021). This network consists of towers set up at different heights to monitor atmospheric components such as CO2, CH4, N2O, as well as meteorological variables like temperature and humidity (Lelandais et al., 2022). This provides valuable datasets for exploring CO2 variation mechanisms and developing predictive models.

Ground station observation, with its core advantage of high temporal resolution, can not only track and accurately depict the overall evolution trend of local CO2 concentration over the long term, but also capture its short-term dynamic changes such as diurnal fluctuations in real time. However, the variation of CO2 concentration is comprehensively influenced by multiple local factors such as traffic emissions, industrial activities, and meteorological conditions. Consequently, the CO2 time series data exhibit pronounced characteristics such as strong seasonality, nonlinearity, and local heterogeneity, which still pose challenges for CO2 prediction. Traditional statistical models and shallow learning methods lack the ability to effectively capture long-term temporal dependencies and complex nonlinear mechanisms, making it difficult to meet the requirements of high-precision prediction. In recent years, deep learning methods have become powerful tools for dealing with nonlinear and long-term dependent sequences (Cui et al., 2023; Gao and Li, 2021; Li R. et al., 2025). Recurrent Neural Network (RNN), as one of the earliest deep learning methods for time series prediction (Amalou et al., 2022; Khaldi et al., 2023), has made some progress in handling sequential data. However, it is prone to the issues of gradient disappearance and explosion (Fang et al., 2021; Wang J. et al., 2022). Long Short-Term Memory (LSTM) network is an improvement based on RNN, which can effectively capture long-range dependencies (Chu et al., 2023; Qadeer et al., 2020) by introducing gating mechanisms (Zaini et al., 2023). However, conventional LSTM network only processes sequential data in one direction, which limits the model’s ability to utilize future information (Zrira et al., 2024). The Bidirectional Long Short-Term Memory network (BILSTM) overcomes this limitation by combining forward and backward propagation, thus enabling a more comprehensive understanding of temporal dependencies and improving prediction robustness (Li et al., 2024; Liu et al., 2023; Wu K. et al., 2023).

In this study, we developed a BILSTM-based CO2 prediction model using multi-factor meteorological inputs from ICOS ground stations. The main objectives are: 1. To analyze the correlations between CO2 concentration and meteorological variables, and assess the impact of feature selection on model performance; 2. To compare the predictive accuracy of BILSTM with GRU, LSTM, and SVR baseline models using standard statistical metrics. The remainder of this paper is organized as follows: Section 2 introduces the data and methods; Section 3 presents the experimental results and discussion; Section 4 summarizes the main conclusions and future research directions.

2 Data and methods

2.1 Data source

The time series data of CO2 concentration used in this paper were obtained from the Integrated Carbon Observation System (ICOS). ICOS is an Observer Intergovernmental Organization (IGO) to the United Nations Framework Convention on Climate Change (UNFCCC). It has a prominent role in some of the working groups of the Global Climate Observing System (GCOS) and is also a part of the Group on Earth Observations (GEO) and Global Atmosphere Watch (GAW) program, with data products accessible through the World Meteorological Organization (WMO). Currently, ICOS comprises approximately 180 stations across 16 European countries, covering three major observational domains, atmosphere, ecosystem, and ocean. All ICOS stations follow unified instrumentation standards, calibration procedures, and quality-control protocols, ensuring high precision, consistency, and long-term stability of the measurements.

Specifically, the ICOS Atmosphere Release 2023-1 Level 2 greenhouse gas dataset (https://www.icos-cp.eu/data-products/atmosphere-release) was utilized (ICOS et al., 2023). This collection includes final quality controlled hourly averaged data for CO2, CH4, N2O, and CO, as well as corresponding meteorological parameters measured at multiple vertical levels. It also contains additional records such as 14C in CO2 in two-weekly integrated samples and flask sample analyses of CO2, CH4, CO, N2O, SF6 and H2, covering the period from September 2015 to April 2023.

For this research, 2 years (1 January 2020 to 31 December 2021) of hourly CO2 concentration and meteorological observations were selected from four ICOS atmospheric stations: HPB, IPR, KRE, and TRN. The meteorological variables considered include air temperature (AT), relative humidity (RH), air pressure (AP), wind direction (WD), and wind speed (WS). The geographical distribution of these stations is illustrated in Figure 1, and detailed site information is provided in Table 1.

Figure 1
Map illustrating Europe with elevation data, using a gradient from green (low) to red (high). ICOS_sites are marked with blue dots. A legend indicates data categories: ICOS_sites, ICOS member country, DEM. A scale bar and compass are included.

Figure 1. Location distribution of HPB, IPR, KRE, and TRN stations.

Table 1
www.frontiersin.org

Table 1. Detailed location of HPB, IPR, KRE, and TRN stations.

2.2 Data preprocessing

During data acquisition, inevitable factors such as instrument maintenance, network interruptions, and power outages may cause missing values in the original time series, which can negatively affect the accuracy and stability of model predictions. To assess the impact of missing data, we systematically analyzed the missingness of CO2 concentration and meteorological variables at the four observation stations from 2020 to 2021, with the results visualized in Figure 2.

Figure 2
Bar chart and heatmap showing missing data rates for 2020 and 2021. Chart (a) compares CO2 missing rates across four categories: HPB, IPR, KRE, and TRN, with the relatively higher missing rates of KRE and TRN than those of HPB and IPR. Heatmap (b) details missing rates by categories AP, RH, AT, WD, and WS for 2020 and 2021. Darker shades represent higher missing rates, with significant values in IPR and KRE for 2021.

Figure 2. Missing rate at each station. (a) CO2. (b) Meteorological variable.

The missing distribution of CO2 observations at each station across different years was shown in Figure 2a, while the missingness of meteorological variables was presented in Figure 2b. It can be seen that although all six types of observational variables at each station exhibit varying degrees of missing data, missing rates remain relatively low, with most variables missing less than 5%, indicating that the dataset quality is sufficient for model training. To minimize the impact of missing data, the linear interpolation method was applied to fill the data gaps. The calculation formula is shown in Equation 1:

yt=xj+xixjijtj(1)

where t represents the time point with missing observation data, i and j are the time points without missing data before and after time t, xi and xj are the observed values at time points i and j, and yt is the final interpolation calculation result.

Additionally, since the measurement scales of the six observation variables are different, it is necessary to normalize them and generate a unified value range of [0,1] so that improve the calculation and convergence speed of the prediction model. The calculation formula for normalization is shown in Equation 2:

y=xminxmaxxminx(2)

where x represents the original observed value, minx and maxx represent the minimum and maximum values of the observed data respectively, and y represents the normalized result.

Finally, to prepare the data for model training, the continuous time series from each station were converted into a supervised learning format. Specifically, the sliding window method was employed to segment the time series into input-output pairs, enabling the model to learn temporal dependencies from historical observations. As illustrated in Figure 3, when the window size is L and the time step is 1, each input sequence consists of L consecutive historical data points, which are used to predict the value at the next time step.

Figure 3
Diagram depicting a sequence of input and output windows on a timeline. Three rows illustrate overlapping sections where the input window is marked in blue and the output window in green. Each row shifts rightward, showing progression from \(X_{t-L}\) to \(X_{t+3}\).

Figure 3. Sliding window technology.

2.3 Methods

The overall methodological framework for CO2 concentration prediction is illustrated in Figure 4. It mainly consists of two main stages: data analysis and model training. In the first stage, time series data of CO2 concentration and meteorological observations from the HPB, IPR, KRE, and TRN stations were collected and preprocessed. Missing values were filled using linear interpolation, and all variables were normalized to enhance the numerical stability of subsequent model training. The preprocessed data were then divided into training set and testing set. In the second stage, the correlation between meteorological variables and CO2 concentration was analyzed to identify the most influential factors. The highly correlated features were selected as model inputs to reduce noise and improve predictive accuracy. Finally, the BILSTM network was constructed for time series prediction, and its performance was evaluated by statistical metrics and three baseline models, LSTM, GRU, and SVR.

Figure 4
Flowchart illustrating a machine learning model process. The data section includes dataset collection (carbon dioxide, meteorology), preprocessing (interpolation, normalization), and splitting (sliding window for train/test sets, feature selection using Pearson correlation). The model section outlines construction (input, BILSTM, dropout, dense, output layers) and an evaluation with GRU, LSTM, SVR, RMSE, MAE, and R-squared. A BILSTM diagram shows inputs, outputs, and hidden layers' interactions.

Figure 4. Methodological framework.

2.3.1 Long short-term memory

Long Short-Term Memory, first proposed by Sepp Hochreiter and Jurgen Schmidhuber in 1997, can solve the gradient disappearance or explosion problem that exists in the traditional Recurrent Neural Network model (Schmidhuber, 1997). As shown in Figure 5, its basic structure contains a memory cell ct and three gates (forget gate ft, input gate it, and output gate ot).

Figure 5
Diagram illustrating a recurrent neural network with Long Short-Term Memory (LSTM) units. Three sections show the flow of data through input, hidden, and output layers. Each section contains components like sigmoid and tanh activation functions, illustrating input \( x \), cell state \( c \), and hidden state \( h \) connections across time steps \( t-1 \), \( t \), and \( t+1 \).

Figure 5. The structure of LSTM model.

The core idea of LSTM is the adaptive gate mechanism, which determines whether and to what extent the cell state of LSTM is updated (Gers and Cummins, 2000). The input gate it contains the sigmod function and tanh function, which together control how input information is passed to the memory cell. As the most important component of LSTM, the forget gate ft determines how to retain and forget historical information by calculating the weight, so it can effectively overcome the problems of gradient disappearance and gradient explosion (Wang et al., 2021). The output gate ot determines which amount of information is to be output in the current state. The complete calculation process of the LSTM network from input to output is as follows:

sigmodx=11+ex(3)
tanhx=exexex+ex(4)
ft=σWf·ht1,xt+bf(5)
it=σWi·ht1,xt+bi(6)
ct=tanhWc·ht1,xt+bc(7)
ct=ft·ct1+it·ct(8)
ot=σWo·ht1,xt+bo(9)
ht=ot·tanhct(10)

In Equations 310, Wf, Wi, Wc and Wo denote the weight vectors of forget gate, input gate, output gate and memory cell respectively. bf, bi, bi and bo represent the bias vectors of forget gate, input gate, output gate and memory cell respectively. ft, it, ct and ot denote the vectors of forget gate, input gate, memory cell and output gate respectively. ht1 and ht represent the output vectors at time steps t and t1; xt denotes the input at time t.

2.3.2 Bidirectional long short-term memory

As shown in Figure 6, Bidirectional Long Short-Term Memory (BILSTM) network is an improvement based on LSTM model, consisting of two independent LSTM networks, forward and backward propagation (Alex Graves and Jürgen, 2005; Graves and Schmidhuber, 2005).

Figure 6
Diagram showing a Bidirectional LSTM neural network architecture. It includes input, forward, and backward layers with LSTM units. Each unit processes time steps \(x_{t-1}\), \(x_t\), and \(x_{t+1}\). An activation layer with tanh functions connects LSTM outputs \(h_{t-1}\), \(h_t\), and \(h_{t+1}\) to the output layers \(y_{t-1}\), \(y_t\), and \(y_{t+1}\). Solid and dashed lines indicate data flow and dependencies between layers.

Figure 6. The structure of BILSTM model.

The hidden state of the forward LSTM network ht is:

ht=LSTMxt,ht1(11)

In Equation 11, xt represents the input at time t, LSTM denotes the LSTM unit, ht1 represents the hidden state of the forward LSTM network at time t1.

Similarly, the hidden state of the backward LSTM network ht is:

ht=LSTMxt,ht1(12)

In Equation 12, xt represents the input at time t, LSTM denotes the LSTM unit, ht1 represents the hidden state of the backward LSTM network at time t1.

The hidden state ht of the BILSTM network is composed of the hidden states​ of forward LSTM network ht and the hidden states of backward LSTM network ht. The output value is obtained by activation function.

2.3.3 Parameter setting

The design of model parameters has a significant impact on prediction performance. After multiple rounds of testing and adjustment, the final parameter configuration is summarized in Table 2. Specifically, the CO2 concentration and meteorological data from the HPB, IPR, KRE, and TRN stations were divided into training and testing sets according to the time dimension using a 7:3 ratio. The first 70% of the data were used for model training, while the remaining 30% were used for testing. The historical sequence length was set to 24, and the prediction length was set to 1, meaning that the CO2 concentration of next hour was predicted based on observations of past 24 h from each station. The BILSTM network consisted of one BILSTM layer, one Dropout layer, and one Dense layer. The BILSTM layer contained 128 neurons, while the dropout rate was set to 0.2 to prevent overfitting. The model was trained using the Mean Squared Error (MSE) as the loss function, with a learning rate of 0.001 and a batch size of 64. The Adaptive Moment Estimation (Adam) optimizer, known for its high computational efficiency and low memory demand, was employed for parameter optimization and loss minimization. The number of epochs was set to 100. All experiments were implemented in Python 3.9.18, utilizing the TensorFlow 2.7.0 and Keras 2.7.0 frameworks. Both training and testing were conducted on a system equipped with a GeForce RTX 3060 GPU and 16 GB of memory.

Table 2
www.frontiersin.org

Table 2. Parameter Settings of BILSTM model.

2.3.4 Model evaluation

To quantitatively evaluate the performance of the BILSTM model in predicting CO2 concentration, RMSE (represent overall prediction accuracy), MAE (represent the average distance between predicted values and true values), and R2 (represent the degree of model fit) were employed (Yan et al., 2025). The formulas for each metric are shown in Equations 1316:

RMSE=t=1Tyty^t2/T(13)
MAE=1Tt=1Tyty^t(14)
R2=1t=1Tyty^t2/t=1Tyty¯2(15)
y¯=1Tt=1Tyt(16)

In the above equation, yt and y^t are the observed and predicted values at time t respectively; T is the amount of data in the test set.

3 Results and discussion

3.1 Input feature selection

Selecting appropriate input features is crucial for constructing a high precision prediction model. Feature selection not only directly affects the predictive performance of the model but also influences its generalization ability and computational efficiency. An excessive number of input variables may introduce redundancy, reduce interpretability, and increase computational cost, thereby limiting the model’s performance. To identify the key variables influencing CO2 concentration, the correlations between meteorological factors and CO2 concentration at the HPB, IPR, KRE, and TRN stations were analyzed using Pearson’s correlation coefficient. According to conventional criteria, the absolute value of the Pearson correlation coefficient less than 0.3 indicates a low correlation, between 0.3 and 0.7 indicates a moderate correlation, and above 0.7 indicates a high correlation.

As shown in Figure 7, the correlations between CO2 concentration and five meteorological factors, air temperature (AT), relative humidity (RH), air pressure (AP), wind direction (WD), and wind speed (WS), were analyzed. Among these factors, air temperature exhibits the strongest correlation with CO2 concentration at all four stations, with correlation coefficients exceeding 0.6. Relative humidity is identified as the second most influential factor, showing a relatively strong correlation at the IPR, KRE, and TRN stations. Although the correlation at the HPB station is slightly weaker, RH still ranks as the second most important variable after temperature. In contrast, air pressure, wind direction, and wind speed display consistently low correlations across all stations, with average correlation coefficients below 0.1 (Table 3). Therefore, based on the correlation analysis, air temperature (AT) and relative humidity (RH), which show relatively strong and consistent correlations with CO2 concentration across all stations, were selected as the key input features for the subsequent BILSTM prediction model.

Figure 7
Four correlation matrices labeled (a) to (d) show relationships between variables CO2, AT, RH, AP, WD, and WS. Each matrix displays correlation values in circles colored from blue (negative correlation) to red (positive correlation), with a gradient bar indicating correlation strength from -1 to 1. The matrices highlight different correlation patterns among the variables.

Figure 7. Correlation between CO2 concentration and meteorological factors at four stations. (a) HPB. (b) IPR. (c) KRE. (d) TRN.

Table 3
www.frontiersin.org

Table 3. Correlation values between CO2 concentration and meteorological factors at four stations.

3.2 Feature selection effect

To verify the influence of feature selection on the model’s predictive performance, the results of the BILSTM model using all meteorological factors as input were compared with using only air temperature (AT) and relative humidity (RH). The predicted values and the true observed values for the four stations under these two different input feature configurations are compared in Figure 8. From the prediction results, it can be observed that for all four stations, when the model input features are restricted to air temperature and relative humidity, the predicted values are closer to the actual observations, indicating a higher level of agreement between predictions and measurements. These findings suggest that simplifying the input feature set does not compromise prediction accuracy. On the contrary, it effectively eliminates redundant and weakly correlated information, thereby improving the model’s ability to capture key patterns and enhancing its overall predictive stability.

Figure 8
Four line graphs labeled (a) to (d) compare actual and predicted CO2 concentrations over time in hours. Each graph uses different data features: model predictions with atmospheric temperature and relative humidity, and predictions with all features. The CO2 levels fluctuate similarly across graphs, with lines closely following each other. Graph (b) has a higher range of CO2 levels compared to the others.

Figure 8. Comparison results before and after feature selection for each station. (a) HPB station. (b) IPR station. (c) KRE station. (d) TRN station. The black line represents the true observations; the red line represents the predicted values using only air temperature and relative humidity and the blue line represents the predicted values obtained using all meteorological factors as input.

For further quantifying the prediction performance of before and after input feature selection, the RMSE, MAE, and R2 for each station under the two feature configurations are calculated in Table 4. Additionally, the fitting relationships between predicted values and actual observations are visualized in Figure 9 through density scatter plots. As can be seen, when input features are limited to air temperature and relative humidity, the RMSE and MAE of the BILSTM model are decreased and the R2 is increased at all stations. The overall prediction accuracy is improved compared to using all meteorological factors as input features. Specifically, the average RMSE at the four stations decreases from 2.426 to 2.178, and the average MAE decreases from 3.715 to 3.429, both of which decrease by about 0.3. Meanwhile, the average R2 increases from 0.899 to 0.915. This indicates that selecting features that are highly correlated with CO2 mole fraction will be more effective than using all of them. Removing low-correlation meteorological factors can not only enhance prediction performance but also reduce computational complexity and overfitting risk.

Table 4
www.frontiersin.org

Table 4. Prediction performance of the BILSTM model under different input feature configurations.

Figure 9
Scatter plots in a two-by-four grid comparing predicted values versus actual values for different variables. Top row (AT+RH) includes graphs for HPB, IPR, KRE, and TRN with respective equations and R-squared values. Bottom row (All features) mirrors the top row variables with slight differences in equations and R-squared values. Data points are color-coded by density, ranging from purple (low) to red (high).

Figure 9. Density scatter plots between predicted values and actual observations under the two feature configurations. (a) HPB_(AT+RH). (b) HPB_(All features). (c) IPR_(AT+RH). (d) IPR_(All features). (e) KRE_(AT+RH). (f) KRE_(All features). (g) TRN_(AT+RH). (h) TRN_(All features).

It is observed that after feature selection, the prediction performance has improved. However, there are still differences among all four stations, which may be closely related to the geographical location characteristics (Figure 10). The HPB station, located in a high-altitude mountainous region in Germany, serving as a regional background station, is far from major anthropogenic emission sources. Its CO2 observations primarily reflect large-scale background signals with relatively smooth temporal variability and minimal local disturbances, making the time series easier for the model to learn and resulting in the highest prediction accuracy. In contrast, the IPR station exhibits the highest prediction error, as it is located near industrial zone in northern Italy. The CO2 observations at IPR is strongly influenced by highly variable local anthropogenic emissions, including industrial activities and traffic, making it challenging to accurately capture short-term fluctuations and leading to greater uncertainty. KRE and TRN stations show relatively similar prediction accuracies, as both are dominated by natural ecosystem processes. KRE, located in an agricultural area, is influenced by crop growth and soil respiration, while TRN, as a forest station, exhibits CO2 variations dependent on vegetation photosynthesis and respiration. Overall, the observed differences in model performance across stations are not incidental but a direct reflection of varying environmental conditions. The selected stations in this study encompass diverse geographical and environmental backgrounds, demonstrating the strong environmental adaptability of the BILSTM model in CO2 prediction. Furthermore, this analysis highlights that future prediction targeting complex environments should incorporate external indicators of human activity, such as traffic indices or emission inventories, which may further enhance prediction accuracy.

Figure 10
Satellite images of four regions labeled (a) Hohenpeissenberg, Germany, characterized by green countryside and forested areas; (b) Ispra, Italy, showing urban development and forest; (c) Trainou, France, a mix of fields and rural land; (d) Křešín u Pacova, Czech Republic, with patchwork of fields and wooded regions.

Figure 10. The locations of the four stations. (a) HPB. (b) IPR. (c) KRE. (d) TRN.

3.3 Model performance evaluation

In order to explore the accuracy and effectiveness of BILSTM model in predicting CO2 concentration, the prediction results were compared with those of three baseline models, LSTM, GRU, and SVR. The parameter settings for each model are shown in Table 5.

Table 5
www.frontiersin.org

Table 5. Parameter Settings of the three baseline models.

3.3.1 Short-term prediction performance

The performance comparison of the four models in predicting CO2 concentration is illustrated in Figure 11. It is evident that the three neural network models, BILSTM, GRU, and LSTM, demonstrate similar prediction performance, all of which outperform the SVR model, a traditional machine learning approach. Among them, BILSTM achieves the best overall performance, with RMSE and MAE values consistently lower than those of GRU and LSTM across all stations, indicating higher prediction accuracy. Although the performance of GRU and LSTM is slightly inferior to BILSTM, they still provide relatively accurate predictions. In contrast, the SVR model performs the worst among the four, exhibiting notably higher RMSE and MAE values at all stations, particularly at HPB and TRN, where prediction errors are significantly larger. This demonstrates that the generalization capability of SVR is much weaker when dealing with complex nonlinear time series data compared to neural network approaches.

Figure 11
Two bar charts compare error metrics for different models: BILSTM, GRU, LSTM, and SVR across four datasets labeled HPB, IPR, KRE, and TRN. The left chart shows RMSE values, and the right chart shows MAE values. In both charts, BILSTM generally performs similarly to GRU and LSTM, while SVR shows higher values across most datasets.

Figure 11. Comparison of prediction performance of the four models.

The quantitative evaluation metrics (RMSE, MAE, and R2) for each model and station are summarized in Table 6. As shown, the R2 of BILSTM consistently exceed those of the other three models, reflecting its superior ability to capture temporal variability in CO2 concentration. BILSTM also exhibits stronger explanatory power and a better overall fit, confirming its robustness and reliability in CO2 prediction.

Table 6
www.frontiersin.org

Table 6. MAE, RMSE, and R2 of the four models at each station.

To further analyze model performance at a detailed level, part of the prediction results from the TRN station were selected for visualization and discussion. The comparison between the true observations and the predicted results of BILSTM, GRU, LSTM, and SVR is presented in Figure 12. As shown, all four models can capture the overall temporal trend of CO2 concentration. In Figure 12a, both GRU and BILSTM accurately reproduce the detailed variations of CO2 concentration, with BILSTM showing slightly higher precision, particularly in peak regions. This improvement can be attributed to the bidirectional propagation mechanism of BILSTM, which allows the model to incorporate both past and future information for more comprehensive temporal representation. The comparison between BILSTM and LSTM (Figure 12b) further highlights this advantage. LSTM tends to produce larger prediction errors, especially around CO2 concentration peaks, whereas BILSTM more effectively captures subtle fluctuations in the time series, resulting in predictions closer to the observed values. Finally, Figure 12c compares the prediction performance of BILSTM and SVR. It is clear that BILSTM not only captures the overall trend more accurately but also provides better predictions during periods of complex CO2 fluctuations. In contrast, SVR tends to lose detailed information, leading to relatively larger errors and poorer performance.

Figure 12
Three line graphs (a, b, c) compare actual CO2 levels with predicted levels using different algorithms over time in hours. Graph (a) compares actual data with BIDirectional Long Short-Term Memory (BILSTM) and Gated Recurrent Unit (GRU) predictions, highlighting discrepancies with red circles. Graph (b) compares BILSTM and Long Short-Term Memory (LSTM), and graph (c) compares BILSTM and Support Vector Machine (SVM) predictions. Each graph includes a zoomed inset showing finer detail of variations. CO2 levels range from 400 to 460 ppm. Each graph's x-axis is time in hours, and the y-axis is CO2 concentration in ppm.

Figure 12. Comparison of the prediction results of BILSTM, GRU, LSTM, and SVR at TRN station. (a) Comparison of BILSTM and GRU prediction results. (b) Comparison of BILSTM and LSTM prediction results. (c) Comparison of BILSTM and SVR prediction results. The black line represents the true observed values, the red line represents the BILSTM prediction results, the blue line represents the GRU prediction results, the green line represents the LSTM prediction results, and the purple line represents the SVR prediction results.

Overall, both the quantitative metrics and visual analysis confirm that the BILSTM model achieves the highest prediction accuracy among the four models. Its bidirectional learning structure significantly enhances the model’s capacity to represent temporal dependencies, yielding more precise and stable prediction values compared with GRU, LSTM, and SVR.

3.3.2 Long-term prediction performance

To comprehensively verify the long-term prediction capability of the BILSTM model, we extended the prediction horizon and conducted multi-step prediction experiments to evaluate its temporal generalization performance. Specifically, the prediction intervals were expanded to 3, 6, 12, 24, and 48 h, and the average MAE, RMSE, and R2 across the four stations were calculated for each prediction step. The results are summarized in Table 7.

Table 7
www.frontiersin.org

Table 7. Multi-step prediction results of different models.

As can be seen from the Table 7, all models exhibit a continuous increase in MAE and RMSE, accompanied by a gradual decrease in R2 as the prediction horizon becomes longer. This pattern indicates that long-term prediction tasks have higher uncertainty, so the prediction accuracy decreases over time. Despite this, BILSTM consistently demonstrates superior performance compared with GRU, LSTM, and SVM across all prediction steps. Especially when the prediction step reaches 24 h and 48 h, BILSTM still maintains the lowest error and highest R2, highlighting its stronger stability and robustness in long-term prediction.

Furthermore, as illustrated in Figure 13, the performance differences between neural network models (BILSTM, GRU, LSTM) and machine learning model (SVM) is particularly significant as the prediction step increases. When the prediction step reaches 6 h, the R2 of SVM drops to only 0.59, whereas all three neural network models retain R2 above 0.7. When the prediction step is extended to 24 h, the R2 of SVM sharply decreases to 0.334, while the R2 of neural network models still maintains above 0.5. This rapid degradation in SVM performance is due to its inherent lack of temporal memory capability and inability to model long-term dependencies. In contrast, BILSTM, GRU, and LSTM benefit from gated recurrent units that effectively capture sequential patterns, enabling them to maintain relatively high R2 values. As the prediction horizon becomes longer and cumulative errors increase, the advantages of BILSTM become more evident. Across all prediction steps, BILSTM consistently exhibits the lowest MAE and RMSE, indicating that its bidirectional recurrent structure allows to leverage both forward and backward temporal information during training. This enables BILSTM to more effectively characterize the temporal dependencies of CO2 concentration dynamics, thereby achieving higher prediction reliability.

Figure 13
Three charts comparing SVM, LSTM, GRU, and BILSTM over time steps. Chart (a) shows MAE with BILSTM performing best. Chart (b) shows RMSE with BILSTM lowest. Chart (c) shows R², with BILSTM, GRU, and LSTM outperforming SVM, maintaining higher scores.

Figure 13. Comparison of prediction performance among BILSTM, GRU, LSTM, and SVM models at different time steps. (a) MAE, (b) RMSE, and (c) R2.

4 Conclusion and future work

This study proposed a CO2 concentration prediction method based on a BILSTM and compared its performance with three baseline models, GRU, LSTM, and SVR. The predictive capabilities of all models were evaluated using three statistical metrics, RMSE, MAE, and R2. The main findings are summarized as follows: 1. Among the five meteorological factors (air temperature, relative humidity, air pressure, wind direction, and wind speed) at the HPB, IPR, KRE, and TRN stations, air temperature exhibited the highest correlation with CO2 concentration, followed by relative humidity. In contrast, the correlations of air pressure, wind direction, and wind speed with CO2 concentration were relatively weak. 2. Feature selection played an essential role in improving model prediction accuracy. When only air temperature and relative humidity were used as input variables, the prediction accuracy of the BILSTM model significantly improved compared with using all meteorological factors. This result indicates that selecting features with stronger correlations effectively reduces input noise and enhances the model’s predictive performance. 3. Comparisons with GRU, LSTM, and SVR baseline models revealed that the BILSTM model achieved the best overall performance among the four models. Across all four stations, BILSTM consistently achieved the lowest RMSE and MAE values and the highest R2. The superior performance of BILSTM is attributed to its bidirectional propagation mechanism, which enables the model to capture historical and future information comprehensively so that more detailed information of CO2 concentration could be captured and further improve prediction accuracy.

Although this study demonstrates the capability of the BILSTM model to effectively predict CO2 concentrations at ground-based observation sites, several limitations remain. First, the predictive variables rely on meteorological factors, whereas CO2 is also influenced by anthropogenic emissions, land-cover characteristics, and vegetation activity. Incorporating these additional driver factors in future work would allow for a more comprehensive representation of the spatiotemporal variability and governing mechanisms of atmospheric CO2. Additionally, the present analysis is validated using data from only four ICOS stations in Europe. To further examine the generalizability of the proposed model, future studies should extend the evaluation to a larger number of ICOS sites or global GAW/NOAA observation networks, enabling a more robust assessment of model applicability under diverse atmospheric conditions. Finally, future research could explore more advanced hybrid deep learning approaches, such as introducing attention mechanisms to focus on critical prediction features, or employing graph neural networks to explicitly capture spatial correlations among different monitoring stations thereby further enhancing the predictive capability and interpretability.

Data availability statement

The CO2 and meteorological data for this study can be downloaded from the ICOS Carbon portal, DOI: 10.18160/VXCS-95EV.ICOS - Integrated Carbon Observation System.

Author contributions

YQ: Data curation, Writing – review and editing, Writing – original draft, Methodology. HY: Writing – review and editing, Conceptualization, Methodology, Funding acquisition, Supervision. LC: Formal Analysis, Writing – review and editing, Software. YZ: Formal Analysis, Writing – review and editing. GF: Writing – review and editing, Software. HF: Writing – review and editing, Formal Analysis. QL: Investigation, Writing – review and editing. YY: Investigation, Writing – review and editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This research is funded by the National Natural Science Foundation of China (42571545), the National Natural Science Foundation of China (52478011), the Third Xinjiang Scientific Expedition Program (2022xjkk1006), the Xinjiang Uygur Autonomous Region Key Research and Development Program (2022B01012-1), the Science and Technology Innovation Project of Jiangsu Provincial Department of Natural Resources (2023018), the Fundamental Research Funds for the Central Universities (2024ZDPYCH1002), the Research on Multi-scale Estimation and Simulation Methods for Terrestrial Ecosystem Carbon Storage in China (JSNRM-2025A07), the Graduate Innovation Program of China University of Mining and Technology (2023WLKXJ004), the Postgraduate Research and Practice Innovation Program of Jiangsu Province (KYCX23_2761).

Acknowledgements

We would like to thank the ICOS for providing the CO2 and meteorological data of HPB, IPR, KRE, and TRN stations.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Alex Graves, S. F., and Jürgen, S. (2005). “Bidirectional lstm networks for improved phoneme classification and recognition,” in International conference on artificial neural networks (Springer), 799–804.

Google Scholar

Amalou, I., Mouhni, N., and Abdali, A. (2022). Multivariate time series prediction by RNN architectures for energy consumption forecasting. Energy Rep. 8, 1084–1091. doi:10.1016/j.egyr.2022.07.139

CrossRef Full Text | Google Scholar

Ayugi, B. O., Chung, E. S., Zhu, H., Ogega, O. M., Babousmail, H., and Ongoma, V. (2023). Projected changes in extreme climate events over Africa under 1.5°C, 2.0°C and 3.0°C global warming levels based on CMIP6 projections. Atmos. Res. 292, 106872. doi:10.1016/j.atmosres.2023.106872

CrossRef Full Text | Google Scholar

Chu, Y., Yao, J., Qiao, D., Zhang, Z. Y., Zhong, C. Y., and Tang, L. J. (2023). Three-hourly PM2.5 and O3 concentrations prediction based on time series decomposition and LSTM model with attention mechanism. Atmos. Pollut. Res. 14 (11), 101879. doi:10.1016/j.apr.2023.101879

CrossRef Full Text | Google Scholar

Conil, S., Helle, J., Langrene, L., Laurent, O., Delmotte, M., and Ramonet, M. (2019). Continuous atmospheric CO2, CH4 and CO measurements at the Observatoire Pérenne de l'Environnement (OPE) station in France from 2011 to 2018. Atmos. Meas. Tech. 12 (12), 6361–6383. doi:10.5194/amt-12-6361-2019

CrossRef Full Text | Google Scholar

Cui, B., Liu, M., Li, S., Jin, Z., Zeng, Y., and Lin, X. (2023). Deep learning methods for atmospheric PM2.5 prediction: a comparative study of transformer and CNN-LSTM-attention. Atmos. Pollut. Res. 14 (9), 101833. doi:10.1016/j.apr.2023.101833

CrossRef Full Text | Google Scholar

Curcoll, R., Camarero, L., Bacardit, M., Àgueda, A., Grossi, C., Gacia, E., et al. (2018). Atmospheric carbon dioxide variability at aigüestortes, central pyrenees, Spain. Reg. Environ. Change 19 (2), 313–324. doi:10.1007/s10113-018-1443-2

CrossRef Full Text | Google Scholar

Das, C., Kunchala, R. K., Chandra, N., Chhabra, A., and Pandya, M. R. (2023). Characterizing the regional XCO2 variability and its association with ENSO over India inferred from GOSAT and OCO-2 satellite observations. Sci. Total Environ. 902, 166176. doi:10.1016/j.scitotenv.2023.166176

PubMed Abstract | CrossRef Full Text | Google Scholar

Diffenbaugh, N. S., Singh, D., Mankin, J. S., Horton, D. E., Swain, D. L., Touma, D., et al. (2017). Quantifying the influence of global warming on unprecedented extreme climate events. Proc. Natl. Acad. Sci. 114 (19), 4881–4886. doi:10.1073/pnas.1618082114

PubMed Abstract | CrossRef Full Text | Google Scholar

Fang, Z., Wang, Y., Peng, L., and Hong, H. (2021). Predicting flood susceptibility using LSTM neural networks. J. Hydrology 594, 125734. doi:10.1016/j.jhydrol.2020.125734

CrossRef Full Text | Google Scholar

Ferrarese, S., Apadula, F., Bertiglia, F., Cassardo, C., Ferrero, A., Fialdini, L., et al. (2015). Inspection of high–concentration CO2 events at the Plateau Rosa alpine station. Atmos. Pollut. Res. 6 (3), 415–427. doi:10.5094/apr.2015.046

CrossRef Full Text | Google Scholar

Gao, X., and Li, W. (2021). A graph-based LSTM model for PM2.5 forecasting. Atmos. Pollut. Res. 12 (9), 101150. doi:10.1016/j.apr.2021.101150

CrossRef Full Text | Google Scholar

Gers, F. A. S. J., and Cummins, F. (2000). Learning to forget: continual prediction with LSTM. Neural Computation 12 (10), 2451–2471. doi:10.1162/089976600300015015

PubMed Abstract | CrossRef Full Text | Google Scholar

Graves, A., and Schmidhuber, J. (2005). Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 18 (5-6), 602–610. doi:10.1016/j.neunet.2005.06.042

PubMed Abstract | CrossRef Full Text | Google Scholar

Heymann, J., Reuter, M., Buchwitz, M., Schneising, O., Bovensmann, H., Burrows, J. P., et al. (2017). CO2 emission of Indonesian fires in 2015 estimated from satellite-derived atmospheric CO2 concentrations. Geophys. Res. Lett. 44 (3), 1537–1544. doi:10.1002/2016gl072042

CrossRef Full Text | Google Scholar

ICOS, RI, Apadula, F., Arnold, S., Bergamaschi, P., Biermann, T., Chen, H., et al. (2023). ICOS atmosphere release 2023-1 of level 2 greenhouse gas mole fractions of CO2, CH4, N2O, CO, meteorology and 14CO2, and flask samples analysed for CO2, CH4, N2O, CO, H2 and SF6. ICOS ERIC - Carbon Portal. doi:10.18160/VXCS-95EV

CrossRef Full Text | Google Scholar

Jin, C., Xue, Y., Jiang, X., Zhao, L., Yuan, T., Sun, Y., et al. (2022). A long-term global XCO2 dataset: ensemble of satellite products. Atmos. Res. 279, 106385. doi:10.1016/j.atmosres.2022.106385

CrossRef Full Text | Google Scholar

Jin, C., Xue, Y., Yuan, T., Zhao, L., Jiang, X., Sun, Y., et al. (2024). Retrieval anthropogenic CO2 emissions from OCO-2 and comparison with gridded emission inventories. J. Clean. Prod. 448, 141418. doi:10.1016/j.jclepro.2024.141418

CrossRef Full Text | Google Scholar

Keeling, C. D., Bacastow, R. B., Bainbridge, A. E., Ekdahl, C. A., Jr., Guenther, P. R., Waterman, L. S., et al. (1976). Atmospheric carbon dioxide variations at Mauna Loa observatory, Hawaii. Tellus 28 (6), 538–551. doi:10.1111/j.2153-3490.1976.tb00701.x

CrossRef Full Text | Google Scholar

Khaldi, R., El Afia, A., Chiheb, R., and Tabik, S. (2023). What is the best RNN-cell structure to forecast each time series behavior? Expert Syst. Appl. 215, 119140. doi:10.1016/j.eswa.2022.119140

CrossRef Full Text | Google Scholar

Lelandais, L., Xueref-Remy, I., Riandet, A., Blanc, P., Armengaud, A., Oppo, S., et al. (2022). Analysis of 5.5 years of atmospheric CO2, CH4, CO continuous observations (2014-2020) and their correlations, at the Observatoire de Haute Provence, a station of the ICOS-France national greenhouse gases observation network. Atmos. Environ. 277, 119020. doi:10.1016/j.atmosenv.2022.119020

CrossRef Full Text | Google Scholar

Li, S., Kim, S., Lee, H., Kenea, S. T., Kim, J. E., Chung, C. Y., et al. (2022). Analysis of source distribution of high carbon monoxide events using airborne and surface observations in Korea. Atmos. Environ. 289, 119316. doi:10.1016/j.atmosenv.2022.119316

CrossRef Full Text | Google Scholar

Li, Y., Huang, W., Lou, K., Zhang, X., and Wan, Q. (2024). Short-term PV power prediction based on meteorological similarity days and SSA-BiLSTM. Syst. Soft Comput. 6, 200084. doi:10.1016/j.sasc.2024.200084

CrossRef Full Text | Google Scholar

Li, R., Yang, H., Zhang, X., Xu, X., Shao, L., and Bai, K. (2025a). Near real-time land surface temperature reconstruction from FY-4A satellite using spatio-temporal attention network. Int. J. Appl. Earth Observation Geoinformation 139, 104480. doi:10.1016/j.jag.2025.104480

CrossRef Full Text | Google Scholar

Li, Y., Yan, J., Zhong, L., Bao, D., Sun, L., and Li, G. (2025b). Full-coverage mapping of daily high-resolution XCO2 across China from 2015 to 2020 by deep learning-based spatio-temporal fusion. IEEE Trans. Geoscience Remote Sens., 1. doi:10.1109/tgrs.2025.3540289

CrossRef Full Text | Google Scholar

Liu, B., Wang, S., Liang, X., and Han, Z. (2023). Carbon emission reduction prediction of new energy vehicles in China based on GRA-BiLSTM model. Atmos. Pollut. Res. 14 (9), 101865. doi:10.1016/j.apr.2023.101865

CrossRef Full Text | Google Scholar

Mingwei, Z., Tianxiang, Y., Xingying, Z., Jinglu, S., Ling, J., and Chun, W. (2017). Fusion of multi-source near-surface CO2 concentration data based on high accuracy surface modeling. Atmos. Pollut. Res. 8 (6), 1170–1178. doi:10.1016/j.apr.2017.05.003

CrossRef Full Text | Google Scholar

Qadeer, K., Rehman, W. U., Sheri, A. M., Park, I., Kim, H. K., and Jeon, M. (2020). A long short-term memory (LSTM) network for hourly estimation of PM2.5 concentration in two cities of South Korea. Appl. Sci. 10 (11), 3984. doi:10.3390/app10113984

CrossRef Full Text | Google Scholar

Schmidhuber, S. H. J., and Schmidhuber, J. (1997). Long short-term memory. Neural Computation 9 (8), 1735–1780. doi:10.1162/neco.1997.9.8.1735

PubMed Abstract | CrossRef Full Text | Google Scholar

UNFCCC. (2016). The paris agreement.

Google Scholar

Wang, Z., Zhou, Y., Zhao, R., Wang, N., Biswas, A., and Shi, Z. (2021). High-resolution prediction of the spatial distribution of PM2.5 concentrations in China using a long short-term memory model. J. Clean. Prod. 297, 126493. doi:10.1016/j.jclepro.2021.126493

CrossRef Full Text | Google Scholar

Wang, J., Li, X., Li, J., Sun, Q., and Wang, H. (2022a). NGCU: a new RNN model for time-series data prediction. Big Data Res. 27, 100296. doi:10.1016/j.bdr.2021.100296

CrossRef Full Text | Google Scholar

Wang, W., He, J., Feng, H., and Jin, Z. (2022b). High-coverage reconstruction of XCO2 using multisource satellite remote sensing data in beijing–tianjin–hebei region. Int. J. Environ. Res. Public Health 19 (17), 10853. doi:10.3390/ijerph191710853

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, C., Ju, Y., Yang, S., Zhang, Z., and Chen, Y. (2023a). Reconstructing annual XCO2 at a 1 kmx1 km spatial resolution across China from 2012 to 2019 based on a spatial CatBoost method. Environ. Res. 236 (Pt 2), 116866. doi:10.1016/j.envres.2023.116866

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, K., Peng, X., Chen, Z., Su, H., Quan, H., and Liu, H. (2023b). A novel short-term household load forecasting method combined BiLSTM with trend feature extraction. Energy Rep. 9, 1013–1022. doi:10.1016/j.egyr.2023.05.041

CrossRef Full Text | Google Scholar

Xiang, R., Yang, H., Yan, Z., Mohamed Taha, A. M., Xu, X., and Wu, T. (2022). Super-resolution reconstruction of GOSAT CO2 products using bicubic interpolation. Geocarto Int. 37 (27), 15187–15211. doi:10.1080/10106049.2022.2096699

CrossRef Full Text | Google Scholar

Yan, J., Wang, S., Feng, J., He, H., Wang, L., Sun, Z., et al. (2025). New 30-m resolution dataset reveals declining soil erosion with regional increases across Chinese mainland (1990–2022). Remote Sens. Environ. 323, 114681. doi:10.1016/j.rse.2025.114681

CrossRef Full Text | Google Scholar

Yang, H., Feng, G., Xiang, R., Xu, Y., Qin, Y., and Li, S. (2020). Spatio-temporal validation of AIRS CO2 observations using GAW, HIPPO and TCCON. Remote Sens. 12 (21), 3583. doi:10.3390/rs12213583

CrossRef Full Text | Google Scholar

Yang, H., Li, T., Wu, J., and Zhang, L. (2023). Inter-comparison and evaluation of global satellite XCO2 products. Geo-spatial Inf. Sci. 28 (1), 131–144. doi:10.1080/10095020.2023.2252017

CrossRef Full Text | Google Scholar

Yver-Kwok, C., Philippon, C., Bergamaschi, P., Biermann, T., Calzolari, F., Chen, H., et al. (2021). Evaluation and optimization of ICOS atmosphere station data as part of the labeling process. Atmos. Meas. Tech. 14 (1), 89–116. doi:10.5194/amt-14-89-2021

CrossRef Full Text | Google Scholar

Zaini, N. A., Ahmed, A. N., Ean, L. W., Chow, M. F., and Malek, M. A. (2023). Forecasting of fine particulate matter based on LSTM and optimization algorithm. J. Clean. Prod. 427, 139233. doi:10.1016/j.jclepro.2023.139233

CrossRef Full Text | Google Scholar

Zhang, S., and Chen, W. (2022). China’s energy transition pathway in a carbon neutral vision. Engineering 14, 64–76. doi:10.1016/j.eng.2021.09.004

CrossRef Full Text | Google Scholar

Zhang, L., Li, T., and Wu, J. (2022). Deriving gapless CO2 concentrations using a geographically weighted neural network: china, 2014–2020. Int. J. Appl. Earth Observation Geoinformation 114, 103063. doi:10.1016/j.jag.2022.103063

CrossRef Full Text | Google Scholar

Zhang, D., Huang, X.-D., Zhong, J.-T., Guo, L. F., Guo, S. Y., Wang, D. Y., et al. (2023). A representative CO2 emissions pathway for China toward carbon neutrality under the paris Agreement's 2 °C target. Adv. Clim. Change Res. 14 (6), 941–951. doi:10.1016/j.accre.2023.11.004

CrossRef Full Text | Google Scholar

Zrira, N., Kamal-Idrissi, A., Farssi, R., and Khan, H. A. (2024). Time series prediction of sea surface temperature based on BiLSTM model with attention mechanism. J. Sea Res. 198, 102472. doi:10.1016/j.seares.2024.102472

CrossRef Full Text | Google Scholar

Keywords: bidirectional long short-term memory, carbon dioxide, ground monitoring, integrated carbon observation system, time series prediction

Citation: Qiao Y, Yang H, Cui L, Zhang Y, Feng G, Fan H, Lv Q and Yao Y (2026) Atmospheric CO2 concentration prediction based on bidirectional long short-term memory. Front. Earth Sci. 14:1736569. doi: 10.3389/feart.2026.1736569

Received: 31 October 2025; Accepted: 06 January 2026;
Published: 22 January 2026.

Edited by:

Takafumi Hirata, Hokkaido University, Japan

Reviewed by:

Tongwen Li, Sun Yat-sen University, China
Jining Yan, China University of Geosciences Wuhan, China

Copyright © 2026 Qiao, Yang, Cui, Zhang, Feng, Fan, Lv and Yao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hui Yang, eWFuZ2h1aUBjdW10LmVkdS5jbg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.