- 1CCTEG Xi‘an Transparent Geology Technology Co., Ltd., Xi'an, China
- 2National Key Laboratory of Intelligent Coal Mining and Rock Stratum Control, Beijing, China
- 3CCTEG Xi'an Research Institute (Group) Co., Ltd., Xi'an, China
- 4CCTEG China Coal Research Institute, Beijing, China
The rapid advancement of artificial intelligence (AI) and machine learning (ML) technologies has opened up novel avenues for predicting coal and gas outbursts in coal mines. This study proposes a novel prediction framework that integrates advanced AI methodologies through a multi-model fusion strategy based on ensemble learning and model Stacking. The proposed model leverages the diverse data interpretation capabilities and distinct training mechanisms of various algorithms, thereby capitalizing on the complementary strengths of each constituent learner. Specifically, a Stacking-based ensemble model is constructed, incorporating Support Vector Machines (SVM), Random Forests (RF), and k-Nearest Neighbors (KNN) as base learners. An attention mechanism is then employed to adaptively weight the outputs of these base learners, thereby harnessing their complementary strengths. The meta-learner, primarily built upon the XGBoost algorithm, integrates these weighted outputs to generate the final prediction. The model's performance is rigorously evaluated using real-world coal and gas outburst data collected from a mine in Pingdingshan, China, with evaluation metrics including the F1-score and other standard classification indicators. The results reveal that individual models, such as XGBoost, SVM, and RF, can effectively quantify the contribution of input feature importance using their inherent mechanisms. Furthermore, the ensemble model significantly outperforms single-model approaches, particularly when the base learners are both strong and mutually uncorrelated. The proposed ensemble framework achieves a markedly higher F1-score, demonstrating its robustness and effectiveness in the complex task of coal and gas outburst prediction.
1 Introduction
Coal and gas outbursts remain a significant threat to the safety of coal mine workers in China. In geologically complex regions, the deepening of mining activities further aggravates subsurface conditions, increasing both the frequency and severity of outburst-related incidents (Ou et al., 2023; Fu et al., 2022). Therefore, the development of accurate and reliable predictive models for coal and gas outbursts is critically important.
Extensive research has been conducted on predictive models for coal and gas outbursts, resulting in the development of a variety of approaches. These include the initial velocity method for borehole outbursts (Wang et al., 2020a), the drilling cuttings index method (Wang et al., 2020b), mathematical evaluation models (Soleimani et al., 2023; Hassan et al., 2017; Zhou et al., 2019; Rudakov and Sobolev, 2019; Yang et al., 2023), and AI-based models (Qiao et al., 2019; Anani et al., 2024; Li et al., 2024; Zhu et al., 2023; Song et al., 2021; Wang et al., 2023), all of which have demonstrated varying degrees of effectiveness. The rapid progress in AI technology has provided new opportunities for enhancing prediction accuracy. For instance, Fan et al. improved the SVM model using the firefly algorithm (FA) to predict coal and gas outbursts and validated its overall performance (Fan et al., 2023). Liu et al. used a least squares SVM optimized with the particle swarm optimization (PSO) algorithm, confirming its effectiveness using gas outburst data from the Jiulishan Coal Mine in Jiaozuo City, China (Liu et al., 2021). Furthermore, Zheng et al. (2023) used XGBoost to predict and analyze the contribution rate distribution of coal and gas outburst indicators. However, these aforementioned studies tend to treat coal and gas outburst prediction as an isolated task. Given the inherent uncertainties and complex underlying mechanisms of such predictions, multiple hypotheses may perform well on the training set. Relying on a single model may suffer from poor generalization due to its susceptibility to randomness and overfitting. To address these limitations, we proposed a novel multi-model fusion prediction method that integrates an attention mechanism for analyzing the contribution rates of coal and gas data (Zhao et al., 2024a; Lin et al., 2020). Initially, Pearson's correlation analysis was conducted to identify and select strongly correlated features as model inputs. Subsequently, within the Stacking ensemble framework, a coal and gas outburst prediction model that integrates multiple learners was constructed to capture a more comprehensive data observation space. Finally, the efficacy of the proposed model was rigorously validated using real-world data from the Pingdingshan Coal Mine in China. The results unequivocally demonstrate that the Stacking-based ensemble method with multi-model fusion achieves robust predictive performance for coal and gas outburst events.
2 Data analysis
2.1 Research overview and data sources
Coal and gas outbursts are influenced by four main factors: geological conditions, coal seam characteristics, gas-related factors, and operational practices. Based on field observations, these factors are further subdivided into 14 specific elements (He et al., 2010) (Figure 1). The risk level (L) of coal and gas outbursts is classified into five categories based on the amount of ejected coal (Table 1).
In this study, we used coal and gas outburst data collected from a coal mine in Pingdingshan over a period spanning from 1984 to 2009 (Xie et al., 2018). The first 50 data points were selected as the training set, while the final 10 data points were reserved for testing purposes. The coal and gas outburst data from the Pingdingshan Coal Mine are summarized in Table 2. The geographical location of Pingdingshan is shown in Figure 2.
2.2 Data process
Before feeding the training data into the model, it is crucial to carefully preprocess the dataset by identifying and addressing any anomalies.
Step 1: The Pauta criterion was applied to establish the outlier threshold. Data points that deviated from the mean by more than three standard deviations were considered outliers. These outliers were then removed, and their corresponding entries were set to null values.
Step 2: Missing values were then filled by extracting the five data points preceding and following each missing entry. In this study, Lagrange interpolation was used to estimate the missing data, as shown in Equations 1, 2.
where li(x) represents the interpolation basis function; Ln(x) denotes the interpolated value of the missing data; yi is a known (non-missing) value; x is the index corresponding to the missing value; xi denotes the index of the known data point; and xj is the interpolation node.
2.3 Correlation analysis
The primary indicator for determining the severity of a coal and gas outburst was the quantity of coal thrown (Table 1). As shown in Figure 3, Pearson's correlation analysis was performed to rigorously investigate both the interrelationships among the various influencing factors and their individual correlations with the quantity of coal thrown, as defined by Equation 3.
where cov(X, Y) is covariance and σX and σY are the standard deviations of X and Y, respectively.
Following the strong and weak correlation partitions established by Zhang et al. (2022), the correlation results presented in Figure 3 were subsequently classified. The outcomes of this classification are comprehensively detailed in Table 3. As demonstrated in Table 3, this study meticulously selected six factors exhibiting medium to high correlation levels as input variables for the model, specifically including A5, B3, D2, A4, B2, and A3.
3 Methods
3.1 XGBoost principle
XGBoost is a prominent and highly efficient boosting ensemble learning algorithm, representing an advanced evolution of the Gradient Boosted Decision Tree (GBDT) model (Yao et al., 2022; Xiong et al., 2024; Utkarsh, 2024). The predictive output of the XGBoost model is formulated as shown in Equation 4:
where ŷi denotes the predicted value for the i-th sample; K represents the number of trees; F signifies the function space of the tree; xi is the feature vector of the i-th data point; and fk refers to the function learned by the k-th tree, which is characterized by its structure q and leaf weights w.
The loss function of the XGBoost model comprises two components, as shown in Equation (5):
where the first term represents the training error between the predicted value ŷi and the true target value yi; the second term denotes the sum of tree complexities, which serves as a regularization term to control the model's complexity, as presented in Equation 6:
where γ and λ are the penalty coefficients.
During the minimization process of the objective function defined in Equation 5, the incremental function ft(xi) is added at each iteration to reduce the loss function. The objective function at the t-th iteration is presented in Equation 7:
For Equation 7, the The sample set is defined in each leaf of t objective function is approximated using a second-order Taylor expansion. The j-th tree as Ij = {i|q(xi = j)}. Here, and represent the first and second derivatives of the loss function, respectively. From these definitions, Equation 8 can be derived as follows:
Defining and leads to Equation 9:
The partial derivative with respect to wj yields Equation 10:
Substituting the weights into the objective function yields Equation 11:
A smaller loss function signifies enhanced model performance. A greedy algorithm is used to partition the subtree by enumerating feasible split points: a new split is added to existing leaves at each step, and the maximum gain is computed accordingly. The gain is then calculated as shown in Equation 12.
where the first and second terms represent the gains achieved by splitting the left and right subtrees, respectively, whereas the third term corresponds to the gain obtained without any split.
3.2 Attention mechanism
The attention mechanism receives the output from the Stacking model as its input and adaptively assigns weights to its input features, thereby emphasizing the most relevant ones and suppressing less important features, consequently facilitating more accurate feature selection (Zhu et al., 2021; Wankhade et al., 2023). The structure of the attention mechanism is shown in Figure 4.
In Figure 4, x1, x2, …, xt represents the inputs from multi-source fusion data; h1, h2, …, ht corresponds to the output state values generated by the ensemble model, while ati signifies the adaptive attention weight assigned to each output; st denotes the final output.
By calculating the correlation between h1, h2, ⋯ , ht and the current decoding time, the et, i for each influencing factor is obtained. These updated values are presented in Equation 13.
According to the computed probability et, i of each influencing factor within the population, this value is then used to compute the attention weight for each output of the ensemble model. The updated output is presented in Equation 14.
The hidden states h1, h2, ⋯ , ht are weighted by their corresponding attention values and then linearly combined. The updated output is shown in Equation 15.
St represents the final output derived through the attention mechanism, as shown in Equation 16.
where V, W, and U represent the trainable parameters, which are continuously updated during model training.
3.3 Multi-model fusion for coal and gas outburst prediction
The performance of the Stacking model is directly influenced by the number of base models used. Using too few base models may not provide adequate diversity to effectively support the meta-model, while using too many could result in redundancy, higher computational costs, and a more intricate tuning process. Typically, 3–5 base models are recommended (Kumar et al., 2024; Zhao et al., 2024b).
Based on the predictive capabilities of various base learners, this article selected high-performing models as the first-layer base learners in the Stacking model. This selection is driven by the fact that base models with strong learning abilities contribute to improving the overall predictive accuracy of the ensemble. Specifically, RF, which uses the bagging technique, is preferred for its robust learning capacity and well-established theoretical foundation, making it applicable across a wide range of domains. SVM is selected for its unique strengths in handling small datasets, non-linear relationships, and high-dimensional regression problems. KNN is included due to its solid theoretical background and efficient training process, delivering strong practical performance. For the second layer, models with robust generalization capabilities are chosen to aggregate and correct biases from the multiple base learners in the training set while mitigating overfitting through ensemble strategies. Consequently, the Stacking ensemble model incorporates RF, KNN, and SVM as the first-layer base learners, with Attention-XGBoost serving as the meta-learner in the second layer. The overall architecture is shown in Figure 5.
It is important to note that the training set for the meta-learner is derived from the outputs of the base learners. Directly utilizing the base learners' training data to form the meta-learner's training set could result in significant overfitting. To prevent the data from being redundantly learned by both layers and to avoid overfitting, an appropriate data usage strategy must be implemented. The dataset is first split into training and testing sets using cross-validation, with the three base learners making independent predictions. For each base learner, the original training dataset is partitioned into six mutually exclusive subsets, ensuring that no data IDs are repeated across subsets. For each base learner, one data subset is reserved as the validation set, while the remaining five subsets serve as the training set. Each base learner produces prediction results on its own validation subset. These predictions from the three base learners are then combined to form a new dataset, equal in size to the original dataset, as illustrated in Figure 6. This approach facilitates a comprehensive feature transformation from the original input features to the meta-learner's input features. Since each base learner's predicted data subset was excluded from its own training, this method guarantees that every data point is used only once during training, effectively preventing overfitting.

Figure 6. Coal thrown quantity prediction method based on multi-model fusion within a Stacking framework.
The training and prediction process of the AXGBoost-Stacking model is shown in Figure 6, and the detailed training procedure is outlined as follows:
Step 1: The coal and gas outburst dataset is defined as presented in Equation 17.
where xn represents the feature vector of the n-th sample, yn is the corresponding target (predicted) value, and p is the number of features, meaning that each feature vector can be expressed as (x1, x2,..., xp). Next, the dataset is partitioned into Z equally sized subsets: S1, S2,..., Sz. The cross-validation between datasets as presented in Equation 18.
where Sz denotes the z-th test set and represents the corresponding training set.
Step 2: The training set is fed into the first layer of the XGBoost-Stacking ensemble model, where three base learners are trained to obtain the base model L. Simultaneously, each sample xn in the cross-validation test set Sz is passed through the trained base model L to generate the corresponding predictions.
Step 3: The output predictions from the three base learners are concatenated to form a new data sample, which is then used as the input for the second layer of the Stacking model. At this stage, a prediction algorithm that integrates the attention mechanism with XGBoost is used to aggregate these outputs and finalize the prediction of coal ejection volume.
In this study, the AXGBoost-Stacking model is implemented using the scikit-learn library in Python. A detailed description of the algorithm is provided in Table 4.
4 Experimentation and evaluation
4.1 Model evaluation indicators
The multi-model Stacking prediction framework proposed in this study adopts the AXGBoost-Stacking model, which uses SVM, RF, and KNN as base learners and an attention-enhanced XGBoost model as the meta-learner. The Stacking ensemble learning algorithm enables a two-layer fusion of the SVM, RF, KNN, and XGBoost models. In addition to AXGBoost-Stacking model, three alternative Stacking models can also be constructed for comparative analysis:
1) The SVM-Stacking model uses RF, KNN, and AXGBoost as base learners, with SVM serving as the meta-learner.
2) The RF-Stacking model uses SVM, KNN, and AXGBoost as the base learners, with RF acting as the meta-learner.
3) The KNN-Stacking model uses SVM, RF, and AXGBoost as the base learners, with KNN acting as the meta-learner.
To evaluate the predictive performance of the AXGBoost-Stacking model and compare it with the individual predictive capabilities of the other three Stacking models, this study uses mean squared error (MSE), mean error (ME), and the F1-score [as defined in reference Xie et al. (2018)] as evaluation metrics. The formulas for calculating MSE and ME are provided in Equations 19, 20.
where yi is true data, and ŷi is prediction data.
The calculation formula for the F1-score is presented in Equation 21.
where Precision indicates the model's accuracy, and Recall represents the model's recall rate.
4.2 Comparison of prediction results
4.2.1 Input feature contribution analysis
As previously mentioned, this study uses the following features as model inputs: A5, B3, D2, A4, B2, and A3. The model's output is the coal thrown quantity, corresponding to the classification levels outlined in Table 2. Figure 7 illustrates the contribution analysis of input features for the SVM, RF, and XGBoost models. Additionally, the comparison of prediction performance among single models is provided in Figure 8.

Figure 7. Contribution of input features to the prediction model. (A) SVM model input feature contribution. (B) RF model input feature contribution. (C) XGBoost model input feature contribution.
As shown in Figure 7, A5, B3, and D2 exhibit high feature importance across different models. This finding is consistent with the Pearson correlation results presented in Table 3, indirectly validating that A5, B3, and D2 exert a greater influence on the model's predictive performance than other factors.
Based on the AUC values used for parameter tuning of the SVM, RF, KNN, and XGBoost models, the optimal parameter settings are listed in Table 5. As shown in Table 5, the prediction performance of the individual models, assessed by MSE and ME, is also compared. Combined with the results in Figure 8, it is clear that the XGBoost and SVM models exhibit superior predictive performance.
4.2.2 Performance analysis of Stacking model prediction
To evaluate the predictive performance of the Stacking ensemble model, SVM, RF, KNN, and XGBoost were used as meta-learners for comparative analysis. The selected parameters for SVM, RF, and KNN are consistent with those listed in Table 6. The resulting prediction results are shown in Figure 9, Tables 6, 7. The results highlight that the selection of base learners significantly impacts the final predictive performance.
As shown in Tables 6, 7, the method proposed in this study achieves high prediction accuracy. Moreover, a comparison between Tables 5, 7 reveals that the Stacking model outperforms the individual models in terms of prediction accuracy. Compared with the prediction results reported in Xie et al. (2018), the approach utilized in this study demonstrates superior predictive performance.
5 Conclusion
This study incorporates advanced algorithmic techniques from the fields of AI and ML. In contrast to previous studies, particularly Xie et al. (2018), this study, within the Stacking ensemble framework, leverages multiple algorithms to interpret the data space and structure from diverse perspectives, enabling complementary strengths among models and yielding optimal prediction outcomes. Experimental results demonstrate that conducting feature contribution analysis before model construction effectively quantifies the importance of each feature. The Stacking ensemble learning algorithm exhibits strong predictive accuracy and holds significant application in coal and gas outburst prediction. The main contributions of this study are summarized as follows:
• Through Pearson's correlation analysis and feature importance evaluation, coal seam angle, coal seam solidity coefficient, and slag falling situation are identified as key factors contributing significantly to the prediction outcomes.
• Compared with individual models, the Stacking ensemble model effectively integrates the strengths of each base learner, thereby enhancing overall prediction accuracy.
• Due to the complexity of the model and the risk of overfitting caused by the small data size, cross-validation was adopted to prevent overfitting from occurring. In future research, adversarial learning or large-scale models will be introduced to effectively expand and validate the dataset.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author contributions
QX: Writing – original draft, Writing – review & editing. JY: Writing – review & editing. ZD: Writing – review & editing. WD: Funding acquisition, Writing – review & editing. XW: Resources, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This study was supported by National Science and Technology Major Project of the Ministry of Science and Technology of China (2024ZD1700102) and the Key R&D Program of Shaanxi Province (2024GX-YBXM-492).
Conflict of interest
QX, JY, ZD, and WD were employed by CCTEG Xi'an Transparent Geology Technology Co., Ltd. QX, JY, ZD, WD, and XW were employed by CCTEG Xi'an Research Institute (Group) Co., Ltd.
Generative AI statement
The author(s) declare that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Anani, A., Adewuyi, S., Risso, N., and Nyaaba, W. (2024). Advancements in machine learning techniques for coal and gas outburst prediction in underground mines. Int. J. Coal Geol. 285:104471. doi: 10.1016/j.coal.2024.104471
Fan, C., Lai, X., Wen, H., and Yang, L. (2023). Coal and gas outburst prediction model based on principal component analysis and improved support vector machine. Geohazard Mech. 1, 319–324. doi: 10.1016/j.ghm.2023.11.003
Fu, H., Shi, H., Xu, Y., and Shao, J. (2022). Research on gas outburst prediction model based on multiple strategy fusion improved snake optimization algorithm with temporal convolutional network. IEEE Access 10, 117973–117984. doi: 10.1109/ACCESS.2022.3220765
Hassan, M., Umar, M., and Bermak, A. (2017). Computationally efficient weighted binary decision codes for gas identification with array of gas sensors. IEEE Sens. J. 17, 487–497. doi: 10.1109/JSEN.2016.2631476
He, X., Chen, W., Nie, B., and Zhang, M. (2010). Classification technique for danger classes of coal and gas outburst in deep coal mines. Saf. Sci. 48, 173–178. doi: 10.1016/j.ssci.2009.07.007
Kumar, J. P., Singh, A., and Singh, A. K. (2024). Explainable BERT-LSTM Stacking for sentiment analysis of COVID-19 vaccination. IEEE Transac. Comput. Soc. Syst. 12, 1296–1306 doi: 10.1109/TCSS.2023.3329664
Li, Y., Sun, C., and Li, Q. (2024). Enhanced deep learning method for natural gas pipeline flow prediction based on integrated learning. IEEE Access 12:83822–83829. doi: 10.1109/ACCESS.2024.3406733
Lin, T., Pan, Y., Xue, G., Song, J., and Qi, C. (2020). A novel hybrid spatial-temporal attention-LSTM model for heat load prediction. IEEE Access 8:159182–159195. doi: 10.1109/ACCESS.2020.3017516
Liu, H., Dong, Y., and Wang, F. (2021). Prediction model for gas outburst intensity of coal mining face based on improved PSO and LSSVM,energy engineering. Energy Eng. 118, 679–689. doi: 10.32604/EE.2021.014630
Ou, J., Wang, E., Li, Z., Li, N., Liu, H., and Wang, X (2023) Experimental study of coal gas outburst processes influenced by gas pressure ground stress coal properties. Front. Earth Sci. 11:1303996. doi: 10.3389/feart.2023.1303996
Qiao, W., Huang, K., Azimi, M., and Han, S. (2019). A novel hybrid prediction model for hourly gas consumption in supply side based on improved whale optimization algorithm and relevance vector machine. IEEE Access, 7, 88218–88230. doi: 10.1109/ACCESS.2019.2918156
Rudakov, D., and Sobolev, V. (2019). A mathematical model of gas flow during coal outburst initiation. Int. J. Min. Sci. Technol. 29, 791–796. doi: 10.1016/j.ijmst.2019.02.002
Soleimani, F., Si, G., Roshan, H., and Zhang, J. (2023). Numerical modelling of gas outburst from coal: a review from control parameters to the initiation process. Int. J. Coal Sci. Technol. 10:657. doi: 10.1007/s40789-023-00657-7
Song, S., Li, S., Zhang, T., Ma, L., Zhang, L., and Pan, S. (2021). Research on time series characteristics of the gas drainage evaluation index based on lasso regression. Sci. Rep. 11:20593. doi: 10.1038/s41598-021-00210-z
Utkarsh, and Jain, P. (2024). Predicting bentonite swelling pressure: optimized XGBoost versus neural networks. Sci. Rep. 14:17533. doi: 10.1038/s41598-024-68038-x
Wang, C., Li, X., Xu, C., Niu, Y., Chen, Y., and Yang, S. (2020b). Study on factors influencing and the critical value of the drilling cuttings weight: an index for outburst risk prediction. Process Saf. Environ. Protect. 140, 356–366. doi: 10.1016/j.psep.2020.05.027
Wang, L., Lu, Z., Chen, D., Liu, Q., Chu, P., Shu, L., et al. (2020a). Safe strategy for coal and gas outburst prevention in deep-and-thick coal seams using a soft rock protective layer mining. Saf. Sci. 129:104800. doi: 10.1016/j.ssci.2020.104800
Wang, Z., Xu, J., Ma, J., and Cai, Z. (2023). A novel combined intelligent algorithm prediction model for the risk of the coal and gas outburst. Sci. Rep. 13:15988. doi: 10.1038/s41598-023-43013-0
Wankhade, M., Annavarapu, C. S. R., and Abraham, A. (2023). CBMAFM: CNN-BiLSTM multi-attention fusion mechanism for sentiment classification. Multimed. Tools Appl. 83, 51755–51786. doi: 10.1007/s11042-023-17437-9
Xie, X., Fu, G., Xue, Y., Zhao, Z., Chen, P., Lu, B., et al. (2018). Jiang. Risk prediction and factors risk analysis based on IFOA-GRNN and apriori algorithms: application of artificial intelligence in accident prevention. Process Saf. Environ. Protect. 122, 169–184. doi: 10.1016/j.psep.2018.11.019
Xiong, G., Zhang, J., Fu, X., Chen, J., and Moharmed, A. (2024). Seasonal short-term photovoltaic power prediction based on GSK–BiGRU–XGboost considering correlation of meteorological factors. J. Big Data 11:164. doi: 10.1186/s40537-024-01037-x
Yang, W., Wang, W., Jia, R., Walton, G., Sinha, S., Chen, Q., et al. (2023). Parameter optimization of coal face blasting for coal and gas outburst control. Bull. Eng. Geol. Environ. 82:80. doi: 10.1007/s10064-023-03086-7
Yao, X., Fu, X., and Zong, C. (2022). Short-term load forecasting method based on feature preference strategy and LightGBM-XGboost. IEEE Access 10, 75257–75268. doi: 10.1109/ACCESS.2022.3192011
Zhang, Z., Ye, Y., Luo, B., Chen, G., and Wu, M. (2022). nvestigation of microseismic signal denoising using an improved wavelet adaptive thresholding method. Sci. Rep. 12:22186. doi: 10.1038/s41598-022-26576-2
Zhao, D., Fang, K., and Lian, Z. (2024b). Mechanical and vibrational behaviors of bilayer hexagonal boron nitride in different Stacking modes. Sci. Rep. 14:10619. doi: 10.1038/s41598-024-61486-5
Zhao, R-. X., Shi, J., and Li, X. (2024a). QKSAN: a quantum kernel self-attention network. IEEE Transac. Pattern Anal. Mach. Intell. 46, 10184–10195. doi: 10.1109/TPAMI.2024.3434974
Zheng, X., Lai, W., Zhang, L., and Sheng, X. (2023). Quantitative evaluation of the indexes contribution to coal and gas outburst prediction based on machine learning. Fuel 338:127389. doi: 10.1016/j.fuel.2023.127389
Zhou, B., Yang, S., Wang, C., Cai, J., Xu, Q., and Sang, N. (2019). Experimental study on the influence of coal oxidation on coal and gas outburst during invasion of magmatic rocks into coal seams. Process Saf. Environ. Protect. 123, 213–222. doi: 10.1016/j.psep.2019.02.017
Zhu, J., Zheng, H., Yang, L., Li, S., Sun, L., and Geng, J. (2023). Evaluation of deep coal and gas outburst based on RS-GA-BP. Nat. Hazards 115, 2531–2551. doi: 10.1007/s11069-022-05652-w
Keywords: artificial intelligence, coal and gas outbursts prediction, multi-model fusion, XGBoost, attention mechanism
Citation: Xie Q, Yan J, Dai Z, Du W and Wu X (2025) Study on coal and gas outburst prediction technology based on multi-model fusion. Front. Big Data 8:1623883. doi: 10.3389/fdata.2025.1623883
Received: 06 May 2025; Accepted: 24 September 2025;
Published: 20 October 2025.
Edited by:
Emad Maher Natsheh, An-Najah National University, PalestineReviewed by:
Hongjian Zhu, Yanshan University, ChinaPengfei Shan, Xi'an University of Science and Technology, China
Zhengzheng Cao, Henan Polytechnic University, China
Mohammad Azarafza, University of Tabriz, Iran
Muhammad Shahab, King Saud University, Saudi Arabia
Arifuggaman Arif, Huaiyin Institute of Technology, China
Copyright © 2025 Xie, Yan, Dai, Du and Wu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Qian Xie, eGllLnFpYW4xOTkwQDE2My5jb20=