<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Archiving and Interchange DTD v2.3 20070202//EN" "archivearticle.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="methods-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frai.2023.1283741</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Artificial Intelligence</subject>
<subj-group>
<subject>Methods</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Application of hybrid fuzzy interval-based machine learning models on financial time series &#x02014; A case study of Taiwan biotech index during the epidemic period</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Lin</surname> <given-names>Hsio-Yi</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2394087/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Hsu</surname> <given-names>Bin-Wei</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2346164/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Finance, Chien Hsin University of Science and Technology</institution>, <addr-line>Taoyuan</addr-line>, <country>Taiwan</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Business Administration, Chien Hsin University of Science and Technology</institution>, <addr-line>Taoyuan</addr-line>, <country>Taiwan</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Paolo Giudici, University of Pavia, Italy</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Jinran Wu, Australian Catholic University, Australia</p>
<p>Shahram Rahimi, Mississippi State University, United States</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Bin-Wei Hsu <email>hsudomingo&#x00040;gmail.com</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>08</day>
<month>01</month>
<year>2024</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>6</volume>
<elocation-id>1283741</elocation-id>
<history>
<date date-type="received">
<day>27</day>
<month>08</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>13</day>
<month>12</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2024 Lin and Hsu.</copyright-statement>
<copyright-year>2024</copyright-year>
<copyright-holder>Lin and Hsu</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license></permissions>
<abstract>
<p>In recent years, the use of machine learning to predict stock market indices has emerged as a vital concern in the FinTech domain. However, the inherent nature of point estimation in traditional supervised machine learning models leads to an almost negligible probability of achieving perfect predictions, significantly constraining the applicability of machine learning prediction models. This study employs 4 machine learning models, namely BPN, LSTM, RF, and ELM, to establish predictive models for the Taiwan biotech index during the COVID-19 period. Additionally, it integrates the Gaussian membership function MF from fuzzy theory to develop 4 hybrid fuzzy interval-based machine learning models, evaluating their predictive accuracy through empirical analysis and comparing them with conventional point estimation models. The empirical data is sourced from the financial time series of the &#x0201C;M1722 Listed Biotechnology and Medical Care Index&#x0201D; compiled by the Taiwan Economic Journal during the outbreak of the COVID-19 pandemic, aiming to understand the effectiveness of machine learning models in the face of significant disruptive factors like the pandemic. The findings demonstrate that despite the influence of COVID-19, machine learning remains effective. LSTM performs the best among the models, both in traditional mode and after fuzzy interval enhancement, followed by the ELM and RF models. The predictive results of these three models reach a certain level of accuracy and all outperform the BPN model. Fuzzy-LSTM effectively predicts at a 68% confidence level, while Fuzzy-ELM and Fuzzy-RF yield better results at a 95% confidence level. Fuzzy-BPN exhibits the lowest predictive accuracy. Overall, the fuzzy interval-based LSTM excels in time series prediction, suggesting its potential application in forecasting time series data in financial markets to enhance the efficacy of investment analysis for investors.</p></abstract>
<kwd-group>
<kwd>hybrid fuzzy interval-based machine learning model</kwd>
<kwd>BPN</kwd>
<kwd>LSTM</kwd>
<kwd>Random Forest</kwd>
<kwd>ELM</kwd>
<kwd>financial time series</kwd>
</kwd-group>
<counts>
<fig-count count="8"/>
<table-count count="7"/>
<equation-count count="24"/>
<ref-count count="61"/>
<page-count count="17"/>
<word-count count="9332"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Artificial Intelligence in Finance</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1 Introduction</title>
<p>Predictability studies of stock market indices have a long history within the field of finance (Bacchetta et al., <xref ref-type="bibr" rid="B2">2009</xref>; Chong et al., <xref ref-type="bibr" rid="B17">2017</xref>). Researchers often analyze historical financial data as time series and establish models such as linear regression methods, Autoregressive (AR), Autoregressive Moving Average (ARMA), and Autoregressive Integrated Moving Average (ARIMA) (Li et al., <xref ref-type="bibr" rid="B35">2015</xref>; Zhang et al., <xref ref-type="bibr" rid="B59">2016</xref>; Kili&#x000E7; and Ugur, <xref ref-type="bibr" rid="B30">2018</xref>) to predict trends in stock market indices (Marsza&#x00142;ek and Burczy&#x00144;ski, <xref ref-type="bibr" rid="B43">2014</xref>). However, due to the non-stationarity, nonlinearity, and extended lag in the time series data of stock market indices, significant patterns are often not precisely captured using conventional methods (Bildirici and Ersin, <xref ref-type="bibr" rid="B8">2014</xref>; Chong et al., <xref ref-type="bibr" rid="B17">2017</xref>; He et al., <xref ref-type="bibr" rid="B23">2023</xref>). Consequently, in recent years, many researchers have sought to employ AI machine learning models to discover more robust predictive models for stock market indices (Sunny et al., <xref ref-type="bibr" rid="B50">2020</xref>; Lim and Zohren, <xref ref-type="bibr" rid="B37">2021</xref>).</p>
<p>Machine learning is an algorithm that uses artificial neural networks as its framework for data representation learning. The primary logic involves transforming data into a structured representation of a multi-layered neural network to facilitate learning. This enables the extraction of complex features from the data, allowing for prediction and interpretation. One advantage is that it eliminates the need for significant human effort in feature engineering and model design (Bengio et al., <xref ref-type="bibr" rid="B7">2013</xref>), and it has been widely applied in recent years for extracting features from time series data. Among various machine learning models, the Back-propagation Neural Network (BPN) introduced by Rumelhart and McClelland (<xref ref-type="bibr" rid="B47">1986</xref>) can be considered one of the most representative and commonly used models. BPN belongs to the supervised learning framework, primarily built upon a multi-layer perceptron structure and utilizing error backpropagation. It is frequently applied in fields such as diagnostics and predictions (Rumelhart and McClelland, <xref ref-type="bibr" rid="B47">1986</xref>). Qu (<xref ref-type="bibr" rid="B46">2003</xref>) employed BPN for infectious disease prediction, demonstrating that its predictive performance outperformed traditional multiple regression models. However, there are still studies pointing out that the effectiveness of using BPN for time series data prediction is limited. The main reason is that the learning objective of the BPN model is to establish a mapping relationship between inputs and outputs, neglecting the mutual influences among outputs. Moreover, BPN predictions are based on a sample comparison approach, rather than truly learning the relationships between time series data. Therefore, when there are mutual influences among time series, the effectiveness of computation becomes constrained (Chen et al., <xref ref-type="bibr" rid="B13">2015</xref>).</p>
<p>Another commonly encountered machine learning model, the Recurrent Neural Network (RNN), is regarded as one of the most potent frameworks for processing temporal sequence data. The principal drawback of RNN lies in its neglect of memory capability, rendering it inadequate for capturing long-range dependencies between nodes when sequences are distantly separated. Moreover, the straightforward architecture of the RNN fails to address challenges such as the vanishing gradient problem, wherein gradients can vanish or explode due to the iterative recurrence of weights, ultimately impeding effective training. Consequently, practical instances wherein superior predictive outcomes are solely achieved via the vanilla RNN model are relatively rare. Literature frequently proposes remedies for the conventional RNN model through the design of enhanced gradient descent algorithms or the incorporation of superior activation functions within neural units. In 1997, Hochreiter and Schmidhuber (<xref ref-type="bibr" rid="B24">1997</xref>) introduced a groundbreaking enhancement to the RNN model known as the Long Short-Term Memory Network (LSTM). LSTM innovatively introduces memory mechanisms to augment long-term dependencies, featuring three essential steps within its neurons: forget, update, and output. This formulation substantially bolsters long-term memory performance. Additionally, LSTM partially mitigates the vanishing gradient issue encountered in RNN. Over the years, LSTM has emerged as one of the most commonly employed RNN variants. In the realm of financial time series forecasting, Di Persio and Honchar (<xref ref-type="bibr" rid="B20">2016</xref>) explored the suitability and effectiveness of LSTM. Selvin et al. (<xref ref-type="bibr" rid="B48">2017</xref>) applied LSTM along with CNN-sliding window methods for stock price prediction. Chen K. et al. (<xref ref-type="bibr" rid="B14">2015</xref>) highlighted the enhanced accuracy of the LSTM model in comparison to other regression models. Liu et al. (<xref ref-type="bibr" rid="B40">2018</xref>) specifically pointed out that LSTM-based feature extraction for time series forecasting attains an accuracy of approximately 72%, indicative of its commendable performance. Nevertheless, there remains room for refinement (Liu et al., <xref ref-type="bibr" rid="B40">2018</xref>). Furthermore, LSTM encounters difficulty when handling sequences with a magnitude of 1,000 or more, and the presence of four fully connected layers (Multilayer Perceptron, MLP) within each LSTM cell can lead to computational inefficiency and time consumption as the LSTM spans longer temporal ranges or deeper network configurations. To optimize the LSTM model, scholars have proposed several enhancement strategies. Di Persio and Honchar (<xref ref-type="bibr" rid="B20">2016</xref>) utilized a hybrid LSTM to enhance the precision of time series predictions. Zhao et al. (<xref ref-type="bibr" rid="B60">2017</xref>) highlighted the superior predictive accuracy of LSTM when incorporating a time-weighted function, outperforming other deep learning models in time series forecasting.</p>
<p>In recent years, the Random Forest (RF) model has also been commonly employed for financial time series forecasting. It is regarded as an ensemble learning technique based on decision tree algorithms. RF employs a Bagging approach to generate multiple decision trees and then combines the predictive outcomes of these trees. The final prediction is determined through a voting mechanism, where the most frequent class is selected. However, in comparison to individual decision tree algorithms, RF exhibits stronger generalization capabilities, can handle a larger number of input variables, and is able to assess the importance of each variable (Pal, <xref ref-type="bibr" rid="B45">2005</xref>). Particularly for datasets with imbalanced classes, RF can reduce errors and is less prone to overfitting issues. Lee et al. (<xref ref-type="bibr" rid="B33">2019</xref>) have reported an accuracy of 54.12% for stock market prediction using RF. In the analysis of stock prices within a single industry, RF demonstrates effectiveness in predicting stock prices that possess inherent randomness, thus overcoming subjective empirical judgments and the interference of emotional factors (Khaidem et al., <xref ref-type="bibr" rid="B29">2016</xref>; Nana and Jiangtao, <xref ref-type="bibr" rid="B44">2018</xref>). In contrast to other machine learning models, Basak et al. (<xref ref-type="bibr" rid="B6">2019</xref>) trained RF and XGBoost using exponential smoothing data. The accuracy of trend prediction for these two classifiers improved with an extended time window. The experimentation suggests that RF holds more advantages than XGBoost in this context. Leveraging technical indicators from the stock market, Khaidem et al. (<xref ref-type="bibr" rid="B29">2016</xref>) employed RF to predict stock trends. Their findings indicate that RF outperforms Support Vector Machines (SVM) and Logistic Regression (LR) in terms of obtaining more effective trend prediction results.</p>
<p>In addition to the aforementioned models, Huang et al. (<xref ref-type="bibr" rid="B26">2006</xref>) have introduced a Single-hidden Layer Feedforward Neural Network (SLFNN) known as Extreme Learning Machine (ELM). ELM has been proven to possess high learning efficiency and strong generalization capabilities, making it widely applicable to problems such as classification, regression, clustering, and feature learning (Cao et al., <xref ref-type="bibr" rid="B11">2016</xref>). The number of neurons and activation function in ELM must be regulated, as the input weights and hidden layer biases are fixed during its application. These characteristics contribute to ELM&#x00027;s reputation for achieving enhanced generalization performance with rapid learning. Cheng et al. (<xref ref-type="bibr" rid="B16">2009</xref>) have elegantly demonstrated ELM&#x00027;s superiority over SVM in predicting petroleum reservoir permeability. Huang et al. (<xref ref-type="bibr" rid="B25">2011</xref>) have successfully implemented ELM for regression and classification tasks across various domains. Over the past decade, ELM has consistently shown its advantages over traditional techniques in the realm of stock market forecasting (Sun et al., <xref ref-type="bibr" rid="B49">2014</xref>; Li et al., <xref ref-type="bibr" rid="B36">2016</xref>). Due to the fact that traditional feedforward neural networks (such as BPN) require manual configuration of a significant number of network training parameters, ELM stands out for its simplicity and ease of use. Unlike BPN, ELM only requires the setting of the network&#x00027;s structure and doesn&#x00027;t necessitate the adjustment of other parameters. The weights from the input layer to the hidden layer are determined in a single random iteration and do not need further tuning during execution. Similarly, the weights from the hidden layer to the output layer are determined by solving a linear system of equations, generally contributing to improved computational speed.</p>
<p>However, from a statistical perspective, if the predicted values of the aforementioned supervised machine learning models are only point estimates with binary outcomes (such as binary classification or single-point prediction), it will lead to the problem of the estimated probability of perfect correctness approaching zero. This is due to the fact that in continuous random variables, single points hold no probability value. Therefore, the point estimation prediction method greatly restricts the usability of machine learning models (Lowe and Zapart, <xref ref-type="bibr" rid="B41">1999</xref>). In contrast to point forecasting, probabilistic forecasting describes the variation of the value by providing outputs in terms of probabilistic density function, confidential intervals of the distribution. It can better describe the uncertainty of values (Gan et al., <xref ref-type="bibr" rid="B22">2017</xref>). In reality, the best predictions should include estimated probability distribution intervals for a future time period to better align with real-world situations. In related studies, Quantile regression is utilized in Liu et al. (<xref ref-type="bibr" rid="B38">2017</xref>) to generate multiple forecasting results. In Yang et al. (<xref ref-type="bibr" rid="B55">2006</xref>), Liu et al. (<xref ref-type="bibr" rid="B38">2017</xref>), and Xie et al. (<xref ref-type="bibr" rid="B54">2017</xref>) simulation of historical-error distribution was implemented to convert point loads into intervals. Zadeh (<xref ref-type="bibr" rid="B57">1965</xref>) introduced Fuzzy Logic in his publication &#x0201C;Information and Control,&#x0201D; aiming to utilize fuzzy phenomena to address the reasoning model of uncertainty in the real world. Fuzzy logic has since been widely applied in artificial intelligence fields such as automatic control, pattern recognition, and decision analysis. Whether fuzzy logic can be applied to predictive models in machine learning is a topic worth exploring. Ballings et al. (<xref ref-type="bibr" rid="B3">2015</xref>), in comparing traditional models and integrated models in the machine learning domain, demonstrated that integrated models perform better than single models in predicting financial data based on time series. This study will build upon the existing BPN, LSTM, RF, and ELM models, which are commonly used machine learning models. Initially, point estimation predictions for stock price indices will be computed. Subsequently, by integrating the Gaussian membership function (MF) of fuzzy theory, interval calculations will be performed to develop a fuzzy interval-based machine learning model. Empirical analysis will further investigate whether these models can achieve more accurate predictions of stock price indices. It is anticipated that the outcomes of this study will enhance the practicality and predictive capabilities of machine learning models in real-world scenarios.</p>
<p>In addition, due to the outbreak of the COVID-19 epidemic in recent years, various countries have implemented city closures and restricted crowd activities, which has had a significant impact on the economy and financial markets. Frequent phenomena such as stock market crashes, plummeting commodity prices, and declining global demand have created greater uncertainty for investors. Therefore, during the epidemic period, whether in the economic field or social field, many linear or machine learning prediction model-related research Generated in large quantities, for example: Wu et al. (<xref ref-type="bibr" rid="B53">2022</xref>) once used a time series prediction model to predict half-hourly electricity demand in Victoria. Zhao et al. (<xref ref-type="bibr" rid="B61">2023</xref>) once constructed a deep learning framework, combining time autocorrelation with Spatially correlated combination, reflecting the impact of neighboring cities and historical data on air quality during COVID-19. Cui et al. (<xref ref-type="bibr" rid="B18">2023</xref>) propose a deep learning framework with a COVID-19 adjustment for electricity demand forecasting. In summary, when the market is faced with the noise and interference of the epidemic, what impact will the machine learning model have on the forecast accuracy of the Taiwan stock market index? It is also one of the topics that this study is interested in exploring.</p>
<p>This study employs the highly representative dataset compiled by the Taiwan Economic Journal (TEJ), specifically the &#x0201C;M1722 Listed Biotechnology and Medical Care Index&#x0201D; (hereinafter referred to as the Taiwan TEJ Biotech Index), as empirical data to represent the performance of Taiwan&#x00027;s listed biotechnology and medical care stock market industry. Furthermore, due to the severe stock market fluctuations caused by the COVID-19 pandemic (Baret et al., <xref ref-type="bibr" rid="B4">2020</xref>; Uddin et al., <xref ref-type="bibr" rid="B52">2021</xref>), which exhibit dynamics different from non-pandemic periods, the empirical period of this study is set to the outbreak of the COVID-19 pandemic (from January 2020 to the end of June 2022). The aim is to understand the extent to which the accuracy of machine learning model predictions is affected when the stock market experiences significant turmoil due to pandemic-related disturbances. In summary, this study has three main objectives: (1) To establish predictive models for the Taiwan biotech index after the COVID-19 outbreak using four machine learning models: BPN, LSTM, RF, and ELM. (2) To integrate fuzzy theory to modify the existing point estimation approach of machine learning models and thus develop a fuzzy interval-based machine learning model, while comparing it with traditional point estimation models. (3) To understand whether machine learning models are suitable for predicting stock indices when the stock market faces significant disturbances and substantial fluctuations (such as during the outbreak of the COVID-19 pandemic).</p></sec>
<sec sec-type="materials and methods" id="s2">
<title>2 Materials and methods</title>
<sec>
<title>2.1 Variables</title>
<p>This study focuses on the &#x0201C;Taiwan TEJ Biotechnology Index&#x0201D; as the research subject, with the closing prices of the Taiwan biotechnology index during the COVID-19 outbreak period as the research object. The primary data source is the TEJ database. The study aims to compare 8 machine learning models: BPN, LSTM, RF, ELM, as well as fuzzy interval-based BPN (fuzzy-BPN), fuzzy interval-based LSTM (fuzzy-LSTM), fuzzy interval-based RF (fuzzy-RF), and fuzzy interval-based ELM (fuzzy-ELM), in terms of their predictive accuracy.</p>
<p>Initially, literatures were gathered to collect various variable data used for index prediction. These variables include Taiwan index data, international index data, futures prices, sentiment indicators, macroeconomic analysis, and 23 other variables. The study then employed factor analysis to identify significant variables affecting the Taiwan biotechnology index during the COVID-19 outbreak period. Subsequently, a model was established using MATLAB to predict the closing prices of the Taiwan TEJ Biotechnology Index, thereby validating the feasibility of the proposed research methodology.</p>
<p>The research framework consists of 9 steps, which are elaborated as follows:</p>
<list list-type="order">
<list-item><p>Data Collection: Involves the collection of various variable indicators, including seven technical indicators, five variables related to Taiwan&#x00027;s biotechnology and healthcare sector&#x00027;s net buying and selling, as well as trading volume, 8 variables related to Taiwan and international index market trends, one sentiment indicator, one futures price index, and one macroeconomic analysis, total 23 variable indicators (see <xref ref-type="table" rid="T1">Table 1</xref>).</p></list-item>
<list-item><p>Data Preprocessing: Due to variations in trading holidays for international stock market, if trading data cannot be obtained due to market closures or other reasons on certain days, the entire set of data will be removed in advance.</p></list-item>
<list-item><p>Removal of Ineffective Variables: Confirmatory Factor Analysis (CFA) is employed to select appropriate variables as input indicators for the research model. Ineffective variables are eliminated to enhance the predictive accuracy of the model.</p></list-item>
<list-item><p>Normalization: The data is subjected to normalization, scaled to a range between 0 and 1.</p></list-item>
<list-item><p>Model Construction: Eight machine learning models are individually established using the TEJ listed biotechnology and healthcare sector index (see <xref ref-type="table" rid="T2">Table 2</xref>).</p>
</list-item>
<list-item><p>Setting Training and Testing Parameters: Divide the data in a 7:3 ratio, setting it as training data and testing data, respectively.</p></list-item>
<list-item><p>Train the model.</p></list-item>
<list-item><p>Validate Predictive Results: If the predicted results are not as expected, repeat steps 5 to 7.</p></list-item>
<list-item><p>Model Comparison: Compare the predictive accuracy of MODEL1 to MODEL8 from step (5), and perform a comparison of predictive capabilities using indicators such as the Mean Absolute Percentage Error (MAPE).</p></list-item>
</list>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Variables in this study.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919497;color:#ffffff">
<th valign="top" align="left"><bold>Category</bold></th>
<th valign="top" align="left"><bold>Variables</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Technical indicators</td>
<td valign="top" align="left">MA_5, MA_20, K_9, D_9, RSI_6, RSI_12, MACD_9</td>
</tr> <tr>
<td valign="top" align="left">Biotech and medical sector indicators</td>
<td valign="top" align="left">Listed biotech and medical sector turnover rate, biotech and medical sector mutual fund net buy/sell, biotech and medical sector proprietary trading net buy/sell, biotech and medical sector foreign investor net buy/sell, biotech and medical sector equity-to-debt ratio.</td>
</tr> <tr>
<td valign="top" align="left">Composite index</td>
<td valign="top" align="left">Taiwan Weighted Index, U.S. Dow Jones Industrial Average, U.S. S and P 500 Index, U.S. Nasdaq Biotechnology Index, Shanghai Composite Index, Hong Kong Hang Seng Index, South Korea Composite Index, Japan Nikkei 225 Index.</td>
</tr> <tr>
<td valign="top" align="left">News-based index</td>
<td valign="top" align="left">VIX Fear Index</td>
</tr> <tr>
<td valign="top" align="left">Futures index</td>
<td valign="top" align="left">CRB Index</td>
</tr> <tr>
<td valign="top" align="left">Macroeconomic analysis</td>
<td valign="top" align="left">Business Cycle Indicators</td>
</tr></tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Machine learning models in this study.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919497;color:#ffffff">
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Model</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">BPN</td>
<td valign="top" align="left">MODEL 1</td>
</tr> <tr>
<td valign="top" align="left">LSTM</td>
<td valign="top" align="left">MODEL 2</td>
</tr> <tr>
<td valign="top" align="left">RF</td>
<td valign="top" align="left">MODEL 3</td>
</tr> <tr>
<td valign="top" align="left">ELM</td>
<td valign="top" align="left">MODEL 4</td>
</tr> <tr>
<td valign="top" align="left">Fuzzy-BPN</td>
<td valign="top" align="left">MODEL 5</td>
</tr> <tr>
<td valign="top" align="left">Fuzzy-LSTM</td>
<td valign="top" align="left">MODEL 6</td>
</tr> <tr>
<td valign="top" align="left">Fuzzy-RF</td>
<td valign="top" align="left">MODEL 7</td>
</tr> <tr>
<td valign="top" align="left">Fuzzy-ELM</td>
<td valign="top" align="left">MODEL 8</td>
</tr></tbody>
</table>
</table-wrap>
</sec>
<sec>
<title>2.2 Data processing</title>
<p>The research data were sampled from the month of the first confirmed COVID-19 case in Taiwan in January 2020 until the end of June 2022, excluding market holidays for Taiwan and international index markets. A total of 513 samples were collected, with 359 samples used for learning and 154 samples used for testing. This study conducted confirmatory factor analysis on 23 variables. Factors with eigenvalues &#x0003E;1 were extracted (Kaiser, <xref ref-type="bibr" rid="B28">1960</xref>), followed by rotation using the maximum variance rotation method. The results indicated that nine variables should be excluded from the initial 23 variables. These variables are: economic policy signals, South Korea composite index, Shanghai composite index, biotech and medical stock-to-asset ratio, CRB index, VIX panic index, US Nasdaq biotechnology NBI index, listed biotech and medical stock turnover rate, and mutual fund net buying and selling. The remaining 14 variables (MA_5, MA_20, K_9, D_9, RSI_6, RSI_12, MACD_9, biotech and medical proprietary net buying and selling, biotech and medical foreign net buying and selling, Taiwan weighted index, US Dow Jones Industrial Average, US S&#x00026;P 500 index, Hong Kong Hang Seng index, Nikkei 225 index) were selected as the final input variable indicators for this study, explaining a total of 80.48% of the variance. After adjusting the aforementioned indicator variables, their suitability was tested with a Kaiser-Meyer-Olkin measure of 0.679, exceeding the recommended threshold of 0.6 (Tabachnick and Fidell, <xref ref-type="bibr" rid="B51">1996</xref>), and a Bartlett&#x00027;s sphericity test approximate chi-square distribution value of 9668.408 with a <italic>p</italic>-value of 0.000 for 91 degrees of freedom. These values suggest that the data of these 14 input variable indicators are suitable for subsequent analysis. To standardize the residuals between data points, the study normalized the variable data. Normalization was performed using the sampled data and the maximum value (X<sub>max</sub>) and minimum value (X<sub>min</sub>) within a range. Depending on whether the initial value of the variable was &#x02265;0 or had negative values, <xref ref-type="disp-formula" rid="E1">equations (1)</xref> and <xref ref-type="disp-formula" rid="E2">(2)</xref> were used separately to obtain the normalized value (X<sub>nom</sub>). This normalized value serves as the input variable data for the deep learning model in this study.</p>
<p>If the initial values of the variables are all &#x02265;0:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>o</mml:mi><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>X</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo class="qopname">min</mml:mo></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo class="qopname">max</mml:mo></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo class="qopname">min</mml:mo></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>If the initial values of the variables have negative values:</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>o</mml:mi><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo class="qopname">max</mml:mo><mml:mo>|</mml:mo><mml:mi>X</mml:mi><mml:mo>|</mml:mo></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>(Denominator is the maximum absolute value of X).</p></sec>
<sec>
<title>2.3 Machine learning model</title>
<sec>
<title>2.3.1 BPN</title>
<p>BPN, introduced by Rumelhart and McClelland (<xref ref-type="bibr" rid="B47">1986</xref>), is a supervised learning feedforward multilayer network architecture, incorporating the concept of hidden layers and bias weights. The network architecture is illustrated in <xref ref-type="fig" rid="F1">Figure 1</xref>. It consists of 3 main layers: the input layer, hidden layer, and output layer, each containing multiple processing units. Units in different layers are interconnected through threshold values and weight values. Input variables are transmitted from the input layer to the hidden layer, computed, and then propagated to the output layer.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>Architecture diagram of BPN (Chen and Lin, <xref ref-type="bibr" rid="B12">2007</xref>).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-06-1283741-g0001.tif"/>
</fig>
<p>After comparing the differences between actual values and output variables, the results are propagated back to the hidden layer. Based on this, the weight values of the connecting links are adjusted. This iterative training process employs the steepest descent method. Whenever a training sample is input, the network adjusts the weights by an amount &#x00394;<italic>w</italic><sub><italic>ij</italic></sub> [expressed using <xref ref-type="disp-formula" rid="E3">equations (3)</xref> and <xref ref-type="disp-formula" rid="E4">(4)</xref>], continuing until the error converges under predetermined conditions.</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mi>&#x003B7;</mml:mi><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>&#x003B7; represents the learning rate, which serves the purpose of controlling the magnitude of each step in the steepest descent method to minimize the error function. E represents the error function.</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:mo>&#x02211;</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p><italic>T</italic><sub><italic>j</italic></sub> represents the target output value of the j-th unit in the output layer.</p>
<p><italic>A</italic><sub><italic>j</italic></sub> represents the inferred output value of the j-th unit in the output layer.</p></sec>
<sec>
<title>2.3.2 LSTM</title>
<p>LSTM is a model derived from recurrent neural networks (RNN) that incorporates memory units. It was introduced by Hochreiter and Schmidhuber (<xref ref-type="bibr" rid="B24">1997</xref>). LSTM primarily employs the Sigmoid activation function and dot product operations to control the switches of three gates (input gate, output gate, forget gate), determining which data can be stored in the memory unit. The input gate mainly controls whether input values flow into the memory unit; the output gate regulates whether data computed through the Tanh activation function should be output; the forget gate&#x00027;s main purpose is to decide whether the stored information from the previous time step should be forgotten or retained in the memory unit. Due to its memory units, LSTM is capable of recording longer information compared to RNN and addresses the issue of poor performance in long-term memory of RNN. Hence, it is more frequently used than RNN. <xref ref-type="fig" rid="F2">Figure 2</xref> shows the architecture of the LSTM model (Liu and Wei, <xref ref-type="bibr" rid="B39">2022</xref>).</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Architecture diagram of LSTM.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-06-1283741-g0002.tif"/>
</fig>
<p>The function of LSTM at time t is given by <xref ref-type="disp-formula" rid="E5">equations (5)</xref> to <xref ref-type="disp-formula" rid="E11">(11)</xref>.</p>
<disp-formula id="E5"><label>(5)</label><mml:math id="M5"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E6"><label>(6)</label><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E7"><label>(7)</label><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>o</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>o</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>o</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E8"><label>(8)</label><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E9"><label>(9)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x0007E;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>h</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E10"><label>(10)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:msub><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x0007E;</mml:mo></mml:mover></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E11"><label>(11)</label><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>o</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>h</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p><italic>x</italic><sub><italic>t</italic></sub> represents the input data at time t</p>
<p><italic>h</italic><sub><italic>t</italic></sub> represents the output of the hidden layer at time t</p>
<p><italic>w</italic><sub><italic>f</italic></sub>, <italic>w</italic><sub><italic>o</italic></sub>, <italic>w</italic><sub><italic>c</italic></sub>, <italic>w</italic><sub><italic>i</italic></sub>, <italic>U</italic><sub><italic>f</italic></sub>, <italic>U</italic><sub><italic>o</italic></sub>, <italic>U</italic><sub><italic>c</italic></sub>, <italic>U</italic><sub><italic>i</italic></sub> represents the weight function</p>
<p><italic>b</italic><sub><italic>f</italic></sub>, <italic>b</italic><sub><italic>o</italic></sub>, <italic>b</italic><sub><italic>c</italic></sub>, <italic>b</italic><sub><italic>t</italic></sub> are bias parameters</p>
<p>&#x003C3;<sub><italic>g</italic></sub> is sigmoid function, &#x003C3;<sub><italic>h</italic></sub> is tanh function</p>
<p><italic>f</italic><sub><italic>t</italic></sub>, <italic>i</italic><sub><italic>t</italic></sub>, <italic>o</italic><sub><italic>t</italic></sub> are the forget, input, and output gates, respectively.</p></sec>
<sec>
<title>2.3.3 ELM</title>
<p>ELM is a feedforward neural network devised by Professor Guang-Bin Huang from Nanyang Technological University in Singapore. Unlike conventional artificial neural networks such as BPN that necessitate the configuration of numerous network training parameters, ELM solely requires the specification of the network&#x00027;s structure, omitting the necessity for additional parameters. Hence, it has gained renown for its straightforwardness and user-friendliness (Cao et al., <xref ref-type="bibr" rid="B11">2016</xref>). In this investigation, we embrace the structure of a single-layer feedforward neural network (SLFN) for ELM. This structure comprises an input layer, a hidden layer, and an output layer. The output function FL of the hidden layer is delineated as <xref ref-type="disp-formula" rid="E12">equation (12)</xref>:</p>
<disp-formula id="E12"><label>(12)</label><mml:math id="M12"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>L</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>h</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Within the equation, the symbol <italic>x</italic> signifies the input variable, while l denotes the number of nodes in the hidden layer. &#x003B2; corresponds to the output weight, and <italic>h</italic>(<italic>x</italic>) embodies the activation function responsible for transforming data from the input layer into the feature space of ELM. This expression is depicted as <xref ref-type="disp-formula" rid="E13">equation (13)</xref>:</p>
<disp-formula id="E13"><label>(13)</label><mml:math id="M13"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>h</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>In the provided equation, the variables <italic>a</italic><sub><italic>i</italic></sub> and <italic>b</italic><sub><italic>i</italic></sub> represent feature mapping parameters, often referred to as node parameters. Specifically, <italic>a</italic><sub><italic>i</italic></sub> denotes the input weight or input weights in this context. This investigation utilizes the widely used Sigmoid function, as depicted in <xref ref-type="disp-formula" rid="E14">equation (14)</xref>:</p>
<disp-formula id="E14"><label>(14)</label><mml:math id="M14"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>G</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x0002B;</mml:mo><mml:mo class="qopname">exp</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>a</mml:mi><mml:mo>&#x000B7;</mml:mo><mml:mi>x</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mi>b</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The goal of training a neural network with a single hidden layer revolves around the minimization of output errors. By undergoing the process of learning and training, we can derive the values of &#x003B2; that lead to the achievement of minimal and distinct error.</p></sec>
<sec>
<title>2.3.4 RF</title>
<p>Breiman (<xref ref-type="bibr" rid="B10">2001</xref>) introduced RF in 2001. RF operates based on the concept of ensemble learning, where it amalgamates several decision trees to create a more resilient learning model. This, in turn, addresses the challenge of overfitting, leading to enhanced predictive accuracy within the domain of machine learning.</p>
<p>Breiman&#x00027;s definition of RF, delineated in <xref ref-type="disp-formula" rid="E15">equation (15)</xref>, depicts an assembly of tree-like structures that collectively shape a classifier:</p>
<disp-formula id="E15"><label>(15)</label><mml:math id="M15"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mi>h</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mtext>&#x000A0;&#x000A0;</mml:mtext><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x02026;</mml:mo></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>In this definition, {k} represents an array of independently and identically distributed random vectors. This conglomerate of classifiers converges through their amalgamation, as illustrated in <xref ref-type="disp-formula" rid="E16">equation (16)</xref>:</p>
<disp-formula id="E16"><label>(16)</label><mml:math id="M16"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x02026;</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>By creating the training set in a random manner from the probability distributions of random vectors X and Y, the margin function is established as outlined in <xref ref-type="disp-formula" rid="E17">equation (17)</xref>:</p>
<disp-formula id="E17"><label>(17)</label><mml:math id="M17"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>m</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>=</mml:mo><mml:mi>&#x003B1;</mml:mi><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x02260;</mml:mo><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>a</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mi>&#x003B1;</mml:mi><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Here, <italic>I</italic> represents the indicator function utilized for the accurate classification of X and Y. The magnitude of the margin function directly corresponds to an elevated correct classification score. The generalization error is precisely defined as depicted in <xref ref-type="disp-formula" rid="E19">equation (18)</xref>:</p>
<disp-formula id="E19"><label>(18)</label><mml:math id="M19"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:msup><mml:mrow><mml:mi>E</mml:mi></mml:mrow><mml:mrow><mml:mo>*</mml:mo></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0003C;</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>In this context, X and Y stand as representations of probabilities. The effectiveness of the RF model is commonly evaluated based on the subsequent considerations:</p>
<list list-type="order">
<list-item><p>A more robust growth trajectory for each tree corresponds to an enhanced overall performance of the forest.</p></list-item>
<list-item><p>Improved independence and reduced correlation among individual trees within the forest lead to superior classification performance.</p></list-item>
<list-item><p>The quantity of decision trees stands as the sole parameter for RF execution and serves as the pivotal determinant for achieving the RF model with the minimum error.</p></list-item>
</list></sec></sec>
<sec>
<title>2.4 Hybrid fuzzy interval-based machine learning model</title>
<p>Supervised machine learning models and deep learning models share a common characteristic: the predicted values of the output variable are point estimates. Despite the advantages of highly predictive interpretability and low-error precision in machine learning models, the drawback of single-point probabilistic estimation still exists. To address this, this study attempts to propose a fuzzy membership function to enhance and intervalize machine learning models, aiming to mitigate the shortcomings of point estimation while retaining the ability of machine learning models to handle dynamic and complex data.</p>
<p>The traditional approach to modeling financial time-series data heavily relied on normal distributions until 1963 when Mandelbrot (<xref ref-type="bibr" rid="B42">1963</xref>) challenged this norm. He noticed leptokurtosis in the empirical distributions of price changes and suggested using symmetric stable distributions to account for this excess kurtosis. Subsequent developments by researchers such as Ali and Giaccotto (<xref ref-type="bibr" rid="B1">1982</xref>), Kon (<xref ref-type="bibr" rid="B31">1984</xref>), Bookstaber and McDonald (<xref ref-type="bibr" rid="B9">1987</xref>), and Barinath and Chaterjee (<xref ref-type="bibr" rid="B5">1988</xref>) advanced the use of various non-normal distributions for modeling financial data. Despite these advancements in characterizing financial data with non-normal distributions, there remains a lack of techniques to fully explain their distribution.</p>
<p>Addressing this gap, this paper introduces a fuzzy-interval architecture to enhance machine learning models, referred to as Fuzzy machine learning models. These models utilize fuzzy sets, defined by a membership function (MF), to overcome the limitations of single-point predictions inherent in traditional machine learning models. Specifically, the Gaussian MF, a common assumption in normal distribution, is adopted, characterized by two parameters: the center {c, &#x003C3;} <xref ref-type="disp-formula" rid="E20">equation (19)</xref>:</p>
<disp-formula id="E20"><label>(19)</label><mml:math id="M20"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>;</mml:mo><mml:mi>c</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x003C3;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:mn>2</mml:mn><mml:mtext>&#x000A0;</mml:mtext><mml:mi>&#x003C0;</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:mi>&#x003C3;</mml:mi></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:msup><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>x</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:mo>-</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003C3;</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where is the Gaussian MF&#x00027;s center and &#x003C3; determines MF&#x00027;s width. In this paper, indicates the mean of n-days returns or indices, &#x003C3; denotes the standard deviation of n-day returns or indices, and the MF of fuzzy-interval is also decided completely by and &#x003C3;. The Gaussian MF in this approach is essentially an extension of the normal distribution, a fundamental concept in probability theory. Central to our methodology is the placement of the fuzzy-interval MF around a central point <italic>c</italic>, allowing for a range of variance that includes 1.68&#x003C3; (representing a 95% probability) and 1.96&#x003C3; (accounting for a 99% probability). This strategic inclusion of 1.68&#x003C3; and 1.96&#x003C3; within the interval significantly enhances traditional machine learning models by addressing their inherent limitation of relying solely on single-point predictions. By adopting this approach, the model not only adheres to the principles of Gaussian distribution but also substantially improves prediction accuracy by accommodating a wider range of outcomes, thereby rectifying the shortcoming of single-point forecasting prevalent in conventional machine learning models.</p>
<p><xref ref-type="fig" rid="F3">Figure 3</xref> illustrates the Gaussian MF of the fuzzy-interval approach. Building on this foundation, the paper endeavors to identify the parameters and &#x003C3; using machine learning models. <xref ref-type="fig" rid="F4">Figure 4</xref> will demonstrate the framework for generating the fuzzy-interval MF, allowing for the retention of the nonlinear characteristics of machine learning models while simultaneously enhancing them by addressing their single-point prediction constraints. This novel framework is termed Fuzzy machine learning models in our study.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Gaussian MF of the fuzzy-interval approach.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-06-1283741-g0003.tif"/>
</fig>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Conceptual diagram of Gaussian membership functions (MFs).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-06-1283741-g0004.tif"/>
</fig>
<p>In the context of this study, <italic>c</italic> represents the center of the Gaussian MFs, while &#x003C3; governs the width of these MFs. Within the scope of this paper, <italic>c</italic> denotes the average value of the weekly Taiwan Biotechnology Index, while &#x003C3; signifies the standard deviation of the same index on a weekly basis. Moreover, the characteristics of the fuzzy-interval MF are entirely determined by the values of <italic>c</italic> and &#x003C3;. It&#x00027;s worth noting that the Gaussian Membership Function is a straightforward extension of the normal distribution employed in probability theory. In the case of the fuzzy-interval MF, its center is aligned with <italic>c</italic>, and its spread around <italic>c</italic> is defined by adding and subtracting 1.68 or 1.96 times the value of &#x003C3;, representing the corresponding confidence intervals.</p>
<p>The fuzzy-machine learning models proposed in this study utilize the characteristics of financial time series data, taking the dynamic N-day average and standard deviation to determine the center and width of the interval. In application, it does not require extensive mathematical derivations and computations to complete the dynamic estimation interval.</p></sec>
<sec>
<title>2.5 Evaluation index</title>
<p>The evaluation indices used employed in this research for gauging the effectiveness of the trained models consist of RMSE (Root Mean Square Error), MAPE (Mean Absolute Percentage Error), and MAE (Mean Absolute Error). The formulas are presented as follows <xref ref-type="disp-formula" rid="E21">equations (20</xref>&#x02013;<xref ref-type="disp-formula" rid="E23">22</xref>):</p>
<disp-formula id="E21"><label>(20)</label><mml:math id="M21"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>R</mml:mi><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:msqrt><mml:mrow><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:mfrac><mml:mo>&#x000D7;</mml:mo><mml:msubsup><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msubsup><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msqrt></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E22"><label>(21)</label><mml:math id="M22"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>M</mml:mi><mml:mi>A</mml:mi><mml:mi>P</mml:mi><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>100</mml:mn><mml:mi>%</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:mfrac><mml:msubsup><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msubsup><mml:mo>|</mml:mo><mml:mfrac><mml:mrow><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>|</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E23"><label>(22)</label><mml:math id="M23"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>M</mml:mi><mml:mi>A</mml:mi><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo>&#x02211;</mml:mo><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>-</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p><italic>Y</italic><sub><italic>i</italic></sub> : Actual value</p>
<p>&#x00176;<sub><italic>i</italic></sub>: Predicted output value from the network</p>
<p>n: Number of test examples</p>
<p>Among the aforementioned indicators, RMSE is a statistical metric that measures the difference between predicted and actual values, and commonly used to assess model accuracy. It calculates the mean of the squared prediction errors and then takes the square root, providing a measure in the same units as the original data. A smaller RMSE indicates higher accuracy of the predictive model. However, as it is influenced by the data range, it&#x00027;s suitable for comparing predictive errors of specific variables among different models. MAPE is a relative measure that determines the degree of difference between estimated and actual values, independent of unit influences. It calculates the absolute percentage error for each predicted value and then takes the average of these errors. A lower MAPE indicates higher accuracy in the predictive model. Generally, a MAPE% value below 10 is considered highly accurate, between 10 and 20 signifies good accuracy, between 20 and 50 suggests reasonable accuracy, and values exceeding 50 are deemed inaccurate (Lewis, <xref ref-type="bibr" rid="B34">1982</xref>). MAE, a metric used to assess the error of a predictive model, representing the sum of absolute differences between target and predicted values, measures the average length of prediction errors without considering their direction. It ranges from 0 to positive infinity. A lower MAE indicates smaller errors in the predictive model, meaning less average difference between predicted and actual values.</p>
<p>Furthermore, when evaluating interval-based ML, this study also employs the Accuracy (ACC) metric to assess model prediction performance. ACC gauges the accurate prediction ratio of the model and is calculated as shown in <xref ref-type="disp-formula" rid="E24">equation (23)</xref>.</p>
<disp-formula id="E24"><label>(23)</label><mml:math id="M24"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>A</mml:mi><mml:mi>C</mml:mi><mml:mi>C</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>t: Number of the actual values that fall within the predicted CIs.</p>
<p>n: Total number of test examples.</p></sec></sec>
<sec sec-type="results" id="s3">
<title>3 Results</title>
<p>In this study, a total of 359 training examples and 154 testing examples were used for the eight models, resulting in a total of 513 examples. The parameter settings and empirical performance of the BPN, LSTM, RF, and ELM models are detailed as follows.</p>
<sec>
<title>3.1 Empirical analysis of BPN</title>
<p>In this study, the BPN model was implemented using Matlab 2021 software. Regarding the determination of BPN model parameters, Zhang et al. (<xref ref-type="bibr" rid="B58">1998</xref>) indicated that the most commonly used number of neural network layers is 1 or 2, with usually 1 hidden layer achieving highly effective prediction performance. Yoon et al. (<xref ref-type="bibr" rid="B56">1993</xref>) found through empirical research that a 2-layer hidden layer configuration provides better predictive capabilities for time series. Therefore, this study will test parameters with 1 to 2 hidden layers.</p>
<p>Furthermore, for determining the number of nodes in the hidden layers, Davies (<xref ref-type="bibr" rid="B19">1994</xref>) stated that the suitable number of nodes for each hidden layer can only be found through a trial-and-error approach. Lawrence and Petterson (<xref ref-type="bibr" rid="B32">1991</xref>), on the other hand, recommended that the number of nodes in each hidden layer should be tested based on 50 to 75% of the sum of input and output variables. Therefore, this study intends to test a range of approximately 7 to 12 nodes, considering the total of 15 nodes (14 input variables and 1 output variable) as 50 to 75% of the range, using a trial-and-error method. As for the learning rate, Freeman and Skapura (<xref ref-type="bibr" rid="B21">1992</xref>) explained that the learning rate of an artificial neural network should be &#x0003C;1 to achieve optimal learning state and convergence. Therefore, this study plans to test a learning rate of 0.5 and set the training cycles to 1,000 or terminate the learning process if the RMSE has converged for the BPN model. Regarding the empirical execution of training and testing the BPN model in this study, the evaluation was conducted using RMSE, MAE, and MAPE metrics, and the results are summarized in the <xref ref-type="table" rid="T3">Table 3</xref>.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Empirical Performance of BPN.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919497;color:#ffffff">
<th valign="top" align="left"><bold>Hidden layer parameters</bold></th>
<th valign="top" align="center"><bold>Data set</bold></th>
<th valign="top" align="center"><bold>RMSE</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>MAPE</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" rowspan="2">10</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">1.6292</td>
<td valign="top" align="center">1.1643</td>
<td valign="top" align="center">1.3060%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.1983</td>
<td valign="top" align="center">0.8613</td>
<td valign="top" align="center">1.2709%</td>
</tr> <tr style="color:#ff0000">
<td valign="top" align="left" rowspan="2">12</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">1.4818</td>
<td valign="top" align="center">1.0725</td>
<td valign="top" align="center">1.3371%</td>
</tr>
 <tr style="color:#ff0000">
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.1131</td>
<td valign="top" align="center">0.8303</td>
<td valign="top" align="center">1.2252%</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">7<sup>&#x0002A;</sup>10</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">1.7902</td>
<td valign="top" align="center">1.4344</td>
<td valign="top" align="center">1.4509%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.4707</td>
<td valign="top" align="center">1.2094</td>
<td valign="top" align="center">1.7845%</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">7<sup>&#x0002A;</sup>12</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">1.7450</td>
<td valign="top" align="center">1.2701</td>
<td valign="top" align="center">1.3992%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.3223</td>
<td valign="top" align="center">1.0784</td>
<td valign="top" align="center">1.5912%</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">10<sup>&#x0002A;</sup>10</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">1.6006</td>
<td valign="top" align="center">1.1983</td>
<td valign="top" align="center">1.3767%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.2500</td>
<td valign="top" align="center">0.9607</td>
<td valign="top" align="center">1.4176%</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">10<sup>&#x0002A;</sup>12</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">1.6787</td>
<td valign="top" align="center">1.2231</td>
<td valign="top" align="center">1.3844%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.3392</td>
<td valign="top" align="center">1.0107</td>
<td valign="top" align="center">1.4914%</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>Red and bold text indicate the best performance of BPN model.</p>
</table-wrap-foot>
</table-wrap>
<p>After experimentation through trial and error, it was found that the optimal model configuration is with a single hidden layer and 12 nodes.</p></sec>
<sec>
<title>3.2 Empirical analysis of LSTM</title>
<p>The empirical implementation of the LSTM model in this study was also conducted using Matlab2021 software. Currently, there is no definitive standard for setting the parameters of the LSTM model, and adjustments are often determined through a trial-and-error approach (Chen et al., <xref ref-type="bibr" rid="B15">2019</xref>). In this study, the number of hidden layers was tested between 2 and 3, and the number of nodes in each hidden layer was adjusted from 60 to 256. The range for the number of iterations was set between 100 and 1,000, and the dropout rate for hidden layer weights ranged from 0.2 to 0.5. The learning rate was tested within the range of 0.005 to 0.01, and the learning rate decay factor was set to 0.02. These parameter ranges were consolidated from various literature sources. The parameter values tested through the trial-and-error method are summarized in the following <xref ref-type="table" rid="T4">Table 4</xref>.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Trial-and-error Parameter List for the LSTM Model.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919497;color:#ffffff">
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Hidden layer</bold></th>
<th valign="top" align="center"><bold>Weight loss</bold></th>
<th valign="top" align="center"><bold>Epoch time</bold></th>
<th valign="top" align="center"><bold>Learning rate</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">LSTM 1</td>
<td valign="top" align="center">60<sup>&#x0002A;</sup>180<sup>&#x0002A;</sup>60</td>
<td valign="top" align="center">0.2/0.3/0.2</td>
<td valign="top" align="center">300</td>
<td valign="top" align="center">0.005</td>
</tr> <tr>
<td valign="top" align="left">LSTM 2</td>
<td valign="top" align="center">128<sup>&#x0002A;</sup>256<sup>&#x0002A;</sup>64</td>
<td valign="top" align="center">0.3/0.3/0.3</td>
<td valign="top" align="center">300</td>
<td valign="top" align="center">0.005</td>
</tr> <tr>
<td valign="top" align="left">LSTM 3</td>
<td valign="top" align="center">256<sup>&#x0002A;</sup>256<sup>&#x0002A;</sup>64</td>
<td valign="top" align="center">0.5/0.5/0.5</td>
<td valign="top" align="center">1,000</td>
<td valign="top" align="center">0.01</td>
</tr> <tr>
<td valign="top" align="left">LSTM 4</td>
<td valign="top" align="center">128<sup>&#x0002A;</sup>128</td>
<td valign="top" align="center">0.2/0.2</td>
<td valign="top" align="center">300</td>
<td valign="top" align="center">0.005</td>
</tr> <tr>
<td valign="top" align="left">LSTM 5</td>
<td valign="top" align="center">128<sup>&#x0002A;</sup>256</td>
<td valign="top" align="center">0.3/0.3</td>
<td valign="top" align="center">1,000</td>
<td valign="top" align="center">0.005</td>
</tr> <tr>
<td valign="top" align="left">LSTM 6</td>
<td valign="top" align="center">180<sup>&#x0002A;</sup>180</td>
<td valign="top" align="center">0.5/0.5</td>
<td valign="top" align="center">100</td>
<td valign="top" align="center">0.01</td>
</tr></tbody>
</table>
</table-wrap>
<p>Through implementation in Matlab2021, the results for the LSTM model are presented in the <xref ref-type="table" rid="T5">Table 5</xref>.</p>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>Performance of LSTM model validation.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919497;color:#ffffff">
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Data set</bold></th>
<th valign="top" align="center"><bold>RMSE</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>MAPE</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" rowspan="2">LSTM 1</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">0.6889</td>
<td valign="top" align="center">0.5315</td>
<td valign="top" align="center">0.8023%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.5719</td>
<td valign="top" align="center">1.2889</td>
<td valign="top" align="center">1.9601%</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">LSTM 2</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">1.9659</td>
<td valign="top" align="center">1.6023</td>
<td valign="top" align="center">2.4702%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">2.3904</td>
<td valign="top" align="center">1.8512</td>
<td valign="top" align="center">2.9331%</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">LSTM 3</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">4.5501</td>
<td valign="top" align="center">3.2104</td>
<td valign="top" align="center">4.9538%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">4.8382</td>
<td valign="top" align="center">3.4790</td>
<td valign="top" align="center">5.5118%</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">LSTM 4</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">0.6200</td>
<td valign="top" align="center">0.4804</td>
<td valign="top" align="center">0.7268%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">0.7646</td>
<td valign="top" align="center">0.5811</td>
<td valign="top" align="center">0.8989%</td>
</tr> 
<tr>
<td valign="top" align="left" rowspan="2" style="color:#ff0000">LSTM 5</td>
<td valign="top" align="center" style="color:#ff0000">Training</td>
<td valign="top" align="center" style="color:#ff0000">0.4009</td>
<td valign="top" align="center" style="color:#ff0000">0.3088</td>
<td valign="top" align="center" style="color:#ff0000">0.4637%</td>
</tr>
 <tr>
<td valign="top" align="center" style="color:#ff0000">Testing</td>
<td valign="top" align="center" style="color:#ff0000">0.6486</td>
<td valign="top" align="center" style="color:#ff0000">0.4880</td>
<td valign="top" align="center" style="color:#ff0000">0.7603%</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">LSTM 6</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">1.2653</td>
<td valign="top" align="center">0.9750</td>
<td valign="top" align="center">1.5053%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.5645</td>
<td valign="top" align="center">1.1268</td>
<td valign="top" align="center">1.7844%</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>Red and bold text indicate the best performance among all the models.</p>
</table-wrap-foot>
</table-wrap>
<p>From the <xref ref-type="table" rid="T5">Table 5</xref>, it can be observed that based on the performance evaluation using RMSE, MAPE, and MAE values, the optimal parameter configuration for the LSTM model is the 5th set (LSTM 5) model.</p></sec>
<sec>
<title>3.3 Empirical analysis of ELM and RF</title>
<p>The ELM model, as a single hidden layer neural network, offers advantages such as not requiring the setup of numerous parameters and having strong learning capabilities, compared to the traditional BPN. In this study, the parameter settings for the ELM model are as follows: a single hidden layer with 30 nodes, determined through trial and error to achieve good convergence. The activation function used is the commonly used Sigmoid function, and the remaining parameters are set to their default values. For the RF classification model in this study, the number of decision trees is set to 20 using the TreeBagger function. RF is specified to operate in regression mode. The feature selection method is set to &#x0201C;curvature,&#x0201D; which selects split points based on the curvature of features. The other parameters are set to their default values in the program.</p>
<p>Based on the aforementioned execution results, the RMSE, MAE, and MAPE values of the empirical models in this study have all converged to reasonable standards. Among the 4 machine learning tools, the LSTM model exhibits the best convergence state. The differences in error values between ELM and RF are not significant, and the RMSE, MAE, and MAPE values of these 3 models are superior to those of the BPN model.</p></sec>
<sec>
<title>3.4 Empirical summary: fuzzy-BPN, fuzzy-LSTM, fuzzy-ELM, and fuzzy-RF</title>
<p>In summary, to address the limitation of traditional &#x0201C;single-point without probability&#x0201D; point estimation, this study proposes the incorporation of Fuzzy Gaussian Membership Function for interval estimation improvement, aiming to enhance the accuracy of predicting actual values. Under the assumption that confidence intervals are extended by 1 standard deviation for 68%, 1.96 standard deviations for 95%, and 2.58 standard deviations for 99%, the predictive results are presented in the <bold>Table 7</bold>.</p>
<p>This study compares the actual values with the intervals formed based on the predictions and standard deviation multiples of the 4 models to determine if the actual values fall within the predicted intervals. <xref ref-type="fig" rid="F5">Figures 5</xref>&#x02013;<xref ref-type="fig" rid="F8">8</xref> illustrate the predictive values, upper and lower bounds, as well as the actual qualitative chart (for the first 50 data points) of the four fuzzy interval-based machine learning models during both the training set and testing set phases. Thus, <xref ref-type="table" rid="T6">Tables 6</xref>, <xref ref-type="table" rid="T7">7</xref>, it is evident that the LSTM model exhibits the best convergence state in both training and testing examples among all machine learning models in this study. It achieves the smallest errors in terms of RMSE, MAE, and MAPE, followed by the RF, ELM, and BPN models. Additionally, as shown in <xref ref-type="table" rid="T7">Table 7</xref>, compared to other models, the Fuzzy-LSTM, in terms of &#x000B1;1&#x003C3;, &#x000B1;1.68&#x003C3;, and &#x000B1;1.96&#x003C3;, interval levels, maintains an effective predictive accuracy of over 97%. In contrast, the Fuzzy-BPN shows the least ideal prediction results, with a maximum accuracy of only 70.56%. The empirical results demonstrate the superior performance of the Fuzzy-LSTM model. At the 99% confidence level of the prediction interval, all models except Fuzzy-BPN achieve an accuracy of at least 85%, including Fuzzy-LSTM, Fuzzy-ELM, and Fuzzy-RF. Notably, the Fuzzy-LSTM model even achieves a 100% accuracy rate in predicting the actual values of the test dataset. It is clear that as the interval size and coverage increase, the predictive accuracy also improves. The second-best performing model is Fuzzy-ELM, which achieves an accuracy rate of nearly 76% at the 68% confidence interval level. In summary, the Hybrid Fuzzy interval-based machine learning models (LSTM, ELM, RF) indeed are capable of effectively capturing the time-series characteristics of stock price data in financial market time series and accurately predicting their values.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Fuzzy-BPN Interval chart (for the first 50 data). <bold>(A)</bold> Training data; <bold>(B)</bold> Testing data.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-06-1283741-g0005.tif"/>
</fig>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Fuzzy-LSTM Interval chart (for the first 50 data). <bold>(A)</bold> Training data; <bold>(B)</bold> Testing data.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-06-1283741-g0006.tif"/>
</fig>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Fuzzy-ELM Interval chart (for the first 50 data). <bold>(A)</bold> Training data; <bold>(B)</bold> Testing data.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-06-1283741-g0007.tif"/>
</fig>
<fig id="F8" position="float">
<label>Figure 8</label>
<caption><p>Fuzzy-RF Interval chart (for the first 50 data). <bold>(A)</bold> Training data; <bold>(B)</bold> Testing data.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-06-1283741-g0008.tif"/>
</fig>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>Performance of ELM, RF, BPN, and LSTM models.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919497;color:#ffffff">
<th valign="top" align="left"><bold>Evaluation index</bold></th>
<th valign="top" align="center"><bold>Data set</bold></th>
<th valign="top" align="center"><bold>ELM</bold></th>
<th valign="top" align="center"><bold>RF</bold></th>
<th valign="top" align="center"><bold>BPN<sup>&#x0002A;</sup></bold></th>
<th valign="top" align="center"><bold>LSTM<sup>&#x0002A;</sup></bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" rowspan="2">RMSE</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">0.6007</td>
<td valign="top" align="center">0.6110</td>
<td valign="top" align="center">1.488</td>
<td valign="top" align="center">0.4009</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.0014</td>
<td valign="top" align="center">1.0605</td>
<td valign="top" align="center">1.1131</td>
<td valign="top" align="center">0.6486</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">MAE</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">0.4623</td>
<td valign="top" align="center">0.4180</td>
<td valign="top" align="center">1.0725</td>
<td valign="top" align="center">0.3088</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">0.8146</td>
<td valign="top" align="center">0.8782</td>
<td valign="top" align="center">0.8303</td>
<td valign="top" align="center">0.4880</td>
</tr> <tr>
<td valign="top" align="left" rowspan="2">MAPE</td>
<td valign="top" align="center">Training</td>
<td valign="top" align="center">0.7017%</td>
<td valign="top" align="center">0.6432%</td>
<td valign="top" align="center">1.3371%</td>
<td valign="top" align="center">0.4637%</td>
</tr>
 <tr>
<td valign="top" align="center">Testing</td>
<td valign="top" align="center">1.2517%</td>
<td valign="top" align="center">1.3433%</td>
<td valign="top" align="center">1.2252%</td>
<td valign="top" align="center">0.7603%</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p><sup>&#x0002A;</sup>Means that the performance of the model is based on the optimal parameter combination identified through trial and error.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T7">
<label>Table 7</label>
<caption><p>Performance of hybrid fuzzy interval-based machine learning model.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919497;color:#ffffff">
<th valign="top" align="left" rowspan="2"><inline-graphic xlink:href="frai-06-1283741-i0001.tif"/></th>
<th valign="top" align="center" colspan="2"><bold>Fuzzy-BPN</bold></th>
<th valign="top" align="center" colspan="2"><bold>Fuzzy-LSTM</bold></th>
<th valign="top" align="center" colspan="2"><bold>Fuzzy-ELM</bold></th>
<th valign="top" align="center" colspan="2"><bold>Fuzzy-RF</bold></th>
</tr>
</thead>
<tbody>
<tr style="background-color:#919497;color:#ffffff">
<td valign="top" align="center"><bold>Training</bold></td>
<td valign="top" align="center"><bold>Testing</bold></td>
<td valign="top" align="center"><bold>Training</bold></td>
<td valign="top" align="center"><bold>Testing</bold></td>
<td valign="top" align="center"><bold>Training</bold></td>
<td valign="top" align="center"><bold>Testing</bold></td>
<td valign="top" align="center"><bold>Training</bold></td>
<td valign="top" align="center"><bold>Testing</bold></td>
</tr> <tr>
<td valign="top" align="left">68% (&#x000B1; 1&#x003C3;)</td>
<td valign="top" align="center">41.67%</td>
<td valign="top" align="center">39.22%</td>
<td valign="top" align="center"><bold>97.18%</bold><sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center"><bold>97.00%</bold><sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">82.78%<sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">76.47%</td>
<td valign="top" align="center">71.94%</td>
<td valign="top" align="center">50.33%</td>
</tr> <tr>
<td valign="top" align="left">95% (&#x000B1; 1.68&#x003C3;)</td>
<td valign="top" align="center">56.11%</td>
<td valign="top" align="center">46.41%</td>
<td valign="top" align="center"><bold>99.44%</bold><sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center"><bold>99.33%</bold><sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">95.83%<sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">94.12%<sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">91.67%<sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">80.39%<sup><bold>&#x0002A;</bold></sup></td>
</tr> <tr>
<td valign="top" align="left">99% (&#x000B1; 1.96&#x003C3;)</td>
<td valign="top" align="center">70.56%</td>
<td valign="top" align="center">63.40%</td>
<td valign="top" align="center"><bold>99.27%</bold><sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center"><bold>100%</bold><sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">97.78%<sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">97.39%<sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">95.00%<sup><bold>&#x0002A;</bold></sup></td>
<td valign="top" align="center">85.62%<sup><bold>&#x0002A;</bold></sup></td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>Red and bold text indicate the best performance among the 4 models. <sup>&#x0002A;</sup>The performance of prediction accuracy is higher than 80%.</p>
</table-wrap-foot>
</table-wrap>
</sec></sec>
<sec sec-type="discussion" id="s4">
<title>4 Discussion</title>
<p>This study presents an interval estimation principle to address the limitations of traditional machine learning models&#x00027; point estimates and aims to enhance the capability of time series prediction. The summarized findings are as follows:</p>
<list list-type="order">
<list-item><p>Both the traditional BPN and LSTM models require trial-and-error methods to find optimal parameter combinations. After comparing with literature recommendations and trial-and-error adjustments, the error values of the LSTM model in this study are consistently lower than those of the BPN model. On the other hand, the ELM and RF models require less trial and error to adjust parameters, resulting in faster training and testing processes. Although their error values are larger than those of the LSTM model, their predictive results are acceptable and outperform the BPN model. Overall, the results suggest that the LSTM model is more suitable for predicting time series data of the biotech and medical index during the COVID-19 period.</p></list-item>
<list-item><p>Despite the impact of the COVID-19 pandemic, the volatility of Taiwan&#x00027;s biotech index did not experience significant fluctuations compared to industries like the service or tourism sector. It remained relatively stable, allowing for effective learning by machine learning models despite the relatively short sampling period.</p></list-item>
<list-item><p>As described in (2), empirical evidence suggests that the Fuzzy-LSTM model with a 68% confidence level estimation can provide effective and reasonable predictions. The Fuzzy-ELM and Fuzzy-RF models perform better with a 95% confidence level estimation, while the Fuzzy-BPN model exhibits the lowest predictive accuracy among all models.</p></list-item>
<list-item><p>The proposed hybrid fuzzy interval LSTM model (LSTM, ELM, RF) in this study achieves high predictive accuracy for time series data. It implies that they are indeed capable of effectively capturing the time-series characteristics of stock price data in financial market time series and accurately predicting their values. Future applications of this approach in predicting time series data in other financial markets are recommended, as it could enhance the effectiveness of investment analysis measures for relevant financial decision-makers.</p></list-item>
<list-item><p>When making investment forecasts with financial data, investors not only focus on potential profits but also pay close attention to the risk management of their investment portfolios. If an inadvertent investment error leads to losses, the ability to reasonably estimate the maximum possible loss can make investors more willing to fund investment activities. An improved interval estimation machine learning tool can incorporate Value at Risk (VaR) (Jorion, <xref ref-type="bibr" rid="B27">1996</xref>) to estimate the maximum potential loss for effective risk management. This represents a suggested direction for future research in this thesis.</p></list-item>
</list></sec>
<sec sec-type="data-availability" id="s5">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p></sec>
<sec sec-type="author-contributions" id="s6">
<title>Author contributions</title>
<p>H-YL: Conceptualization, Data curation, Formal analysis, Methodology, Software, Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing. B-WH: Data curation, Investigation, Project administration, Validation, Visualization, Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing.</p></sec>
</body>
<back>
<sec sec-type="funding-information" id="s7">
<title>Funding</title>
<p>The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s8">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ali</surname> <given-names>M. N.</given-names></name> <name><surname>Giaccotto</surname> <given-names>C.</given-names></name></person-group> (<year>1982</year>). <article-title>The identical distribution hypothesis for stock market prices: location and scale shift alternatives</article-title>. <source>J. Am. Stat. Assoc</source>. <volume>77</volume>, <fpage>19</fpage>&#x02013;<lpage>28</lpage>. <pub-id pub-id-type="doi">10.1080/01621459.1982.10477762</pub-id></citation>
</ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bacchetta</surname> <given-names>P.</given-names></name> <name><surname>Mertens</surname> <given-names>E.</given-names></name> <name><surname>van Wincoop</surname> <given-names>E.</given-names></name></person-group> (<year>2009</year>). <article-title>Predictability in financial markets: what do survey expectations tell us?</article-title> <source>J. Int. Money Finance</source> <volume>28</volume>, <fpage>406</fpage>&#x02013;<lpage>426</lpage>. <pub-id pub-id-type="doi">10.1016/j.jimonfin.2008.09.001</pub-id></citation>
</ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ballings</surname> <given-names>M.</given-names></name> <name><surname>Van den Poel</surname> <given-names>D.</given-names></name> <name><surname>Hespeels</surname> <given-names>N.</given-names></name> <name><surname>Gryp</surname> <given-names>R.</given-names></name></person-group> (<year>2015</year>). <article-title>Evaluating multiple classifiers for stock price direction prediction</article-title>. <source>Expert Syst. Appl</source>. <volume>42</volume>, <fpage>7046</fpage>&#x02013;<lpage>7056</lpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2015.05.013</pub-id></citation>
</ref>
<ref id="B4">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Baret</surname> <given-names>S.</given-names></name> <name><surname>Celner</surname> <given-names>A.</given-names></name> <name><surname>O&#x00027;Reilly</surname> <given-names>M.</given-names></name> <name><surname>Shilling</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <source>COVID-19 Potential Implications for the Banking and Capital Markets Sector</source>. <publisher-loc>London</publisher-loc>: <publisher-name>Deloitte Center for Financial Services</publisher-name>.</citation>
</ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Barinath</surname> <given-names>S. G.</given-names></name> <name><surname>Chaterjee</surname> <given-names>S.</given-names></name></person-group> (<year>1988</year>). <article-title>On measuring skewness and elongation in common stock return distributions, the case of market index</article-title>. <source>J. Business</source> <volume>61</volume>, <fpage>451</fpage>&#x02013;<lpage>472</lpage>. <pub-id pub-id-type="doi">10.1086/296443</pub-id></citation>
</ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Basak</surname> <given-names>S.</given-names></name> <name><surname>Kar</surname> <given-names>S.</given-names></name> <name><surname>Saha</surname> <given-names>S.</given-names></name> <name><surname>Khaidem</surname> <given-names>L.</given-names></name> <name><surname>Roy Dey</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Predicting the direction of stock market prices using tree-based classifiers</article-title>. <source>North Am. J. Econ. Finance</source> <volume>47</volume>, <fpage>552</fpage>&#x02013;<lpage>567</lpage>. <pub-id pub-id-type="doi">10.1016/j.najef.2018.06.013</pub-id></citation>
</ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bengio</surname> <given-names>Y.</given-names></name> <name><surname>Courville</surname> <given-names>A. C.</given-names></name> <name><surname>Vincent</surname> <given-names>P.</given-names></name></person-group> (<year>2013</year>). <article-title>Representation learning: a review and new perspectives</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell</source>. <volume>35</volume>, <fpage>1798</fpage>&#x02013;<lpage>1828</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2013.50</pub-id></citation>
</ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bildirici</surname> <given-names>M.</given-names></name> <name><surname>Ersin</surname> <given-names>&#x000D6;. &#x000D6;.</given-names></name></person-group> (<year>2014</year>). <article-title>Nonlinearity volatility and fractional integration in daily oil prices: smooth transition autoregressive ST-FI (AP) GARCH models</article-title>. <source>Rom. J. Econ. Forecast</source> <volume>3</volume>, <fpage>108</fpage>&#x02013;<lpage>135</lpage>.</citation>
</ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bookstaber</surname> <given-names>R. M.</given-names></name> <name><surname>McDonald</surname> <given-names>J. B.</given-names></name></person-group> (<year>1987</year>). <article-title>A general distribution for describing security price returns</article-title>. <source>J. Business</source> <volume>60</volume>, <fpage>401</fpage>&#x02013;<lpage>424</lpage>. <pub-id pub-id-type="doi">10.1086/296404</pub-id></citation>
</ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Breiman</surname> <given-names>L.</given-names></name></person-group> (<year>2001</year>). <article-title>Random Forests</article-title>. <source>Mach. Learn</source>. <volume>45</volume>, <fpage>5</fpage>&#x02013;<lpage>32</lpage>. <pub-id pub-id-type="doi">10.1023/A:1010933404324</pub-id></citation>
</ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cao</surname> <given-names>J. W.</given-names></name> <name><surname>Zhang</surname> <given-names>K.</given-names></name> <name><surname>Luo</surname> <given-names>M. X.</given-names></name> <name><surname>Yin</surname> <given-names>C.</given-names></name> <name><surname>Lai</surname> <given-names>X. P.</given-names></name></person-group> (<year>2016</year>). <article-title>Extreme learning machine and adaptive sparse representation for image classification</article-title>. <source>Neural Netw</source>. <volume>81</volume>, <fpage>91</fpage>&#x02013;<lpage>102</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2016.06.001</pub-id><pub-id pub-id-type="pmid">27389571</pub-id></citation></ref>
<ref id="B12">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>A. P.</given-names></name> <name><surname>Lin</surname> <given-names>H. Y.</given-names></name></person-group> (<year>2007</year>). <article-title>&#x0201C;Exchange rates forecasting using a hybrid fuzzy and neural network model,&#x0201D;</article-title> in <source>Proceedings of IEEE Symposium on Computational Intelligence and Data Mining (CIDM)</source> (<publisher-loc>Honolulu</publisher-loc>), <fpage>758</fpage>&#x02013;<lpage>763</lpage>. <pub-id pub-id-type="doi">10.1109/CIDM.2007.368952</pub-id></citation>
</ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>C.-C.</given-names></name> <name><surname>Kuo</surname> <given-names>C.</given-names></name> <name><surname>Kuo</surname> <given-names>S.-Y.</given-names></name> <name><surname>Chou</surname> <given-names>Y.-H.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Dynamic normalization BPN for stock price forecasting,&#x0201D;</article-title> <italic>in Proceedings of the 2015 IEEE International Conference on Systems, Man, and Cybernetics</italic> (Hong Kong), <fpage>2855</fpage>&#x02013;<lpage>2860</lpage>. <pub-id pub-id-type="doi">10.1109/SMC.2015.497</pub-id></citation>
</ref>
<ref id="B14">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>K.</given-names></name> <name><surname>Zhou</surname> <given-names>Y.</given-names></name> <name><surname>Dai</surname> <given-names>F.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;ALSTM-based method for stock returns prediction: a case study of China stock market,&#x0201D;</article-title> in <source>Proceedings of the 2015 IEEE International Conference on Big Data (Big Data)</source> (<publisher-loc>Santa Clara, CA</publisher-loc>), 2823&#x02212;2824. <pub-id pub-id-type="doi">10.1109/BigData.2015.7364089</pub-id></citation>
</ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>Z. H.</given-names></name> <name><surname>Chen</surname> <given-names>Y. L.</given-names></name> <name><surname>Chang</surname> <given-names>W. Y.</given-names></name> <name><surname>Tsai</surname> <given-names>C. W.</given-names></name></person-group> (<year>2019</year>). <article-title>A hybrid classification algorithm for intrusion detection system</article-title>. <source>Commun. CCISA</source> <volume>25</volume>, <fpage>14</fpage>&#x02013;<lpage>27</lpage>.</citation>
</ref>
<ref id="B16">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Cheng</surname> <given-names>G. J.</given-names></name> <name><surname>Cai</surname> <given-names>L.</given-names></name> <name><surname>Pan</surname> <given-names>H. X.</given-names></name></person-group> (<year>2009</year>). <article-title>&#x0201C;Comparison of extreme learning machine with support vector regression for reservoir permeability prediction,&#x0201D;</article-title> in <source>Proceedings of the 2009 International Conference on Computational Intelligence and Security (CIS)</source> (<publisher-loc>Beijing</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>173</fpage>&#x02013;<lpage>176</lpage>. <pub-id pub-id-type="doi">10.1109/CIS.2009.124</pub-id></citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chong</surname> <given-names>E.</given-names></name> <name><surname>Han</surname> <given-names>C.</given-names></name> <name><surname>Park</surname> <given-names>F. C.</given-names></name></person-group> (<year>2017</year>). <article-title>Deep learning networks for stock market analysis and prediction: methodology, data representations, and case studies</article-title>. <source>Expert Syst. Appl</source>. <volume>83</volume>, <fpage>187</fpage>&#x02013;<lpage>205</lpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2017.04.030</pub-id></citation>
</ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cui</surname> <given-names>Z.</given-names></name> <name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Lian</surname> <given-names>W.</given-names></name> <name><surname>Wang</surname> <given-names>Y. -G.</given-names></name></person-group> (<year>2023</year>). <article-title>A novel deep learning framework with a COVID-19 adjustment for electricity demand forecasting</article-title>. <source>Energy Rep</source>. <volume>9</volume>, <fpage>1887</fpage>&#x02013;<lpage>1895</lpage>. <pub-id pub-id-type="doi">10.1016/j.egyr.2023.01.019</pub-id></citation>
</ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Davies</surname> <given-names>P. C.</given-names></name></person-group> (<year>1994</year>). <article-title>Design issues in neural network development</article-title>. <source>Neurovest J</source>. <volume>5</volume>, <fpage>21</fpage>&#x02013;<lpage>25</lpage>.</citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Di Persio</surname> <given-names>L.</given-names></name> <name><surname>Honchar</surname> <given-names>O.</given-names></name></person-group> (<year>2016</year>). <article-title>Artificial neural networks architectures for stock price prediction: comparisons and applications</article-title>. <source>Int. J. Circ. Syst. Signal Process</source>. <volume>10</volume>, <fpage>403</fpage>&#x02013;<lpage>413</lpage>.</citation>
</ref>
<ref id="B21">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Freeman</surname> <given-names>J. A.</given-names></name> <name><surname>Skapura</surname> <given-names>D. M.</given-names></name></person-group> (<year>1992</year>). <source>Neural Networks Algorithms, Applications, and Programming Techniques.</source> <publisher-loc>CA, USA</publisher-loc>: <publisher-name>Addison-Wesley</publisher-name>.</citation>
</ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gan</surname> <given-names>D.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>N.</given-names></name> <name><surname>Zhu</surname> <given-names>W.</given-names></name></person-group> (<year>2017</year>). <article-title>Enhancing short-term probabilistic residential load forecasting with quantile long-short-term memory</article-title>. <source>J. Eng</source>. <volume>14</volume>, <fpage>2622</fpage>&#x02013;<lpage>2627</lpage>. <pub-id pub-id-type="doi">10.1049/joe.2017.0833</pub-id></citation>
</ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>He</surname> <given-names>K.</given-names></name> <name><surname>Yang</surname> <given-names>Q.</given-names></name> <name><surname>Ji</surname> <given-names>L.</given-names></name> <name><surname>Pan</surname> <given-names>J.</given-names></name> <name><surname>Zou</surname> <given-names>Y.</given-names></name></person-group> (<year>2023</year>). <article-title>Financial time series forecasting with the deep learning ensemble model</article-title>. <source>Mathematics</source> <volume>11</volume>:<fpage>1054</fpage>. <pub-id pub-id-type="doi">10.3390/math11041054</pub-id></citation>
</ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hochreiter</surname> <given-names>S.</given-names></name> <name><surname>Schmidhuber</surname> <given-names>J.</given-names></name></person-group> (<year>1997</year>). <article-title>Long short-term memory</article-title>. <source>Neural. Comput</source>. <volume>9</volume>, <fpage>1735</fpage>&#x02013;<lpage>1780</lpage>. <pub-id pub-id-type="doi">10.1162/neco.1997.9.8.1735</pub-id></citation>
</ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>G. B.</given-names></name> <name><surname>Zhou</surname> <given-names>H.</given-names></name> <name><surname>Ding</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>R.</given-names></name></person-group> (<year>2011</year>). <article-title>Extreme learning machine for regression and multiclass classification</article-title>. <source>IEEE Trans. Syst. Man. Cybern. B Cybern</source>. <volume>42</volume>, <fpage>513</fpage>&#x02013;<lpage>529</lpage>. <pub-id pub-id-type="doi">10.1109/TSMCB.2011.2168604</pub-id><pub-id pub-id-type="pmid">21984515</pub-id></citation></ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>G. B.</given-names></name> <name><surname>Zhu</surname> <given-names>Q. Y.</given-names></name> <name><surname>Siew</surname> <given-names>C. K.</given-names></name></person-group> (<year>2006</year>). <article-title>Extreme learning machine: theory and applications</article-title>. <source>Neurocomputing</source> <volume>70</volume>, <fpage>489</fpage>&#x02013;<lpage>501</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2005.12.126</pub-id></citation>
</ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jorion</surname> <given-names>P.</given-names></name></person-group> (<year>1996</year>). <article-title>Risk - measuring the risk in value at risk</article-title>. <source>Financ. Anal. J</source>. <volume>52</volume>, <fpage>47</fpage>&#x02013;<lpage>56</lpage>. <pub-id pub-id-type="doi">10.2469/faj.v52.n6.2039</pub-id></citation>
</ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kaiser</surname> <given-names>H. F.</given-names></name></person-group> (<year>1960</year>). <article-title>The application of electronic computers to factor analysis</article-title>. <source>Educ. Psychol. Meas</source>. <volume>20</volume>, <fpage>141</fpage>&#x02013;<lpage>151</lpage>. <pub-id pub-id-type="doi">10.1177/001316446002000116</pub-id></citation>
</ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Khaidem</surname> <given-names>L.</given-names></name> <name><surname>Saha</surname> <given-names>S.</given-names></name> <name><surname>Dey</surname> <given-names>S. R.</given-names></name></person-group> (<year>2016</year>). <source>Predicting the Direction of Stock Market Prices using Random Forest.</source></citation>
</ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kili&#x000E7;</surname> <given-names>D. K.</given-names></name> <name><surname>Ugur</surname> <given-names>&#x000D6;.</given-names></name></person-group> (<year>2018</year>). <article-title>Multiresolution analysis of SandP500 time series</article-title>. <source>Ann. Oper Res</source>. <volume>260</volume>, <fpage>197</fpage>&#x02013;<lpage>216</lpage>. <pub-id pub-id-type="doi">10.1007/s10479-016-2215-3</pub-id></citation>
</ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kon</surname> <given-names>S. J.</given-names></name></person-group> (<year>1984</year>). <article-title>Models of stock returns - a comparison</article-title>. <source>J. Finance</source> <volume>39</volume>, <fpage>147</fpage>&#x02013;<lpage>165</lpage>. <pub-id pub-id-type="doi">10.1111/j.1540-6261.1984.tb03865.x</pub-id></citation>
</ref>
<ref id="B32">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Lawrence</surname> <given-names>M.</given-names></name> <name><surname>Petterson</surname> <given-names>A.</given-names></name></person-group> (<year>1991</year>). <source>Getting Started with Brain Maker: Neural Network Simulation Software User&#x00027;s Guide and Reference Manual/Introduction to Neural Networks and Disk.</source> <publisher-loc>MA, USA</publisher-loc>: <publisher-name>California Scientific Software</publisher-name>.</citation>
</ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>T. K.</given-names></name> <name><surname>Cho</surname> <given-names>J. H.</given-names></name> <name><surname>Kwon</surname> <given-names>D. S.</given-names></name> <name><surname>Sohn</surname> <given-names>S. Y.</given-names></name></person-group> (<year>2019</year>). <article-title>Global stock market investment strategies based on financial network indicators using machine learning</article-title> <source>Techniques</source> <volume>117</volume>, <fpage>228</fpage>&#x02013;<lpage>242</lpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2018.09.005</pub-id></citation>
</ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lewis</surname> <given-names>E. B.</given-names></name></person-group> (<year>1982</year>). <article-title>Control of body segment differentiation in drosophila by the bithorax gene complex</article-title>. <source>Embryo. Dev</source>. <volume>1</volume>, <fpage>383</fpage>&#x02013;<lpage>417</lpage>. <pub-id pub-id-type="doi">10.1007/978-1-4419-8981-9_15</pub-id><pub-id pub-id-type="pmid">7111279</pub-id></citation></ref>
<ref id="B35">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>P.</given-names></name> <name><surname>Jing</surname> <given-names>C.</given-names></name> <name><surname>Liang</surname> <given-names>T.</given-names></name> <name><surname>Liu</surname> <given-names>M.</given-names></name> <name><surname>Chen</surname> <given-names>Z.</given-names></name> <name><surname>Guo</surname> <given-names>L.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Autoregressive moving average modeling in the financial sector,&#x0201D;</article-title> in <source>Proceedings of the 2nd International Conference on Information Technology Computer and Electrical Engineering (ICITACEE)</source> (<publisher-loc>Semarang</publisher-loc>), <fpage>68</fpage>&#x02013;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1109/ICITACEE.2015.7437772</pub-id></citation>
</ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>X.</given-names></name> <name><surname>Xie</surname> <given-names>H.</given-names></name> <name><surname>Wang</surname> <given-names>R.</given-names></name> <name><surname>Cai</surname> <given-names>Y.</given-names></name> <name><surname>Cao</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Empirical analysis: stock market prediction via extreme learning machine</article-title>. <source>Neural Comput. Appl</source>. <volume>27</volume>, <fpage>67</fpage>&#x02013;<lpage>78</lpage>. <pub-id pub-id-type="doi">10.1007/s00521-014-1550-z</pub-id></citation>
</ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lim</surname> <given-names>B.</given-names></name> <name><surname>Zohren</surname> <given-names>S.</given-names></name></person-group> (<year>2021</year>). <article-title>Time series forecasting with deep learning: a survey</article-title>. <source>Philos. Trans. R. Soc</source>. <volume>379</volume>, <fpage>202</fpage>&#x02013;<lpage>209</lpage>. <pub-id pub-id-type="doi">10.1098/rsta.2020.0209</pub-id><pub-id pub-id-type="pmid">33583273</pub-id></citation></ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>B.</given-names></name> <name><surname>Nowotarski</surname> <given-names>J.</given-names></name> <name><surname>Hong</surname> <given-names>T.</given-names></name> <name><surname>Weron</surname> <given-names>R.</given-names></name></person-group> (<year>2017</year>). <article-title>Probabilistic load forecasting via quantile regression averaging on sister forecasts</article-title>. <source>IEEE Trans. Smart Grid</source>. <volume>8</volume>, <fpage>730</fpage>&#x02013;<lpage>737</lpage>. <pub-id pub-id-type="doi">10.1109/TSG.2015.2437877</pub-id></citation>
</ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>D.</given-names></name> <name><surname>Wei</surname> <given-names>A.</given-names></name></person-group> (<year>2022</year>). <article-title>The performance of hybrid artificial neural network models for option pricing during financial crises</article-title>. <source>J. Data Sci</source>. <volume>14</volume>, <fpage>1</fpage>&#x02013;<lpage>18</lpage>. <pub-id pub-id-type="doi">10.6339/JDS.201601_14(1)0.0001</pub-id></citation>
</ref>
<ref id="B40">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>S.</given-names></name> <name><surname>Liao</surname> <given-names>G.</given-names></name> <name><surname>Ding</surname> <given-names>Y.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;Stock transaction prediction modeling and analysis based on LSTM,&#x0201D;</article-title> in <source>Proceedings of the IEEE Conference on Industrial Electronics and Applications (ICIEA)</source> (<publisher-loc>Wuhan</publisher-loc>), <fpage>2787</fpage>&#x02013;<lpage>2790</lpage>. <pub-id pub-id-type="doi">10.1109/ICIEA.2018.8398183</pub-id></citation>
</ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lowe</surname> <given-names>D.</given-names></name> <name><surname>Zapart</surname> <given-names>K.</given-names></name></person-group> (<year>1999</year>). <article-title>Point-wise confidence interval estimation by neural networks: a comparative study based on automotive engine calibration</article-title>. <source>Neural Comput. Appl</source>. <volume>8</volume>, <fpage>77</fpage>&#x02013;<lpage>85</lpage>. <pub-id pub-id-type="doi">10.1007/s005210050009</pub-id></citation>
</ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mandelbrot</surname> <given-names>B. B.</given-names></name></person-group> (<year>1963</year>). <article-title>The variation of certain speculative prices</article-title>. <source>J. Business</source> <volume>36</volume>, <fpage>394</fpage>&#x02013;<lpage>419</lpage>. <pub-id pub-id-type="doi">10.1086/294632</pub-id></citation>
</ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Marsza&#x00142;ek</surname> <given-names>A.</given-names></name> <name><surname>Burczy&#x00144;ski</surname> <given-names>T.</given-names></name></person-group> (<year>2014</year>). <article-title>Modeling and forecasting financial time series with ordered fuzzy candlesticks</article-title>. <source>Inf. Sci</source>. <volume>273</volume>, <fpage>144</fpage>&#x02013;<lpage>155</lpage>. <pub-id pub-id-type="doi">10.1016/j.ins.2014.03.026</pub-id></citation>
</ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nana</surname> <given-names>L.</given-names></name> <name><surname>Jiangtao</surname> <given-names>Q.</given-names></name></person-group> (<year>2018</year>). <article-title>Research on A-share stock rise and fall prediction based on Random Forest</article-title>. <source>J. Shanghai Univ. Technol.</source> <fpage>267</fpage>&#x02013;<lpage>273</lpage>.</citation>
</ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pal</surname> <given-names>M.</given-names></name></person-group> (<year>2005</year>). <article-title>Random Forest classifier for remote sensing classification</article-title>. <source>Int. J. Remote Sens</source>. <volume>26</volume>, <fpage>217</fpage>&#x02013;<lpage>222</lpage>. <pub-id pub-id-type="doi">10.1080/01431160412331269698</pub-id></citation>
</ref>
<ref id="B46">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Qu</surname> <given-names>B.</given-names></name></person-group> (<year>2003</year>). <source>The research of the effect and forecast of meteorological factors on epidemic situation of common infectious diseases in drought area</source> (Master&#x00027;s thesis). TaiChung: China Medical University.</citation>
</ref>
<ref id="B47">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Rumelhart</surname> <given-names>D. E.</given-names></name> <name><surname>McClelland</surname> <given-names>J. L.</given-names></name></person-group> (<year>1986</year>). <source>Parallel Distributed Processing, Explorations in the Microstructure of Cognition. Vol. 1: Foundations.</source> <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>. <pub-id pub-id-type="doi">10.7551/mitpress/5236.001.0001</pub-id><pub-id pub-id-type="pmid">25087578</pub-id></citation></ref>
<ref id="B48">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Selvin</surname> <given-names>S.</given-names></name> <name><surname>Vinayakumar</surname> <given-names>R.</given-names></name> <name><surname>Gopalakrishnan</surname> <given-names>E.</given-names></name> <name><surname>Menon</surname> <given-names>V. K.</given-names></name> <name><surname>Soman</surname> <given-names>K.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Stock price prediction using LSTM RNN and CNN-sliding window model,&#x0201D;</article-title> in <source>Proceedings of the 2017 International Conference on Advances in Computing, Communications and Informatics (ICACCI)</source> (<publisher-loc>Udupi</publisher-loc>), <fpage>1643</fpage>&#x02013;<lpage>1647</lpage>. <pub-id pub-id-type="doi">10.1109/ICACCI.2017.8126078</pub-id></citation>
</ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>F.</given-names></name> <name><surname>Toh</surname> <given-names>K. A.</given-names></name> <name><surname>Romay</surname> <given-names>M. G.</given-names></name> <name><surname>Mao</surname> <given-names>K.</given-names></name></person-group> (<year>2014</year>). <source>Extreme Learning Machines: Algorithms and Applications</source>. Berlin: Springer International Publishing. <pub-id pub-id-type="doi">10.1007/978-3-319-04741-6</pub-id></citation>
</ref>
<ref id="B50">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Sunny</surname> <given-names>M. A. I.</given-names></name> <name><surname>Maswood</surname> <given-names>M. M. S.</given-names></name> <name><surname>Alharbi</surname> <given-names>A. G.</given-names></name></person-group> (<year>2020</year>). <article-title>&#x0201C;Deep learning-based stock price prediction using LSTM and bi-directional LSTM model,&#x0201D;</article-title> in <source>Proceedings of the 2nd Novel Intelligent and Leading Emerging Sciences Conference (NILES)</source> (<publisher-loc>Giza</publisher-loc>), <fpage>87</fpage>&#x02013;<lpage>92</lpage>.</citation>
</ref>
<ref id="B51">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Tabachnick</surname> <given-names>B. G.</given-names></name> <name><surname>Fidell</surname> <given-names>L. S.</given-names></name></person-group> (<year>1996</year>). <source>Using Multivariate Statistics (3rd edn.)</source>. <publisher-loc>Giza</publisher-loc>: <publisher-name>Harper Collins</publisher-name>.</citation>
</ref>
<ref id="B52">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Uddin</surname> <given-names>M.</given-names></name> <name><surname>Chowdhury</surname> <given-names>A.</given-names></name> <name><surname>Anderson</surname> <given-names>K.</given-names></name> <name><surname>Chaudhuri</surname> <given-names>K.</given-names></name></person-group> (<year>2021</year>). <article-title>The effect of COVID&#x02212;19 pandemic on global stock market volatility: can economic strength help to manage the uncertainty?</article-title> <source>J. Bus. Res</source>. <volume>128</volume>, <fpage>31</fpage>&#x02013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1016/j.jbusres.2021.01.061</pub-id><pub-id pub-id-type="pmid">36540352</pub-id></citation></ref>
<ref id="B53">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Levi</surname> <given-names>N.</given-names></name> <name><surname>Araujob</surname> <given-names>R.</given-names></name> <name><surname>Wang</surname> <given-names>Y. -G.</given-names></name></person-group> (<year>2022</year>). <article-title>An evaluation of the impact of COVID-19 lockdowns on electricity demand</article-title>. <source>Electr. Power Syst. Res</source>. <volume>216</volume>:<fpage>109015</fpage>. <pub-id pub-id-type="doi">10.1016/j.epsr.2022.109015</pub-id></citation>
</ref>
<ref id="B54">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xie</surname> <given-names>J.</given-names></name> <name><surname>Hong</surname> <given-names>T.</given-names></name> <name><surname>Laing</surname> <given-names>T.</given-names></name> <name><surname>Kang</surname> <given-names>C.</given-names></name></person-group> (<year>2017</year>). <article-title>On normality assumption in residual simulation for probabilistic load forecasting</article-title>. <source>IEEE Trans. Smart Grid</source>. <volume>8</volume>, <fpage>1046</fpage>&#x02013;<lpage>1053</lpage>. <pub-id pub-id-type="doi">10.1109/TSG.2015.2447007</pub-id></citation>
</ref>
<ref id="B55">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>W.</given-names></name> <name><surname>Kang</surname> <given-names>C.</given-names></name> <name><surname>Xia</surname> <given-names>Q.</given-names></name> <name><surname>Liu</surname> <given-names>R.</given-names></name> <name><surname>Tang</surname> <given-names>T.</given-names></name> <name><surname>Wang</surname> <given-names>P.</given-names></name></person-group> (<year>2006</year>). <article-title>Short-term probabilistic load forecasting based on statistics of probability distribution of forecasting errors</article-title>. <source>Autom. Electr. Power Syst</source>. <volume>19</volume>, <fpage>11</fpage>. <pub-id pub-id-type="doi">10.1109/TPWRS.2005.860937</pub-id></citation>
</ref>
<ref id="B56">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yoon</surname> <given-names>Y.</given-names></name> <name><surname>Swales</surname> <given-names>G.</given-names></name> <name><surname>Margavio</surname> <given-names>T. M.</given-names></name></person-group> (<year>1993</year>). <article-title>A comparison of discriminant analysis versus artificial neural networks</article-title>. <source>J. Oper. Res. Soc</source>. <volume>44</volume>, <fpage>51</fpage>&#x02013;<lpage>60</lpage>. <pub-id pub-id-type="doi">10.1057/jors.1993.6</pub-id></citation>
</ref>
<ref id="B57">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zadeh</surname> <given-names>L. A.</given-names></name></person-group> (<year>1965</year>). <article-title>Fuzzy sets</article-title>. <source>Inf. Control</source> <volume>8</volume>, <fpage>338</fpage>&#x02013;<lpage>353</lpage>. <pub-id pub-id-type="doi">10.1016/S0019-9958(65)90241-X</pub-id></citation>
</ref>
<ref id="B58">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>G.</given-names></name> <name><surname>Patuwo</surname> <given-names>B. E.</given-names></name> <name><surname>Hu</surname> <given-names>M. Y.</given-names></name></person-group> (<year>1998</year>). <article-title>Forecasting with artificial neural networks: the state of the art</article-title>. <source>Int. J. Forecast</source>. <volume>14</volume>, <fpage>35</fpage>&#x02013;<lpage>62</lpage>. <pub-id pub-id-type="doi">10.1016/S0169-2070(97)00044-7</pub-id></citation>
</ref>
<ref id="B59">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>G.</given-names></name> <name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Feng</surname> <given-names>H.</given-names></name></person-group> (<year>2016</year>). <article-title>Forecasting financial time series using a methodology based on autoregressive integrated moving average and Taylor expansion</article-title>. <source>Expert. Syst</source>. <volume>33</volume>, <fpage>501</fpage>&#x02013;<lpage>516</lpage>. <pub-id pub-id-type="doi">10.1111/exsy.12164</pub-id></citation>
</ref>
<ref id="B60">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>Z.</given-names></name> <name><surname>Rao</surname> <given-names>R.</given-names></name> <name><surname>Tu</surname> <given-names>S.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Time-weighted LSTM model with redefined labeling for stock trend prediction,&#x0201D;</article-title> in <source>Proceedings of the IEEE 29th International Conference on Tools with Artificial Intelligence (ICTAI)</source> (<publisher-loc>Boston, MA</publisher-loc>), <fpage>1210</fpage>&#x02013;<lpage>1217</lpage>. <pub-id pub-id-type="doi">10.1109/ICTAI.2017.00184</pub-id></citation>
</ref>
<ref id="B61">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>Z.</given-names></name> <name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Cai</surname> <given-names>F.</given-names></name> <name><surname>Zhang</surname> <given-names>S.</given-names></name> <name><surname>Wang</surname> <given-names>Y. -G.</given-names></name></person-group> (<year>2023</year>). <article-title>A hybrid deep learning framework for air quality prediction with spatial autocorrelation during the COVID-19 pandemic</article-title>. <source>Sci. Rep</source>. <volume>13</volume>, <fpage>1015</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-023-28287-8</pub-id><pub-id pub-id-type="pmid">36653488</pub-id></citation></ref>
</ref-list>
</back>
</article>