Skip to main content

ORIGINAL RESEARCH article

Front. Appl. Math. Stat., 26 May 2021
Sec. Mathematics of Computation and Data Science
Volume 7 - 2021 | https://doi.org/10.3389/fams.2021.601105

Sentiment-Guided Adversarial Learning for Stock Price Prediction

  • 1Beijing Institute of Technology, Haidian District, Beijing, China
  • 2Kamakura Corporation, Honolulu, HI, United States
  • 3Department of Applied Mathematics, Illinois Institute of Technology, Chicago, IL, United States

Prediction of stock prices or trends have attracted financial researchers’ attention for many years. Recently, machine learning models such as neural networks have significantly contributed to this research problem. These methods often enable researchers to take stock-related factors such as sentiment information into consideration, improving prediction accuracies. At present, Long Short-Term Memory (LSTM) networks is one of the best techniques known to learn knowledge from time-series data and to predict future tendencies. The inception of generative adversarial networks (GANs) also provides researchers with diversified and powerful methods to explore the stock prediction problem. A GAN network consists of two sub-networks known as generator and discriminator, which work together to minimize maximum loss on both actual and simulated data. In this paper, we developed a sentiment-guided adversarial learning and predictive models of stock prices, adopting a popular variation of GAN called conditional GAN (CGAN). We adopted an LSTM network in the generator and a multilayer perceptron (MLP) network in the discriminator. After extensively pre-processing historical stock price datasets, we analyzed the sentiment information from daily tweets and computed sentiment scores as an additional model feature. Our experiments demonstrated that the average forecast accuracies of the CGAN models were improved with sentiment data. Moreover, our GAN and CGAN models outperformed LSTM and other traditional methods on 11 out of 36 processed stock price datasets, potentially playing a part in ensemble methods.

1 Introduction

Stock prediction has been attached with great importance in the financial world. While stock fluctuation is very unpredictable, researchers have made every effort to simulate the stock variation because a relatively reasonable prediction can create massive profits and help reduce risks. Stock prediction researchers often consider two kinds of solution methods: classification and regression. Classification methods predict stock movement, while regression methods predict stock prices. Stock movement prediction can be seen as a simple classification task since it only predicts whether the stock will be up or down by a certain amount, or remain almost unchanged. However, many researchers focus their efforts on stock price prediction–a regression task that forecasts future prices with past values – since it could yield more profits than simple movement prediction. In a regression task, researchers have to tackle very complex situations to forecast the future price accurately. Like other time-series prediction problems, many factors should also be taken into consideration besides historical prices. Investors’ sentiments, economic conditions, and public opinions are all critical variables that should be added into a stock prediction system. Particularly, sentiment analysis has been identified as a useful tool in recent years. Not just price and price-related values, researchers have paid more attention to the extraction of sentiment information from newspapers, magazines, and even social media. For example, Ref. [1] examined the relation between social media sentiment and stock prices, in addition to economic indicators in their research on stock market. Ref. [2] Focused on combining stock-related events with the sentiment information from reviews of financial critics. They analyzed sentiments and predicted the fluctuation of the stock market with a matrix and tensor factorization framework. Besides mathematical methods, the development of natural language processing also provides additional techniques for analyzing text and sentiment information. Machine learning, data mining, and semantic comprehension have made extracting large amounts of stock sentiments possible. With the development of social media, people are increasingly inclined to exchange information through the Internet platform. Real-time stock reviews contain a wealth of financial information that reflects the emotional changes of investors. Works such as [3, 4]; and [5] analyzed sentiment information from large numbers of real-time tweets, which were related to stocks and corresponding companies, then investigated the correlation between stock movements and public emotions with supervised machine learning principles. In particular, Ref. [3] utilized two mood tracking tools, OpinionFinder and Google-Profile of Mood States (GPOMS), to analyze the text content of daily tweets, which succeeded in measuring varying degrees of mood. These research results proved that modern techniques are mature enough to handle mountains of sentiment data and that Twitter is a valuable text resource for sentiment analysis.

In our work, we used VADER (Valence Aware Dictionary and sEntiment Reasoner) [6] as a mood tracking tool to help us analyze the sentiment information from tweets. We extracted sentiment features from the past several days of tweets and input them into stock prediction models to predict future stock prices. Applying sentiment analysis to nine stocks, we trained the models with two-month-long training sets with tweets and tested the model performance with the final five days’ data. We applied several linear and nonlinear models such as LSTM to this regression task. Moreover, referring to the theories of recurrent neural networks and generative adversarial networks, we designed a sentiment-guided model to improve the accuracy of stock prediction further.

In the remainder of this article, the organization is set in the following order: First, in Section 2 we review existing stock prediction methods and the development of GANs. Then, in Section 3, we introduce our methods in data collection and processing. In Sections 4 and 5, we propose our sentiment-guided adversarial learning model as well as comparisons between our models and baselines. Finally, conclusions and future work are discussed at the end of this paper in Section 6.

2 Related Works

In this section, we first present an overview of the relatively recent generative adversarial networks, followed by a brief review of some traditional methods for stock price prediction.

2.1 Adversarial Learning of Neural Networks

Generative adversarial networks (GANs) [7], which try to fool a classification model in an adversarial minimax game, have shown high potential in obtaining more robust results compared to traditional neural networks [8]. In a GAN framework, a generator produces fake data based on noisy samples and attempts to minimize the difference between real and fake distribution, which is maximized by a discriminator oppositely. The GAN framework and GAN-based research have attracted huge attention in various fields recently. Existing GAN works mainly focus on computer vision tasks like image classification [9] and natural language processing like text analysis [10]. With the inception and extensive applications of knowledge representation of natural or complex data such as languages and multi-media, the GAN framework is also widely applied to classification missions for data in representation spaces (e.g., vector spaces of matrices) rather than just for the original data in feature spaces. Some researchers also extended applications of GAN to more challenging problems such as recommendation systems [11] or social network alignment problems [12].

Conditional GANs (CGANs) [13] add auxiliary information to input data, guiding the training process to acquire expected results. Additional information (like class labels or extra data) is fed into both generator and discriminator to perform the conditioning. This architecture is mostly used in image generation [14] and translation [15]. However, we noticed that the CGAN framework has rarely been applied to time-series prediction, specifically, stock prediction problem. In our work, we added sentiment labels to guide the training process of our model and achieved a stronger performance.

Previous work [16] has shown the capacity of GANs for generating sequential data and to fulfill the adversarial training with discrete tokens (e.g., words). Recurrent neural networks (RNNs) have also been widely used to solve problems based on sequential data, such as speech recognition and image generation. Ref. [17] first combined RNNs with a GAN framework and successfully produced many kinds of continuous sequential data, including polyphonic music. Inspired by this work, researchers paid more attention to the potential of RNNs in adversarial training. Refs. [18, 19] succeeded in producing realistic real-valued multi-dimensional time-series data and concentrated their work on medical applications. They built models capable of synthesizing realistic medical data and developed new approaches to create predictive systems in the medical domain. Ref. [20] utilized prerecorded human motion data to train their models and applied their neural network in random motion synthesis, online or offline motion control, and motion filtering. Our work drew on the idea of the RNN-GAN framework and applied time-series analysis to the financial problem of stock prediction. Although similar application in stock prediction has been raised recently [21], the main difference was that our work also utilized text data from Twitter to create sentiment scores, which could gauge the market mood and guide our models to achieve more accurate prediction.

2.2 Stock Prediction Methods

The stock market prediction problem has a long history and is regarded as an essential issue in the field of financial mathematics. According to the development of research over the years, we can divide the prediction techniques into two categories: statistical and machine learning methods. Frequently used statistical techniques include the autoregressive method (AR), the moving average model (MA), the autoregressive moving average model (ARMA), and the autoregressive integrated moving average (ARIMA) [22]. These methods adopt the principles of random processes and essentially rely on past values of the sequences to analyze and predict the future data. Another technique is generalized autoregressive conditional heteroskedasticity (GARCH) [22], which takes the fluctuation of variance into account and nicely simulates the variation of financial variables. However, both kinds of techniques depend on strong prerequisites such as stationarity and independent and identically distributed (iid) random variables, which may not be satisfied by data in the real-world financial markets.

Recently, machine learning methods (including linear regression, random forest, and support vector machine) stimulate the interest of financial researchers and are applied to forecasting problems. Among them, support vector machine (SVM) and its application in the financial market revealed superior properties compared to other individual classification methods [23]. Since the emergence of deep learning networks, researchers have proved that neural networks can obtain better performance than linear models. Their capacity to extract features from enormous amount of raw data from diverse sources without prior knowledge of predictors makes deep learning a preferred technique for stock market prediction. Artificial neural networks and other advanced neural models like convolution neural networks (CNNs) are evidenced to be good at capturing the non-linear hidden relationships of stock prices without any statistical or econometric assumption [24]. Furthermore, neural networks have also been found to be more efficient in solving non-linear financial problems compared to traditional statistical methods like ARIMA [25]. Nowadays, RNNs are one of the most popular tools in time-series prediction problems. Notably, the LSTM network has been a great success due to its ability to retain recent samples and forget earlier ones [26]. Each LSTM unit has three different gates: forget gate, update gate, and output gate. LSTM units can change their states by controlling their inner operant, namely, their three gates. In our model, we utilized the LSTM network to extract features according to the timeline and to generate fake stock prices from real past price values. We also used sentiment measures to enhance the robustness of our prediction models and make the generative results approach the real distribution.

3 Data and Methods

In this section, we document our data sources and preprocessing techniques. The latter part includes feature engineering, imputation of missing data, fast Fourier transform for denoising data, and last but not least, Isolation Forest for anomaly or outlier detection.

3.1 Data Collection and Sentiment Analysis

Before we started our work, we had the conviction that a collection of tweets related to the stocks could make comparatively accurate models of the investors’ mood and thus reach better predictions of the stock market prices. Generally speaking, tweets have neutral, positive, or negative emotions; and we focused on those that contains one or several cashtags (unique identifiers for businesses), which could influence the stock’s trend in the following day. If negative sentiment dominated a day, then the next day’s stock prices would be expected to fall. The number of followers on one’s Twitter account would also be a significant factor. The more followers of an account, the higher the influence of tweets from the account, and the more significant their impact would likely have on stock prices. Cashtags system is a particularly convenient feature of Twitter, allowing users to see what everyone is saying about public companies. The way this system works is similar to the well-known #hashtags of Twitter, except that a cashtag requires “$” followed by a stock symbol (e.g., $GOOG for Google, LLC; $FB for Facebook, Inc.; and $AAPL for Apple Inc.).

For our work, financial tweets and their retweets were downloaded from the website, https://data.world/kike/nasdaq-100-tweets. The time span of these tweets is 71 days from Apr 1–June 10, 2016. We could get all tweets mentioning any NASDAQ 100 companies from this source. The key daily stock price data includes Open prices (O), High prices (H), Low prices (L), and Close prices (C); they were subsequently crawled from Yahoo Finance with the Python package, pandas_datareader. We produced nine datasets, each one including tweets and price values for the nine companies as listed in Table 1. These companies are from different sectors/industry groups by the Global Industry Classification Standard (GICS). They could be classified into three price-trend categories: descending prices, mildly fluctuating prices, and ascending prices. Figure 1 shows the Open price curves of nine stocks with the three distinct trends.

TABLE 1
www.frontiersin.org

TABLE 1. Stock prices of nine companies selected for experiments in this study.

FIGURE 1
www.frontiersin.org

FIGURE 1. The nine stocks classified into three classes with distinct price trends. The variations of Open price from April 1 to June 10, 2016 are displayed. The first column presents the stocks with descending prices; the second represents the stocks with mildly fluctuating prices; and the third represents the stocks with ascending prices.

To analyze the sentiment of each tweet, we used VADER [6], which is available from vader-sentiment, a ready-made Python machine learning package for natural language processing. VADER is able to assign sentiment scores to various words and symbols (punctuations), ranging from extremely negative (−1) to extremely positive (+1), with neutral as 0. In this way, VADER could get an overall sentiment score for a whole sentence by combining different tokens’ scores and analyzing the grammar frames. We assumed that the neutral emotion plays a much weaker role in the overall market’s mood since neural sentiment tends to regard the stock market as unchanged in a specific period. Thus, we excluded the neutral sentiment scores in our analysis and only took the negative and positive moods into consideration. VADER places emphasis on the recognition of uppercase letters, slang, exclamation marks, and the most common emojis. Tweet contents are not written academically or formally, so VADER is suitable for social media analysis. However, we needed to add some new words to the original dictionary of VADER, because VADER missed or misestimated some important words in financial world and therefore caused inaccuracy. For example, “bully” and “bullish” are negative words in the VADER lexicon, but they are positive words in the financial market. We updated the VADER lexicon with the financial dictionary Loughran-McDonald Financial Sentiment Word Lists [27], which include many words used in the stock market. This dictionary has seven categories, and we adopted the “negative” and “positive” lists. We further deleted about 400 existing words in VADER that overlap the Loughran-McDonald Financial Sentiment Word Lists. Finally, we added new negative and positive words to the VADER dictionary and attached sentiment scores to them, with +1.5 to each positive word and −1.5 to each negative word from the Loughran-McDonald lists.

To get one day’s overall sentiment score for one stock, we first analyzed all related tweets with VADER and gained the scores of each tweet. Considering that the more followers the bigger influence, we further regarded the number of followers as weights and calculated the weighted average of sentiment scores. The daily percentage change of this average was taken as a comprehensive factor, called compound_multiplied, for 1 day. One problem of our data was that we have only tweets’ data but no price data on the non-trading days. To make full use of the tweets’ data, we filled the gaps up with the price values from past trading days. We utilized moving averages to fill in the missing values.

We also normalized the compound_multiplied variable as neural networks generally perform better and more efficiently on scaled data. Figure 2 shows the box plots of compound_multiplied and per_change (see Section 5.1.2 for details) for the nine stocks before the data was scaled.

FIGURE 2
www.frontiersin.org

FIGURE 2. Left: Box plots of changes in daily sentiment scores aggregated from tweets that contained cashtags of selected stocks over the time period Apr 2–Jun 10, 2016. Outliers are omitted to declutter the graph. The triangles mark the mean values. Right: Box plots of per_change with outliers shown as circles.

3.2 Data Processing for Stock Prices

As discussed before, we collected stock prices from nine stock symbols, each of which contains four columns of time-series price data for each trading day: Open, High, Low, and Close. To change the sequential data into suitable input for our models, we did data cleaning and transformation for the training datasets, and approximate normalization for our both training and testing data. We present further details of these transformations below.

3.2.1 Data Cleaning

In consideration of the volatility of the stock market, we had to tackle the price data on days that showed abnormal trends or fluctuation. To accomplish the anomaly detection, we utilized the Isolation Forest (iForest) algorithm [28] to find such data points before treating them. Isolation Forest can directly describe the degree of isolation of data points based on binary trees without using other quantitative indicators. The main steps of distinguishing outliers were as follows: For a group of data with multiple dimensions, we attempted to build a binary search tree (BST). We first picked several dimensions, and then randomly selected one dimension and a value between the maximum and the minimum in that dimension as the root of the BST. Next, we divided the data into two groups (i.e., subtrees of the BST) according to the selected value. Likewise, we continued to subdivide the data according to another randomly selected value in another feature dimension—this step was repeated until the BST could not be further subdivided. Finally, we detected the anomaly data according to the path lengths of the nodes in the BST.

More specifically, in our work, we adopted only two dimensions in the iForest algorithm and detected the outliers within two steps. (I) We selected High price and Low price as the two dimensions to detect anomaly situations that stock prices fluctuated intensely in one day. Figure 3 shows the anomaly detection results when we applied the iForest to the AAPL dataset. (II) For each one of the four price series (O, H, L, C), we respectively selected the differences ytyt1 and yt+1yt, in which yt was the price data on a particular date t, as the two input dimensions to iForest. In this way, we were able to detect anomaly local trends on the timeline. Figure 4A shows the detection result of ‘Open Price’ from the AAPL dataset, and Figure 4B displays the anomaly in the form of time series. After we identified each outlier yt by considering its distances from yt±1 using iForest, we replaced yt with the rolling average 13(yt2+yt1+yt). As a result, we prevented our model from being excessively influenced by outliers during the training process.

FIGURE 3
www.frontiersin.org

FIGURE 3. Detecting outliers in the dimensions of High price and Low price with Isolation Forest on AAPL-Open in the training dataset. Only four points of outliers were flagged out by the algorithm.

FIGURE 4
www.frontiersin.org

FIGURE 4. (A) Detecting outliers according to (ytyt1) and (yt+1yt) with Isolation Forest using the AAPL-Open training set as an illustration. (B) Anomaly detection results in the form of time series on the same price values in (A).

3.2.2 Data Transformation

We noticed the successful applications of RNNs in wave forms like sinusoids, so we further transformed our time-series datasets into a wave-like form in order to improve the training effect. In our prediction problem, we assumed that the changes in stock data were periodic, which gave us the opportunity to introduce Fourier Transform into our data processing work. Fourier transform can decompose a periodic function into a linear combination of orthogonal functions (such as sinusoidal and cosine functions) [29]. With Fast Fourier Transform (FFT), we could extract both global and local trends of our stock datasets and thus reduce their noise. For each of the four prices (O, H, L, C of a stock dataset), we respectively used FFT to create a series of sinusoidal waves (with different amplitudes and frames) and then combined these sinusoidal waves to approximate the original curve. In Figure 5, we can see that the pink curve, which was made by a combination of 32 components, closely approximates the original price graph of AAPL-Open training set. By using the derived curve, namely, the denoised sequential data, we could enhance the smoothness of our training set, and reduce noisiness in our data. Consequently, we smoothened the time-series data to make our training data more predictable.

FIGURE 5
www.frontiersin.org

FIGURE 5. Fourier transform curves with 5, 15, 25, 30, and 32 components on AAPL-Open training set.

3.2.3 Approximate Normalization

To make sure that our models could learn the features of time-series variables, we also modified the sequential price data to make it satisfy (roughly) normal distribution. In financial research, there is a traditional assumption that the simple daily returns of time-series data are iid normal samples, as long as the sequential data has constant mean and same variance over time; see [30]; for example. We calculated Rt, defined as one-period simple returns of the price values before inputting them into our supervised learning models as targets, where yt represents one of the four prices for a given stock on day t that we are interested to predict:

Rt=ytyt1yt1.(1)

Let R^t be an estimate of Rt returned by a machine learning model trained with predictors including historical returns, Rt1,Rt2,,Rs for some s<t. Subsequently, we can obtain an estimate of the stock price, y^t=yt1(1+R^t).

4 Model Theory

In this section, we present the architecture of our CGAN framework in relation to the sub-networks of our choice.

4.1 Long Short-Term Memory Networks

Introduced by [31], LSTM is a special kind of RNN, which can learn both short-term and long-term correlations effectively from sequential data. An LSTM network usually consists of many repeating modules, the LSTM units, and we can tune the number of the units to improve the performance of the network. These LSTM units concatenate with each other in line with the information transmitting from one to another. Each unit contains three kinds of gates: Forget gate (decides what kind of old information to be discarded), Update gate (decides what kind of new information to be added), and Output gate (decides what to be output). Figure 6 shows how the past information flows through an LSTM unit between two time steps, and how the LSTM unit transforms the data in both its long-term and short-term memory. The information transmission process is from Ct1 to Ct, which will be changed by inner operants. Ct is a latent variable; χt is the input; and yt is the output. The mathematical operations in the three gates are defined as follows, where W. and b. are parameters to be estimated for minimizing some loss functions:

FIGURE 6
www.frontiersin.org

FIGURE 6. Information flow through an LSTM unit.

I. Forget gate:

Ft=sigmoid(Wf[yt1,χt]+bf).(2)

II. Update gate:

Ut=sigmoid(Wu[yt1,χt]+bu),C˜t=tanh(WC[yt1,χt]+bC),Ct=Ft*Ct1+Ut*C˜t.(3)

III. Output gate:

Ot=sigmoid(Wo[yt1,χt]+bo),yt=Ot*tanh(Ct).(4)

Note that sigmoid and tanh are activation functions applied to an input vector elemenwise, whereas * represents elementwise multiplication of vectors. These three gates cooperate with each other and together determine the final information that is output from an individual unit. In this work, we took advantage of the memory property of LSTM and improved its accuracy with adversarial learning framework.

Nowadays, LSTM has been widely applied in many research fields such as machine translation, language modeling, and image generation.

4.2 Adversarial Learning Model

As the conditional GAN framework [13] has proved to be a great success, we adopted this idea in our GAN-based model to improve the training effect. Figure 7 illustrates the architecture of our GAN-based model. Let Xt={Xtn+1,,Xt1,Xt} represent the input data in the form of time series, in which each term is the historical data from the corresponding day from n1 days ago to end of current day t. Note that Xt is a vector that includes all the daily factors on day t. In addition, let Zt be the sentiment label on day t. Our generator is a multiple-input single-output system, which inputs n historical days’ data with m factors from the stock market each day (plus, the label) and outputs a specific kind of future price data (Open, High, Low, or Close) as the prediction. Since the stock data is typical time-series data, we adopted an LSTM network in our generative model denoted by G. LSTM was trained to extract the features of price data from the past n days, and then used to predict stock prices on day t+1. The generator model G generally consists of three kinds of layers: embedding layer, LSTM layer, and fully connected (FC) layer. An embedding layer first embeds a label in the latent space and turns it into a dense vector, then combines the latent vector with the input matrix, which is the price data from the past in our application. Afterward, an LSTM layer generates some prediction result. A dropout layer could be added between two LSTM layers to make the network have a better generalization and less likely to overfit. Lastly, an FC layer with the Leaky Rectified Linear Unit (ReLU) activation function outputs simulated or fake data X˜t+1, which approximates the real target data Xt+1. The output of our generator is defined as follows:

X˜t+1=G(Xt|Zt).(5)

In our work, m=4 and n=5. The four factors were historical Open price, High price, Low price and Close price, which could be seen as four features of the stock data for the price prediction problem. Specifically, we utilized all four kinds of stock prices in the past five business days to forecast each kind of price (O, H, L, or C) on the next day.

FIGURE 7
www.frontiersin.org

FIGURE 7. The architecture of our conditional adversarial learning model.

The purpose of our discriminator D is to classify input data as real or fake. This classification model is expected to output 0 when it receives real input data or 1 when it receives fake input data. For the input, we concatenated Xt with X˜t+1 and Xt+1 respectively to get the fake data X˜t+1={Xtn+1,,Xt,X˜t+1} and the real data Xt+1={Xtn+1,,Xt,Xt+1}. In this way, we could make the discriminator learn the features of sequential data better. Our discriminator D consists of two parts: two embedding layers and a multilayer perceptron (MLP) network. The embedding layers worked in the way same as those in G, i.e., the embedding layers transformed the data and labels into suitable input for MLP. The MLP model then projected the data into higher dimensional space and effectively accomplished the classification task. The Leaky ReLU activation function was used in the hidden layers of the MLP and the sigmoid function was used in the output layer. Returning either 0 and class 1, the output of a discriminator while making a correct decision is as follows:

0=D(Xt+1|Zt),(6a)
1=D(X˜t+1|Zt).(6b)

Specifically, if we select Open price to be predicted by G, we would concatenate the prediction result to Open price values from the past n days. Then, we distinguished this sequence from the corresponding one sampled from real data with D.

We alternatively trained G and D with binary cross-entropy loss (i.e., log loss), L(y˜,y)=ylog(y˜)(1y)log(1y˜), where y is the target value and y˜ is a predicted value. On the one hand, our generator G simulated the fluctuations of the stock market and to minimize the difference between its prediction and the real data. On the other hand, our discriminator D tried to distinguish and maximize that difference. The training of G focused on making the discriminator D confused; it attempted to minimize the adversarial loss so that D could not discriminate the prediction easily. The adversarial loss at each data point on day t+1 for the generator was

Gloss(X˜t+1)=L(D(X˜t+1|Zt),0).(7)

The training of D focused on improving its ability to distinguish the difference between Xt+1 and X˜t+1 , so it attempted to maximize the adversarial loss at each data point on day t+1:

Dloss(Xt+1,X˜t+1)=L(D(Xt+1|Zt),0)+L(D(X˜t+1|Zt),1).(8)

In general, the generator tried to make D(X˜t+1|Zt)=0 while the discriminator tried to achieve D(Xt+1|Zt)=0 and D(X˜t+1|Zt)=1. [7] defined the loss function of this specific binary classification task as cross-entropy loss. To achieve the training process described by Eq. (7) and Eq. (8), we updated D by maximizing Expreal(x)[logD(x|z)]+Ex˜pfake(x˜)[log(1D(G(x˜|z)))], and then updated G by minimizing Ex˜pfake(x˜)[log(1D(G(x˜|z)))]. In this way, the total loss function of the minimax game over all training samples was as follows:

minGmaxDLRMSprop (D,G)=Expreal(x)[logD(x|z)]+Ex˜pfake(x˜)[log(1D(G(x˜|z)))],(9)

where preal(x) was the distribution of real data, pfake(x˜) was the distribution of fake data, and lastly, z represented the label vector.

The two models G and D were trained iteratively with the RMSprop optimizer, a stochastic gradient descent algorithm with mini batches [32]. In each iteration, we first trained the discriminator r times and then trained the generator one time, in which r was seen as a hyper-parameter to be tuned in our training process. The reason for training D first was that a trained discriminator would help improve the training effect of the generator. Besides, we also adopted gradient clipping (ensuring the norm of a gradient not too large in magnitude) for both G and D at the end of each iteration to avoid the potential problem of gradient explosion.

If Zt is omitted from the discussion in this subsection, then we have simply a GAN model. In practice, we used Keras [33] with TensorFlow [34] version 2 as the backend to implement our LSTM, GAN, and CGAN models.

5 Experiment

In this section, we summarize the numerical results of our learning and prediction models in multiple error measures in response to hyperparameter search when applicable.

5.1 Experimental Settings

We collected price data and related tweets of nine companies from April 1st to June 10th, 2016. The data before June 5th was taken as the training set and the last five days’ data as the test set. For each dataset, we respectively experimented on Open price, High price, Low price, and Close price, so we totally built and tested 9×4=36 models and datasets (36 groups of experiments).

We have already illustrated the main data processing work in Section 3. Here, we elaborate on specific configuration of our models.

5.1.1 Baseline Models

Linear Multiple Regression (LMR) is a classical statistical approach for analyzing the relationship between one dependent variable and one or more independent variables. Its simple formula in matrix notation is Y=Xβ+ε, where ε is iid normal with mean 0 and variance matrix, σ2I, and I is the identity matrix of order the same as number of independent observations in Y.

K-Nearest Neighbors (KNN) [35] can be used in both classification and regression problems. In our time-series regression problem, the input is the k closest training sequences from the feature space according to the timeline, and the output is the average of the values of the k nearest neighbors.

Autoregressive Integrated Moving Average (ARIMA) is composed of autoregression (AR), an integrated (I) model that calculates differences, and moving average (MA). Auto ARIMA [36, 37] can automatically perform grid search with parallel processing to find an optimal combination of p, q, and d, which are the parameters associated with order of AR, degree of differencing with I, and order of MA, respectively.

Long Short-Term Memory is a special kind of recurrent neural network. We tuned three hyper-parameters to improve its training effect: the number of LSTM units, step size, and training epochs.

5.1.2 Data Pre-Processing Techniques

Besides Open price, High price, Low price, and Close price, we also added two more engineered factors to the dataset as input to predict the price on day t+1. One was the sentiment variable, compound_multiplied, which was obtained with the method illustrated in Section 3.1. The other was per_change, which was obtained by the equation: per_change :=100%×[Close(t)Open(t)]/Open(t). Table 2 shows the AAPL dataset sample with six columns, where the compound_multiplied was the sentiment variable.

TABLE 2
www.frontiersin.org

TABLE 2. Stock price data and two engineered features on the first five days of the AAPL dataset, where NaN represents Not a Number.

Transformations for baseline models (LMR, KNN, and auto ARIMA) When predicting one type of stock price mentioned above, we utilized the two derived factors, per_change and compound_multiplied, as predictors.

Techniques for neural-network models (LSTM, GAN, and CGAN) (I) As for LSTM, we chose only three factors as input. For instance, we used ‘Open price’, per_change and compound_multiplied from day t as input when we predicted the Open price on day t+1. However, for our adversarial learning model, we chose four kinds of prices as input and utilized the other two factors to create labels. The details about the labels were discussed below. (II) We converted time-series data to data that was fitting in supervised learning problems. Specifically, if we predicted the price on day t+1 with the prices from the past 5 days, we would create a six-term sequence Xt+1={xt4,xt3,xt2,xt1,xt,xt+1}. In other words, we used the first five time-lagged columns as input and the last column as the real target price on day t+1.

5.1.3 Settings for Adversarial Learning Models

Labels We utilized the two variables, compound_multiplied and per_change, to create our final sentiment label in the CGAN models. The reason for adding the per_change was that we would like to take the local trend of past stock variation into account. Finally, the label was set to three classes: 0, 1, and 2. The three classes respectively represent three different tendencies of forecasting prices: up, down, and almost unchanged. Our model was trained to learn the categories from labels with the training set and enhance its prediction accuracy. Let Zt represent the label on day t. The process of creating the labels could be captured by the following min-max scaling of each vt[0,3] before mapping it to Zt{0,1,2}, meaning [0,1) ↦ 0, [1,2) ↦ 1, and [2,3]2:

vt=λ1×compound_multiplied(t)+λ2×per_change(t),(10)
Zt=min(2,3vtmin(vt)max(vt)min(vt)),

where λ1+λ2=1 and 0λ2,λ21. Empirically, tuning λ1 and λ2 could change the training accuracies.

We also did experiments to evaluate the impact of the sentiment analysis had made in our work. Specially, we modified our CGAN models by taking out the embedding layers and not inputting the sentiment labels Zt. Without the labels Zt, we have GAN models. In this way, we could compare the training effect of our models with or without sentiment labels.

Hyper-parameters As discussed in Section 4, we alternately trained the D and the G. The epoch ratio, r, defined as epochs of D to epochs of G, was in the range of [1,4].

The learning rates for D and G, αD and αG, were searched in the range, [5×105,8×104], respectively, with the same decay, 10−8. We also tuned αD and αG in our experiments to improve the training process. Besides, the gradient clipping thresholds were set to be the same at 0.8 for both the generator and the discriminator.

5.2 Results and Discussions

We used mean absolute percentage errors (MAPE) as the main metric to evaluate the performance of our model. The metric equation is

MAPE=100%Nt=1N|yty˜tyt|,

where yt is the real price on day t, y˜t is the prediction price by a model, and N is number of data points. In our experiments, we first trained the CGAN model until the loss curves of both the generator and the discriminator converged. The training process was described in Section 4. After that, we tested the generator “LSTM model” to get the prediction results of testing data. Figure 8 displays the comparison of prediction curves on the AAPL-Open and KHC-Low test sets. As our test sets contained five-day-long data, we calculated the average MAPEs for the five days’ OHLC prices as the performance metric of the model.

FIGURE 8
www.frontiersin.org

FIGURE 8. Comparison of the prediction curves of different models on the AAPL-Open and KHC Low test sets. For AAPL Open, among all models, KNN achieved the minimum MAPE over the five-day test period; see Table 3. However, on each of the last three days, CGAN actually performed better than KNN. For KHC Low, GAN predictions were closest to the market values and CGAN came next.

Empirically, we tuned three kinds of hyper-parameters to improve the performance of our adversarial learning model: the learning rates of the sub-networks, αD and αG; the epoch ratio r; weightage λ1 and λ2 in Eq. (10). Table 3 show the MAPE results of the nine test sets. We did four groups of experiments, respectively on O, H, L, C, for each test set and then calculated the average MAPE to compare the performances of different models effectively. KNN achieved minimum mean MAPE for AAPL, SBUX, CSCO, SWKS, and KHC; linear models for EA and ENDP; GAN and CGAN for NVDA and EBAY, respectively. While KNN achieved minimum errors most of the time, GAN and CGAN were the best for 11 of the 36 stock prices and the second best for 18. In terms of overall average of all nine MAPEs (last column of Table 4), KNN was the best and CGAN came as a close second. Without sentiment labels, GAN models on average had a higher average of all MAPEs than CGAN, showing that our sentiment labels help generally in improving price prediction accuracy.

TABLE 3
www.frontiersin.org

TABLE 3. Tables 3a–i present the MAPEs of forecasted O, H, L, and C prices from various models for nine stock symbols. The smallest MAPE is highlighted in bold across the models in each column for every stock price; the second smallest is underlined. The last column contains the average values of the MAPEs of the four price targets.

TABLE 4
www.frontiersin.org

TABLE 4. Counts number of times a model is the best and second best in predicting the stock prices, and lists the mean of all mean values in the last column of Tables 3a–i.

Root mean square errors (RMSE), mean square errors (MSE), mean absolute errors (MAE), and symmetric mean absolute percentage errors (SMAPE) were also used to verify the training results. Their defining equations are recapped here:

RMSE=1Nt=1N(yty˜t)2,MSE=1Nt=1N(yty˜t)2,MAE=1Nt=1N|yty˜t|,SMAPE=100%Nt=1N|yty˜t|(|yt|+|y˜t|)/2.

To evaluate our model in a more comprehensive view, we selected AAPL and EBAY to illustrate these four metrics. We respectively obtained the results of O, H, L, C from all models and then calculate the average errors to compare the performance more carefully. Table 5 for AAPL shows that KNN performed best on average while CGAN came second. Table 5 for EBAY demonstrates that the CGAN models outperformed all the baselines not only in MAPE but also in RMSE, MSE, MAE, and SMAPE.

TABLE 5
www.frontiersin.org

TABLE 5. Comparison of prediction results in different metrics on the AAPL and EBAY test sets using the average values of model errors on O, H, L, C prices. The smallest errors are displayed in bold font across the models we have built; the second smallest errors are underlined.

As discussed above, we treated λ1 and λ2 as hyper-parameters. Considering the importance of condition labels, adjusting the proportion of sentiment information would be necessary every time we trained the CGAN model. For short-term stock data, we need to train the model only once, since people’s attitude toward a stock often remains more or less the same in a short period. However, readjusting λ1 and λ2 every month maybe useful. We collected another tweet dataset about Apple Inc. with a time span from Jan. 1, 2014 to Dec. 31, 2015 (available from stocknet-dataset by [5]. We selected three periods in these two-year-long datasets and calculated the MAPEs of prediction results for the Open price. In the three groups of experiments, we chose three different (λ1, λ2) pairs in our CGAN model with λ2:=1λ1. We also applied GAN as the baseline to highlight the influence of condition labels. As shown in Table 6, for the first test period, which follows immediately after the time span of the training data, the MAPEs from CGAN and GAN were least (see the first row of each group in Table 6). However, both models performed worse with increasingly larger MAPEs during the subsequent test periods except for GAN in the last training scenario. Thus, the learning models should be trained periodically, or whenever the test errors are larger than user-desired tolerances, especially when they are used in real-time prediction tasks.

TABLE 6
www.frontiersin.org

TABLE 6. Comparison of the MAPEs for AAPL-Open dataset from CGAN and GAN over three different testing periods for each training time span. The smallest error in each test group for each model is highlighted in bold.

6 Conclusions and Future Work

For individual investors and investment banks, rational prediction through statistical modeling helps decide stock trading schemes and increase their expected profits. However, it is well known that human trading decisions are not purely rational and the irrational drivers are often hard to observe. Introducing proxies of irrational decision factors such as relevant online discussion from Twitter followed by sentiment analysis has become an advanced approach in financial price modeling.

In this work, we have successfully built a sentiment-conditional GAN network in which LSTM served as the generator. We encoded Twitter messages, extracted sentiment information from the tweets, and utilized it to conduct our stock prediction experiments. The experiments showed, for about a third of our models, superior properties of our GAN and CGAN models compared to LMR and ARIMA models, KNN method, as well as LSTM networks. Our GAN and CGAN models could better adapt to the variations in the stock market and fluctuation trends of some real price data. Hence they could play a role in ensemble methods that combine strong models with small prediction errors to achieve better accuracies than any individual model when no one model would always perform best; see, for example, [38].

Even though our neural network models could outperform the simpler baseline models at times, it could still be enhanced. The main obstacle in our work was that we failed to find tweets’ dataset that was longer than 70 days, which might largely weaken the training effect. For instance, we initially planned to add more labels to our input data, since it would help represent varying degrees of sentiment tendencies instead of only ‘up’, ‘down’, and ‘almost unchanged’. Nevertheless, the GAN-based models could not learn better with more labels in such a short time-series dataset. Therefore, we finally chose only three classes to make sure that our model would be fully trained. The same consideration went when we created the sentiment variable, which was the reason why we only selected ‘positive’ and ‘negative’ lists when adding new words to the dictionary. If we could get a dataset with a longer time coverage, we would be able to do experiments with sentiment labels in more categories and potentially further improve our models.

Another shortcoming in our current approach is that we only took sentiment factors into account. The stock market is very complicated and there are potentially a great many factors to be considered. There are some other factors like economic growth, interest rates, stability, investor confidence and expectations. We also noticed that political events would have a huge impact on the variation of the stock market. We can extract political information from newspapers and news websites. If we added these factors to the input, our model may better learn the features of the stock market and make the prediction tendencies close to that in the real world.

In this work, we have assumed that the tweets are more or less truthful. However, social media sources could be contaminated with fake news or groundless comments, that are hard to be distinguished from the good ones with real signals. A fruitful area of future research is to look into GAN models for alleviating the problem.

It is also known that neural network models often need much bigger datasets to beat simpler models. To this end, there are multiple ways we could explore: consider more than nine stocks; join all stock prices into one single dataset, i.e., to train one model on panel data grouped by stocks and prices (O, H, L, C) instead of a single time series; sample our datasets at hourly frequency.

Lastly, many properties of neural networks are still active areas of research. For example, in LSTM, GAN, and CGAN models, the loss function values and the solution quality often appear to be sensitive to small changes in inputs, hyperparameters, or stopping conditions. Creating stable yet efficient numerical algorithms are necessary for reliable solutions. In addition, the success of neural networks are often associated with computer vision problems. Adopting them in finance may require different techniques or transformations.

In the future, we would like to explore the topic further in the following directions:

1) Obtaining larger datasets and creating labels in more classes to improve our model by examining hourly data, for example.

2) Building GAN models for detecting overhyped or deceitful messages about a stock before incorporating them into our models.

3) Attempting to extract more stock-related factors and adding them as predictors in our models.

4) Experimenting with more stocks or other financial assets, and considering having one model for all stock price data.

5) Utilizing more sophisticated natural-language processing (NLP) methods to analyze the financial or political information from news media and assessing the impact and the role they play in the stock market.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://data.world/kike/nasdaq-100-tweets.

Author Contributions

YZ: Drafted the manuscript; developed code for data preprocessing and sentiment analysis; designed modeling methods and experiments. JL: Modified the manuscript and cooperated in the design of data processing. HW: Modified the manuscript and cooperated in the design of sentiment analysis. S-CC: Provided directions and supervision on scientific principles and methodology for the whole project; reviewed and modified code for computational scalability and reproducibility; edited manuscript for content accuracy and clarity.

Conflict of Interest

Author S-CC was employed by the company Kamakura Corporation.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

The authors would like to thank the editor, Prof. Qingtang Jiang, and reviewers, Prof. Qi Ye and Prof. Qiang Wu, for their invaluable support and feedback during the anonymous review process. In addition, we are grateful to the following seminar speakers for inspiration during the development of this work in the 2019 IIT Elevate Summer Program’s research course, SCI 498-106 Machine Learning Algorithms on Heterogeneous Big Data: Victoria Belotti, Adam Ginensky, Joshua Herman, Rebecca Jones, Sunit Kajarekar, Hongxuan Liu, Lawrence K. H. Ma, Qiao Qiao, Jagadeeswaran Rathinavel, Aleksei Sorokin, Huayang (Charlie) Xie, and Zhenghao Zhao. In addition, the last author would like to thank the following colleagues for discussion and encouragement: Prof. Robert Ellis, Prof. Robert Jarrow, Jackson Kwan, Wilson Lee, Prof. Xiaofan Li, Prof. Lek-Heng Lim, Aleksei Sorokin, Don van Deventer, and Martin Zorn.

References

1. Devi, KN, and Bhaskaran, VM. Impact of social media sentiments and economic indicators in stock market prediction. Int J Comput Sci Eng Technol (2015). 6.

Google Scholar

2. Zhang, X, Zhang, Y, Wang, S, Yao, Y, Fang, B, and Yu, PS. Improving stock market prediction via heterogeneous information fusion. Knowledge-Based Syst (2018). 143:236–47. doi:10.1016/j.knosys.2017.12.025

CrossRef Full Text | Google Scholar

3. Bollen, J, Mao, H, and Zeng, X. Twitter mood predicts the stock market. J Comput Sci (2011). 2:1–8. doi:10.1016/j.jocs.2010.12.007

CrossRef Full Text | Google Scholar

4. Pagolu, VS, Reddy, KN, Panda, G, and Majhi, B. Sentiment analysis of Twitter data for predicting stock market movements. In: International Conference on Signal Processing, Communication, Power and Embedded System (SCOPES); 2016 Oct 3–5; Paralakhemundi, India. New York, US: IEEE (2016). p. 1345–50.

Google Scholar

5. Xu, Y, and Cohen, SB. Stock movement prediction from tweets and historical prices. In Proceedings of the 56th annual meeting of the association for computational linguistics (Vol. 1: Long Papers); 2018 July; Melbourne, Australia. Stroudsburg, Pennsylvania: Association for Computational Linguistics (2018). 1970–9.

Google Scholar

6. Hutto, CJ, and Gilbert, E. VADER: A parsimonious rule-based model for sentiment analysis of social media text. In Eighth International AAAI Conference on Weblogs and Social Media; 2014 June 1-4; Ann Arbor, MIThe AAAI Press (2014).

Google Scholar

7. Goodfellow, I, Pouget-Abadie, J, Mirza, M, Xu, B, Warde-Farley, D, and Ozair, S. Generative adversarial nets. Adv Neural Inf Process Syst (2014). 2672–80.

Google Scholar

8. Goodfellow, IJ, Shlens, J, and Szegedy, C. Explaining and harnessing adversarial examples. In: Bengio A, LeCun Y, editors. 3rd International Conference on Learning Representations, {ICLR} 2015; 2015 May 7–9; San Diego, CA (2015). Available from: http://arxiv.org/abs/1412.6572.

Google Scholar

9. Kurakin, A, Goodfellow, I, and Bengio, S. (2017). Adversarial Machine Learning at Scale, 5th International Conference on Learning Representations, ICLR 2017, Conference Track Proceedings, Toulon, France, Apr 24–26, 2017, OpenReview.net. Available from: https://openreview.net/forum?id=BJm4T4Kgx

Google Scholar

10. Miyato, T, Dai, AM, and Goodfellow, I. Adversarial Training Methods for Semi-Supervised Text Classification, 5th International Conference on Learning Representations, ICLR 2017, Conference Track Proceedings Toulon, France, Apr 24–26, 2017, OpenReview.net. Available from: https://arxiv.org/abs/1605.07725

Google Scholar

11. He, X, He, Z, Du, X, and Chua, TS. Adversarial personalized ranking for recommendation. In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval; 2008 June. New York, US: ACM (2018). p. 355–64.

Google Scholar

12. Li, C, Wang, S, Wang, Y, Yu, P, Liang, Y, and Liu, Y. Adversarial learning for weakly-supervised social network alignment. In: Proceedings of the AAAI Conference on Artificial Intelligence; 2019 Jan 27-Feb 1; Honolulu, Hawaii. Menlo park, California: AAAI (2019). 33:996–1003. doi:10.1609/aaai.v33i01.3301996

CrossRef Full Text | Google Scholar

13. Mirza, M, and Osindero, S. Conditional generative adversarial nets. CoRR abs/1411.1784 (2014). Available from: http://arxiv.org/abs/1411.1784

Google Scholar

14. Antipov, G, Baccouche, M, and Dugelay, JL. Face aging with conditional generative adversarial networks. In: IEEE International Conference on Image Processing (ICIP); 2017 Sept 17–20; Beijing, China. New York, US: IEEE (2017). p. 2089–93.

Google Scholar

15. Yang, Z, Chen, W, Wang, F, and Xu, B. Improving neural machine translation with conditional sequence generative adversarial nets. CoRR abs/1703.04887 (2017). Available from: http://arxiv.org/abs/1703.04887

Google Scholar

16. Yu, L, Zhang, W, Wang, J, and Yu, Y. SeqGAN: Sequence generative adversarial nets with policy gradient. In: Thirty-First AAAI conference on artificial intelligence. New York, US: ACM (2017). 2852–8.

Google Scholar

17. Mogren, O. C-RNN-GAN: Continuous recurrent neural networks with adversarial training. Constructive Machine Learning Workshop at NIPS 2016 in Barcelona, Spain. Available from: https://arxiv.org/pdf/1611.09904.

Google Scholar

18. Esteban, C, Hyland, SL, and Rätsch, G. Real-valued (medical) time series generation with recurrent conditional GANs. CoRR (2017). Available from: http://arxiv.org/abs/1706.02633.

Google Scholar

19. Zhu, F, Ye, F, Fu, Y, Liu, Q, and Shen, B. Electrocardiogram generation with a bidirectional LSTM-CNN generative adversarial network. Sci Rep (2019). 9:6734. doi:10.1038/s41598-019-42516-z

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Wang, Z, Chai, J, and Xia, S. Combining recurrent neural networks and adversarial training for human motion synthesis and control. IEEE Trans Visualization Comput Graphics (2019). 27:14–28. doi:10.1109/TVCG.2019.2938520

CrossRef Full Text | Google Scholar

21. Zhou, X, Pan, Z, Hu, G, Tang, S, and Zhao, C. Stock market prediction on high-frequency data using generative adversarial nets. Math Probl Eng (2018). 2018:1–11. doi:10.1155/2018/4907423

CrossRef Full Text | Google Scholar

22. Wang, JJ, Wang, JZ, Zhang, ZG, and Guo, SP. Stock index forecasting based on a hybrid model. Omega (2012). 40:758–66. doi:10.1016/j.omega.2011.07.008

CrossRef Full Text | Google Scholar

23. Huang, W, Nakamori, Y, and Wang, S-Y. Forecasting stock market movement direction with support vector machine. Comput Operations Res (2005). 32:2513–22. doi:10.1016/j.cor.2004.03.016

CrossRef Full Text | Google Scholar

24. Chong, E, Han, C, and Park, FC. Deep learning networks for stock market analysis and prediction: methodology, data representations, and case studies. Expert Syst Appl (2017). 83:187–205. doi:10.1016/j.eswa.2017.04.030

CrossRef Full Text | Google Scholar

25. Adebiyi, AA, Adewumi, AO, and Ayo, CK. Comparison of ARIMA and artificial neural networks models for stock price prediction. J Appl Math (2014). 2014:1–7. doi:10.1155/2014/614342

CrossRef Full Text | Google Scholar

26. Nelson, DM, Pereira, AC, and de Oliveira, RA. Stock market’s price movement prediction with LSTM neural networks. In: International Joint Conference on Neural Networks (IJCNN); 2017 May 14–19; Anchorage, AK, USA. New York, US: IEEE (2017). p. 1419–26.

Google Scholar

27. Loughran, T, and McDonald, B. The use of word lists in textual analysis. J Behav Finance (2015). 16:1–11. doi:10.1080/15427560.2015.1000335

CrossRef Full Text | Google Scholar

28. Li, FT, Ting, KM, and Zhou, ZH. Isolation-based anomaly detection. ACM Trans Knowledge Discov Data (TKDD) (2012). 6:3. doi:10.1145/2133360.2133363

CrossRef Full Text | Google Scholar

29. Bracewell, RN. The Fourier transform and its applications. New York: McGraw-Hill (1986).

30. Tsay, RS. Analysis of financial time series. New York, US: John Wiley & Sons (2005). Vol. 543.

31. Hochreiter, S, and Schmidhuber, J. Long short-term memory. Neural Comput (1997). 9:1735–80. doi:10.1162/neco.1997.9.8.1735

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Tieleman, T, and Hinton, G. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. COURSERA: Neural Networks Machine Learn (2012). 4:26–31.

Google Scholar

33. Chollet, F. Keras (2015). Available from: https://keras.io.

34. Abadi, M, Agarwal, A, Barham, P, Brevdo, E, Chen, Z, and Citro, C. TensorFlow: large-scale machine learning on heterogeneous distributed systems (2015). Available from: tensorflow.org.

Google Scholar

35. Cover, T, and Hart, P. Nearest neighbor pattern classification. IEEE Trans Inform Theor (1967). 13:21–7. doi:10.1109/tit.1967.1053964

CrossRef Full Text | Google Scholar

36. Hyndman, RJ, and Khandakar, Y. Automatic time series forecasting: The forecast Package for R. J Stat Soft (2008). 27:1–22. doi:10.18637/jss.v027.i03

CrossRef Full Text | Google Scholar

37. Smith, TG. pmdarima: ARIMA estimators for Python (2017).

38. Clemen, RT. Combining forecasts: A review and annotated bibliography. Int J Forecast (1989). 5:559–83. doi:10.1016/0169-2070(89)90012-5

CrossRef Full Text | Google Scholar

Keywords: stock prediction, data processing, labels, regression, long short-term memory, sentiment variable, conditional generative adversarial net, adversarial learning

Citation: Zhang Y, Li J, Wang H and Choi S-CT (2021) Sentiment-Guided Adversarial Learning for Stock Price Prediction. Front. Appl. Math. Stat. 7:601105. doi: 10.3389/fams.2021.601105

Received: 31 August 2020; Accepted: 01 February 2021;
Published: 26 May 2021.

Edited by:

Qingtang Jiang, University of Missouri–St. Louis, United States

Reviewed by:

Qi Ye, South China Normal University, China
Qiang Wu, Middle Tennessee State University, United States

Copyright © 2021 Zhang, Li, Wang and Choi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Sou-Cheng T. Choi, schoi32@iit.edu

Download