This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Novel Distributed Representation of News (DRNews) for Stock Market Predictions thanks: We acknowledge the support by 2016 Jiangsu Science and Technology Programme: Young Scholar Programme (No. BK20160391)

1st Ye Ma Department of Financial Mathematics
School of Science
Xi’an Jiaotong-Liverpool University
Suzhou, China
[email protected]
   2nd Lu Zong Department of Financial Mathematics
School of Science
Xi’an Jiaotong-Liverpool University
Suzhou, China
[email protected]
   3rd Peiwan Wang Department of Financial Mathematics
School of Science
Xi’an Jiaotong-Liverpool University
Suzhou, China
[email protected]
Abstract

In this study, a novel Distributed Representation of News (DRNews) model is developed and applied in deep learning-based stock market predictions. With the merit of integrating contextual information and cross-documental knowledge, the DRNews model creates news vectors that describe both the semantic information and potential linkages among news events through an attributed news network. Two stock market prediction tasks, namely the short-term stock movement prediction and stock crises early warning, are implemented in the framework of the attention-based Long Short Term-Memory (LSTM) network. It is suggested that DRNews substantially enhances the results of both tasks comparing with five baselines of news embedding models. Further, the attention mechanism suggests that short-term stock trend and stock market crises both receive influences from daily news with the former demonstrates more critical responses on the information related to the stock market per se, whilst the latter draws more concerns on the banking sector and economic policies.

Index Terms:
News network, news embedding, attributed network embedding, Subword model, stock movement prediction, crises warning

I Introduction

The stock market, one of the most knowledge-driven place, absorbs information from all channels and releases instant responses through the price oscillation. As the major source of public information, news is considered to be a type of textual data that is voluminous and continuously updated, incorporating vast interactive information between news and various labels from manual annotation or advanced natural language processing (NLP) models. Regarded as one of the driving factors in the stock market, news places major influences on the price evolution in various aspects including the domestic/global economy and financial situations as well as the investors’ sentiment. With the aim of facilitating the performance of news-driven stock market predictions, this study proposes a novel Distributed Representation of News (DRNews) model that embeds news as continuous vectors that describe the collective information of its inherited features as well as the inter-textual knowledge across different news articles.

Attention has been drawn to the recent literature on textual representations of news articles and their applications to the deep predictive network of stock movements. For example, the Paragraph Vector model [23] is proposed to train news embeddings that facilitate the decision-making process of stock trading strategies [1]. Further, Ding et al. (2015) [12] suggest to extract event tuples from headlines for news event embedding and associated event-driven stock prediction, whilst Hu et al. (2018) [22] use a deep learning framework for news-oriented stock trend prediction by representing news as the average of word vectors. In addition to those unsupervised learning methods, DeepClue [39] represents news in a supervised manner by constructing an end-to-end stock price forecasting model using a neural network as the news encoder and words in the headline as the input. Instead of using solely textual information to predict stock trends, some studies incorporate additional news features, such as news sentiment [37, 6, 40], public mood [26] and news views [33], into their models for greater reliability and robustness in the predictions. Although the above-mentioned news embedding approaches share the merit that similar news are represented with similar vectors, their news embeddings are limited to the representation of the literal information brought by the news article of interest, rather than the inter-textual knowledge and cross-documental linkages inherited from the news corpus. According to Long’s study (2019) [29] that uses the graphic kernel method to predict stock movements based on financial news semantic and structural similarities, it is found that both the content and structure information of news contribute to the prediction with higher weights assigned to the structure information. Moreover, no mechanism has been developed in the existing literature to project additional news features simultaneously into the same vector space of news texts in order to enhance the integrity of the embedding with both contextual and exogenous information.

In this paper, a Distributed Representation of News (DRNews) model is developed to incorporate the structural linkage between news in addition to a variety of news features. The novelty of DRNews is mainly reflected in two aspects. First, distributed relations between different news are represented in the form of a network so that news articles are embedded with both the semantic and inter-textual knowledge (See Figure 1 for a simplified example of the DRNews network). In particular, different from other established text/sentence embedding models that focus on document-level relationships, such as coherence among sentences [20, 24, 27] or the subordination of words to text [23], the proposed DRNews embeddings are trained based on the cross-documental relationship between news stories. According to Figure 1, the three pieces of news Ali strategically stakes in Focus with 15 billion, innovating digital marketing together. & Focus Media limits up today with five institutes jointly selling more than 70 million. & Overseas buying rises from institutes’ game on Focus Media. are connected by sharing the same contextual element Focus Media. The verbal elements, such as stake in, limits up, rises, of the three different news are thus linked through their common nodes and become the latent contextual elements of each other. In this way, the DRNews network describes not only the co-existence of news and its own elements, but also the potential linkage among news through its connected elements. Second, the Subnode model, inspired by the Subword model [2], is proposed to treat each news node as a bag of node attributes (i.e. news features including semantic features, category features, time features, etc.), thus the news vector is computed as the sum of its feature vectors. In the framework of the biased random walk [15] and the skip-gram architecture [31], the Subnode model enriches news embeddings with additional news features, at the same time allowing fast inferences on the embedding of unseen news based on its feature vectors.

Refer to caption
Figure 1: A simplified example of DRNews network

The major contribution of this study is threefold. First, the DRNews model is developed and applied for two news-driven stock market prediction tasks, i.e. the short-term movement prediction and the stock crises warning, in comparison with other five embedding models. As the results show, DRNews achieves state-of-art performances in both tasks demonstrating its effectiveness in capturing more useful information than the baselines. In particular, a test-set accuracy of 46.24%46.24\% is achieved in the ternary prediction task of stock movement, whilst a test-set accuracy of 94.61%94.61\% and an average of 2.52.5-days forewarned period are obtained in the early warning task of stock crises. Second, the Subnode mechanism offers a straightforward solution to enrich node vectors with node attributes and compute embeddings of unseen nodes quickly. Last, the attention-based LSTM (Long Short-Term Memory) [21] network is adopted as the predictive model, which explicitly identifies the distinguishing dynamics of short-term stock movement and stock crises warning in response to news of different topics and keywords.

II Related work

II-A News embedding

To represent news articles as vectors, two possible ways are emerged. As news articles are essentially textual data, one could use directly text/sentence embedding models to embed news. Alternatively, news event embedding models are developed with the argument that news stories describe events.

In terms of Text/Sentence Embedding, one of the simplest but robust baseline model [8] is to perform numerical operations on the existing contextual word vectors [32]. SDAE [42] uses an auto-encoder to project text into a low-dimensional vector space, whilst Paragraph Vector [23] learns the distributed representation of sentences and documents by predicting their contexts, i.e., words in the sentence/document. Other unsupervised models include Skip-Thought [24], FastSent [20] and Quick-Though [27] and learn distributed sentence-level representations regarding to the coherence among sentences across the text. BERT [10] is a powerful pre-trained sentence encoder trained on unsupervised data sets based on the masked language model and next-sentence prediction. As for supervised learning models, most studies use a sentence encoder such as LSTM or Transformer [41] to convert word vectors to sentence vectors. The sentence encoder is usually trained on NLP tasks, which could be a single supervised learning task, such as InferSent [7] which uses only the language inference data, or multiple tasks [4] which involve both unsupervised and labeled corpus.

On the other hand, Event Embedding models emphasize the mining of event tuples and their projections onto the vector space. Ding et al.’s Event Embedding model [12, 13] extracts event tuples E=(Actor,Predicate,Object)E=(Actor,Predicate,Object) from headlines and learn dense vector representations by scoring correct tuples higher than corrupted tuples. Basically, the event embedding model is built on the knowledge graph embedding after decomposing the knowledge graph into tuples of entities, relations, etc. Following Ding et al.’s idea, news events could then be embedded with other knowledge graph embedding models such as TransE [3], which exploits scoring functions based on the distance between h+rh+r and tt. In addition to knowledge graph embedding, Event2vec [38] employs network embedding to learn distributed representations of summarized news events by proposing an event network where each news event node is connected to the associated entities, the event type and the year of the event. The major limitations of Event2vec are threefold. First, it only works for artificially organized database of already extracted events. Second, the heterogeneous network it constructs raises difficulties in determining the weights of element nodes of different types which leads to further imposed assumptions on the importance of the elements. Third, it lacks a solution to represent unseen events based on the trained network.

In addition to converting news into continuous vectors, structured Event Extraction from news also allows statistical analysis based on the event type [19, 30]. In the early studies of news data mining, the textual information of news is often regarded as high-dimensional term vectors [36, 43, 45], which are then considered to lead to a great loss of information [25].

II-B Stock market prediction based on deep learning models

As the central pillar of the financial system, stock market demonstrates great significance as a major channel of capitalization and maintaining liquidity. The applications of deep learning models in stock market predictions have become an emerging research field due to the rapid development and maturation of relevant techniques. With the merit of high feature learning capacity, deep learning enhances the traditional machine learning models and achieves higher level of accuracy as the volume of data increases.

In the field of stock market prediction, stock trend prediction, return/volatility prediction and stock crisis warning are the three mainstream tasks that attract the most attention. Among the deep learning models that have been adopted in these tasks, classic neural networks, such as long short-term memory (LSTM) [35, 14, 44] and convolutional Neural Networks (CNN) [1, 12] are proven to be the most reliable and widely-adopted techniques handling financial time series [11]. Due to recent advances in textual feature engineering which allows the extraction of public events and investor sentiments, integrated predictive methods, such as ensemble learning and attention-based neural networks, have also gained significant popularity in the field. For instance, Chatzis et al. (2018) [5] develop a stacked model combining classification trees, support vector machines, random forests, neural networks, extreme gradient boosting, and deep neural networks to forecast crisis episodes in stock, bond and currency markets. Long et al. (2019) [28] develop an end-to-end model, multi-filters neural network (MFNN), to implement feature extraction and price movement prediction. As one of the core techniques in deep learning, the attention mechanism learns the data in the way that allows the machine to assign heavier weights to the more important features in order to improve the forecasting power of the model. Hu et al. (2018) [22] apply the Hierarchical Attention Networks [46] on the stock trend prediction, whilst Vaswani et al. (2017) [47] propose the Self-attention Networks, inspired by Transformer [41], for the stock volatility prediction.

This study applies an attention-based LSTM to perform two news-driven stock prediction tasks, i.e. stock trend prediction and crisis early warning. With the established predictive model, predictions obtained from the DRNews embeddings and five baseline embedding models are compared and discussed. The primary objectives are (I) to evaluate the effectiveness of DRNews presentations against other news embedding paradigms in the context of news-driven stock market prediction, at the same time (II) to extract key information that drives stock market evolution from the news corpus.

In remaining part of the paper is organized as follows. Section III describes the formation of the proposed DRNews model in terms of the construction of the news network and the implementation of news embeddings using the Subnode model. Section IV provides an overview of the attention-based LSTM model for stock predictions. In Section V, the two stock market prediction experiments are presented. And Section VI summarizes the conclusion.

III The DRNews Model

In this section, the formation of the DRNews model is discussed. Figure 2 shows the overall flow chart of creating news embeddings with the DRNews model. With the text data of news articles as inputs, we first extract the key elements of news according to their term frequency-inverse document frequency (tf-idf) scores. The event elements are then collected as the news features together with some labeled features, such as publish date, words count, type and sentiment polarity, that are obtained from either directly the database or a trained classifier if there is a lack of ready-made labels. Based on these news elements and their tf-idfs, the news network is thus built whereas node sequences are sampled by the biased random walk. News nodes in these sequences are then represented as bags of extracted news attributes, a.k.a. features, and associated vectors are assigned to each features using the Subnode model.

Refer to caption
Figure 2: Flow chart of the DRNews model

In the rest of this section, a more explicit explanation is given in terms of the news network construction and the Subnode model for attributed network embedding.

III-A Constructing the news network

The DRNews model determines news embeddings based on an attributed news network, which is constructed by connecting the news node with its element nodes and news features as the news node attributes. With respect to the argument that a piece of news often describes one or more events that could be represented as a collection of event elements, the DRNews model extracts news elements as entities, actions and nouns by ranking the words/phrases by their tf-idf scores and select those with the 1%1\% highest tf-idfs. To comprise a news event, an entity is usually a person’s name, an organization, or a location, whilst an action is often the trigger of an event at the same time contains key verbs that possibly reflect the type of the event. To enrich the news representation and address the problem that some verbs are often used as gerunds, we also consider nouns with high tf-idf scores as news elements. The tf-idf score is expressed as:

tfidfe,k=ne,kwnw,k×log|D||k:edk|tfidf_{e,k}=\frac{n_{e,k}}{\sum_{w}n_{w,k}}\times\log\frac{\left|D\right|}{\left|{k:e\in d_{k}}\right|} (1)

where ne,kn_{e,k} denotes the number of the news element ee in the news story kk and its denominator is the total number of words/phrases in the news document, |D|\left|D\right| is the total number of documents, the denominator represents the number of stories containing the element ee. Overall, the tf-idf assigns a higher score to the element which appears more frequently in the news story than in others.

Figure 1 gives a glance on the fundamental structure of the DRNews news network. Specifically, news nodes kVkk\in V_{k} are represented by their titles and the published time with no edges connecting any pair of element nodes eVee\in V_{e} or news nodes. The weights of edges (k,e)(k,e) are assigned based on the importance of the event element to the news. The tf-idf is then used to measure the importance of a word/phrase inside the document and to determine the weight of the edge (k,e)(k,e):

Wk,e=tfidfe,k(title)Z1+tfidfe,k(content)Z2W_{k,e}=\frac{tfidf_{e,k}(title)}{Z_{1}}+\frac{tfidf_{e,k}(content)}{Z_{2}} (2)

The weight is simply the summation of the normalized tf-idfs of the event elements in the news title and text. Since there are fewer event elements in the title, the normalized weights tend to be higher than those in the content. This result is also true because headlines often contain the main information of an entire story.

III-B News network embedding with the Subnode Model

III-B1 Network embedding by random walks

To determine the news vectors, our network embedding approach shares a similar idea with the Node2vec model [15]. The goal is to learn a latent feature vector F(v)F(v) for each node vVv\in V, that maximizes the probability of predicting the node vsv^{\prime}s network neighborhood N(v)N(v). Hence, the objective function is written as:

maxvVlogPr(N(v)|F(v))max\sum_{v\in V}\log Pr(N(v)|F(v)) (3)

Given the feature vector of the node vv, it is assumed that the prediction probabilities of its neighborhoods are independent of one another, so

logPr(N(v)|F(v))=uN(v)logPr(u|F(v))logPr(N(v)|F(v))=\sum_{u\in N(v)}logPr(u|F(v)) (4)

where network neighborhoods N(v)N(v) for each node are sampled by the biased random walk and concatenated into sequences. These sequences are processed by extending the skip-gram model [31] from dealing with word sequences to node sequences. Specifically, negative sampling [32] is used to approximate the conditional probability:

logPr(u|F(v))logPr(1|F(v),u)+i=1m𝐄uiP(v)[logPr(0|F(v),ui)]\begin{split}logPr(u|F(v))\approx logPr(1|F(v),u)+\\ \sum_{i=1}^{m}\mathbf{E}_{u^{{}^{\prime}}_{i}\sim P(v)}[logPr(0|F(v),u^{{}^{\prime}}_{i})]\end{split} (5)

where uu is sampled from the context of node vv in the sequence while uu^{{}^{\prime}} (uuu^{{}^{\prime}}\neq u) are sampled mm times from the noise distribution P(v)P(v), which is typically the Unigram Distribution. Each sampling is treated as a simple binary classification to be solved by the corresponding logistic regression classifier.

Different from uniform random walks of DeepWalk [34], Node2vec introduces two classic search strategies, namely the Breadth-first Search (BFS) and Depth-first Search (DFS). Suppose that the walk just transitioned from tt to vv and is now walking to its next step node xx, in our network, there are two possible situations for the next step. One is revisiting tt, i.e., dtx=0d_{tx}=0, whereas the other is keeping going deeper and further away from the node tt, i.e., dtx=2d_{tx}=2. Node2vec defines a search bias α\alpha to control the search preference:

αvx={1p,ifdtx=01q,ifdtx=2\alpha_{vx}=\left\{\begin{array}[]{lr}\frac{1}{p},&if\ d_{tx}=0\\ \frac{1}{q},&if\ d_{tx}=2\end{array}\right. (6)

Next, the transition probability of edge (v,x)(v,x) is defined as the product of the weight of the edge and the bias, i.e., αvxwvxZ\frac{\alpha_{vx}*w_{vx}}{Z} where ZZ is the normalization constant. If p<qp<q, the walk is width-first associated with a greater probability to revisit the original node. Therefore, it is easier to sample nodes around the original node, that is, semantically similar news as they share many same news elements. If p>qp>q, the walk is depth-first, and tends to visit nodes that have not been passed before. As such it is more likely to explore news with latent relations. During random walks, each node is used as a starting vertex for a fixed-length walk, producing a group of sequences. Based on the skip-gram model, node representations are finally trained on these sequences by optimizing the objective function using the Stochastic Gradient Descent (SGD).

III-B2 The Subnode model

To address the major limitations that Node2vec can neither infer unseen nodes easily nor integrate node attributes into node representations, we propose a Subnode model to embed the DRNews network.

Essentially, the Subnode model is inspired by the Subword model [2], which represents each word using a bag of character nn-grams. As an example, the word wherewhere is represented as G=<wh,whe,her,ere,re>G=<wh,whe,her,ere,re>. Based on the skip-gram architecture, each character nn-gram can learn a vector representation and the word vector is the summation of its character nn-grams.

By the biased random walk, sequences of news and event element nodes (such as newsAelementAnewsBelementBnews_{A}\to element_{A}\to news_{B}\to element_{B}\to\cdots) are obtained. Similar to Subword, the proposed Subnode rewrites each node as a bag of node attributes. With the focus on the news nodes, the node attributes are news features including semantic features such as event elements with high tf-idf scores, text structure features such as words count and paragraphs count and side information such as published date, news type and news emotion. Each news vertex eVee\in V_{e} is thus represented as:

G=<entityA,entityB,action,,G=<entity_{A},entity_{B},action,\cdots,
wordscount,paragraphcount,words\ count,paragraph\ count,
month,day,week,type,sentiment,>month,day,week,type,sentiment,\cdots>

where words count, paragraph count and sentiment are ordinal data rather than real values. Similarly, the representation of a news node no longer directly corresponds to a distinct vector, but is instead the sum of the vectors of all its features. The objective function is now updated to:

maxkVklogPr(N(k)|gGkF(g))max\sum_{k\in V_{k}}\log Pr(N(k)|\sum_{g\in G_{k}}F(g)) (7)

where F(g)F(g) denotes the vector representation of a news feature gg. It is worth-mentioning that nodes with only one edge are removed to reduce the sparsity of the network. The formulation of Eq. (7) is similar to Eq. (4), (5) and the vector representations of news features are trained by optimizing the updated objective function using SGD with negative sampling [32]. The pseudo-code of the Subnode algorithm is given in Algorithm 1.

In this way, the Subnode representation of unseen news nodes can be computed easily by extracting its news features, looking up for the corresponding vectors in the featuretovectorfeature\ to\ vector dictionary, and summing the vectors up. The algorithm code is open source and avaiblable at Github111https://github.com/yema2018/News2vec.

Input: SS, the node sequence S=[k1,e1,k2,e2,]S=[k^{1},e^{1},k^{2},e^{2},\cdots] generated by random walks
Output: FF, the embedding matrix with respect to news feature gg
1 Rewrite the sequence as S=[(g1,g2,),e1,(g3,g4,),e2,]S^{{}^{\prime}}=[(g_{1},g_{2},\cdots),e^{1},(g_{3},g_{4},\cdots),e^{2},\cdots];
2 Initialize FF;
3 for ss in SS^{\prime}, ses\neq e do
4      for uu in Context(s)Context(s) do
5            NegativeSamplinguNegativeSampling\ u^{\prime} from S/{s}S^{\prime}/\{s\};
6             J(F)=logPr(u|gGsF(g))J(F)=-logPr(u|\sum_{g\in G_{s}}F(g));
7             F:=FηtialJtialFF:=F-\eta*\frac{tialJ}{tialF}
8       end for
9      
10 end for
Return FF
Algorithm 1 Subnode

III-C Visualization of DRNews embeddings: dimension reduction plots

To give a more straightforward view on the DRNews embeddings, we train news vectors on the THUCTC222http://thuctc.thunlp.org (THU Chinese Text Classification) news data set with one labeled feature, namely typetype, beside the semantic features. There are 14 types of news in the news corpus and we randomly take 6000/1000 pieces of news from each type to make up the training set/test set. Based on the news training set, a vector representation of 128 dimensions is learned for each news feature. Figure 3 shows the dimension reduction plots of news vectors in the test set using the t-Distributed Stochastic Neighbor Embedding (t-SNE).

Refer to caption
Figure 3: Dimension reduction plots of test set news without and with the type feature

The right panel of Figure 3 shows the news vectors trained with the type feature, in which clustering patterns are distinctly evident, suggesting that DRNews allows the embedding of the type information into news vectors after incorporating the corresponding feature. Furthermore, it is observed that the distance between two clusters is positively correlated with the actual relevancy between the topics. In particular, neighboring relationships are detected between pairs of related topics, such as sports & lottery ticket, current affairs & society, stock & finance, and fashion & entertainment, whilst the distances between irrelevant clusters, such as constellation & finance, game & education, are larger. Nevertheless, the left panel of Figure 3 shows that news embeddings significantly lose classification information when vectors of type features are removed from the news vector. In a nutshell, the dimension reduction visualization of news vectors shows that our DRNews embedding model enables news vectors to represent multiple attributes simultaneously.

IV The Deep Predictive Model

This section discusses the deep predictive model that is used in the stock market prediction tasks by mining information from past news and market indicators. Since both news and market indicators can be viewed as daily time series, so a sequence model based on the long short-term memory (LSTM) network [21]) network is employed to encode the time series and to learn the effect of past news and market data on the future market. As different news stories tend to have different relevance to the market, an attention mechanism [46] is adopted to assign a weight to each news automatically by gradient descent. The prediction structure is shown in Figure 4 as a bi-input neural network for two types of data, namely news texts and the numeral data of historical returns. In particular, the news data is encoded by the attention-based LSTM, while the market data is encoded by a sole LSTM model. Their outputs at the final time step are concatenated and fed into the output layer.

Refer to caption
Figure 4: The deep predictive model

Attention model: To make news-driven stock predictions, each news vector nt,in_{t,i} is first fed into a single-layer network, whose outputs ut,iu_{t,i} are normalized by a softmax activation function and weighted (αt,i\alpha_{t,i}) by the corresponding news. Specifically,

ut,i=tanh(Wnnt,i+bn)u_{t,i}=tanh(W_{n}n_{t,i}+b_{n}) (8)
αt,i=exp(ut,iTuw)jexp(ut,jTuw)\alpha_{t,i}=\frac{exp(u_{t,i}^{T}u_{w})}{\sum_{j}{exp(u_{t,j}^{T}u_{w})}} (9)
dt=iαt,int,id_{t}=\sum_{i}{\alpha_{t,i}n_{t,i}} (10)

where WnW_{n} are initialized at the beginning and trained by back propagation. After determining weights of all the news, the daily news vector dtd_{t} is computed as the weighted sum of all the news vectors in that day, integrating the daily news information.

LSTM model: The LSTM model is a variant of the recurrent neural network that uses updating Γu\Gamma_{u} and forgetting Γf\Gamma_{f} factors to update the memory cell c<t>c^{<t>}. Specifically,

c~<t>=tanh(Wc[a<t1>,x<t>]+bc)\widetilde{c}^{<t>}=tanh(W_{c}[a^{<t-1>},x^{<t>}]+b_{c}) (11)
Γu=sigmoid(Wu[a<t1>,x<t>]+bu)\Gamma_{u}=sigmoid(W_{u}[a^{<t-1>},x^{<t>}]+b_{u}) (12)
Γf=sigmoid(Wf[a<t1>,x<t>]+bf)\Gamma_{f}=sigmoid(W_{f}[a^{<t-1>},x^{<t>}]+b_{f}) (13)
Γo=sigmoid(Wo[a<t1>,x<t>]+bo)\Gamma_{o}=sigmoid(W_{o}[a^{<t-1>},x^{<t>}]+b_{o}) (14)
c<t>=Γu×c~<t>+Γf×c<t1>c^{<t>}=\Gamma_{u}\times\widetilde{c}^{<t>}+\Gamma_{f}\times c^{<t-1>} (15)
a<t>=Γo×tanh(c<t>)a^{<t>}=\Gamma_{o}\times tanh(c^{<t>}) (16)

where a<t1>a^{<t-1>} is both the output at the time step t1t-1 and one of the inputs of the next LSTM cell. Another input of the LSTM cell is possibly x<t>x^{<t>} which is an integration dtd_{t} of news set {nt,1,,nt,N}\{n_{t,1},\cdots,n_{t,N}\} for the news-driven LSTM, or a vector ItI_{t} of historical returns for the market-driven LSTM. In addition to the output of a<:>a^{<:>}, the memory factor c<:>c^{<:>} is passed through the LSTM cells along with the output being constantly updated to ensure that the most important information is retained.

Training: Encoded News and market data are integrated by a concatenation layer and fed into an output layer with JJ neurons. Note the number JJ depends on how many classifications a task needs to discriminate. We train the JJ classifier by optimizing the cross entropy loss with the Adam optimization algorithm,

L(y^,y)=j=1Jyjlogy^j+λQ22L(\hat{y},y)=-\sum_{j=1}^{J}{y_{j}log\hat{y}_{j}}+\lambda\left\|Q\right\|_{2}^{2} (17)

where the QQ is the set of all trainable parameters and λ\lambda is the L2L_{2} regularization weight. Through the concatenation, parameters from both news-driven and market-driven model are updated simultaneously by the gradient descent.

Performance measures: To investigate the forecasting power of the attention-based LSTM model, two performance measures, i.e. the Accuracy (Acc) and the Matthews Correlation Coefficient (MCC), are hired. In particular, the accuracy of a binary predictive model takes value in the interval between 0 and 11 and is computed as:

Acc=TP + TNTP + TN + FP + FN,\text{Acc}=\frac{\text{TP + TN}}{\text{TP + TN + FP + FN}}, (18)

where TP = True positive, FP = False positive, TN = True negative, FN = False negative. As a result, models with higher values of accuracy are preferred as a higher level of predicting power is indicated. On the other hand, the MCC measure is adopted as the the quality measure of the predictive model, which returns the correlation coefficient between the observed and predicted classifications:

MCC=TP×TNFP×FN(TP + FP)(TP + FN)(TN + FP)(TN + FN).\text{MCC}=\frac{\text{TP}\times\text{TN}-\text{FP}\times\text{FN}}{\sqrt{\text{(TP + FP)(TP + FN)(TN + FP)(TN + FN)}}}. (19)

The MCC measure is considered as one of the most reliable criteria in the selection of classification models for its strength in dealing with imbalanced data. The value of the MCC ranges from 1-1 to 11, where a perfect classification model is indicated by 11, and a completely wrong model is indicated by 1-1.

V Experiments

V-A Stock Movement Prediction

V-A1 Experimental settings

We use financial news (2009/10/19 to 2016/10/31) from Sohu333https://www.jianshu.com/p/370d3e67a18f to predict the daily movement of Shanghai Securities Composite (SSE) index.

To obtain the news embeddings, vectors of the news features are first obtained according to the news network from 2009/10/19 to 2015/12/31 whereas news vectors are then computed by summing up its associated feature vectors. In particular, the news network is trained by setting the length of the biased random walk as 100100, the context size as 1010, the return hyper-parameter (pp in Eq. (6)) and the input hyper-parameter (qq in Eq. (6)) both as 11. Beside the semantic features, five additional features, that attach the basic information of time (three features: monthmonth, dayday, weekdayweekday), news sentiment polarity (one feature: sentimentsentiment)444In this study, the sentiment polarity, is determined by the number of positive words over the number of negative words. and words count (one feature: wordscountwords\ count) are incorporated for each piece of news considered.

Next, we use the previous 20 days (T=20T=20) of news and market returns to predict the next day’s movement of the SSE stock index. The output layer is a ternary classifier to indicate whether the index moves UP (daily return >> 0.33%0.33\%), DOWN (daily return << 0.29%-0.29\%) or PRESERVES (0.29%-0.29\% << daily return << 0.33%0.33\%). We employ a rolling window with size=20size=20 and strider=1strider=1 to augment the sample.

In this study, news and market data before the year 2016 is used as the training set, whilst the rest is used as the test-validation set.

V-A2 Stock movement prediction results

As the baseline models, we include the following news embedding approaches that are used for stock movement predictions. Note that to ensure the validity of comparison, the performance measures of all models are computed from the same input data of daily news text. The other two event embedding models, namely Event2vec [38] and knowledge-enhanced event embedding [13], are not considered as the former could only be used in an organized event data set and the latter requires additional knowledge graphs.
- Average BOW: News is represented as the average of word vectors in the news titles and bodies [22, 9]. Word embedding is obtained by training the skip-gram Word2vec model [31].
- Doc2vec [23]: News text is represented as dense vectors by the Paragraph Vector model [1].
- Event embedding [12]: Event tuples are extracted from headlines and represented as vector representations by scoring the correct tuples higher than the corrupted ones.
- Deepclue [39]: Different from the two-stage prediction of DRNews that returns the news representation and the prediction separately, Deepclue adopts an end-to-end architecture to predict next day’s stock price using words in headlines from the day before. To use Deepclue as a baseline of creating news embeddings, we modify the model to a two-stage task. Specifically, we first use Deepclue to train a news encoder, instead of predicting stock prices. After the news representations are obtained from the trained encoder, we then feed the news vectors into our attention-based LSTM model to make stock market predictions. In this study, we only include two out of three news encoders proposed by Deepclue, namely the CNN and LSTM, as the third encoder based on the bi-gram produces no trainable parameters.
In addition, we apply a sentence-level encoder BERT [10] to encode news headlines as another baseline model:
- BERT: Average pooling is used to get a fixed representation of a headline based on the Chinese pre-trained BERT model555https://github.com/google-research/bert#pre-trained-models.
The proposed DRNews model is implemented with and without the five additional, i.e. monthmonth, weekweek, sentimentsentiment, wordscountwords\ count and dayday, features:
- DRNews-: News vectors are the summation of solely the semantic features (i.e., event elements).
- DRNews+: News vectors are the summation of all news features including the semantic and the five additional features.

Acc MCC
Average  BOW (d=128d=128) 41.91%41.91\% 0.11370.1137
Doc2vec (d=128d=128) 43.90%43.90\% 0.14840.1484
Event  embedding (d=128d=128) 43.76%43.76\% 0.14390.1439
Deepclue-LSTM (d=128d=128) 41.04%41.04\% 0.10340.1034
Deepclue-CNN (d=128d=128) 41.43%41.43\% 0.10660.1066
BERT (d=764d=764) 45.83%45.83\% 0.14420.1442
DRNews- (d=128d=128) 44.68%44.68\% 0.16640.1664
DRNews+ (d=128d=128) 46.24%\mathbf{46.24\%} 0.1936\mathbf{0.1936}
Market Return 41.04%41.04\% 0.10500.1050
DRNews+ & Return 44.22%44.22\% 0.16370.1637
TABLE I: Acc and MCC results of stock movement predictions in the test set

In Table I, two performance measures, i.e. the Acc and the MCC, are given to describe the predicting power of DRNews in comparison to other news embedding approaches. Specifically, the first panel lists the Acc and MCC of predictions based on each individual embedding model, the second panel gives the prediction result solely using the historical data of return, whilst the last panel shows the prediction with the joint information of the historical return and the DRNews embedding. Major implications of Table I are threefold.

First, it is shown that stock movement predictions solely based on the embedded information of news (Panel 1) outperform the predictions taking into account the historical data of market return (Panel 2 and 3). This finding suggests that daily news embeddings contain sufficient information for short-term stock movement prediction, at the same time indicates that the SSE stock market complies the weak-form of efficiency as the historical price data makes no contribution to the prediction of its future movements.

Second, the DRNews model outperforms the baseline models with the most prominent Acc and MCC results in a lower dimension of d=128d=128 666In the experiment of stock movement prediction, the three best-performing models are DRNews+ (higest Acc and MCC), BERT (Second highest Acc) and DRNews- (Second highest MCC). The dimension of BERT is 764764.. As for news-driven stock price prediction, it is more reasonable to take into account potential connections between news, such as sharing similar background story or leading to the same event, and reflect the connection by embedding similar values in certain dimensions of the news vector. Through the biased random walk in the news network, DRNews embeddings capture inter-textual knowledge, i.e. latent connections between news, which enriches the news vectors with vital information that contributes to the high predictive power of the model. On the other hand, Doc2vec, Event embedding and BERT produce stock movement predictions of similar qualities in terms of MCC, as the news vectors generated by the three approaches uniformly limit to the textual information within news articles.

Last, the DRNews algorithm is competent to enrich the news vector representation with additional information brought by supplementary features, as the Acc and MCC results of DRNews+ improve by 1.56%1.56\% and 2.72%2.72\% in comparison to DRNews- after incorporating the five labeled features (monthmonth, weekweek, dayday sentimentsentiment, and wordscountwords\ count). It is worth-mentioning that in this study, the New2vec model is performed in an unsupervised manner as all features are extracted from common news text without any artificial label or trained classifier. Therefore, we believe that there is a large chance to further improve the result once higher quality labels, such as news sentiment from trained classifier or subcategories of financial news, are incorporated. Despite that the performance of Deepclue is hardly satisfactory, we believe that this supervised news embedding model has great potential once a stronger news encoder and supervised tasks are placed to enhance the capacity of feature extraction and the universality of embedding.

V-B Early Warning of Stock Market Crises

V-B1 Experimental settings

In this section, we use an attention-based LSTM network to construct an early warning model for stock market crises. To classify the crisis/non-crisis dates of the SSE index, the switching ARCH (SWARCH) model [18] is hired through the modeling of high/low volatility regimes that reflect respectively the tranquility and turmoil [44]. To determine whether a crisis occurs, a SWARCH(2,1) process is written as:

yt\displaystyle y_{t} =u+θ1yt1+ϵt,ϵt|t1N(0,ht);\displaystyle=u+\theta_{1}y_{t-1}+\epsilon_{t},\quad\epsilon_{t}|\mathcal{I}_{t-1}\sim N(0,h_{t}); (20)
ht2γst\displaystyle\frac{h_{t}^{2}}{\gamma_{s_{t}}} =α0+α1ϵt12γst1,st={1,2}.\displaystyle=\alpha_{0}+\alpha_{1}\frac{\epsilon_{t-1}^{2}}{\gamma_{s_{t-1}}},s_{t}=\{1,2\}. (21)

Eq.(20) describes an AR(1) process with a normal error term ϵt\epsilon_{t} of variance hth_{t}. The regime switching structure of the residual variance hth_{t} is given by Eq.(21) where the αs\alpha^{\prime}s are non-negative, the γs\gamma^{\prime}s are scaling parameters that capture the change in each regime, sts_{t} is the state variable that st=1s_{t}=1 indicates the low volatility state of non-crisis, and st=2s_{t}=2 indicates the high volatility state of crisis.

The classification of high/low volatility regimes can be implemented on the basis of the so-called filtering probability, which is a byproduct of the maximum likelihood estimation. The filtering probability, written as:

P(st=i|Yt;𝜽)\displaystyle P(s_{t}=i|Y_{t};\boldsymbol{\theta}) (22)

where 𝜽\boldsymbol{\theta} is the vector of all model parameters to be estimated, could be interpreted as the conditional probability based on the current information of time tt. The crisis classifier is defined as the following binary function with a probability threshold 0.50.5 [17, 16]:

Crisis={1,SWARCH filtering probability0.50,otherwise.\text{Crisis}=\begin{cases}1,&\text{SWARCH filtering probability}\geq 0.5\\ 0,&\text{otherwise.}\end{cases} (23)
TABLE II: Summary of test-set forecasting.
True crises Market return DRNews+ & return
predictions predictions
Total onsets 2 2 2
Forewarned onsets 1 2
% of forewarned onsets 50 100
Avg. days-ahead onsets 1.0 2.5

By inputting news embeddings and historical price data, we aim to investigate the performance of DRNews in news-driven crisis predictions in comparison with other embedding models. Similar to the stock trend prediction, we use the news vectors and market return of the first 20 days to predict whether a stock market turbulence will occur in the next day. The difference is that the crash warning system is basically a binary classifier. Besides, a rolling window with size=20size=20 and strider=1strider=1 is employed to augment the sample. Finally, the time interval for training is between 2009/10/19 and 2013/4/25 and the interval for testing is from 2013/4/26 to 2016/10/31.

V-B2 Crises early warning results

Acc MCC
Average  BOW (d=128d=128) 66.23%66.23\% 0.45170.4517
Doc2vec (d=128d=128) 68.74%68.74\% 0.49920.4992
Event  embedding (d=128d=128) 69.82%69.82\% 0.49850.4985
Deepclue-LSTM (d=128d=128) 49.40%49.40\% 0.08280.0828
Deepclue-CNN (d=128d=128) 70.06%70.06\% 0.44430.4443
BERT (d=764d=764) 74.61%74.61\% 0.46350.4635
DRNews- (d=128d=128) 69.58%69.58\% 0.5120.512
DRNews+ (d=128d=128) 77.13%\mathbf{77.13\%} 0.5681\mathbf{0.5681}
Market Return 94.25%94.25\% 0.87760.8776
DRNews+ & Return 94.61%\mathbf{94.61\%} 0.8861\mathbf{0.8861}
TABLE III: Acc and MCC results of stock crisis predictions in the test set

According to the Acc and MCC results of the six embedding approaches shown in Panel 1, DRNews+ remains as the best-performing embedding model in terms of both Acc and MCC, whilst BERT produces the second highest Acc and DRNews- produces the second highest MCC. As a concluding remark, the DRNews model demonstrates robust and strong performances in the completion of the two financial prediction tasks, i.e. stock movement prediction and stock crises warning, that requires the representation of latent linkages between news events.

On the other hand, Panel 2 and 3 show that the results of crises warning is significant improved after historical return data is employed. Despite that the market-driven model already offers a fairly good prediction with 94.25%94.25\% Acc and 87.76%87.76\% MCC, the joint prediction of DRNews embedding and historical return generates the highest Acc and MCC of 94.64%94.64\% and 88.61%88.61\%, respectively.

Refer to caption
Figure 5: The SWARCH filter probability of crash warning from historical return
Refer to caption
Figure 6: The SWARCH filter probability of crash warning from historical return and DRNews+
Important nouns
Stock Trend (Ranking) Crisis Warning
A股/A-share (1) 银行/Bank (3)
创业板/GEM (Growth Enterprise Market) (4) 利率/Interest rate (5)
基金/Fund (5) CPI/CPI (9)
ST/Special Treatment stock (8) 存款/Deposit (11)
上市公司/Listed company (10) 房地产/Real estate (14)
期货/Futures (11) 不良贷款/Non-performing loans (24)
券商/Brokerage (18) 准备金率/Reserve ratio (35)
沪指/Shanghai stock index (20) 公积金/Provident fund (40)
CPI/CPI (25) 信托/Trust (41)
限售股/Restricted shares (33) 上市公司/Listed companies (46)
IPO/IPO (36) 光伏/Photovoltaic (53)
融资融券/Margin financing (45) 货币政策/Monetary policy (55)
ETF/Exchange Traded Funds(47) 流动性/Liquidity (65)
市值/Market capitalization (51) PMI/Purchase Managers’ Index (73)
新三板/New OTC(Over the Counter) Market (64) 中小企业/Middle and small-sized enterprises (80)
TABLE IV: Nouns with High Attention (Partial Selection)
Top companies and institutes
Stock Trend (Ranking) Crisis Warning
证监会/China Securities Regulatory Commission (13) 央行/People’s Bank of China(8)
光大证券/Everbright Securities (89) 万科/Vanke (43)
深交所/Shenzhen Stock Exchange (104) 银监会/China Banking Regulatory Commission (49)
上交所/Shanghai Stock Exchange (114) 保利地产/Poly Real Estate (101)
广发证券/GF Securities (116) 浦发银行/Shanghai Pudong Development Bank (106)
中信证券/CITIC Securities (126) 农行/Agricultural Bank of China (112)
中金所/China Financial Futures Exchange (128) 民生银行/China Minsheng Bank (143)
华泰证券/Huatai Securities (154) 中行/Bank of China (155)
兴业证券/Industrial Securities (159) 建行/China Construction Bank (162)
招商证券/China Merchants Securities (160) 中信银行/China CITIC Bank (170)
纳斯达克/NASDAQ (169) 国美/Gome (181)
西南证券/Southwest Securities (170) 平安银行/Ping An Bank (182)
金融时报/Finance Times (176) -
海通证券/Haitong Securities (183) -
纽约商品交易所/New York Mercantile Exchange (194) -
TABLE V: Company and Institute Names with High Attention

Figure 5 and 6 show the predicted SWARCH filter probability of crises based on the historical return with and without DRNews embeddings, respectively. Despite that the market-driven crises warning based on solely the historical return achieves impressive results in terms of Acc and MCC measures, it creates a delay of warning (see Oct 26, 2014 to Jan 26, 2015, red line in Figure 5) in comparison to the actual occurrence of the crisis (blue line in Figure 5). On the other hand, Figure 6 suggests that the joint prediction of DRNews embedding and historical return generates true ’early’ warning signals with predicted probability of crisis gradually raises at the end of 2014 before the actual crash. It appears that DRNews embeddings contain key information that serves as the driving force of crisis alarming during the pre-turmoil episode. In the next section, the key information inherited in the news will be discussed explicitly according to the attention-based LSTM model based on the DRNews embeddings.

Table II confirms the findings in Figure 5 and 6. During the test-set period between April 2013 to October 2016, two stock crisis onsets are detected by the SWARCH model. The joint predictions based on the historical return and DRNews embeddings forewarn both onsets with an average of 2.52.5-days forewarned period, whereas predictions with solely the historical return demonstrate a delay in warning the second onset that take place at the end of 2014.

V-C News Attention

Top actions (verb & gerund)
Stock Trend (Ranking) Crisis Warning
增长/Growth (3) 增长/Growth (1)
披露/Disclosure (16) 贷款/Loan (2)
交易/Transaction (17) 下降/Decline (4)
上涨/Raising (21) 信贷/Credit (9)
解禁/Unban (24) 融资/Financing (16)
退市/Delisting (26) 上涨/Rising (23)
下跌/Falling (27) 回落/Fallback (32)
复牌/Resumption (28) 投资/Investment (33)
停牌/Suspension (32) 调控/Regulate and control (37)
投资/Investment (34) 降息/Interest rate cuts (50)
上市/Listing (41) 投放Launch (61)
公开发行/Public offering (68) 降准/Reserve requirement ratio cuts (68)
反弹/Rebound (80) 出口/Export (71)
亏损/Loss (82) 成交/Close the deal (88)
重组/Restructuring (88) 加息/Interest rate increase (97)
TABLE VI: Actions with High Attention (Partial Selection)

The DRNews embeddings offer a mechanism that allows quantitative investigation of news-driven market predictions via the attention model. In this section, we select the daily news that obtains the highest attention weights in the stock movement prediction and crises warning models, respectively. The top 200 elements of the selected news are then extracted based on their tf-idf scores and listed in Table VI - V as actions, nouns, and company&institute names. By comparing the key elements in the two financial tasks, i.e. stock movement and crisis predictions, we aim to gain insights into the dynamics of the stock market as well as the information transmission channel between media and the financial sector.

With respect to Table VI - V, it is found that the predictions of short-term stock movement and stock market turbulence exhibit different concentrations, with the former receiving more impacts directly related to the secondary market whereas the latter exposed under wider influences from the economical/regulatory developments as well as the banking system. In particular,

Table VI shows the key actions in news that are highly correlated with the stock movement (left panel) and the stock market turbulence (right panel). In the case of stock movements, Disclosure appears to be rather frequently quoted in high-attention news, indicating that the stock dynamics is information-driven with a great level of sensitivity. In addition, operations on listing companies, such as Delisting & Listing and Resumption & Suspension, Unban and Public offering, are closely related to the stock price oscillation. By contrast, stock market crashes are greatly connected with news that has to do with the interest rate and the banking sector. Major keywords that drives crisis predictions include Loan, Credit, Financing, as well as phrases related to the regulatory changes in the interest rate and the reserve requirement, such as Interest rate cuts, Reserve requirement ratio cuts and Interest rate increase.

Clear differences between the important nouns in the stock movement and crisis predictions are observed in Table IV. Basically, the stock movement shows great responses to news directly related to the stock market with key mentions including the A-share, Growth Enterprise Market, Shanghai stock index, Futures, ETF, Restricted shares, New OTC Market, etc. On the other hand, the occurrence of the stock market turmoil is mainly reflected by the macroeconomic indicators and policies, such as the Interest rate, CPI, Reserve ratio, Provident fund, Liquidity, PMI and Monetary policy.

Table V of important company and institute names supports the claims made from Table VI and IV. Specifically, the stock movement is found to be heavily affected by three types of institutions, namely securities companies, stock exchanges and securities regulators. On the contrary, stock market crashes are shown to be largely influenced by news related to the banking sector. Among the monetary-oriented institutes, two government agencies, i.e. the People’s Bank of China (the Central Bank) and the China Banking Regulatory Commission, are involved demonstrating their leading effects on the systematic risk of the stock market, followed by seven commercial Banks including Shanghai Pudong Development Bank, the Agricultural Bank of China, China Minsheng Bank, etc. Additionally, two real estate enterprises (i.e. Vanke & Ploy Real Estate) are frequently mentioned in news with high attention weights.

VI Conclusions

DRNews, a news embedding model with distributed representations is developed and applied for stock market predictions in this study. The DRNews model enhances the established news embedding methods by representing news articles in an attributed network that integrates the semantic information with the inter-textual knowledge. The information contained in the network covers not only the natural co-occurrence relationship between news and its internal event elements, but also latent connections between cross-documental news events. To determine the news embeddings in the framework of a network, a Subnode algorithm is then proposed to enrich the news vectors with various features at the same time be able to produce fast inferential embeddings of latest news outside the network. Note that despite the focus of Subnode model in this paper is its application in the news network, this algorithm could be widely applied to other networks inheriting its merit of providing an effective solution to embed unseen nodes and integrate all node attributes in the embedding.

Using the attention-based LSTM network, the DRNews embeddings are applied in two stock market prediction tasks, namely the short-term price movement prediction and stock crises early warning. By comparing with five baseline models, it is shown that DRNews achieves state-of-art performance in all aspects, mainly due to its comprehensive description of various news features and potential linkages among news events on the vector representations. In particular, DRNews produces a test-set accuracy of 46.24%46.24\% in the ternary prediction task of stock movement. As for the stock crisis early warning, a test-set accuracy of 94.61%94.61\% is obtained with an average of 2.52.5-days forewarned period. Further, our experiments in the stock market show that daily news plays substantial roles on affecting both the short-term stock price oscillation and the systematic evolution of the stock market. In particular, the short-term stock movement is more easily influenced by secondary market-related information, whilst the stock crises are mainly merged from matters related to the banking sector and the economic system.

acknowledgements

We acknowledge the support by XJTLU Key Program Special Fund-Applied Technology Research Programmme (No. KSF-A-14).

References

  • [1] Akita, R., Yoshihara, A., Matsubara, T., Uehara, K.: Deep learning for stock prediction using numerical and textual information. In: 2016 IEEE/ACIS 15th International Conference on Computer and Information Science (ICIS), pp. 1–6 (2016). DOI 10.1109/ICIS.2016.7550882
  • [2] Bojanowski, P., Grave, E., Joulin, A., Mikolov, T.: Enriching Word Vectors with Subword Information. arXiv e-prints (2016)
  • [3] Bordes, A., Usunier, N., Garcia-Durán, A., Weston, J., Yakhnenko, O.: Translating embeddings for modeling multi-relational data. In: Proceedings of the 26th International Conference on Neural Information Processing Systems - Volume 2, NIPS’13, pp. 2787–2795. Curran Associates Inc., USA (2013). URL http://dl.acm.org/citation.cfm?id=2999792.2999923
  • [4] Cer, D., Yang, Y., Kong, S., Hua, N., Limtiaco, N., John, R.S., Constant, N., Guajardo-Cespedes, M., Yuan, S., Tar, C., Sung, Y., Strope, B., Kurzweil, R.: Universal sentence encoder. CoRR abs/1803.11175 (2018). URL http://arxiv.org/abs/1803.11175
  • [5] Chatzis, S.P., Siakoulis, V., Petropoulos, A., Stavroulakis, E., Vlachogiannakis, N.: Forecasting stock market crisis events using deep and statistical machine learning techniques. Expert systems with applications 112, 353–371 (2018)
  • [6] Chen, H.: Predicting market movements: From breaking news to emerging social media. In: A. Datta, S. Shulman, B. Zheng, S.D. Lin, A. Sun, E.P. Lim (eds.) Social Informatics, pp. 5–5. Springer Berlin Heidelberg, Berlin, Heidelberg (2011)
  • [7] Conneau, A., Kiela, D., Schwenk, H., Barrault, L., Bordes, A.: Supervised learning of universal sentence representations from natural language inference data. CoRR abs/1705.02364 (2017). URL http://arxiv.org/abs/1705.02364
  • [8] Conneau, A., Kruszewski, G., Lample, G., Barrault, L., Baroni, M.: What you can cram into a single vector: Probing sentence embeddings for linguistic properties. CoRR abs/1805.01070 (2018). URL http://arxiv.org/abs/1805.01070
  • [9] Das, S., Mishra, A., Berberich, K., Setty, V.: Estimating event focus time using neural word embeddings. In: Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, CIKM ’17, pp. 2039–2042. ACM, New York, NY, USA (2017). DOI 10.1145/3132847.3133131. URL http://doi.acm.org/10.1145/3132847.3133131
  • [10] Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. CoRR abs/1810.04805 (2018). URL http://arxiv.org/abs/1810.04805
  • [11] Di Persio, L., Honchar, O.: Artificial neural networks architectures for stock price prediction: Comparisons and applications. International journal of circuits, systems and signal processing 10, 403–413 (2016)
  • [12] Ding, X., Zhang, Y., Liu, T., Duan, J.: Deep learning for event-driven stock prediction. In: Proceedings of the 24th International Conference on Artificial Intelligence, IJCAI’15, pp. 2327–2333. AAAI Press (2015). URL http://dl.acm.org/citation.cfm?id=2832415.2832572
  • [13] Ding, X., Zhang, Y., Liu, T., Duan, J.: Knowledge-driven event embedding for stock prediction. In: COLING (2016)
  • [14] Fischer, T., Krauss, C.: Deep learning with long short-term memory networks for financial market predictions. European Journal of Operational Research 270(2), 654–669 (2018)
  • [15] Grover, A., Leskovec, J.: Node2vec: Scalable feature learning for networks. In: Proceedings of the 22Nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, pp. 855–864. ACM, New York, NY, USA (2016). DOI 10.1145/2939672.2939754. URL http://doi.acm.org/10.1145/2939672.2939754
  • [16] Hamilton, J., Gang, L.: Stock Market Volatility and the Business Cycle. Journal of Applied Econometrics 11(5), 573–93 (1996). URL https://EconPapers.repec.org/RePEc:jae:japmet:v:11:y:1996:i:5:p:573-93
  • [17] Hamilton, J., Susmel, R.: Autoregressive conditional heteroskedasticity and changes in regime. Journal of Econometrics 64(1-2), 307–333 (1994). URL https://EconPapers.repec.org/RePEc:eee:econom:v:64:y:1994:i:1-2:p:307-333
  • [18] Hamilton, J.D.: A New Approach to the Economic Analysis of Nonstationary Time Series and the Business Cycle. Econometrica 57(2), 357–384 (1989). URL http://www.jstor.org/stable/1912559
  • [19] Han, S., Hao, X., Huang, H.: An event-extraction approach for business analysis from online chinese news. Electronic Commerce Research and Applications 28 (2018). DOI 10.1016/j.elerap.2018.02.006
  • [20] Hill, F., Cho, K., Korhonen, A.: Learning distributed representations of sentences from unlabelled data. CoRR abs/1602.03483 (2016). URL http://arxiv.org/abs/1602.03483
  • [21] Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Computation 9(8), 1735–1780 (1997). DOI 10.1162/neco.1997.9.8.1735. URL https://doi.org/10.1162/neco.1997.9.8.1735
  • [22] Hu, Z., Liu, W., Bian, J., Liu, X., Liu, T.Y.: Listening to chaotic whispers: A deep learning framework for news-oriented stock trend prediction. In: Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining, WSDM ’18, pp. 261–269. ACM, New York, NY, USA (2018). DOI 10.1145/3159652.3159690. URL http://doi.acm.org/10.1145/3159652.3159690
  • [23] Le, Q.V., Mikolov, T.: Distributed representations of sentences and documents. CoRR abs/1405.4053 (2014). URL http://arxiv.org/abs/1405.4053
  • [24] Li, J., Hovy, E.: A model of coherence based on distributed sentence representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 2039–2048. Association for Computational Linguistics, Doha, Qatar (2014). DOI 10.3115/v1/D14-1218. URL https://www.aclweb.org/anthology/D14-1218
  • [25] Li, Q., Chen, Y., Wang, J., Chen, Y., Chen, H.: Web media and stock markets : A survey and future directions from a big data perspective. IEEE Transactions on Knowledge and Data Engineering 30(2), 381–399 (2018). DOI 10.1109/TKDE.2017.2763144
  • [26] Li, Q., Wang, T., Li, P., Liu, L., Gong, Q., Chen, Y.: The effect of news and public mood on stock movements. Information Sciences 278, 826 – 840 (2014). DOI 10.1016/j.ins.2014.03.096
  • [27] Logeswaran, L., Lee, H.: An efficient framework for learning sentence representations. CoRR abs/1803.02893 (2018). URL http://arxiv.org/abs/1803.02893
  • [28] Long, W., Lu, Z., Cui, L.: Deep learning-based feature engineering for stock price movement prediction. Knowledge-Based Systems 164, 163–173 (2019)
  • [29] Long, W., Song, L., Tian, Y.: A new graphic kernel method of stock price trend prediction based on financial news semantic and structural similarity. Expert Systems with Applications 118, 411–424 (2019)
  • [30] Luss, R., D’Aspremont, A.: Predicting abnormal returns from news using text classification. Quantitative Finance 15(6), 999–1012 (2015). DOI 10.1080/14697688.2012.672762. URL https://doi.org/10.1080/14697688.2012.672762
  • [31] Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. CoRR abs/1301.3781 (2013). URL http://arxiv.org/abs/1301.3781
  • [32] Mikolov, T., Sutskever, I., Chen, K., Corrado, G., Dean, J.: Distributed representations of words and phrases and their compositionality. CoRR abs/1310.4546 (2013). URL http://arxiv.org/abs/1310.4546
  • [33] Moat, H.S., Curme, C., Avakian, A., Kenett, D.Y., Stanley, H.E., Preis, T.: Quantifying wikipedia usage patterns before stock market moves. Scientific Reports 3, 1801 EP – (2013). URL https://doi.org/10.1038/srep01801
  • [34] Perozzi, B., Al-Rfou, R., Skiena, S.: Deepwalk: Online learning of social representations. In: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’14, pp. 701–710. ACM, New York, NY, USA (2014). DOI 10.1145/2623330.2623732. URL http://doi.acm.org/10.1145/2623330.2623732
  • [35] Rather, A.M., Agarwal, A., Sastry, V.: Recurrent neural network and a hybrid model for prediction of stock returns. Expert Syst. Appl. 42(6), 3234–3241 (2015). DOI 10.1016/j.eswa.2014.12.003. URL http://dx.doi.org/10.1016/j.eswa.2014.12.003
  • [36] Schumaker, R., Chen, H.c.: Textual analysis of stock market prediction using breaking financial news: The azfin text system. ACM Trans. Inf. Syst. 27 (2009). DOI 10.1145/1462198.1462204
  • [37] Schumaker, R., Zhang, Y., Huang, C., Chen, H.: Evaluating sentiment in financial news articles. Decision Support Systems 53(3), 458–464 (2012). DOI 10.1016/j.dss.2012.03.001
  • [38] Setty, V., Hose, K.: Event2vec: Neural embeddings for news events. In: The 41st International ACM SIGIR Conference on Research &#38; Development in Information Retrieval, SIGIR ’18, pp. 1013–1016. ACM, New York, NY, USA (2018). DOI 10.1145/3209978.3210136. URL http://doi.acm.org/10.1145/3209978.3210136
  • [39] Shi, L., Teng, Z., Wang, L., Zhang, Y., Binder, A.: Deepclue: Visual interpretation of text-based deep stock prediction. IEEE Transactions on Knowledge and Data Engineering PP, 1–1 (2018). DOI 10.1109/TKDE.2018.2854193
  • [40] Si, J., Mukherjee, A., Liu, B., Pan, S.J., Li, Q., Li, H.: Exploiting social relations and sentiment for stock prediction. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1139–1145. Association for Computational Linguistics, Doha, Qatar (2014). DOI 10.3115/v1/D14-1120. URL https://www.aclweb.org/anthology/D14-1120
  • [41] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I.: Attention is all you need. CoRR abs/1706.03762 (2017). URL http://arxiv.org/abs/1706.03762
  • [42] Vincent, P., Larochelle, H., Lajoie, I., Bengio, Y., Manzagol, P.A.: Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. J. Mach. Learn. Res. 11, 3371–3408 (2010). URL http://dl.acm.org/citation.cfm?id=1756006.1953039
  • [43] Wang, B., Huang, H., Wang, X.: A novel text mining approach to financial time series forecasting. Neurocomput. 83, 136–145 (2012). DOI 10.1016/j.neucom.2011.12.013. URL http://dx.doi.org/10.1016/j.neucom.2011.12.013
  • [44] Wang, P., Zong, L., Ma, Y.: An integrated early warning system for stock market turbulence. arXiv preprint arXiv:1911.12596 (2019)
  • [45] Wuthrich, B., Cho, V., Leung, S., Permunetilleke, D., Sankaran, K., Zhang, J.: Daily stock market forecast from textual web data. In: SMC’98 Conference Proceedings. 1998 IEEE International Conference on Systems, Man, and Cybernetics (Cat. No.98CH36218), vol. 3, pp. 2720–2725 vol.3 (1998). DOI 10.1109/ICSMC.1998.725072
  • [46] Yang, Z., Yang, D., Dyer, C., He, X., Smola, A., Hovy, E.: Hierarchical attention networks for document classification. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1480–1489. Association for Computational Linguistics, San Diego, California (2016). DOI 10.18653/v1/N16-1174. URL https://www.aclweb.org/anthology/N16-1174
  • [47] Zheng, J., Xia, A., Shao, L., Wan, T., Qin, Z.: Stock volatility prediction based on self-attention networks with social information. In: 2019 IEEE Conference on Computational Intelligence for Financial Engineering Economics (CIFEr), pp. 1–7 (2019). DOI 10.1109/CIFEr.2019.8759115