Skip to main content

ORIGINAL RESEARCH article

Front. Artif. Intell., 07 June 2022
Sec. AI in Finance
This article is part of the Research Topic Knowledge Discovery from Unstructured Data in Finance View all 5 articles

Forecasting Stock Price Trends by Analyzing Economic Reports With Analyst Profiles

\nMasahiro Suzuki
Masahiro Suzuki1*Hiroki SakajiHiroki Sakaji1Kiyoshi IzumiKiyoshi Izumi1Yasushi IshikawaYasushi Ishikawa2
  • 1Department of Systems Innovation, School of Engineering, The University of Tokyo, Tokyo, Japan
  • 2Nikko Asset Management Co., Ltd., Tokyo, Japan

This article proposes a methodology to forecast the movements of analysts' estimated net income and stock prices using analyst profiles. Our methodology is based on applying natural language processing and neural networks in the context of analyst reports. First, we apply the proposed method to extract opinion sentences from the analyst report while classifying the remaining parts as non-opinion sentences. Then, we employ the proposed method to forecast the movements of analysts' estimated net income and stock price by inputting the opinion and non-opinion sentences into separate neural networks. In addition to analyst reports, we input analyst profiles to the networks. As analyst profiles, we used the name of an analyst, the securities company to which the analyst belongs, the sector which the analyst covers, and the analyst ranking. Consequently, we obtain an indication that the analyst profile effectively improves the model forecasts. However, classifying analyst reports into opinion and non-opinion sentences is insignificant for the forecasts.

1. Introduction

Investors need to search for market information such as stock prices and fundamental information such as companies' sales, earnings, and business conditions. A search engine can provide a variety of information about a company. A company's website also provides investor relation (IR) information on its business performance and financial information. A newspaper or a news site provides stock price charts and new information about companies. The trends of individual investors can be seen on social networking sites and message boards. However, IR information and news articles are limited to factual information, and social networking services (SNS) and message boards are not highly reliable. It is not easy to gather useful information for investment decisions. In this situation, analyst reports have been attracting attention. Analyst reports are written by analysts affiliated with securities companies that evaluate each stock, taking into account earnings guidance, valuation of stock price, press releases, macroeconomy trends, etc. Analyst report includes facts, such as financial results, stock prices, and information announced by the company. It also includes analysts' forecasts of future earnings and stock price performance and objective information which analysts think is worth including. Although analyst report contains useful information and unique information, it is difficult for investors, who manage many stocks, to read all analyst reports because many reports are issued during a period of the earnings announcements. Thus, Several studies applied text mining to analyst reports. Huang et al. (2014) applied Naive Bayes to words extracted from analyst reports to analyze the market's reaction to analyst reports. Suzuki et al. (2020) input the text of analyst reports into long short-term memory (LSTM) to predict stock price movements.

Although the text of an analyst report is used, the information of an analyst who writes text and evaluates stocks is not considered. For large-cap stocks, analyst reports are issued by several analysts, and it is essential to know which analyst writes the report. Some previous studies focused on the text. Amir et al. (2016) constructed user-specific distributed representations from past postings of users to detect sarcasm in text posted on Twitter. They showed that inputting distributed representations in addition to texts improves detection accuracy. Wang et al. (2019) proposed a method to construct representations of users and topics from financial postings on Weibo. They applied the representations to sentiment analysis and showed that the accuracy was improved compared to the baseline. They are indexed and characterized from the textual information; however, information other than the original texts is not utilized.

Figure 1 shows the overview of this research. In this study, with the aim to forecast excess return of stock prices and the change rate of analysts' estimated net income, we propose a method to utilize unique information from the analyst reports using opinion and non-opinion sentences extracted from analyst reports and analyst's profiles. In addition to the main texts of analyst reports, the information (profile) of the analyst (who is the author) is used for input, and the effectiveness is confirmed. We use four types of information below as the analyst profile.

• Analyst name

• Securities company to which an analyst belongs

• Industries in which an analyst is in charge of

• Analyst ranking

FIGURE 1
www.frontiersin.org

Figure 1. Overview of this research. In this study, we conduct two experiments: opinion sentence extraction and the movement forecasts of analysts' estimated net income and excess return.

The securities company for which an analyst works is also used as one of the analyst profile information.

2. Related Study

Several studies have been conducted on financial text mining to forecast financial markets. Regarding the research using financial news, Wuthrich et al. (1998) predicted daily movements of five indices using news articles published on the Internet. They constructed the rule to predict by combining news articles, index values, and some keywords. They found that textual information with bag-of-words in addition to numeric time-series data increases the quality of the input. Low et al. (2001) proposed a semantic expectation-based knowledge extraction methodology for extracting causal relations using WordNet as a thesaurus to extract terms that represent movement concepts. Koppel and Shtrimberg (2006) proposed a method for classifying the news stories of a company according to their apparent impacts on the performance of the company's stock price. Tetlock et al. (2008) showed that negative words in reports of financial institutions predict the low profits and the market incorporates the information in negative words with a slight delay. Although negative words related to fundamentals are particularly useful in predicting earnings and returns, they showed that the market underestimates the information. Bar-Haim et al. (2011) proposed a framework to identify expert investors and used it to predict stock price rise from stock tweets using a support vector machine (SVM) classifier. They trained the classifier that directly learned the relationship between the content of a tweet and stock prices. The user who was writing tweets that could discriminate the rise/fall of the stock price was learned as a specialist. Consequently, they constructed the classifier trained only by the set of tweets of the identified experts.

Regarding the research using the documents published by financial institutions and companies, Milea et al. (2010) predicted the Morgan Stanley Capital International (MSCI) euro index (upwards, downwards, or constant) based on fuzzy grammar fragments extracted from a report published by the European Central Bank. Brown and Tucker (2011) showed that the market price reaction weakens despite the increase in disclosed information, and the usefulness of management's discussion and analysis (MD&A) has declined. They also showed that analysts do not revise their earnings forecasts and equity investors are more responsive to the disclosure of MD&A. Loughran and Mcdonald (2014) showed that the text file size is more effective as an explanatory variable in the readability of the disclosure documents of financial status than Fog Index, which is one of the readability indicators; it is calculated from the number of words per sentence and the ratio of words with three or more syllables. They obtained that the larger the file size of a 10-K document, the greater the volatility of abnormal returns after disclosure and the greater the variance of analyst expectations by regression analysis.

Regarding the research using the texts on financial SNS, Bollen et al. (2011) showed that mood states obtained from tweets are relevant for forecasting Dow Jones Industrial Average (DJIA). They applied Opinion Finder and Google-Profile of Mood States (G-POMS), a mood analysis tool, to extract six types of mood from tweets: Calm, Alert, Sure, Vital, Kind, and Happy. They also used self-organizing fuzzy neural networks to predict rises and drops with an accuracy of more than 80%. They found that mood states in terms of positive or negative mood are ineffective in forecasting; meanwhile, those labeled “Calm” are effective. Chen et al. (2014) showed that opinions posted on popular social media with investors strongly predict future stock returns and market surprise reactions to earnings.

Regarding the study of applying sentiment to financial documents, Schumaker and Chen (2009) proposed a machine learning method to forecast stock prices by analyzing financial news articles. Their model forecasted indicators and stock prices. Schumaker et al. (2012) united their approach using sentiment analysis. The estimated stock prices after releasing financial news articles using SVM. Ito et al. proposed a neural network model for visualizing online financial textual data (Ito et al., 2018a,b). Their proposed model acquired word sentiment and its category. Guijarro et al. (2019) analyzed the impact of investors' moods on market liquidity. They conducted a sentiment analysis of tweets related to the S&P 500 Index. Vu et al. (2012) proposed a method using a decision tree classifier to predict the daily price movements of four famous tech stocks. They applied sentiment analysis, semantic orientation, and movements of previous days as features for tweets. They obtained an accuracy of more than 75%. Oliveira et al. (2017) constructed sentiment and attention indicators extracted from microblogs and used machine learning-based methods for financial tweets sentiment classification of predicting daily stock market variables. They tested five machine learning-based methods for financial tweets sentiment classification using the indicators. Ranco et al. (2015) analyzed the effects of sentiments of tweets about companies on DJIA 30 prices using SVM. They found a dependence between stock price returns and Twitter sentiments. Smailović et al. (2013) showed causality between sentiment polarity of tweets and daily return of closing prices. They used sentiment derived from an SVM model to classify the tweets into positive, negative, and neutral categories.

Regarding financial text mining for the Japanese language, Sakaji et al. (2008) proposed a method to automatically extract basis expressions that indicate economic trends from newspaper articles using a statistical method. Additionally, Sakaji et al. (2017) proposed an unsupervised approach to discover rare causal knowledge from financial statement summaries. Their method extracted basis expressions and causal knowledge using syntactic patterns. Kitamori et al. (2017) proposed a method for extracting and classifying sentences indicating business performance forecasts and economic forecasts from summaries of financial statements. This classification method was based on a neural network using a semi-supervised approach.

Regarding a study using analyst reports, Asquith et al. (2005) showed that the revision of the target stock price in an analyst report has a significant positive correlation with the market reaction. They also showed that half of the analyst reports contained new information in addition to previously known information. They obtained that the market is most responsive to rating downgrades in analyst reports. Huang et al. (2014) showed that analyst reports play two roles, such as information interpretation in conference calls and information discovery not mentioned in conference calls, using the topic model (LDA).

In this research, we apply the network combining bidirectional encoder representations from transformers (BERT) (Devlin et al., 2019) and multilayer perceptron (MLP). BERT is a language model that uses Transformer (Vaswani et al., 2017), which achieves high performance without using a complicated structure, such as LSTM, in bidirectional ways. BERT achieved higher accuracy than conventional methods, such as LSTM, by performing pre-training using a large corpus. In financial text mining, there are various studies applying BERT. Araci (2019) applied pre-trained BERT for two financial sentiment analysis datasets. Araci showed that the BERT model achieves a higher F1 score than other models such as LSTM. Del Corro and Hoffart (2021) proposed a method to automatically identify financially relevant news by applying BERT. They showed that the method ranks relevant news highly and positively correlated with the accuracy of the initial stock price prediction task. Kim and Yoon (2021) proposed a model to predict impending bankruptcies using BERT. They showed that BERT outperforms dictionary-based predictions and Word2Vec-based predictions. Taguchi et al. (2021) showed analysts' sentiment toward individual stocks is useful in predicting the macroeconomic index. They applied BERT to create polarity indexes from analyst reports.

3. Data

3.1. Analyst Report

We use two types of analyst reports: analyst report set and analyst report sentences. The analyst report set consists of 75,440 analyst reports published from January 2016 to September 2020 by Japanese securities companies. The analyst report sentences consist of 2,204 sentences extracted from 100 analyst reports which are randomly extracted from 10,100 analyst reports published in 2017. These sentences are used for the training data of the experiment of opinion sentence extraction. In this study, an opinion sentence is defined as one expressing an analyst's subjective views, such as ratings for future stock prices, sales or forecasted net earnings for the next year, and backgrounds of current sales. A non-opinion sentence is defined as one expressing objective facts such as financial results. Table 1 presents examples of opinion and non-opinion sentences. After manual tagging by three of the authors, including a fund manager who uses the analyst report for operation, 1,180 sentences are labeled as opinion sentences, whereas the remaining 1,024 sentences are labeled as non-opinion sentences. Tagging was done by each of the three people, and the final decision was made by the fund manager.

TABLE 1
www.frontiersin.org

Table 1. Typical examples of opinion and non-opinion sentences in analyst reports. English follows Japanese.

The sentences contained in these analyst reports are classified into opinion and non-opinion sentences using the model constructed in Section 5.1. They are used for the forecast experiments of analysts' estimated net income and excess return. Among the 1,681,114 sentences in the reports, 9,02,652 and 7,78,462 sentences are classified as an opinion and non-opinion sentences, respectively.

3.2. Analyst Profile

In this study, the analysts' information, who are the authors of analyst reports, is used as the text information of analyst reports. The name of an analyst, an affiliated securities company, industries in charge, and an analyst ranking are used as the analyst profile. The analyst's name is used to indicate which analyst writes the analyst report. From the analyst report set, we extract 345 analysts. All the analysts belong to one of the five securities companies that publish the analyst reports used in this research. We use two types of industries: Tokyo Stock Exchange (TSE) 33 industries and those mapped to 10 industries. The analyst ranking uses the latest information of the first to fifth analyst rankings by Nikkei Veritas as of the publication date of the analyst report. In this research, we use all the information available that accompanies the analyst report. Though other information such as years of experience as an analyst might be, this information is not publicly available and is difficult to collect. Therefore, the information mentioned above is used in this research.

3.3. Data Used for the Forecast of the Analyst's Estimated Net Income

This study uses the rate of change in analyst's estimated net income to forecast the rise and fall of the net income. First, the analyst's estimated net income is calculated. Let NI(t) be the analyst's estimated net income of the target stock for the last 12 months at a point t. NI(t) is calculated by prorating the estimated net income for the current term and that for the next term in the latest report at the time t. Proration is used because using only the current term estimated net income or the next term estimated net income makes a jump when crossing the fiscal year-end and the target period of the estimated net income will not be constant. Consider an example of 31 December 2021. For many companies with March settlements, an analyst estimates net income for the current term from April 2021 to March 2022 and that for the next term from April 2022 to March 2023. Therefore, the estimated net income for the next 12 months from 31 December 2021, is calculated as proration of the estimated net income for the three months (from January 1 to March 31) of the current term until March 2022 and the estimated net income for the nine months (from April 1 to December 31) of the next term until March 2023. Let NI′, NI′′ be the current term estimated net income and the next term estimated net income. Then NI(t) is calculated as

NI(t)=NI×312+NI×912.    (1)

Let the change rate of the analyst's estimated net income for 60 business days (about three months) from the day after the analyst report is published be FR(60). We apply the term of 60 business days (about three months) because analyst reports are usually published quarterly (approximately every three months) in line with quarterly financial settlements. Then, it is calculated as

FR(60)={NI(t+60)-NI(t)|NI(t)|(NI(t)0)sgn(NI(t+60)-NI(t))(NI(t)=0).    (2)

Here, sgn(·) is a sign function, which returns 1 when the input is positive, 0 when the input is 0, and –1 when the input is negative. In the experiment, binary classification is conducted according to the change rate of the analyst's estimated net income FR(60). The median value of the training data is used as the threshold value of FR(60).

3.4. Data Used for Forecast of Stock Price Movement

In the stock price movement forecast, the excess returns, which are the returns of stocks exceeding the benchmark, are forecasted as positive or negative. Therefore, the stock prices of stocks and the benchmark Tokyo Stock Price Index (TOPIX) are used. Let the stock price and TOPIX on the day after the analyst report issued date be C0 and T0, respectively. An analyst report is published after the end of the trading day to avoid an impact on the market. Then, the information from the analyst report is incorporated into the market the day after the issue date. Therefore, the value on the day after the issue date is used in this study. Similarly, let those 60 business days (about three months) after the issue date be C60 and T60, respectively. The closing price is used for each index. Let the excess return after 60 business days (about three months) be ER(60), which is calculated as

ER(60)=(C60-C0)C0-(T60-T0)T0.    (3)

For institutional investors, who are relatively valued compared to benchmarks such as TOPIX, predictability of excess returns is more important than that of simple returns. For this reason, the excess return is used. We classify excess returns into positive and negative values for each analyst report. Among the 75,440 reports, 36,163 and 39,277 were classified as positive and negative excess returns, respectively.

4. Method

4.1. Opinion Sentence Extraction

Bidirectional Encoder Representations from Transformers (BERT) (Devlin et al., 2019) is used to construct the opinion sentence extraction model. Natural language models such as BERT apply unsupervised learning to a large number of documents for pre-training. Pre-training of BERT consists of the masked language model and the next sentence prediction. In the masked language model, the model predicts a masked word. However, in the next sentence prediction, the model predicts whether two sentences are continuous or not. These tasks make it possible to capture the meaning of words and phrases and the structure of sentences. Using a pre-trained model constructed by a large corpus and fine-tuning with target data, the model can be more effective than one training with the target data from scratch. This strategy has resulted in significant performance improvements in natural language processing tasks. The pre-trained Japanese BERT model by Tohoku University 1.

4.2. Forecasting Movements of Net Income and Excess Return

Figure 2A shows the overview of the network used in this study. In forecasting movements of net income and stock price, the text of analyst reports and profiles are input. The model outputs the rate of change in a two-class classification of estimated net income and excess returns. The name, affiliated securities company, industry in charge, and the analyst ranking described in Section 3.2 are used as the analyst profile. For each of this information, we create a vector for each analyst report, shown in Figure 3, where the corresponding parts are 1 and the other parts are 0. For the names of analysts, indexes are given to all 345 analysts who are the authors of the analyst reports. Then we construct 345-dimensional vectors with 1 for the index of the analyst and 0 for the others. We also construct five-dimensional vectors to indicate the securities companies since this study uses analyst reports issued by five securities companies. Vectors with 345 dimensions for the name, five dimensions for the affiliated securities company, 33 and ten dimensions for the industry in charge, and five dimensions for the analyst ranking are constructed. A total of 398 dimension vectors is used as the analyst profile.

FIGURE 2
www.frontiersin.org

Figure 2. Overview of the method used in this study. (A) The body text of the analyst report is processed using BERT and output as htext. The analyst profile is processed using MLP and output as hanalyst. Here, h is obtained from the concatenation of htext and hanalyst. MLP and softmax functions are applied to output the probability of each label. (B) Overview of the case where the opinion and non-opinion sentences in an analyst report are input separately. The opinion and non-opinion sentences are processed using BERT and output as hopinion, hnon−opinion, respectively. These are weighted by the weights of attention mechanisms such as α1, α2, respectively, and htext is obtained from the sum.

FIGURE 3
www.frontiersin.org

Figure 3. An example of a vector constructed using an analyst profile. All elements are composed of 0 or 1. In the analyst index, one of the 345 dimensions is 1, and the others are 0. We use two types of industries: TSE 33 industries and those mapped to 10 industries.

Let the analyst report and profile of the author of the report be xtext and xanalyst, respectively. Then, htext is obtained by inputting xtext to BERT as

htext=BERT(xtext).    (4)

Here, BERT(·) is referred to as the output of [CLS] token (short for “classification”) which is inserted at the beginning of the sentence to output a vector used for classification. Furthermore, hanalyst is obtained by inputting xanalyst to MLP consisting of three layers as

u=ReLU(LayerNorm(Wanalyst·xanalyst+banalyst)),    (5)
hanalyst=LayerNorm(Wu·u+bu).    (6)

Here, Wanalyst, Wu are weight matrices; banalyst, bu are bias vectors; LayerNorm(·) is layer normalization (Ba et al., 2016), which is represented using the input vector x as

LayerNorm(x)=x-x¯Var[x]+ϵ·γ+β.    (7)

Here, β, γ are the learnable parameters; ϵ is a constant; Var[x] is the variance of x; x¯ is the average of x. We adopt ReLU2 and Layer Normalization in this research since it is empirically known that these contribute to the stability of training. The vector h is obtained by concatenating htext and hanalyst as

h=[htexthanalyst].    (8)

Multilayer perceptron is used to the obtained vector h, and the output y is obtained as

v=ReLU(LayerNorm(Wh·h+bh))    (9)
y=softmax(Wv·v+bv).    (10)

Here, Wh, Wv are weight matrices, and bh, bv are bias vectors.

Figure 2B shows an overview of inputting opinion and non-opinion sentences in an analyst report. Let the opinion and non-opinion sentences be xopinion and xnon−opinion, respectively. The opinion and non-opinion sentences are input to the BERT models, BERTopinion(·) and BERTnon−opinion(·), respectively, as given as

hopinion=BERTopinion(xopinion)    (11)
hnon-opinion=BERTnon-opinion(xnon-opinion).    (12)

Here, hopiniond,hnon-opiniond, and d is the dimension of the hidden layer of BERT. H ∈ ℝd×2 is defined as the concatenation of hopinion and hnon−opinion as

H=[hopinion;hnon-opinion].    (13)

The weight of the attention mechanism αi(i = 1, 2) is obtained as

αi=esij=12esj,    (14)

where

s=tanh(ws·H+bs),    (15)

Here, wsd is a weight vector, and bs2 is a bias vector. h is weighted by the weight αi, and htext is calculated as

htext=α1·hopinion+α2·hnon-opinion.    (16)

Here, htext and hanalyst are concatenated according to Equation (8). The output y is obtained from Equations (9) and (10).

5. Experiments and Results

5.1. Opinion Sentence Extraction

The analyst report sentences described in Section 3.1 are used as the input data. The probabilities of the opinion sentences and the non-opinion sentence are output. Among the analyst report sentences, 68%, 12%, and 20% are used as the training, validation, and test data, respectively. We experiment with our method and comparative method of LSTM and MLP (Suzuki et al., 2020). We fix batch size with 32 in our method and 64 in the comparative method. We search hyperparameters from 1 to 60 for the number of epochs and from 10−6 to 10−4 for the learning rate. We used PyTorch3 (version 1.8.1) for implementation, optuna4 (version 2.0.0) for hyperparameter optimization, cross-entropy as the loss function, and Adam as the optimization algorithm. These settings are the same in the experiments of net income and excess return.

The result is shown in Table 2. The results show that the proposed model outperforms the comparison methods. We use the proposed model to distinguish between opinion and non-opinion sentences for each sentence in the analyst report set used as input in Section 5.2.

TABLE 2
www.frontiersin.org

Table 2. Result of opinion sentence extraction.

5.2. Estimating Net Income and Excess Return

We perform the forecast of the magnitude of the increase rate of the analyst's estimated net income and the positive or negative prediction of the excess return using the analyst report set and the analyst profile. There are four types of input sentences in the analyst reports such as full analyst report, opinion sentences only, non-opinion sentences only, and opinion and non-opinion sentences separately. For each input, we conduct experiments with and without the analyst profile. In the experiment without the analyst profile, h = htext is input in the subsequent MLP instead of the concatenation of htext and hanalyst in Equation (8). As the comparison method, the combination of LSTM and MLP is used as in Section 5.1. In the comparison method of the combination of LSTM and MLP using the analyst profile, the output of the attention mechanism and MLP are concatenated and input to the subsequent MLP. For each input data, 68%, 12%, and 20% are used as the training, validation, and test data, respectively. Tables 3, 4 present the result of forecasting the movement of net income and excess return, respectively. We perform Welch's t-test of confidence of the model in the correct labels of the test data. * in the tables indicates that there is a significant difference at 1% between the model with and without the analyst profile. † in the tables shows that there is a significant difference of 1% compared with this method using the full text using the analyst profile.

TABLE 3
www.frontiersin.org

Table 3. Result of forecasting the magnitude of the change rate of the analyst's estimated net income. The evaluation method is Macro-F1.

TABLE 4
www.frontiersin.org

Table 4. Result of forecasting the movement of excess return. The evaluation method is Macro-F1.

6. Discussion

In both Tables 3, 4, our method of the full text input using the analyst profile achieves the highest F1. It achieves a higher F1 than our method of the full text input without the analyst profile. This seems to be effective to add the analyst profile. It is thought that the analyst profile reflected the analyst's experience and popularity and contributes to the improvement of accuracy. In this experiment, it is not effective to divide into opinion sentences and non-opinion sentences.

In the net income forecast, the opinion sentences are more effective than non-opinion sentences, whereas, in the excess return forecast, the non-opinion sentences are more effective than the opinion sentences. This reason would be that the analyst's opinion is strongly reflected in the target value of the net income forecast because it is set by the analyst. On the other hand, in the excess return forecast, non-opinion sentences that state facts may be more effective than opinion sentences. In both experiments, the method with the full texts achieves a higher F1 than the method with the opinion sentences and non-opinion sentences. It is considered that the loss of the order (context) of the input sentences is a disadvantage.

We conduct an additional experiment using the analyst profile for each securities company. Tables 5, 6 present the result of forecasting net income and excess return by a securities company, respectively. In many cases, the full-text input achieves a higher F1 than the other input. However, depending on the securities company, the input method other than the full-text achieves higher F1, such as the opinion and non-opinion sentences input in the securities company C in Tables 5, 6.

TABLE 5
www.frontiersin.org

Table 5. Result of forecasting the magnitude of the change rate of the analyst's estimated net income by securities company using our method with the analyst profile. The evaluation method is Macro-F1.

TABLE 6
www.frontiersin.org

Table 6. Result of forecasting excess return by securities company with our method using the analyst profile. The evaluation method is Macro-F1.

7. Conclusion

In this study, we apply the analyst report texts and profiles to the model combining BERT and MLP and forecast the magnitude of the change rate of the analyst's estimated net income and the positive/negative excess return. We first divide the analyst reports into opinion and non-opinion sentences, assuming that the analyst's opinion is effective in forecasting change in the estimated net income. Consequently, we achieve F1 exceeding 0.8 in distinguishing between opinion and non-opinion sentences.

Next, we forecast the magnitude of the change rate of the analyst's estimated net income and positive or negative excess return using the opinion and non-opinion sentences extracted from the analyst reports and the analyst profile. For each forecasting, the full-text method using the analyst profile achieves the highest F1. Although there is no effect of separating opinion and non-opinion sentences, the analyst reports texts and profiles are effective.

Since we focused on the effectiveness of the analyst profile in this research, we conducted experiments within a fixed period. However, changes in results when the period is changed may not be negligible. Therefore, the experiment with the different periods is a future task. It is also conceivable to use analyst reports and analyst profiles before the time of forecasting. Using the information from analyst reports and profiles written about the same analyst and stock brand in the past could improve the prediction accuracy. Furthermore, weighting the analyst reports and profiles to obtain reliability instead of weighting equally might be useful for portfolio construction. This research showed an example of effective utilization of information other than the text information of the analyst report. In the future, we will use it with other numerical information and text information, such as tick data.

Data Availability Statement

The data analyzed in this study is subject to the following licenses/restrictions: Data used in this study is limited to authorized researchers. Requests to access these datasets should be directed to MS, YjIwMTltc3V6dWtpJiN4MDAwNDA7c29jc2ltLm9yZw==.

Author Contributions

MS organized and executed the experiments and wrote most of the articles. HS contributed to the concept and helped to write the article. KI is a supervisor and administrated the project. YI curated the data and reviewed the article. All authors contributed to the article and approved the submitted version.

Conflict of Interest

YI is employed by Nikko Asset Management Co., Ltd.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

1. ^https://github.com/cl-tohoku/bert-japanese/tree/v1.0

2. ^We tried other activation functions such as tanh. However, ReLU had higher performance and stability in training.

3. ^https://github.com/pytorch/pytorch

4. ^https://github.com/optuna/optuna

References

Amir, S., Wallace, B. C., Lyu, H., Carvalho, P., and Silva, M. J. (2016). “Modelling context with user embeddings for sarcasm detection in social media,” in CoNLL 2016-20th SIGNLL Conference on Computational Natural Language Learning, Proceedings (Berlin), 167–177.

Google Scholar

Araci, D.. (2019). Finbert: financial sentiment analysis with pre-trained language models. arXiv:1908.10063. doi: 10.48550/arXiv.1908.10063

CrossRef Full Text | Google Scholar

Asquith, P., Mikhail, M. B., and Au, A. S. (2005). Information content of equity analyst reports. J. Financ. Econ. 75, 245–282. doi: 10.1016/j.jfineco.2004.01.002

CrossRef Full Text | Google Scholar

Ba, J. L., Kiros, J. R., and Hinton, G. E. (2016). Layer normalization. doi: 10.48550/arXiv.1607.06450

CrossRef Full Text | Google Scholar

Bar-Haim, R., Dinur, E., Feldman, R., Fresko, M., and Goldstein, G. (2011). “Identifying and following expert investors in stock microblogs,” in Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing (Edinburgh: Association for Computational Linguistics), 1310–1319,.

Google Scholar

Bollen, J., Mao, H., and Zeng, X. (2011). Twitter mood predicts the stock market. J. Comput. Sci. 2, 1–8. doi: 10.1016/j.jocs.2010.12.007

CrossRef Full Text | Google Scholar

Brown, S. V., and Tucker, J. W. (2011). Large-sample evidence on firms' year-over-year md&a modifications. J. Account. Res. 49, 309–346. doi: 10.1111/j.1475-679X.2010.00396.x

CrossRef Full Text | Google Scholar

Chen, H., De, P., Hu, Y. J., and Hwang, B.-H. (2014). Wisdom of crowds: the value of stock opinions transmitted through social media. Rev. Financ. Stud. 27, 1367–1403. doi: 10.1093/rfs/hhu001

CrossRef Full Text | Google Scholar

Del Corro, L., and Hoffart, J. (2021). “From stock prediction to financial relevance: repurposing attention weights to assess news relevance without manual annotations,” in Proceedings of the Third Workshop on Economics and Natural Language Processing (Punta Cana: Association for Computational Linguistics), 45–49.

Google Scholar

Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2019). “BERT: pre-training of deep bidirectional transformers for language understanding,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers) (Minneapolis, MN: Association for Computational Linguistics), 4171–4186.

Google Scholar

Guijarro, F., Moya-Clemente, I., and Saleemi, J. (2019). Liquidity risk and investors' mood: linking the financial market liquidity to sentiment analysis through twitter in the s&p500 index. Sustainability 11, 1–13. doi: 10.3390/su11247048

CrossRef Full Text | Google Scholar

Huang, A. H., Y.zang, A., and Zheng, R. (2014). Evidence on the information content of text in analyst reports. Account. Rev. 89, 2151–2180. doi: 10.2308/accr-50833

CrossRef Full Text | Google Scholar

Ito, T., Sakaji, H., Izumi, K., Tsubouchi, K., and Yamashita, T. (2018a). Ginn: gradient interpretable neural networks for visualizing financial texts. Int. J. Data Sci. Anal. 9, 431–445. doi: 10.1007/s41060-018-0160-8

CrossRef Full Text | Google Scholar

Ito, T., Sakaji, H., Tsubouchi, K., Izumi, K., and Yamashita, T. (2018b). “Text-visualizing neural network model: understanding online financial textual data,” in Advances in Knowledge Discovery and Data Mining, eds D. Phung, V. S. Tseng, G. I. Webb, B. Ho, M. Ganji, and L. Rashidi (Cham: Springer International Publishing), 247–259.

Google Scholar

Kim, A. G., and Yoon, S. (2021). “Corporate bankruptcy prediction with domain-adapted BERT,” in Proceedings of the Third Workshop on Economics and Natural Language Processing (Punta Cana: Association for Computational Linguistics), 26–36.

Google Scholar

Kitamori, S., Sakai, H., and Sakaji, H. (2017). “Extraction of sentences concerning business performance forecast and economic forecast from summaries of financial statements by deep learning,” in 2017 IEEE Symposium Series on Computational Intelligence (SSCI) (Honolulu, HI: IEEE), 1–7.

Google Scholar

Koppel, M., and Shtrimberg, I. (2006). Good News or Bad News? Let the Market Decide. Dordrecht: Springer Netherlands.

Google Scholar

Loughran, T., and Mcdonald, B. (2014). Measuring readability in financial disclosures. J. Finance 69, 1643–1671. doi: 10.1111/jofi.12162

CrossRef Full Text | Google Scholar

Low, B.-T., Chan, K., Choi, L.-L., Chin, M.-Y., and Lay, S.-L. (2001). “Semantic expectation-based causation knowledge extraction: a study on hong kong stock movement analysis,” in Advances in Knowledge Discovery and Data Mining, eds D. Cheung, G. J. Williams, and Q. Li (Berlin; Heidelberg: Springer Berlin Heidelberg), 114–123.

Google Scholar

Milea, V., Sharef, N. M., Almeida, R. J., Kaymak, U., and Frasincar, F. (2010). “Prediction of the msci euro index based on fuzzy grammar fragments extracted from european central bank statements,” in 2010 International Conference of Soft Computing and Pattern Recognition (Paris), 231–236.

Google Scholar

Oliveira, N., Cortez, P., and Areal, N. (2017). The impact of microblogging data for stock market prediction: using twitter to predict returns, volatility, trading volume and survey sentiment indices. Expert. Syst. Appl. 73, 125–144. doi: 10.1016/j.eswa.2016.12.036

CrossRef Full Text | Google Scholar

Ranco, G., Aleksovski, D., Caldarelli, G., Grcar, M., and Mozetic, I. (2015). The effects of twitter sentiment on stock price returns. PLoS ONE 10, e0138441. doi: 10.1371/journal.pone.0138441

PubMed Abstract | CrossRef Full Text | Google Scholar

Sakaji, H., Murono, R., Sakai, H., Bennett, J., and Izumi, K. (2017). “Discovery of rare causal knowledge from financial statement summaries,” in The 2017 IEEE Symposium on Computational Intelligence for Financial Engineering and Economics (CIFEr) (Honolulu, HI: IEEE), 1–7.

Google Scholar

Sakaji, H., Sakai, H., and Masuyama, S. (2008). “Automatic extraction of basis expressions that indicate economic trends,” in Advances in Knowledge Discovery and Data Mining (Berlin; Heidelberg: Springer Berlin Heidelberg), 977–984.

Google Scholar

Schumaker, R. P., and Chen, H. (2009). Textual analysis of stock market prediction using breaking financial news: the azfin text system. ACM Trans. Inf. Syst. 27, 12:1–12:19. doi: 10.1145/1462198.1462204

CrossRef Full Text | Google Scholar

Schumaker, R. P., Zhang, Y., Huang, C.-N., and Chen, H. (2012). Evaluating sentiment in financial news articles. Decis Support Syst. 53, 458–464. doi: 10.1016/j.dss.2012.03.001

CrossRef Full Text | Google Scholar

Smailović, J., Grčar, M., Lavrač, N., and Žnidaršič, M. (2013). “Predictive sentiment analysis of tweets: a stock market application,” in Human-Computer Interaction and Knowledge Discovery in Complex, Unstructured, Big Data, eds A. Holzinger and G. Pasi (Berlin; Heidelberg: Springer Berlin Heidelberg), 77–88.

Google Scholar

Suzuki, M., Sakaji, H., Izumi, K., Matsushima, H., and Ishikawa, Y. (2020). Forecasting net income estimate and stock price using text mining from economic reports. Information 11, 292. doi: 10.3390/info11060292

CrossRef Full Text | Google Scholar

Taguchi, R., Watanabe, H., Hirano, M., Suzuki, M., Sakaji, H., Izumi, K., et al. (2021). “Market trend analysis using polarity index generated from analyst reports,” in 2021 IEEE International Conference on Big Data (Big Data) (Orlando, FL: IEEE), 3486–3494.

Google Scholar

Tetlock, P. C., Saar-Tsechansky, M., and Macskassy, S. (2008). More than words: quantifying language to measure firms' fundamentals. J. Finance 63, 1437–1467. doi: 10.1111/j.1540-6261.2008.01362.x

CrossRef Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need,” in Advances in Neural Information Processing Systems 30 (NeurIPS) (Long Beach, CA: Curran Associates, Inc.), 5998–6008.

Google Scholar

Vu, T. T., Chang, S., Ha, Q. T., and Collier, N. (2012). “An experiment in integrating sentiment features for tech stock prediction in twitter,” in Proceedings of the Workshop on Information Extraction and Entity Analytics on Social Media Data (Mumbai: The COLING 2012 Organizing Committee), 23–38.

Google Scholar

Wang, C., Miao, Z., Lin, Y., and Gao, J. (2019). User and topic hybrid context embedding for finance-related text data mining. IEEE Int. Conf. Data Min. Workshops 2019, 751–760. doi: 10.1109/ICDMW.2019.00112

CrossRef Full Text | Google Scholar

Wuthrich, B., Cho, V., Leung, S., Permunetilleke, D., Sankaran, K., and Zhang, J. (1998). “Daily stock market forecast from textual web data,” inSMC'98 Conference Proceedings. 1998 IEEE International Conference on Systems, Man, and Cybernetics (Cat. No.98CH36218), Vol. 3 (San Diego, CA: IEEE), 2720–2725.

Google Scholar

Keywords: financial report, text mining, profile, BERT, stock prediction, net income

Citation: Suzuki M, Sakaji H, Izumi K and Ishikawa Y (2022) Forecasting Stock Price Trends by Analyzing Economic Reports With Analyst Profiles. Front. Artif. Intell. 5:866723. doi: 10.3389/frai.2022.866723

Received: 31 January 2022; Accepted: 25 April 2022;
Published: 07 June 2022.

Edited by:

Xiaomo Liu, J. P. Morgan AI Research, United States

Reviewed by:

Bowei Chen, University of Glasgow, United Kingdom
Denis Helic, Graz University of Technology, Austria

Copyright © 2022 Suzuki, Sakaji, Izumi and Ishikawa. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Masahiro Suzuki, YjIwMTltc3V6dWtpJiN4MDAwNDA7c29jc2ltLm9yZw==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.