1 Introduction
Macroeconomic developments, such as cyclic downturns or the economic circumstances associated with the U. S. subprime crisis, affect firm operations in multiple ways and represent direct challenges to management (e. g. Demyanyk & Hasan, 2010; Goudie & Meeks, 1982). Examples include changes in the price of goods and raw materials, as well as the impact on overall demand, supply chain utilization and even operational processes (Xu et al., 2017). Therefore, firms are interested in foreseeing the future economic climate in order to manage operations accordingly and hedge potential risks. In this context, operational research (OR) has a long tradition of addressing such risks (Demyanyk & Hasan, 2010). Our discipline has thus contributed to anticipating a variety of developments at a macroeconomic level, including financial distress (Geng et al., 2015), liquidity risks (Shaik, 2015), credit risks (Akkoç, 2012; Desai et al., 1996), financial crises (Demyanyk & Hasan, 2010; Huang et al., 2017), currency crises (Sevim et al., 2014) and bankruptcy (Du Jardin, 2015; McKee & Lensberg, 2002; Sun & Shenoy, 2007), especially in the financial sector (Tam & Kiang, 1992)
Future expectations regarding the macroeconomic environment play a critical role in the decisionmaking process for many organizations (Xu et al., 2017). Hence, decisionmakers across all sectors must analyze the current economic environment and form accurate expectations about future economic trends in order to support the operational strategy of organization and longterm management. As a result, macroeconomic variables, and the accurate prediction thereof, form the basis for a wide array of OR models (e. g. Calabrese et al., 2017; Fethi & Pasiouras, 2010; Gutiérrez & Lozano, 2012).
The importance of accurate longterm forecasts for firm operations has driven the extensive amount of research conducted with respect to macroeconomic predictions. Specific examples from the OR domain include shortterm predictions of assetrelated values, including government bonds (Tay & Cao, 2001) and stock indices (e. g. Huang et al., 2005; Kung & Yu, 2008; Oztekin et al., 2016). Further research focuses on forecasting macroeconomic indicators of single countries (e. g. Mahmoud et al., 1990) or the relationship between countries (Sermpinis et al., 2013). Other works propose agentbased simulations to study the behavior of human forecasters (Bovi & Cerqueti, 2016).
Previous efforts at forecasting macroeconomic indicators have made use of various input features and methodologies. Historic time series data is a staple input for macroeconomic forecasting, and has been applied to make both short and longterm predictions (e. g. Jansen et al., 2016; Mahmoud et al., 1990; Sermpinis et al., 2013). A prevalent alternative is the subjective judgments of professional forecasters (Matsypura et al., 2018) such as those used by the European Central Bank. However, both time series models and subjective forecasts suffer from severe prediction errors (Jansen et al., 2016), possibly because they cannot benefit from predictors that better measure the state and outlook of the economy.
It is plausible that macroeconomic forecasting could also benefit from the advances of the big data era (Mortenson et al., 2015) and potential improvements in predictive modeling. Especially unstructured data, such as news, promises further insights into global systems and the heterogeneous trends and events that occur within them. Hence, news represents a powerful source of information for financial forecasting and is thus likely to serve as an intriguing – yet currently rare – predictor for macroeconomic forecasts.
Our research therefore aims to improve macroeconomic forecasts using financial disclosures. Financial news, in particular, conveys rich information about expected firm performance that often goes beyond pure numeric data (e. g. Tsai & Wang, 2017). Text mining of financial news has been successfully used in conjunction with predictive analytics to infer profitable shortterm investment decisions in capital markets (Nassirtoussi et al., 2014; Feuerriegel & Prendinger, 2016). However, it is unclear whether financial filings – which should essentially mirror the current health of the firm – also provide longterm prognostic capacity.
To make longterm predictions of macroeconomic indicators, we construct an approach as follows: we use financial news and insert corresponding linguistic features into various machine learning methods. Our specific setting represents a major difference from previous newsbased forecasts, in the sense that the outcome variable is reported in monthly or quarterly resolution. This results in fewer observations and thus extremely highdimensional predictor matrices with severe risks of overfitting. To overcome this challenge, we develop a form of feature engineering based on semantic projections and, on top of that, propose a semantic path model that fulfills the demand of practitioners by being fully interpretable. Afterwards, we conduct an outofsample evaluation of the predictive performance. The results demonstrate that our approach is superior in terms of relative performance when compared to common time series models that serve as a benchmark.
The remainder of this paper is structured as follows. Section 2 reviews related research on macroeconomic forecasting in order to outline the aforementioned research gap. As a remedy, Section 3 explains our methodology for making newsbased macroeconomic predictions, with a particular focus on the proposed projections to latent semantic structures. Section 4 reports the datasets, based on which Section 5 compares the predictive performance of both traditional time series models and newsbased forecasts. A discussion of managerial implications follows in Section 6, while Section 7 concludes with an outlook on future research.
2 Background
2.1 Predictive models for macroeconomic forecasts
Beyond human predictions (Matsypura et al., 2018), common approaches to macroeconomic forecasting include various time series models (Allen & Morzuch, 2006)
, for instance, autoregressive movingaverage (ARMA) models and their multivariate variations, such as vector autoregression and vector error correction models. However, their use entails certain challenges, since they must be adapted to cope with structural changes in the underlying systems
(Litterman, 1986). Nevertheless, time series models still appear to be the standard benchmark against which the predictive performance of alternative approaches is measured (de Gooijer & Hyndman, 2006).Since vector autoregression and its variants are limited to a few variables, researchers have proposed alternative models, such as Bayesian vector autoregression (BVAR). The BVAR accounts more accurately for uncertainty regarding the structure of the economy and is utilized to handle instances with dozens of relevant variables (Litterman, 1986). Its use for forecasts requires extensive computational resources, especially when dealing with many predictors (Carriero et al., 2015; Litterman, 1986).
Bridge equations have been a suggested as an alternative means of incorporating relevant explanatory variables into predictive models. This approach is a form of linear regression that timeaggregates economic time series to bridge the information gap between low and high frequency indicators. It is a widely used method for forecasting macroeconomic variables
(Jansen et al., 2016). However, these forecasts are generally limited to one or two periods ahead.A different concept is represented by leading indicators. These are supposed to provide prognostic capabilities for future changes in the economic outlook. Examples include major stock indices, the Ifo Business Climate Index from Germany, or the University of Michigan Consumer Sentiment Index for the United States. These indices can yield important information regarding future cyclical fluctuations of the general economy (Allen & Morzuch, 2006) and they have thus been used to forecast imminent turning points in the business cycle (Layton & Smith, 2007). Predictive models using leading indicators have been applied, for instance, in single variable ARMA models, multivariate models, diffusion indices and BVAR (Allen & Morzuch, 2006; Stock & Watson, 2002). However, it is not clear whether these indices have a prognostic potential for future developments.
Since previous approaches to macroeconomic forecasting have primarily focused on using various forms of economic data as dynamic predictors, the use of qualitative or unstructured data, such as financial news, deserves the attention of researchers. However, the inclusion of larger numbers of predictors in traditional time series models can prove to be problematic, as such models typically struggle with such settings (Jansen et al., 2016; Litterman, 1986). The above literature review thus motivates our choice of models: we specifically decided to utilize different machine learning models that can cope with nonlinear relationships and large numbers of textual features.
2.2 Predictive text mining with financial news
Newsbased predictions of macroeconomic indicators constitute an innovative area of research. Due to the scarcity of related publications, we decided to extend our literature overview and also include newsbased forecasts of stock returns in addition to macroeconomic variables.
Text mining of financial news have been successfully used in conjunction with predictive analytics to infer profitable shortterm investment decisions in capital markets (Feuerriegel & Prendinger, 2016). Previous approaches have largely focused on predicting the response to financial news in the form of shortterm reactions of stock markets (Nassirtoussi et al., 2014), whereas we develop a forecasting methodology for the general macroeconomic environment over longterm forecast horizons. Text mining of financial news for predictions in capital markets has been extensively studied in recent years; for a comprehensive review we refer to (Nassirtoussi et al., 2014).
Financial news contains fundamental and qualitative information that influences the expectations of market participants, and has thus driven the field of market prediction. A broad range of textual materials has been in the course of recent research, including newspaper articles from media sources and regulated filings, such as ad hoc announcements, 8K filings and annual reports (Nassirtoussi et al., 2014). The common feature tying these previous efforts together is the focus on making shortterm, usually sameday, predictions.
The challenge of working with textual data, such as financial news, lies in how to efficiently transform it into a machinefriendly form and then apply a classifier to it. Hence, text mining approaches commonly follow a similar procedure with the goal of extracting and selecting relevant features that best represent the original text
(Nassirtoussi et al., 2014). In this regard, the bagofwords approach breaks up the freerunning text into single words (orgrams) which can then be used as features. Afterwards, feature selection (sometimes joined with dimensionality reduction) aims to reduce the set of features while keeping relevant information. Common approaches in financial text mining
(Nassirtoussi et al., 2014) include frequencybased statistics, i. e. term weighting. This then presents the input to machine learning classifiers. We draw upon the aforementioned approach utilizing highdimensional input matrices, which results in one of our three textbased forecasting routines.3 Methods
3.1 Overview
This section introduces the research framework for forecasting macroeconomic indicators based on financial news. This task entails two critical challenges: feature engineering and interpretability.
First, the input matrix is highdimensional as the number of predictors exceeds the number of observations. For instance, the macroeconomic indicators with quarterly resolution are predicted based on a termdocument matrix with 536 terms, while a single fold during crossvalidation might consist of only 71 samples. We address this issue through unsupervised techniques for dimension reduction and, in addition, pioneer in replacing the statistical approaches to feature generation (e. g. Liang et al., 2016) with a supervised technique that incorporates domainspecific knowledge. More specifically, it aggregates terms according to a semantic classification with projections onto corresponding constructs, thereby yielding a lowdimensional input with only a couple of predictors.
Second, machine learning is often regarded as a blackbox mechanism, whereas practitioners from operational management demand the potential for interpretation as a means to validate the computerbased forecasts with their own beliefs and expectations, as well as to gain trust in the automated decision support. Hence, we are specifically interested in models that attain a beneficial tradeoff between predictive performance and explanatory power. As a remedy, we propose a semantic path model that can offer full accountability of the forecasts but yields a performance comparable to the best blackbox models from machine learning. Here we follow the intuition that the economic climate is linked to the perception (i. e. sentiments) of the involved stakeholders. Given a meaningful choice of dimensions, we can later decompose the overall forecasts in terms of the underlying constructs.
We adapt to the two aforementioned challenges by suggesting a variety of predictive methods for our forecasting task. The underlying approach for textbased forecasts consists of preprocessing of financial text, including feature extraction (cf. online appendix); optional feature engineering, in which we project individual words onto latent semantic structures; model training and selection; and multistep ahead predictions using a test set. The individual approaches can be grouped as follows:

Baseline: Time series models. These build upon autoregressive lags and combine them in linear and nonlinear fashion.

Machine learning models with highdimensional input. A series of machine learning models is tested with highdimensional input as a comparison. The models are later fed with an input matrix that is (optionally) subject to unsupervised dimension reduction techniques.

Models with semantic features. We develop a supervised technique for feature engineering. This approach involves projections of terms onto semantic constructs that then serve as a lowdimensional feature space.

Semantic path model. Similar to the semantic features, we develop a path model that operationalizes the semantic classification of terms into different categories relevant in a macroeconomic context in order to obtain interpretable forecasts. To circumvent overfitting from collinear constructs, we develop a novel estimation technique with an additional regularization.
We note that each of the textbased models is later trained in different variants, i. e. the models are optionally augmented by the same lags as the time series baseline.
3.2 Baseline: Time series models
In order to evaluate the predictive performance of our forecasting methodology, we employ a number of benchmark models that forecast the macroeconomic time series purely on the basis of historic data points. These models include an autoregressive (AR) process based on lagged values. Formally, let denote the current time period. We then predict the value in time step , i. e. that is steps ahead. This results in
(1) 
with coefficients , inputs and where
gives the predicted indicator. In addition, we incorporate nonlinear relationships by applying a random forest (RF) to historic values for each of the macroeconomic indicators. This machine learning model frequently achieves a high predictive power with little need for tuning
(Hastie et al., 2013).For both AR and RF models, we experiment with different specifications that vary in their lag order: on the one hand, we utilize only lag and, on the other hand, all lags of order 1 to 6. By testing both variants, we also adhere to the selection made in previous literature (e. g. Litterman, 1986; Stock & Watson, 2002). We specifically decided to predict the raw values of each of the macroeconomic indicators, instead of applying additional transformation. This choice stems from the need of practitioners who require clear expectations in common and interpretable scales. As a comparison, we also predict a firstdifference term as part of our evaluation.
3.3 Models for textual predictors
All subsequent models involving textual predictors are estimated in three different variants: on the one hand, we forecast macroeconomic indicators based purely on the narrative materials, excluding the history of the outcome variable. On the other hand, we follow the above time series models and augment the predictor matrix by 1 and 6 autoregressive lags of the tobepredicted macroeconomic indicator. Thus the model can further learn from additional seasonal and trend characteristics.
3.3.1 Machine learning models for highdimensional input
As noted earlier, the documentterm matrices are wide, since the number of features (words) far exceeds the number of observations. We thus choose machine learning models that are known for their performance with wide data sets, as they reduce the risk of overfitting via feature selection, regularization and dimension reduction (Hastie et al., 2013)
. These are as follows: least squares absolute shrinkage operator (LASSO), ridge regression, elastic net (ENET), gradient boosting (GBM), principal component regression (PCR), random forest (RF) and partial least squares regression (PLSR).
Following common conventions, the above models are estimated based on the termdocument matrix. The entries are scaled by tfidf weighting in order to reflect how frequently a term appears in a document in the corpus. Beyond that, the experiments are repeated with different unsupervised transformations for dimension reduction that are applied to the input
as part of a feature engineering step. More precisely, we apply a principal component analysis (PCA) and compute a latent semantic analysis (LSA), which are then fed into the different machine learning models. Altogether, we obtain a total of 21 different models (i. e. 7 classifiers, each with raw tfidf, as well as 2 unsupervised dimension reduction techniques).
3.3.2 Models with semantic features
We now present an alternative approach to feature generation. More precisely, we replace the purely statistical approach behind unsupervised transformations for dimension reduction (i. e. PCA and LSA) with a supervised routine that encodes domainspecific knowledge.
Our approach assumes that the overall macroeconomic climate is driven by the perceptions (or sentiments) of individual stakeholders. For instance, a construct “macroeconomic uncertainty” could be indicated by a term “risk”, while “increase” links to a construct “positive outlook”. Hence, we identify a new set of features which represent such meaningful constructs (i. e. perceptions). Each construct is represented by a linear combination of terms or, more specifically, a subset thereof; i. e. with coefficients . The subset then incorporates our knowledge with regard to how terms correspond to the different constructs. Here merely a categorization of words is needed in the form of a simple wordlist that neither has weights nor is mutually exclusive. As a result, the highdimensional input matrix is projected onto a lowdimensional feature space based on the the prespecified semantic structure.
The weights are computed in a supervised manner by making use of the response . That is, the weights are chosen such that the explain as much of the covariance between and as possible. Mathematically, the computation draws upon the existing routine for projections on latent constructs (Hastie et al., 2013). This algorithm computes the coefficients for the each construct similar to least squares, i. e. , with the difference that it involves the orthogonalization step from path modeling, as terms can appear in more than one construct (Tenenhaus et al., 2005). This highlights the differences from traditional PLSR as a predictive model, in which the number of constructs is subjected to tuning and all terms map on all features. Conversely, our semantic features impose a structure where only a subset of terms maps to each feature, thereby incorporating domain knowledge.
We additionally utilize a Bayesian neural network as the underlying predictive model in order to infer
from , thereby essentially yielding a universal structure model, USM for short (Buckler & HennigThurau, 2008). In fact, universal structure modeling has found to be effective in multiple applications from the fields of decision support and operations research (e. g. Oztekin et al., 2011; Turkyilmaz et al., 2013, 2016).3.3.3 Semantic path model
We now present the semantic path model that also involve projections onto lowdimensional feature spaces. These again build upon a meaningful representation of different perceptions of the macroeconomic climate as latent constructs. However, this approach attains full interpretability as the underlying combination construct is linear. Figure 1 illustrates an example in which terms are mapped onto two constructs – negative outlook and economic uncertainty – which then yield the final prediction. Such path modeling techniques are common in the social sciences (cf. AguirreUrreta & Marakas, 2014; Aubert et al., 1996; McIntosh et al., 2014), where the constructs quantify different aspects of human behavior (Tenenhaus et al., 2005; Rigdon et al., 2010). However, their adaptation to semantic structures for linguistic materials reveals an innovative application with considerable potential for interpreting languagebased predictions.
Mathematically, the model builds upon the prescribed paths of which terms relate to construct . Each construct is then computed by a linear combination , i. e. a projection of words onto latent semantic constructs. Then, the final prediction is an additional linear model . Evidently, one can directly interpret the forecast and decompose its value into individual components , , etc. Hence, decisionmakers can assess the underlying reasoning of the model and evaluate the contribution of the different semantic categories to the overall outcome.
The above model can be estimated based on the algorithms for path modeling based on projections to latent structures (Tenenhaus et al., 2005), which we denote in the following:
Beyond the naïve estimation approach to path modeling based on projection to latent structures (Tenenhaus et al., 2005), we propose an alternative strategy: we argue that the potentially collinear relationships between the different constructs could lead to a low capacity for generalization and thus suboptimal forecasts. This differs from the majority of path models in social sciences, where constructs represent distinct concepts (e. g. Aubert et al., 1996) and thus mostly orthogonal vectors (Tenenhaus et al., 2005). However, in our case, a high negativity is likely to correspond to low positivity and vice versa. Hence, this can be addressed by replacing the ordinary least squares (OLS) estimator from the inner model in path modeling with regularized variants, namely, LASSO, ridge regression and elastic net. This results in the general form
(3) 
where the coefficients and
control the amount of shrinkage and are subject to hyperparameter tuning. This merely affects the magnitude of the coefficients
but maintains the linear relationship that lends itself to direct interpretation. In fact, the LASSO can set coefficients directly to zero and, accordingly, could even perform an implicit selection of relevant constructs. Closest to this are different approaches for shrinkage or the inclusion of sparsity in partial least squares regression (e. g. Chun & Keleş, 2010; Lê Cao et al., 2008), but, to the best of our knowledge, penalization has not yet been incorporated in path modeling.The semantic path model and PLSR share similarities in terms of that both implement projections to latent structures. However, a clear difference becomes evident in how these mappings are defined. PLSR follows a datadriven approach where all terms map onto a number of latent constructs that is subject to tuning. Thereby, the latent constructs loose their ability of interpretation and collinearity among features – such as in our highdimensional setting – could impede generalizability. In contrast, our semantic path model focuses on a knowledgedriven methodology as it directly encodes a structure that lends to interpretation. The numerical results later indicate the potential performance improvements in our setting with extremely highdimensional predictor matrices. Moreover, the proposed use of additional regularization can further bolster interpretation from more parsimonious structures and shrink the influence of collinear constructs, thereby reducing the risk of overfitting.
3.4 Model tuning
All of the previous models from both benchmarks and newsbased predictors were tuned by following the default parameter search grid from Kuhn (2008)
. However, time series data does not lend itself to traditional random sampling used for crossvalidation and we instead utilize a timeslice crossvalidation with ten folds. This applies a rolling forecast origin, where the sample used for calibration contains only observations that occurred prior to the observations that form the validation set. The process is then repeated by rolling the origin of the validation set forward, thus incorporating the previous observations into the estimation procedure. The model is then retested using the next window of observations. Afterwards, the best configuration of hyperparameters is retained, based on which the final model is fitted to the complete training data. All models are tuned with the goal of minimizing the root mean squared errors. All models preprocess the predictors by scaling and centering to zero mean and a variance of one.
4 Datasets
4.1 Financial news
To make longterm predictions of macroeconomic indicators, we use financial news in the form of ad hoc announcements. Ad hoc announcements represent regulatory filings that are obligatory for all firms listed on the German stock exchange (cf. German Securities Trade Act for details). This dataset offers a number of advantages. First, German regulations require listed companies to immediately publish any information which could potentially impact the price of their stock via this type of announcement. Second, the content of the announcement is authorized by the executive board and finally checked by the Federal Financial Supervisory Authority. The tight regulations reduce the likelihood of nonrelevant content. Third, standardized distribution channels guarantee realtime access to disclosures. This allows one to derive a live picture of the industrial sector, since corporations selfreport their current performance but also incorporate statements regarding their optimism, future expectations and legal risks. Altogether, we possess a rich source of information to help sense – potentially early – economic developments from a productionbased view. Ad hoc announcements have been subject to extensive research revealing the benefits of their use in both practice and research, for instance, with regard to shortterm investment decisions and risk predictions (e. g. Kraus & Feuerriegel, 2017).
Our corpus is composed of all 80,813 ad hoc announcements which we collected from the website of the EQS Group, the leading publisher of ad hoc announcements. Our filings are linked to publiclytraded companies and were published between July 1996 and April 2016. The dataset is split chronologically into a training set (60 %) and a test set (40 %). Announcements were selected such that our corpus includes those that are published in either English (our default analysis) or German, as well as a combined corpus of both. This facilitates later experiments that analyze the sensitivity of newsbased forecasts with regard to the underlying language.
4.2 Macroeconomic variables
The ad hoc announcements are predominantly from German companies. Therefore, we selected a set of macroeconomic indicators with the aim of reflecting the general German economic climate. To account for the share of international firms, we also included a number of European indicators for comparison. Table 1 reports our macroeconomic indicators with both monthly and quarterly resolution.
4.3 Descriptive statistics
The average length of an ad hoc announcement over the period amounts to words. LABEL:tbl:descriptives lists further descriptives per year. It further indicates an increasing length of the ad hoc announcements over the research period. The average number of ad hoc announcements published per year is to 4,072 (for the period 1997–2015).
4.4 Semantic constructs
The proposed projections require a categorization of terms into different semantic structures. The method makes no a priori assumption regarding the structure and, hence, the final choice can be subject to the intended insights. In fact, the projections support the conventional wordlists from that are widely used in the finance domain or social sciences (the semantic path model merely requires a categorization with neither weights nor mutuallyexclusive assignments). Other semantic categories might be beneficial when adapting our approach to other predictive settings, such as marketing.
In our computational experiments, the constructs should lend themselves to interpretation with regard to perceived sentiments; e. g. the valence of the outlook, the uncertainty of the economic development, and the associated confidence by the reporting companies. For this reason, we follow research on financial markets (Nassirtoussi et al., 2014) and draw upon the LoughranMcDonald financespecific dictionary^{1}^{1}1See https://www3.nd.edu/~mcdonald/Word_Lists.html, last accessed January 12, 2018., which categorizes words into different semantic categories that reflect six such perceptions. The positive and negative categories entail terms that refer to positive and negative assessments; uncertainty encodes terms used to signal vagueness or even risks with regard to macroeconomic development; legal refers to, e. g., potential liabilities in the future; strong modal and tentative (i. e. weak modal) measure the confidence with regard to statements, conclusions and firm developments. The assignment of words is not unique, as several words belong to more than one semantic category. For instance, all tentative (i. e. weak modal) terms are included in the uncertainty list. In total, 132 out of 2,362 entries appear in more than one construct. The choice of this dictionary is beneficial to our setting as the wordlists have been specifically designed to extract qualitative information from regulatory firm disclosures.
5 Results
This section summarizes the results of our outofsample predictive experiments. The predictive performance of the newsbased models are evaluated relative to the benchmark models. Here we conducted two different prediction experiments. In the first experiment, the models were trained to predict the raw value of each indicator over multiple prediction horizons. In the second experiment, we looked to test the ability of our newsbased approach to predict imminent turning points in the business cycle, where it predicts the change =.
LABEL:tbl:results_monthly and 2
compare the prediction performances as a bestofbreed measure. That is, we only report the bestperforming time series baseline, the best result from semantic path models, etc. For each bestofbreed model, the performance is measured in terms of root mean squared error (RMSE). In addition, we validate the robustness of our results by employing the DieboldMariano test. The test statistic estimates whether our newsbased models obtain statistically significant lower forecast errors compared to the benchmark models. We expect this value frequently to be low due to the small size of the test set and the overall difficulty of macroeconomic predictions. Later, the quality of the forecasts is compared across the different macroeconomic indicators; see
Section 5.4. Here we provide the normalized RMSE as it facilitates comparison across different predicted variables.5.1 Prediction performance across macroeconomic indicators
5.1.1 Eurozone GDP
Table 2 lists the RMSE values obtained from the predictive experiments undertaken for the Eurozone GDP. Several newsbased models are able to surpass the accuracy of the benchmarks for a variety of the prediction horizons. First, in the case of the firstdifference and quarter ahead prediction horizon, both the semantic path model and semantic features prove to be more accurate than both the highdimensional input and benchmark models. Relative to best highdimensional input, the RMSE of and represents 23 % and 44.8 % reduction, respectively.
When looking at the longterm predictions, the LASSOtrained highdimensional input model outperforms the benchmark for the quarters ahead prediction horizon, in this case achieving a RMSE of which is a 5 % reduction relative to the baseline. Furthermore, the highdimensional input prove to be more accurate over the medium and longterm prediction horizons than the semantic paths models.
5.1.2 German GDP
The prediction errors for the German GDP are as follows. We find the elastic net trained on the highdimensional input the most accurate for firstdifference prediction horizon, achieving a RMSE of . We observe improvements over the benchmark only fore aforementioned horizons. This is especially interesting, since it differs from the prediction results for the whole Eurozone, where we achieved consistent improvements.
5.1.3 Unemployment
In terms of unemployment rate, the newsbased models outperform the benchmark over the firstdifference prediction. We find both the semantic path model, as well as the semantic features, to be the most accurate with a RMSE of . Furthermore, the DM test results are highly significant for both models derived from semantic inputs.
We observe that newsbased models cannot beat the benchmarks for mediumterm predictions, while we again gain improvements from using news in the longterm. Here the newsbased models consistently outperform the baseline for the months ahead prediction horizon. The highest predictive accuracy stems from gradient boosting applied to highdimensional input with a prediction error of . The best of the benchmarks reported only an RMSE of ; this results in a forecast error reduction of . Furthermore, the DieboldMariano test results are highly statistically significant for the months prediction horizon for all newsbased models. The semantic path model attains only an improvement of , yet while being fully interpretable.
5.1.4 Inflation
We find support for the fact that the newsbased models are superior to the benchmarks for the first difference prediction horizons. The highdimensional input accounts for the lowest RMSE (). Furthermore, the DieboldMariano test is statistically significant. For the prediction horizons to months ahead, our newsbased models attain similar levels of accuracy however are unable to best the leading benchmark. Further improvements in predictive performance are evident for the months ahead prediction horizon. Here the highdimensional input with PCA dimensionality reduction yields the lowest RMSE of , i. e. an improvement of over the time series baseline. The observed results indicate that the financial news stemming from listed corporations can be utilized to capture the rate of change in consumer prices in the shortterm, and that these, continue to be predictive over the longterm.
5.1.5 Industrial production
The newsbased models outperform the benchmarks for a majority of tested prediction horizons. In particular, the newsbased model achieve a lower forecast error for the midrange prediction horizons. The experiments for the firstdifference prediction reveal that the highdimensional input with LSA preprocessing obtains a RMSE of (i. e. an improvement of . The newsbased models outperform the benchmark for the 3–12 months ahead prediction horizons. Both the semantic path model and the semantic features attain the best RMSE (), for the months ahead prediction. The highdimensional input is the most accurate for the 6–12 months ahead prediction horizon. The values from the individual DM tests indicate that predicted values from our newsbased system are significantly more accurate over the beforementioned prediction horizons.
5.1.6 Business confidence
The semantic path model, as well as semantic features, are able to outperform the benchmark and the highdimensional input across a number of prediction horizons. Both methods register a the lowest RMSE for the month (), month () and month () ahead prediction horizons. The latter amounts to an improvement of . Furthermore, these model significantly outperform the highdimensional input across all predictions horizons. The business confidence index is calculated by conducting surveys of business leaders who know the current state of the corporate sector. The semantic path model can adapt at capturing the sentiment of the individual announcements; i. e. the same sentiment would be reflected in the index. The lack of longterm predictive power for the highdimensional inputs could be due to the fact that the news content might already be factored into their survey answers, therefore reducing potential lag affects. Overall, the semantic models seems better suited for predictions of the business confidence index than the highdimensional inputs.
5.1.7 Consumer confidence
We investigate the prognostic capability of financial news for the consumer confidence index. Overall, the predictive performance of the newsbased models does not diverge greatly from those of the benchmarks. Of the newsbased models, both semantic models prove to be more accurate than the highdimensional input. The newsbased models do narrowly outperform the benchmark for prediction horizons greater than months ahead.
5.1.8 Government bonds
The newsbased models perform best relative to the benchmarks for the firstdifference prediction horizons. For instance, the highdimensional input attains the lowest overall RMSE (, i. e. a minus of over the best baseline), closely followed by both the semantic models (, i. e. a minus of ), while the RMSE of the best benchmark is . The highdimensional input is also able to outperform best the benchmark for the months ahead prediction horizon. The DieboldMariano test results for the highdimensional input are all highly statistically significant.
5.2 Interpretation of latent semantic structures
LABEL:fig:interpretation presents the results from the semantic path model and decomposes the forecasts into different constructs in order to facilitate interpretation. That is, we compute the value of each construct – given by predicted value – and compare these figures across the different semantic categories according to the LoughranMcDonald financespecific dictionary. These categories are: (i) positivity/negativity of economic outlook, (ii) uncertainty regarding economic development, (iii) legal risk for operations and (iv) confidence/modality of the disclosing company.
Altogether, we find the positive construct has the strongest impact on both Eurozone and German GDP and, to a lesser extent, on consumer confidence, industrial production and government bonds. The negative construct, on the other hand, affects the unemployment rate, inflation, industrial production, consumer confidence and government bonds. A rise in negative news could see firms downsize operations and cut investment spending, resulting in lower levels of employment. Reduced confidence in the economic outlook could see consumers also cut spending placing downward pressure on prices.
The lag component is strongly predictive for all variables except business confidence. Potentially, a change in perceptions at a corporate level would already be factored into the index by business leaders. The remaining indicators would be slower to adjust. We find the uncertainty construct to be predictive for inflation, the confidence indices and government bonds. Consumers, firms and financial markets all have an aversion to uncertainty. The legal construct seems to be predictive for industrial production, business and consumer confidence. A possibility explanation is that firms facing legal difficulties would reduce output and other business investment impacting both the suppliers and consumers that function within the economic stream. Finally, we find the modal constructs to be a weak predictor across all of the chosen indicators.
5.3 Sensitivity analysis
LABEL:tbl:results_monthly and 2 entail additional sensitivity checks. First, we compare stemming as a default in contrast to lemmatization. While the latter adapts correctly to the semantic context, it suffers from larger predictor matrices and potentially higher risk of overfitting. However, our results suggest no clear indication concerning which approach is preferable with regard to our task. Second, we performed all analyses based on the English corpus and now compare this choice to both a corpus with German articles and a combined one. Again, we find no clear evidence that one consistently outperforms the other. A potential reason for this parity stems from the fact that the majority of firms publish in both languages and any potential differences therefore are relatively minor.
5.4 Comparison
While the previous results compared the use of time series models to a machine learning approach, we now turn our attention to the overall capacity of foreseen macroeconomic development. For this reason, we computed the normalized RMSE for all macroeconomic indicators as it facilitates comparisons between different outcome variables and their respective scales. Formally, it is defined by .
The normalized RMSE results are detailed in Table 3. We find the Eurozone GDP results to have lower prediction errors than German GDP. Potential reasons could be the export orientation of German companies, or the higher volatility of the Eurozone GDP, which can be better explained by productionrelated information from news disclosures. The prediction accuracy for the business confidence index relative to consumer confidence is also better. Given the direct link between the business leaders who are surveyed to estimate the index and the source of the ad hoc announcements, a higher rate of accuracy for the business confidence index could be expected.
In comparing the performance of both the semantic models relative to the highdimensional input, we find the results varying depending on the indicator and prediction horizon. The semantic models consistently attain lower prediction errors for the confidence indicies across all the prediction horizons except for the very long run. In fact, the semantic path model accomplishes the best forecast in 6 prediction experiments, while the semantic features do so in 8 cases and the highdimensional in 12, yet the latter with hardly any interpretability. This is interesting in the light that our semantic path model enforces a linear structure among constructs and thus excludes potentially nonlinear relationships as used by predictive models with semantic features or the universal structure model. While nonlinearity introduces additional flexibility, our setting is extremely overspecified and the smaller parameter space of linear models can often be beneficial. Moreover, we see that traditional approaches to unsupervised dimensionality reduction rarely yield the bestofbreed model; conversely, our domain knowledge can outperform them frequently.
The overall performance of the newsbased models varies depending on the macroeconomic variable and prediction horizon. Over the short and medium term, the benchmark models tend to be superior in terms of consumercentered variables, such as unemployment rate or inflation. However, financial news reflects both the current health and future expectations of corporations. As a consequence, newsbased models appear especially useful for productrelated indicators, such as the actual industrial product rate and confidence indices. Finally, in terms of longterm predictions, the newsbased models tend to outperform the benchmark models. Altogether, this provides ample evidence of the strength of the predictive properties of our newsbased inputs.
6 Managerial implications
The economic climate has longranging effects on firm operations. As a consequence, managers and policymakers alike are required to form expectations, hedge potential operational risks from economic downturns and adapt their operations accordingly. Our method thus assists managers in making such predictions and provides a costeffective alternative to expert forecasts. Firm managers can utilize our predictions in order to, among other things, set longterm strategic plans, decide on investment levels, and manage human resources. Beyond that, managers in public sector institutions may benefit from our longterm predictions when setting fiscal policy, such as government spending, borrowing and taxation, and monetary policy concerning to control of the supply of money within the economy. Our innovative, newsbased methodology also provides investment managers a greater ability to anticipate the impact of shortterm economic fluctuations on their portfolios and, if necessary, take the necessary precautions to hedge against losses.
Professional macroeconomic forecasts, such as those from central banks, usually stem from both quantitative predictions and the judgment of economic experts (Matsypura et al., 2018). For example, the German Bundesbank combines factor models, demand and supplyside bridge equations and time series models for industrial production. Expert knowledge is then applied to these quantitative forecasts to determine the final projections for key macroeconomic indicators.^{2}^{2}2German Bundesbank. Forecasting models in shortterm business cycle analysis: A workshop report. URL: https://www.bundesbank.de/Redaktion/EN/Downloads/Publications/Monthly_Report_Articles/2013/2013_09_forecasting.pdf, last accessed on January 12, 2018. The European Central Bank has a similar process and their projections “may incorporate a fair amount of judgment”missing (European Central Bank, 2016). Their questionnaire concludes that “respondents consider their predictions to be judgmentbased”missing.^{3}^{3}3European Central Bank. Results of a special questionnaire for participants in the ECB Survey of Professional Forecasters. URL: https://www.ecb.europa.eu/stats/prices/indic/forecast/shared/files/quest_summary.pdf, last accessed on January 12, 2018. Hence, it is likely that these forecasts
Numerous studies have analyzed professional forecasts for their predictive performance and ability to identify potential biases (Blanc & Setzer, 2015; Mostard et al., 2011). In the field of macroeconomic forecasting, a recent study (Dovern & Weisser, 2011) finds that the distribution of a forecasts accuracy varies significantly across indicators, forecasters and nations. A further study reveals that consensus estimates are poor predictors of gross domestic product as compared with statistical models (Jansen et al., 2016). However, the consensus estimates tend to perform better in periods of crisis. An intriguing approach would be to combine expert estimations and statistical forecasts (Lessmann et al., 2012).
Predicting macroeconomic variables is difficult, requiring timeintensive collection of economic data, which, as a result, is often out of date. However, our news source provides free and realtime access. Additionally, time series models are excellent at picking up trends, but perform poorly in situations of volatility (Jansen et al., 2016). In contrast, our approach can incorporate additional predictors besides trend, as it extracts qualitative information, as well as overall sentiments, regarding the economic outlook from financial news. At the same time, our semantic path model accounts for interpretable results and can thus be further assessed by decisionmakers.
Our innovative, newsbased methodology adds to previous literature and entails a number of important implications for researchers, as well as for the public and private sectors, by providing longterm forecasts of the macroeconomic climate. Here big data is once again shifting the frontiers of research (Mortenson et al., 2015) and is likely to change the way in which operations adapt to economic developments. As demonstrated by our previously discussed evaluations, the system elicits a high predictive accuracy, when compared with a number of benchmark models. This contributes to a growing body of literature that replaces manmade forecasts with computerized and finetuned predictive models. Our proposed methodology could be extremely valuable for small enterprises or countries in which macroeconomic forecasts are scarce due to the associated costs.
7 Conclusion
The volatile nature of the economy affects the operation of firms in key dimensions, such as prices for goods and services, financial risk, utilization of supply chains and customer demand. Hence, it is important for managers to obtain an accurate prognosis of developments in the economy in order to direct operations accordingly. Hitherto, macroeconomic forecasts have been dominated by expert opinions and rather simple time series models, while the big data era has created new opportunities to enhance the predictive power of macroeconomic forecasts based on unstructured data sources.
This work contributes a newsbased methodology for predicting macroeconomic indicators. First, we experiment with conventional models from machine learning in order to predict macroeconomic outcomes on the basis of word occurrences and historic lags. Backtesting shows that this method outperforms the benchmark models in an outofsample evaluation for various prediction horizons and macroeconomic indicators. Yet textbased predictions represent a challenging undertaking due to potential overfitting for a highdimensional input. Instead, we suggest an alternative approach whereby the words from different semantic categories are mapped onto latent structures as a form of feature engineering. This shrinks the feature space considerably, thereby achieving superior outofsample performance in multiple experiments. Beyond that, we propose semanticallystructured variant of partial least squares that reaches a comparable accuracy, but while fulfilling the demand of practitioners in being fully interpretable. This contributes to a greater understanding of how qualitative information can be used to make longterm predictions for key macroeconomic indicators.
This paper opens avenues for future research in a number of directions. First, the proposed semantic path model can facilitate interpretable predictions in a variety of settings involving natural language. Second, the approach can be seamlessly extended to include alternative data sources, including additional news sources and time series. Third, one could weigh corporate disclosures according to their market capitalization, since larger firms presumably have a great impact on the general economy than smaller firms.
References
References
 AguirreUrreta & Marakas (2014) AguirreUrreta, M. I., & Marakas, G. M. (2014). Research note—partial least squares and models with formatively specified endogenous constructs: A cautionary note. Information Systems Research, 25, 761–778.
 Akkoç (2012) Akkoç, S. (2012). An empirical comparison of conventional techniques, neural networks and the three stage hybrid adaptive neuro fuzzy inference system (ANFIS) model for credit scoring analysis: The case of turkish credit card data. European Journal of Operational Research, 222, 168–178.
 Allen & Morzuch (2006) Allen, P. G., & Morzuch, B. J. (2006). Twentyfive years of progress, problems, and conflicting evidence in econometric forecasting: What about the next 25 years? International Journal of Forecasting, 22, 475–492.
 Aubert et al. (1996) Aubert, B. A., Rivard, S., & Patry, M. (1996). Development of measures to assess dimensions of is operation transactions. Omega, 24, 661–680.
 Blanc & Setzer (2015) Blanc, S. M., & Setzer, T. (2015). Analytical debiasing of corporate cash flow forecasts. European Journal of Operational Research, 243, 1004–1015.
 Bovi & Cerqueti (2016) Bovi, M., & Cerqueti, R. (2016). Forecasting macroeconomic fundamentals in economic crises. Annals of Operations Research, 247, 451–469.
 Buckler & HennigThurau (2008) Buckler, F., & HennigThurau, T. (2008). Identifying hidden structures in marketing’s structural models through universal structure modeling. Marketing Journal of Research and Management, 4, 49–68.
 Calabrese et al. (2017) Calabrese, R., Degl’Innocenti, M., & Osmetti, S. A. (2017). The effectiveness of tarpcpp on the us banking industry: A new copulabased approach. European Journal of Operational Research, 256, 1029–1037.
 Carriero et al. (2015) Carriero, A., Clark, T. E., & Marcellino, M. (2015). Bayesian VARs: Specification choices and forecast accuracy. Journal of Applied Econometrics, 30, 46–73.
 Chun & Keleş (2010) Chun, H., & Keleş, S. (2010). Sparse partial least squares regression for simultaneous dimension reduction and variable selection. Journal of the Royal Statistical Society (Series B), 72, 3–25.
 Demyanyk & Hasan (2010) Demyanyk, Y., & Hasan, I. (2010). Financial crises and bank failures: A review of prediction methods. Omega, 38, 315–324.
 Desai et al. (1996) Desai, V. S., Crook, J. N., & Overstreet, G. A. (1996). A comparison of neural networks and linear scoring models in the credit union environment. European Journal of Operational Research, 95, 24–37.
 Dovern & Weisser (2011) Dovern, J., & Weisser, J. (2011). Accuracy, unbiasedness and efficiency of professional macroeconomic forecasts: An empirical comparison for the G7. International Journal of Forecasting, 27, 452–465.
 Du Jardin (2015) Du Jardin, P. (2015). Bankruptcy prediction using terminal failure processes. European Journal of Operational Research, 242, 286–303.
 European Central Bank (2016) European Central Bank (2016). A guide to the Eurosystem/ECB staff macroeconomic projection exercises. Frankfurt am Main: European Central Bank.

Fethi & Pasiouras (2010)
Fethi, M. D., & Pasiouras, F.
(2010).
Assessing bank efficiency and performance with operational research and artificial intelligence techniques: A survey.
European Journal of Operational Research, 204, 189–198.  Feuerriegel & Prendinger (2016) Feuerriegel, S., & Prendinger, H. (2016). Newsbased trading strategies. Decision Support Systems, 90, 65–74.
 Geng et al. (2015) Geng, R., Bose, I., & Chen, X. (2015). Prediction of financial distress: An empirical study of listed Chinese companies using data mining. European Journal of Operational Research, 241, 236–247.
 de Gooijer & Hyndman (2006) de Gooijer, J. G., & Hyndman, R. J. (2006). 25 years of time series forecasting. International Journal of Forecasting, 22, 443–473.
 Goudie & Meeks (1982) Goudie, A. W., & Meeks, G. (1982). The effects of macroeconomic developments on individual companies’ flows of funds. Omega, 10, 361–371.
 Gutiérrez & Lozano (2012) Gutiérrez, E., & Lozano, S. (2012). A competing risks analysis of the duration of federal target funds rates. Computers & Operations Research, 39, 785–791.
 Hastie et al. (2013) Hastie, T. J., Tibshirani, R. J., & Friedman, J. H. (2013). The Elements of Statistical Learning: Data Mining, Inference, and Prediction. Springer Series in Statistics (2nd ed.). New York, NY: Springer.

Huang et al. (2005)
Huang, W., Nakamori, Y., &
Wang, S.Y. (2005).
Forecasting stock market movement direction with support vector machine.
Computers & Operations Research, 32, 2513–2522.  Huang et al. (2017) Huang, Y., Kou, G., & Peng, Y. (2017). Nonlinear manifold learning for early warnings in financial markets. European Journal of Operational Research, 258, 692–702.
 Jansen et al. (2016) Jansen, W. J., Jin, X., & de Winter, J. M. (2016). Forecasting and nowcasting real gdp: Comparing statistical models and subjective forecasts. International Journal of Forecasting, 32, 411–436.

Kraus & Feuerriegel (2017)
Kraus, M., & Feuerriegel, S.
(2017).
Decision support from financial disclosures with deep neural networks and transfer learning.
Decision Support Systems, 104, 38–48.  Kuhn (2008) Kuhn, M. (2008). Building predictive models in R using the caret package. Journal of Statistical Software, 28, 1–26.
 Kung & Yu (2008) Kung, L.M., & Yu, S.W. (2008). Prediction of index futures returns and the analysis of financial spillovers: A comparison between GARCH and the grey theorem. European Journal of Operational Research, 186, 1184–1200.
 Layton & Smith (2007) Layton, A. P., & Smith, D. R. (2007). Business cycle dynamics with duration dependence and leading indicators. Journal of Macroeconomics, 29, 855–875.
 Lê Cao et al. (2008) Lê Cao, K.A., Rossouw, D., RobertGranié, C., & Besse, P. (2008). A sparse PLS for variable selection when integrating omics data. Statistical Applications in Genetics and Molecular Biology, 7, 35.
 Lessmann et al. (2012) Lessmann, S., Sung, M.C., Johnson, J. E., & Ma, T. (2012). A new methodology for generating and combining statistical forecasting models to enhance competitive event prediction. European Journal of Operational Research, 218, 163–174.
 Liang et al. (2016) Liang, D., Lu, C.C., Tsai, C.F., & Shih, G.A. (2016). Financial ratios and corporate governance indicators in bankruptcy prediction: A comprehensive study. European Journal of Operational Research, 252, 561–572.
 Litterman (1986) Litterman, R. B. (1986). Forecasting with Bayesian vector autoregressions: Five years of experience. Journal of Business & Economic Statistics, 4, 25–38.
 Mahmoud et al. (1990) Mahmoud, E., Motwani, J., & Rice, G. (1990). Forecasting us exports: An illustration using time series and econometric models. Omega, 18, 375–382.
 Matsypura et al. (2018) Matsypura, D., Thompson, R., & Vasnev, A. L. (2018). Optimal selection of expert forecasts with integer programming. Omega, forthcoming.
 McIntosh et al. (2014) McIntosh, C. N., Edwards, J. R., & Antonakis, J. (2014). Reflections on partial least squares path modeling. Organizational Research Methods, 17, 210–251.
 McKee & Lensberg (2002) McKee, T. E., & Lensberg, T. (2002). Genetic programming and rough sets: A hybrid approach to bankruptcy classification. European Journal of Operational Research, 138, 436–451.
 Mortenson et al. (2015) Mortenson, M. J., Doherty, N. F., & Robinson, S. (2015). Operational research from Taylorism to Terabytes: A research agenda for the analytics age. European Journal of Operational Research, 241, 583–595.
 Mostard et al. (2011) Mostard, J., Teunter, R., & de Koster, R. (2011). Forecasting demand for singleperiod products: A case study in the apparel industry. European Journal of Operational Research, 211, 139–147.
 Nassirtoussi et al. (2014) Nassirtoussi, A. K., Aghabozorgi, S., Wah, T. Y., & Ngo, D. C. L. (2014). Text mining for market prediction: A systematic review. Expert Systems with Applications, 41, 7653–7670.
 Oztekin et al. (2016) Oztekin, A., Kizilaslan, R., Freund, S., & Iseri, A. (2016). A data analytic approach to forecasting daily stock returns in an emerging market. European Journal of Operational Research, 253, 697–710.

Oztekin et al. (2011)
Oztekin, A., Kong, Z. J., &
Delen, D. (2011).
Development of a structural equation modelingbased decision tree methodology for the analysis of lung transplantations.
Decision Support Systems, 51, 155–166.  Rigdon et al. (2010) Rigdon, E. E., Ringle, C. M., & Sarstedt, M. (2010). Structural modeling of heterogeneous data with partial least squares. Review of Marketing Research, 7, 255–296.

Sermpinis et al. (2013)
Sermpinis, G., Theofilatos, K.,
Karathanasopoulos, A., Georgopoulos,
E. F., & Dunis, C. (2013).
Forecasting foreign exchange rates with adaptive neural networks using radialbasis functions and particle swarm optimization.
European Journal of Operational Research, 225, 528–540.  Sevim et al. (2014) Sevim, C., Oztekin, A., Bali, O., Gumus, S., & Guresen, E. (2014). Developing an early warning system to predict currency crises. European Journal of Operational Research, 237, 1095–1104.
 Shaik (2015) Shaik, S. (2015). Impact of liquidity risk on variations in efficiency and productivity: A panel gamma simulated maximum likelihood estimation. European Journal of Operational Research, 245, 463–469.
 Stock & Watson (2002) Stock, J. H., & Watson, M. W. (2002). Macroeconomic forecasting using diffusion indexes. Journal of Business & Economic Statistics, 20, 147–162.

Sun & Shenoy (2007)
Sun, L., & Shenoy, P. P.
(2007).
Using bayesian networks for bankruptcy prediction: Some methodological issues.
European Journal of Operational Research, 180, 738–753.  Tam & Kiang (1992) Tam, K. Y., & Kiang, M. Y. (1992). Managerial applications of neural networks: The case of bank failure predictions. Management Science, 38, 926–947.
 Tay & Cao (2001) Tay, F. E., & Cao, L. (2001). Application of support vector machines in financial time series forecasting. Omega, 29, 309–317.
 Tenenhaus et al. (2005) Tenenhaus, M., Vinzi, V. E., Chatelin, Y.M., & Lauro, C. (2005). PLS path modeling. Computational Statistics & Data Analysis, 48, 159–205.
 Tsai & Wang (2017) Tsai, M.F., & Wang, C.J. (2017). On the risk prediction and analysis of soft information in finance reports. European Journal of Operational Research, 257, 243–250.
 Turkyilmaz et al. (2013) Turkyilmaz, A., Oztekin, A., Zaim, S., & Fahrettin Demirel, O. (2013). Universal structure modeling approach to customer satisfaction index. Industrial Management & Data Systems, 113, 932–949.
 Turkyilmaz et al. (2016) Turkyilmaz, A., Temizer, L., & Oztekin, A. (2016). A causal analytic approach to student satisfaction index modeling. Annals of Operations Research, 7, 209.
 Xu et al. (2017) Xu, Y., Pinedo, M., & Xue, M. (2017). Operational risk in financial services: A review and new research opportunities. Production and Operations Management, 26, 426–445.
Comments
There are no comments yet.