An Alternative Data-Driven Prediction Approach Based on Real Option Theories

04/19/2019
by   Abdullah Alshelahi, et al.
0

This paper presents a new prediction model for time series data by integrating a time-varying Geometric Brownian Motion model with a pricing mechanism used in financial engineering. Typical time series models such as Auto-Regressive Integrated Moving Average assumes a linear correlation structure in time series data. When a stochastic process is highly volatile, such an assumption can be easily violated, leading to inaccurate predictions. We develop a new prediction model that can flexibly characterize a time-varying volatile process without assuming linearity. We formulate the prediction problem as an optimization problem with unequal overestimation and underestimation costs. Based on real option theories developed in finance, we solve the optimization problem and obtain a predicted value, which can minimize the expected prediction cost. We evaluate the proposed approach using multiple datasets obtained from real-life applications including manufacturing, finance, and environment. The numerical results demonstrate that the proposed model shows competitive prediction capability, compared with alternative approaches.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

07/01/2013

Gaussian Process Conditional Copulas with Applications to Financial Time Series

The estimation of dependencies between multiple variables is a central p...
09/13/2020

Time-varying auto-regressive models for count time-series

Count-valued time series data are routinely collected in many applicatio...
10/25/2021

Applying Regression Conformal Prediction with Nearest Neighbors to time series data

In this paper, we apply conformal prediction to time series data. Confor...
09/13/2018

Superstatistics with cut-off tails for financial time series

Financial time series have been investigated to follow fat-tailed distri...
03/16/2021

Rollage: Efficient Rolling Average Algorithm to Estimate ARMA Models for Big Time Series Data

We develop a new method to estimate an ARMA model in the presence of big...
02/23/2021

Optimal Prediction Intervals for Macroeconomic Time Series Using Chaos and NSGA II

In a first-of-its-kind study, this paper proposes the formulation of con...
05/11/2021

Modelling and predicting soil carbon sequestration: is current model structure fit for purpose?

Soil carbon accounting and prediction play a key role in building decisi...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

In many applications including manufacturing, energy, and finance, accurate prediction is required to support strategic, tactical and/or operational decisions of organization [chatfield2000time]. When physical information about the underlying mechanism that generates the time series data is limited, data-driven methods can be useful for predicting future observations [zhang2003time]. In general, data-driven forecasting methods predict future observations based on past observations [box2015time]. Several data-driven methods have been proposed in the literature for modeling time series data, among which Auto-Regressive Integrated Moving Average (ARIMA) and its variants such as the ARIMA-General Auto Regressive Conditional Heteroskedasticity (ARIMA-GARCH) have been widely used in many applications due to their flexibility and statistical properties [ruppertd, hahn2009electric, sohn2007hierarchical, lu2014portfolio]

. ARIMA which assumes a constant standard deviation of stochastic noises, whereas ARIMA-GARCH extends it by allowing the standard deviation to vary over time.

The typical ARIMA-based models estimate its model parameters using historical data and uses the estimated time-invariant parameters throughout the prediction period. Using such time-invariant parameters may not capture possible changes in the underlying data generation mechanism. Some studies modify the original ARIMA model to update the parameters using new observations

[tran2004automatic, ledolter1981recursive]. The basic idea of these ARIMA-based models is that the future observation can be predicted by using a linear combination of past observations (and estimated noises). Therefore they assume a linear correlation structure between consecutive observations [brooks2014introductory]. However, when the underlying dynamics exhibits a highly volatile process, such a simple linear structure may provide poor prediction performance [kantz2004nonlinear].

This study aims to provide accurate predictions for a highly volatile and time-varying stochastic process whose underlying dynamics is complicated and possibly nonlinear. As an example, let us consider a prediction problem faced by a contract manufacturer (CM) located in Michigan in the U.S, which motivates this study. The CM is a manufacturing company that produces various automotive parts, such as front and rear bumper beams, for several large automotive companies worldwide. The CM deals with a large number of orders for bumper beams from several automotive companies and the order sizes are time-varying. The CM should plan their production capacity carefully so that it can deliver products promptly when it gets orders. When an actual order size is greater than expected (i.e., when an order size is underestimated), overtime wages must be paid to workers to meet demands. On the other hand, when an order size is smaller than predicted (i.e., when an order size is overestimated), workers and equipment become idle.

As such, CM wants to predict future order sizes accurately, so that it can reduce its operating costs resulting from the discrepancy between its predicted value and actual sizes. Currently, CM uses its own proprietary prediction model, but its prediction performance is not satisfactory. The details of CM’s proprietary model are confidential, so we cannot find reasons for its unsatisfactory performance. When we apply the ARIMA and ARIMA-GARCH models to CM’s datasets, we also do not obtain significantly better prediction results (detailed results will be provided in Section  III). We believe such poor performance of ARIMA-based approaches is because they cannot fully characterize the underlying volatile dynamics. In addition to historical data, the future order size may depend on other factors which possibly make the order process behave nonlinearly. A new prediction approach that can adapt to such time-varying, and possibly nonlinear, dynamics is needed for providing better forecasts.

To this end we develop a new method for predicting future values in highly volatile processes, based on real option pricing theories typically used in financial engineering. One of the popularly used stochastic process models for pricing real options is the Geometric Brownian Motion (GBM) model. Brownian motion is a continuous-time stochastic process, describing random movements in time series variables. The GBM, which is a stochastic differential equation, incorporates the idea of Brownian motion and consists of two terms: a deterministic term to characterize the main trend over time and a stochastic term to account for random variations. In GBM the random variations are represented by Brownian Motion [bjork2009arbitrage]. GBM is useful to model a positive quantity whose changes over equal and non-overlapping time intervals are identically distributed and independent.

The GBM has been applied to represent various real processes in finance, physics, etc. [gardiner1986handbook]. In particular, it becomes a fundamental block for many asset pricing models [bjork2009arbitrage], and recently it has been applied to facilitate the use of a rich area of options theory to solve various pricing problems (see, for example, [whitt1981stationary, thorsen1999afforestation, benninga2002real, nembhard2002real, boomsma2012renewable, chiu2017real]). However, most of the current GBM studies have been limited to solving pricing problems and have not used real options theory for making forecasts.

In this study, by utilizing the full power of real options theory, we present a new approach for predicting future observations when the system’s underlying dynamics follows the GBM process. Specifically, we allow the GBM parameters to adaptively change over time in order to characterize time-varying dynamics. We formulate the prediction problem as an optimization problem and provide a solution using real option theories. To the best of our knowledge, our study is the first attempt to incorporate options theory in the prediction problem.

Our approach provides extra flexibility by allowing overestimation (or over-prediction) to be handled differently from underestimation (or over-prediction). The overestimation and underestimation costs are determined in real life applications, depending on a decision-maker’s (or organization’s) preference. For example, in the aforementioned CM case, overestimation and underestimation of order sizes could cause different costs. The CM may want to put a larger penalty on the demand underestimation than on the overestimation, so that it can avoid extra overtime wages. We incorporate unequal overestimation and underestimation costs into the optimization problem and find the optimal forecast that minimizes the expected prediction cost.

To evaluate the prediction performance, we use three datasets collected from different applications, including the demand for bumper beams in CM (manufacturing), stock prices (finance), and wind speed (environment). We compare the performance of our model with ARIMA and ARIMA-GARCH models (and the proprietary prediction model in the CM case study) with different combinations of overestimation and underestimation costs. In most cases, our model outperforms those alternative models. In particular, we find that when the process is highly time-varying such as stock prices and wind speed, the proposed approach provides much stronger prediction capability than ARMA and ARIMA-GARCH.

The remainder of the paper is organized as follows. The mathematical formulation and solution procedure are discussed in Section II. Section III provides numerical results in three different applications. Section IV concludes the paper.

Ii Methodology

Ii-a Problem Formulation

Consider a real-valued variable which represents a system state at time . For example, the state variable can be a stock market index price, a manufacturer’s order size, or wind speed. This state variable is assumed to follow an inhomogeneous GBM with time-varying parameters.

Let us consider a filtered probability space

, where the filtration is generated by the Brownian motion , i.e. so that contains all information generated by , up to and including time . With GBM, the stochastic process is modeled by the following dynamics.

(1)

where denotes the volatility of and represents a drift process. The stochastic process represents the Brownian motion where the increment during the time interval

is normally distributed with mean 0 and variance

, denoted by , and is assumed to be stationary.

Our objective is to predict in the future time at when the current time is . Solving (1) by using Itô’s lemma [shreve2004stochastic], we obtain

(2)

and

(3)

Let be the predicted value of at time . When the overestimation and underestimation is penalized equally, the quantity that represent the variable’s central tendency, such as mean and median, is commonly used for prediction. But we consider a more general case where overestimation needs to penalized differently from underestimation, as discussed in Section I. When the observed value is , the overestimated quantity becomes , while the underestimated quantity is .

Let and denote the penalties for over/underestimation, respectively. We formulate the optimization problem for estimating that can minimize the expected prediction cost,

(4)

Note that

(5)

If we substitute (5) into (4), the optimal predicted value, denoted by , can be obtained by solving the following objective function.

(6)

or equivalently,

(7)

In the next section we will present a solution procedure to obtain , based on the option theory.

Ii-B Real Option Based Solution Procedure

The optimization problem in (7) can be reformulated by employing the financial pricing theories. Suppose that we want to predict a state at the future time . In pricing theories, can be viewed as the date to maturity, or the expiration date.

A real option, also called contingent claim, with the date to maturity , can be constructed on the state variable . A real option is a stochastic variable that can be expressed as

(8)

where is a contract function.

The contract function is typically set to the payoff of the real option at time . When the predicted value is , can be viewed as the strike value in the option theory, while is the payoff. Therefore, we get

(9)

It is required that ensures that the value of the payoff of the real option is determined at time .

Let the price process for the real option at time be given by a function , i.e.,

(10)

Here is a function which is assumed to be once continuously differentiable in , and twice in .

For a short-term prediction, the time interval between the current time and the future time is small, so we can assume that and are constants during . Then

can be obtained by solving the Black-Scholes Partial Differential Equation (PDE)

[shreve2004stochastic],

(11)

with

(12)

where represents a discounting factor.

The Black-Scholes PDE in (11)-(12) is usually solved numerically. But alternatively, we solve it using the Feyman-Ka stochastic representation formula [shreve2004stochastic], to obtain

(13)

Next, we derive in a closed form, given . Letting and using the fact that , it follows that

. Thus, the probability density function

of is given by

(14)

Consequently, we obtain

(15)
(16)
(17)

To solve (17), let and , respectively, denote the first and second terms in (17). We also let . First, becomes

(18)
(19)
(20)

where and

denotes the cumulative distribution function (CFD) for the standard normal distribution. Next, we obtain

as

(21)
(22)
(23)
(24)
(25)

where we use in (24) and in (25).

For small , we can set . Then, in (13) becomes:

(26)

Note that given and at the current time and , we can obtain .

With the obtained expected payoff where , we can find the optimal in (7). Let denote the ratio of overestimation cost to underestimation cost, i.e.,

(27)

Given the price of the real option, defined in (13), we can reformulate the optimization problem in (7) as

(28)
(29)
(30)

with and . We use (13) with in the first term in the second equality and the last term in the second equality is obtained using (3). By plugging in (26), we get the last equality.

The predictor prefers overestimation when or underestimation when . When overestimation and underestimation are equally penalized, the optimal can be obtained with in (28). The optimization function in (28) is a convex optimization problem that can be solved efficiently by existing numerical optimization softwares. In our implementation, we use Scipy’s (Scientific Python) optimization library in Python.

Ii-C Parameters Estimation

For a volatile stochastic process, the parameters and can be time-varying. We estimate the nonstationary parameters using recent observations. Consider recent observations at the current time , i.e., . Because follows geometric Brownian motion and and are assumed to be constant during the short interval , the discretization scheme of (2) is given by

(31)

Noting that under GBM is normally distributed with mean and variance , we estimate and using maximum likelihood method as

(32)
(33)

respectively.

The estimated parameters and are plugged into (26) and we obtain the optimal predicted value for by solving (30).

Ii-D Implementation Details

We refer our proposed model to as the option prediction model. Figure 1 summarizes the overall procedure of the proposed approach. We also summarize the procedure of the proposed approach in Algorithm 1 below. We set the time step to make the one-ahead step prediction. The data is divided into three sets: training, validation, and testing. The training set starts at and ends at , consisting of about 50% of the entire data set, is used to determine the model parameters as shown in Figure 1. The validation set, consisting of about 20% of the data set, is used for determining the window size . Lastly the testing set consists of the last of the data set and it starts at .

1:Initialization:
2:Choose a window size by validation as shown in Figure 1.
3:Obtain initial estimates for the model parameters and in (32) and (33), respectively.
4:Determine in (13).
5:for  to  do
6:     Prediction:
7:     Obtain by solving (30) to obtain the one-step ahead state prediction.
8:     Update:
9:     Observe .
10:     Obtain and in (32) and (33), respectively, by using recent observations.
11:     Determine in (13).
12:end for
Algorithm 1 Option prediction model

In Algorithm 1 we determine the window size for obtaining the parameters and , we use the validation technique [friedman2001elements]. We fit the model with a different window size and evaluate the prediction performance using data in the validation set and choose the best window size that generates the lowest prediction error in the validation set. The performance of our approach is evaluated using data in the testing set (See Figure 1). We report the prediction performance in the testing set in Section III.

Set a window size ,

Update the model parametersusing sample pointsfrom equations (32)-(33)

Predict one-step aheadin the validation setby solving (29)

Compute the predictionerror WMAE from (34)for the validation set

Choose a windowsize that generatesthe lowest WMAE

Update the modelparameters usingthe selected

Training set (50%)

Validation set (20%)

Testing set (30%)
Fig. 1: Overall procedure of the proposed approach (the dotted lines imply that the model parameters are updated in a rolling-horizon manner using the most recent observations)

In evaluating the prediction performance, we consider that the overestimated and underestimated prediction results need to be evaluated differently for . As such we employ the following two performance measures, namely, Weighted Mean Absolute Error (WMAE) and Weighted Mean Absolute Percentage Error (WMAPE), defined by

(34)

and

(35)

respectively, where denotes the number of data points in the testing set and is the predicted value at time .

Iii Case Studies

This section implements the proposed prediction model using multiple datasets obtained from real-life applications. Specifically we examine the performance of the predictive model in predicting the size of a manufacturer’s order, a stock market index price, and wind speed.

Iii-a Alternative methods

We compare our model with two standard time series models, namely, the ARIMA and the ARIMA-GARCH. We use the Akaika Information Criteria (AIC) to select the model order in both models. For fair comparison, we update the model parameters in a rolling horizon manner, similar to the procedure discussed in Section II-D. That is, we determine the window size using the validation technique and update the model parameters using the most recent observations whenever a new observation is obtained.

With underestimation penalties, Pourhab et al. [pourhabib2015short]

suggest using quantile of the predictive state density. With

denoting the ratio of underestimation cost to overestimation cost, we use the -quantile, given by

(36)

where denotes the estimated predicted mean, is the estimated standard deviation in ARIMA (or ARIMA-GARCH) model, and denotes the inverse of the standard normal CDF. Note that large (small) puts more penalty on () and the quantile prediction provides a larger (smaller) predicted value, so underestimation (overestimation) can be avoided.

Iii-B Manufacturing Data

We first study the prediction problem faced by our industry partner, CM. The historical data obtained from CM includes orders of 10 different types of bumper beams. We use monthly data on those 10 types of bumper beams ordered over a period of 29 consecutive months (the order size varies from 0 to over 36,000 items). When applying the proposed model to this problem, the choice of weight affects the final prediction. By changing the weight, we are able to show a preference for over-capacity (overestimation) or under-capacity (underestimation). We consider different cases for choosing the weight parameter .

Let us first look at the case when is set to be less than one (i.e, ). According to CM, workers and equipment can be shifted from one type of bumper beam to another, but doing so incurs 15 loss of production efficiency. In other words, if one type of bumper beam is overestimated, causing over-capacity, available resources can be assigned to other bumper beam production, but with a reduced efficiency. In this case, underestimation is favored and we set .

Next the weight parameter can be set to be greater than 1 (i.e, ) when the prediction is preferred to be more than the actual order size. According to the labor law in Michigan in the U.S., overtime rate is higher than the regular salary. In this case we set to emphasize the preference of overestimation to underestimation. Finally we also consider , which reflects equal penalties.

The errors in terms of WMAE and WMAPE for all ten types of bumper beams are presented in Tables I-III with three different weights. Overall our option prediction model performs better than the CM’s own prediction, ARIMA and ARIMA-GARCH in both criteria. With the proposed approach provides lower WMAEs (WMAPEs) for 9 (5) types of bumper beams out of 10 types. Similarly, with other values, our approach outperforms the alternative models in most cases.

Weighted Mean Absolute Error (WMAE)
Product No. ARIMA Option Prediction ARIMA-GARCH CM Prediction
1 1196.58 411.97 2105.25 1161.56
2 332.36 127.90 168.72 151.24
3 119.35 105.49 107.43 194.69
4 1476.09 574.52 2185.30 936.94
5 1330.40 1299.00 1327.08 1797.74
6 542.33 64.38 63.24 42.90
7 357.38 24.14 497.54 92.17
8 1776.17 1339.11 3103.77 2520.75
9 1496.62 1305.49 2887.48 2475.71
10 1125.52 516.06 3278.77 1928.58
Weighted Mean Absolute Percent Error (WMAPE)
Product No. ARIMA Option Prediction ARIMA-GARCH CM Prediction
1 1.14 0.19 0.94 0.52
2 0.62 0.43 0.42 0.40
3 28.89 0.44 0.41 0.69
4 1.20 0.69 3.92 0.66
5 0.12 0.11 0.12 0.15
6 45.23 5.30 11.01 7.93
7 215.17 3.66 408.98 9.63
8 0.24 0.26 0.76 0.60
9 0.19 0.22 0.66 0.58
10 0.36 0.20 2.63 0.99
TABLE I: CM Prediction Results for ten types of bumper beams with in the Testing Set (The values in bold indicate the lowest prediction error for each product)
Weighted Mean Absolute Error (WMAE)
Product No. ARIMA Option Prediction ARIMA-GARCH CM Prediction
1 1193.29 537.94 2103.43 1335.79
2 328.39 134.67 164.49 168.93
3 111.78 118.39 103.40 223.89
4 1307.50 609.06 1920.83 1073.16
5 1178.90 1403.17 1200.81 2060.95
6 472.86 74.52 57.08 45.12
7 315.42 34.61 434.33 93.51
8 1691.28 1465.94 2737.90 2590.93
9 1453.32 1397.14 2564.50 2527.94
10 1008.59 569.58 2865.30 1932.47
Weighted Mean Absolute Percent Error (WMAPE)
Product No. ARIMA Option Prediction ARIMA-GARCH CM Prediction
1 1.09 0.24 0.94 0.59
2 0.61 0.45 0.40 0.44
3 25.76 0.50 0.39 0.79
4 1.05 0.72 3.41 0.75
5 0.11 0.12 0.11 0.17
6 39.61 6.17 9.59 7.95
7 187.36 3.87 356.23 9.67
8 0.22 0.28 0.66 0.60
9 0.18 0.23 0.58 0.58
10 0.31 0.21 2.29 0.99
TABLE II: CM Prediction Results for ten types of bumper beams with in the Testing Set (The values in bold indicate the lowest prediction error for each product)
Weighted Mean Absolute Error (WMAE)
Product No. ARIMA Option Prediction ARIMA-GARCH CM Prediction
1 1196.58 411.97 2105.25 1161.56
2 332.36 127.90 168.72 151.24
3 119.35 105.49 107.43 194.69
4 1476.09 574.52 2185.30 936.94
5 1330.40 1299.00 1327.08 1797.74
6 542.33 64.38 63.24 42.90
7 357.38 24.14 497.54 92.17
8 1776.17 1339.11 3103.77 2520.75
9 1496.62 1305.49 2887.48 2475.71
10 1125.52 516.06 3278.77 1928.58
Weighted Mean Absolute Percent Error (WMAPE)
Product No. ARIMA Option Prediction ARIMA-GARCH CM Prediction
1 1.14 0.19 0.94 0.52
2 0.62 0.43 0.42 0.40
3 28.89 0.44 0.41 0.69
4 1.20 0.69 3.92 0.66
5 0.12 0.11 0.12 0.15
6 45.23 5.30 11.01 7.93
7 215.17 3.66 408.98 9.63
8 0.24 0.26 0.76 0.60
9 0.19 0.22 0.66 0.58
10 0.36 0.20 2.63 0.99
TABLE III: CM Prediction Results for ten types of bumper beams with in the Testing Set (The values in bold indicate the lowest prediction error for each product)

Although ARIMA and ARIMA-GARCH provide the lowest errors for some products, their prediction performance is not consistent. For example, for , and product, WMAEs from ARMA are much higher than the proposed approach, whereas ARIMA-GARCH results in pretty poor performance for predicting order sizes for products. On the contrary, our approach provides more stable results. Even when WMAEs and WMAPEs from our approach are higher than other approaches, they are close to the lowest errors. Therefore, we can conclude that our approach is more accurate and reliable. The CM’s proprietary model does not account for unequal weights on overestimation and underestimation. If the company wants to minimize the excess inventory due to overestimation, a small (less than 1) weight parameter should be assigned. If the company goal is to meet customer satisfaction, overestimation should be preferred with a large (larger than 1) weight parameter. In this sense our approach can reflect the company’s management preference more flexibly.

Iii-C Stock Market Index Data

To evaluate the performance of our approach in a highly volatile process, we consider stock market index price time series data. We analyze the daily closing price of the Dow Jones index in three time periods between 2010 and 2015.

Risk averse and risk seeking investors have different preferences in terms of overestimation and underestimation. That being said, in a bull market, stock prices are expected to increase. In such a case, risk seeking investors with aggressive investment strategies would prefer biasing their prediction to overestimation. On the contrary, risk averse investors tend to be less optimistic, making them conservative, preferring underestimation. To reflect different investment preferences, we consider three values of the weight parameter , , , or , to represent the underestimation preference, neutral/no preference, and overestimation preference, respectively.

Table IV summarizes the results with three testing periods. Each testing period includes 100 days. Clearly, our option prediction performs better than ARIMA-GARCH and ARIMA in all cases, alerting for the possibility of a profitable trading strategy. The ARMA and ARMA-GARCH models generate 2.5 to 10 times higher WMAEs and 2 to 11 times higher WMAPEs.

Testing Period Weight () Method WMAE WMAPE
ARIMA-GARCH 595.54 0.04998
1/1.15 Option Prediction 50.40 0.0043
ARIMA 628.69 0.0615
ARIMA-GARCH 594.58 0.0499
Oct 2010- Mar 2011 1 Option Prediction 54.30 0.0046
ARIMA 553.05 0.0541
ARIMA-GARCH 682.72 0.0572
1.15 Option Prediction 58.78 0.0050
ARIMA 559.59 0.0548
ARIMA-GARCH 360.39 0.0237
1/1.15 Option Prediction 70.94 0.0046
ARIMA 455.35 0.0309
ARIMA-GARCH 317.34 0.0209
Aug 2013 - Dec 2013 1 Option Prediction 75.78 0.0049
ARIMA 406.97 0.0276
ARIMA-GARCH 321.71 0.0211
1.15 Option Prediction 81.09 0.0052
ARIMA 418.90 0.0284
ARIMA-GARCH 338.71 0.0195
1/1.15 Option Prediction 97.78 0.0056
ARIMA 182.60 0.0109
ARIMA-GARCH 319.72 0.0184
Oct 2014 - Mar 2015 1 Option Prediction 104.70 0.0060
ARIMA 166.65 0.0099
ARIMA-GARCH 348.25 0.0200
1.15 Option Prediction 113.11 0.0065
ARIMA 175.25 0.0104
TABLE IV: Dow Jones Index Price Prediction Results in the Testing Set (The values in bold indicate the lowest prediction error for each testing period and weight)

Iii-D Wind Speed Data

Finally, we additionally consider another highly volatile process, wind speed. Because of environmental considerations, wind power, as a renewable source of energy, has been increasingly adopted worldwide [byon2015adaptive]. Intermittent output of the farm is considered a challenging issue in terms of integrating the wind power into electric power grids. For reliable supply of power, steady and uninterrupted energy generation is desirable, which is not the case with wind energy. Wind speed is highly variable, depending on weather conditions and geographical factors such as the terrain. Such variability imposes challenges in power grid operations. To overcome the challenges, accurate forecasting of wind speed is required [soman2010review].

We use wind speed data collected from a meteorological tower near a wind farm located in Europe. The whole dataset consists of about 3000 samples, which covers a period of about a month. Due to the data confidentiality required by our industry partner, we omit more detailed description of the dataset studied in this case study. In wind farm operations some operators want to put a higher penalty on overestimation to avoid unsatisfied demand (or unsatisfied commitment), whereas underestimating wind speeds may be preferred when the salvage cost of excessively generated power is high [pourhabib2015short, HG2010_PCE]. To reflect different costs, we use three different values for , , and .

Table V summarizes the prediction results in the testing set from the three models. The proposed option prediction significantly outperform the other methods. The WMAEs and WMAPEs from ARMA and ARIMA-GARCH are higher by one order of magnitude than our approach. It demonstrates the superior prediction performance of our approach in a highly volatile process.

Weight () Method WMAE WMAPE
ARIMA-GARCH 3.364 0.402
1/1.15 Option Prediction 0.318 0.040
ARIMA 8.73 0.957
ARIMA-GARCH 3.31 0.380
1 Option Prediction 0.342 0.043
ARIMA 8.73 0.96
ARIMA-GARCH 3.74 0.422
1.15 Option Prediction 0.365 0.046
ARIMA 10.03 1.100
TABLE V: Wind Speed Prediction Results in the Testing Set (The values in bold indicate the lowest prediction error for each weight)

Iv Conclusion

In this study, we present a new prediction methodology for the time series data, based on option theories in finance when the underlying dynamics is assumed to follow the GBM process. To characterize time-varying patterns, we allow the GBM model parameters to vary over time and update the parameter values using recent observations. We formulate the prediction problem with unequal overestimation and underestimation penalties as the stochastic optimization problem and provide its solution procedure. We demonstrate the prediction capability of the proposed approach in various applications. Our approach appears to work well in the manufacturing application, when the order size varies over time. For more highly volatile processes such as stock prices and wind speeds, the proposed model exhibits much stronger prediction capability, compared to alternative ARIMA-based models.

In the future, we plan to investigate other parameter updating schemes. In this study, we update parameters in a rolling horizon manner using the maximum likelihood estimations. Another possibility is to use the Kalman filtering or its variants. Long-term predictions are beyond the scope of this study, but we plan to extend the approach presented in this study for obtaining accurate long-term predictions. We will also incorporate prediction results into managerial decision-making in several applications such as power grid operation with renewable energy

[Bouffard2008].

References