1. Introduction
1.1. Problem formulation
Consider a onedimensional stochastic differential equation (SDE)
(1) 
where is the drift coefficient, the deterministic dispersion coefficient or volatility, and is a deterministic initial condition. Here is a standard Brownian motion. Assume that standard conditions for existence and uniqueness of a strong solution to (1) are satisfied (see, e.g., [karatzas91]), and observations
are available, where Using a nonparametric Bayesian approach, our aim is to estimate the volatility function . In a financial context, knowledge of the volatility is of fundamental importance e.g. in pricing financial derivatives; see [bjork09] and [musiela05]. However, SDEs have applications far beyond the financial context as well, e.g. in physics, biology, life sciences, neuroscience and engineering (see [allen07], [fuchs13], [hindriks11] and [wong85]). Note that by Itô’s formula, using a simple transformation of the state variable, also an SDE of the form
can be reduced to the form (1), provided the function is known and regular enough; see, e.g., p. 186 in [soulier98]. Some classical examples that fall under our statistical framework are the geometric Brownian motion and the OrnsteinUhlenbeck process. Note also that as we allow the drift in (1) to be nonlinear, marginal distributions of are not necessarily Gaussian and may thus exhibit heavy tails, which is attractive in financial modelling.
A nonparametric approach guards one against model misspecification and is an excellent tool for a preliminary, exploratory data analysis, see, e.g., [silverman86]. Commonly acknowledged advantages of a Bayesian approach include automatic uncertainty quantification in parameter estimates via Bayesian credible sets, and the fact that it is a fundamentally likelihoodbased method. In [mueller13] it has been argued that a nonparametric Bayesian approach is important for honest representation of uncertainties in inferential conclusions. Furthermore, use of a prior allows one to easily incorporate the available external, a priori information into the estimation procedure, which is not straightforward to achieve with frequentist approaches. For instance, this a priori information could be an increasing or decreasing trend in the volatility.
1.2. Literature overview
Literature on nonparametric Bayesian volatility estimation in SDE models is scarce. We can list theoretical contributions [gugu14], [gugu16], [nickl17b], and the practically oriented paper [batz17]. The model in the former two papers is close to the one considered in the present work, but from the methodological point of view different Bayesian priors are used and practical usefulness of the corresponding Bayesian approaches is limited. On the other hand, the models considered in [nickl17b] and [batz17] are rather different from ours, and so are the corresponding Bayesian approaches. The nearest predecessor of the model and the method in our paper is the one studied in [gugu17]. In the sequel we will explain in what aspects the present contribution differs from that one and what the current improvements are. We note in passing that there exists a solid body of literature on nonparametric Bayesian estimation of the drift coefficient, see, e.g., [gugu14b], [papaspiliopoulos12], [pokern13], [ruttor13], [frank13], [frank14] and the review article [harry13], but Bayesian volatility estimation requires use of substantially different ideas. We also note existence of works dealing with parametric Bayesian estimation in discretetime stochastic volatility models, see, e.g., [jacquier94] and [jacquier04], but again, these are not directly related to the problem we study in this paper.
1.3. Approach and results
The main potential difficulties facing a Bayesian approach to inference in SDE models from discrete observations are an intractable likelihood and absence of a closed form expression for the posterior distribution; see, e.g., [elerian01], [fuchs13], [roberts01] and [vdmsestpaper]. Typically, these difficulties necessitate the use of a data augmentation device (see [tanner87]) and some intricate form of a Markov chain Monte Carlo (MCMC) sampler (see [robert04]). In [gugu17], these difficulties are circumvented by intentionally setting the drift coefficient to zero, and employing a (conjugate) histogramtype prior on the diffusion coefficient, that has piecewise constant realisations on bins forming a partition of . Specifically, the (squared) volatility is modelled a priori as a function with independent and identically distributed inverse gamma coefficients ’s, and the prior is defined as the law of . Here are bins forming a partition of . With this independent inverse Gamma (IIG) prior, are independent, conditional on the data, and of inverse gamma type. Therefore, this approach results in a fast and simple to understand and implement Bayesian procedure. A study of its favourable practical performance, as well as its theoretical validation was recently undertaken in [gugu17]. As shown there under precise regularity conditions, misspecification of the drift is asymptotically, as the sample size , harmless for consistent estimation of the volatility coefficient.
Despite a good practical performance of the method in [gugu17], there are some limitations associated with it too. Thus, the method offers limited possibilities for adaptation to the local structure of the volatility coefficient, which may become an issue if the volatility has a wildly varying curvature on the time interval . A possible fix to this would be to equip the number of bins forming a partition of with a prior, and choose the endpoints of bins also according to a prior. However, this would force one to go beyond the conjugate Bayesian setting as in [gugu17], and posterior inference in practice would require, for instance, the use of a reversible jump MCMC algorithm (see [green95]). Even in the incomparably simpler setting of intensity function estimation for nonhomogeneous Poisson processes with histogramtype priors, this is very challenging, as observed in [yang01]. Principal difficulties include designing moves between models of differing dimensions that result in MCMC algorithms that mix well, and assessment of convergence of Markov chains (see [fearnhead06], p. 204). Thus, e.g., the inferential conclusions in [green95] and [green03] are different on the same real data example using the same reversible jump method, since it turned out that in the first paper the chain was not run long enough. Cf. also the remarks on Bayesian histograms in [gelman:book:14], p. 546.
Here we propose an alternative approach, inspired by ideas in [cemgil07] in the context of audio signal modelling different from the SDE setting that we consider; see also [cemgil:proc:08], [cemgil07b], [dikmen08], [dikmen10] and [virtanen08]. Namely, instead of using a prior on the (squared) volatility that has piecewise constant realisations on with independent coefficients ’s, we will assume that the sequence
forms a suitably defined Markov chain. An immediately apparent advantage of using such an approach is that it induces extra smoothing via dependence in prior realisations of the volatility function across different bins. Arguing heuristically, with a large number
of bins it is then possible to closely mimick the local structure of the volatility: in those parts of the interval , where the volatility has a high curvature or is subject to abrupt changes, a large number of (narrow) bins is required to adequately capture these features. However, the grid used to define the bins ’s is uniform, and if are a priori independent, a large may induce spurious variability in the volatility estimates in those regions of where the volatility in fact varies slowly. As we will see in the sequel, this problem may be alleviated using a priori dependent ’s.In the subsequent sections we detail our approach, and study its practical performance via simulation and real data examples. Specifically, we implement our method via a straightforward version of the Gibbs sampler, employing the fact that full conditional distributions of ’s are known in closed form (and are in fact inverse gamma). Unlike [gugu17], posterior inference in our new approach requires the use of MCMC. However, this is offset by the advantages of our new approach outlined above, and in fact the additional computational complexity of our new method is modest in comparison to [gugu17]. The prior in our new method depends on hyperparameters, and we will also discuss several ways of their choice in practice.
1.4. Organisation of this paper
In Section 2 we supply a detailed description of our nonparametric Bayesian approach to volatility estimation. In Section 3 we study the performance of our method via extensive simulation examples. In Section 4 we apply the method on a real data example. Section 5 summarises our findings and provides an outlook on our results. Finally, Section 6 contains some additional technical details of our procedure.
1.5. Notation
We denote the prior distribution on the (squared) volatility function by and write the posterior measure given data as . We use the notation
for the inverse gamma distribution with shape parameter
and scale parameter . This distribution has a density(2) 
For two sequences , , the notation will be used to denote the fact that the sequences are asymptotically (as ) of the same order. Finally, for a density and a function , the notation will mean that is proportional to , with proportionality constant on the righthand side recovered as , where the integral is over the domain of definition of (and of ). The function
can be referred to as an unnormalised probability density.
2. Nonparametric Bayesian approach
2.1. Generalities
Our starting point is the same as in [gugu17]. Namely, we misspecify the drift coefficient by intentionally setting it to zero (see also [martin16] for a similar idea of ‘misspecification on purpose’). The theoretical justification for this under the ‘infill’ asymptotics, with the time horizon staying fixed and the observation times filling up the interval as , is provided in [gugu17], to which we refer for further details (the argument there ultimately relies on Girsanov’s theorem). Similar ideas are also encountered in the nonBayesian setting in the econometrics literature on highfrequency financial data, see, e.g., [mykland12].
Set . With the assumption , the pseudolikelihood of our observations is tractable, in fact Gaussian,
(3) 
where
The posterior probability of any measurable set
of volatility functions can be computed via Bayes’ theorem as
Here the denominator is the normalising constant, the integral over the whole space on which the prior is defined, which ensures that the posterior is a probability measure (i.e. integrates to one).
2.2. Prior construction
Our prior is constructed similarly to [gugu17], with an important difference to be noted below. Fix an integer . Then with , where . Now define bins , , and . Thus the first bins are of length , whereas the last bin has length . The parameter (equivalently, ) is a hyperparameter of our prior. We model as piecewise constant on bins , thus The prior on the volatility can now be defined by assigning a prior to the coefficients ’s.
Let . Since the bins are disjoint,
As the likelihood depends on only through its square , it suffices to assign the prior to the coefficients ’s of . This is the point where we fundamentally diverge from [gugu17]. Whereas in [gugu17] it is assumed that
is an i.i.d. sequence of inverse gamma random variables, here we suppose that
forms a Markov chain. This will be referred to as an inverse Gamma Markov chain (IGMC) prior (see [cemgil07]), and is defined as follows. Introduce auxiliary variables , and define a Markov chain using the time ordering . Transition distributions of this chain are defined as follows: fix hyperparameters , and , and set(4) 
The name of the chain reflects the fact that these distributions are inverse Gamma.
Remark 1.
Our definition of the IGMC prior differs from the one in [cemgil07] in the choice of the initial distribution of , which is important to alleviate possible ‘edge effects’ in volatility estimates in a neighbourhood of . The parameter determines the initial distribution of the inverse Gamma Markov chain. Letting (which corresponds to a vague prior) ‘releases’ the chain at the time origin. ∎
Remark 2.
As observed in [cemgil07], there are various ways of defining an inverse Gamma Markov chain. The point to be kept in mind is that the resulting posterior should be computationally tractable, and the prior on ’s should have a capability of producing realisations with positive correlation structures, as this introduces smoothing among the ’s in adjacent bins. This latter property is not possible to attain with arbitrary constructions of inverse Gamma Markov chains, such as e.g. a natural construction . On the other hand, positive correlation between realisations ’s can be achieved e.g. by setting , but this results in intractable posterior computations. The definition of the IGMC prior in the present work, that employs latent variables ’s, takes care of both these important points. For an additional discussion see [cemgil07]. ∎
Remark 3.
Setting the drift coefficient to zero effectively results in pretending that the process has independent (Gaussian) increments. In reality, since the drift in practical applications is typically nonzero, increments of the process are dependent, and hence all observations contain some indirect information on the value of the volatility at each time point . On the other hand, assuming the IGMC prior on yields a posteriori dependence of coefficients , which should be of help in inference with smaller sample sizes . See Section 4 for an illustration. ∎
2.3. Gibbs sampler
It can be verified by direct computations employing (4) that the full conditional distributions of ’s and ’s are inverse gamma,
(5)  
(6)  
(7)  
(8) 
See Section 6 for details. Next, the effective transition kernel of the Markov chain is given by formula (4) in [cemgil07], and is a scale mixture of inverse gamma distributions; however, its exact expression is of no direct concern for our purposes. As noted in [cemgil07], p. 700, depending on the parameter values , it is possible for the chain to exhibit either an increasing or decreasing trend. We illustrate this point by plotting realisations of in Figure 1 for different values of and . In the context of volatility estimation this feature is attractive, if prior information on the volatility trend is available.
Inference in [cemgil07] is performed using a meanfield variational Bayes approach, see, e.g., [blei16]. Here we describe instead a fully Bayesian approach relying on Gibbs sampling (see, e.g., [gelfand90] and [geman84]), cf. [cemgil07b].
The algorithm is initialised at values , e.g. generated from the prior specification (4). In order to derive update formulae for the full conditionals of the ’s, define
With this notation, the likelihood from (3) satisfies
Using this formula and equation (5), and recalling the form of the inverse gamma density (2), it is seen that the update distribution for , is
whereas by (7) the ones for and are
respectively.
Next, the latent variables ’s will be updated using formula (8). This update step for ’s does not directly involve the data , except through the previous values of ’s.
Finally, one iterates these two Gibbs steps for ’s and ’s a large number of times (until chains can be assessed as reasonably converged), which gives posterior samples of the ’s. Using the latter, the posterior inference can proceed in the usual way, e.g. by computing the sample posterior mean of
’s, as well as sample quantiles, that provide, respectively, a point estimate and uncertainty quantification via marginal Bayesian credible bands for the squared volatility
. Similar calculations on the square roots of the posterior samples can be used to obtain point estimates and credible bands for the volatility function itself.2.4. Hyperparameter choice
Like [gugu17], our Bayesian procedure depends on the hyperparameter . As argued in [gugu17], in practice it is recommended to use the theoretical results in [gugu17] (that suggest to take , if the true volatility function is Hölder smooth) and try several values of simultaneously. Different ’s all provide information on the unknown volatility, but at different resolution levels; see Section 5 in [gugu17] for an additional discussion.
Additionally, we have hyperparameters , and , that govern properties of the Markov chain prior. In [cemgil07], where an IGMC prior was introduced, guidance on the hyperparameter selection is not discussed. In [cemgil:proc:08], the hyperparameters are finetuned by hand in specific problems studied there (audio denoising and single channel audio source separation). Another practical solution is to try several different fixed combinations of the hyperparameters , and , if only to verify sensitivity of inferential conclusions with respect to variations in the hyperparameters. Some further methods for hyperparameter optimisation are discussed in [dikmen10]. In [cemgil:proc:08] optimisation of the hyperparameters via the maximum likelihood method is suggested; practical implementation relies on the EM algorithm (see [dempster77]), and some additional details are given in [dikmen08]. Put in other terms, the proposal in [dikmen08]
amounts to using an empirical Bayes method (see, e.g.,
[efron10], [robbins56] and [robbins64]). The use of the latter is widespread and often leads to good practical results, but the method is still insufficiently understood theoretically, except in toy models like the white noise model (see, however,
[donnet18] and [petrone14] for some results in other contexts). On the practical side, in our case, given that the dimension of the sequences and is rather high, namely with large, and the marginal likelihood is not available in closed form, this approach is expected to be computationally intensive. Therefore, a priori there is no reason not to try instead a fully Bayesian approach by equiping the hyperparameters with a prior, and this is in fact our default approach in the present work. However, the corresponding full conditional distribution turns out to be nonstandard, which necessitates the use of a MetropolisHastings step within the Gibbs sampler (see, e.g., [hastings70], [metropolis53] and [tierney94]). We provide the necessary details in Section 6.3. Synthetic data examples
Computations in this section have been done in the programming language Julia, see [bezanson17]. In order to test the practical performance of our estimation method, we use a challenging example with the blocks function from [donoho95]. As a second example, we consider the case of the CoxRossIngersoll model. Precise details are given in the subsections below.
We used the Euler scheme on a grid with 800 001 equidistant points on the interval to obtain realisations of a solution to (1) for different combinations of the drift and dispersion coefficients. These were then subsampled to obtain observations in each example.
The hyperparameter was set to , whereas for the other two hyperparameters we assumed that and used a diffuse prior, except in specially noted cases below. Inference was performed using the Gibbs sampler from Section 2, with a MetropolisHastings step to update the hyperparameter . The latter used an independent Gaussian random walk proposal with a scaling to ensure the acceptance rate of ca. ; see Section 6. The Gibbs sampler was run for iterations and we used a burnin of samples. In each example we plotted marginal credible bands obtained from the central posterior intervals for the coefficients .
3.1. Blocks function
As our first example, we considered the case when the volatility function was given by the blocks function from [donoho95]. With a vertical shift for positivity, this is defined as follows:
(9) 
where , and
The function serves as a challenging benchmark example in nonparametric regression: it is mostly very smooth, but spatially inhomogeneous and characterised by abrupt changes (cf. Chapter 9 in [wasserman06]). Unlike nonparametric regression, the noise (Wiener process) in our setting should be thought of as multiplicative and proportional to rather than additive, which combined with the fact that takes rather large values further complicates the inference problem. Our main goal here was to compare the performance of the IGMC priorbased approach to the IIG priorbased one from [gugu17]. To complete the SDE specification, our drift coefficient was chosen to be a rather strong linear drift
In Figure 2 we plot the blocks function (9) and the corresponding realisation of the process used in this simulation run.
The left and right panels of Figure 3 contrast the results obtained using the IGMC prior with and versus and . These plots illustrate the fact that increasing has the effect of undersmoothing prior realisations, that can be balanced by increasing the values of , which has the opposite smoothing effect. Because of this, in fact, both plots look quite similar.
The top left and top right panels of Figure 4 give estimation results obtained with the IIG priorbased approach from [gugu17]. The number of bins was again and , and in both these cases we used diffuse independent priors on the coefficients of the (squared) volatility function (see [gugu17] for details). These results have to be contrasted to those obtained with the IGMC prior, plotted in the bottom left and bottom right panels of Figure 4, where we assumed and . The following conclusions emerge from Figure 4:

Although both the IGMC and IIG approaches recover globally the shape of the volatility function, the IIG approach results in much greater uncertainty in inferential conclusions, as reflected in wider marginal confidence bands. The effect is especially pronounced in the case , where the width of the band for the IIG prior renders it almost useless for inference.

The bands based on the IGMC prior look more ‘regular’ than the ones for the IIG prior.

Comparing the results to Figure 3, we see the benefits of equipping the hyperparameters with a prior: credible bands in Figure 3 do not adequately capture two dips of the function right before and after the point , since completely falls outside the credible bands there. Thus, an incorrect amount of smoothing is used in Figure 3.

The method based on the IIG prior is sensitive to the bin number selection: compare the top left panel of Figure 4 using bins to the top right panel using bins, where the credible band is much wider. On the other hand, the method based on the IGMC prior automatically rebalances the amount of smoothing it uses with different numbers of bins
, thanks to the hyperprior on the parameters
; in fact, the bottom two plots in Figure 4 look similar to each other.
3.2. CIR model
Our core estimation procedure, as described in the previous sections, assumes that the volatility function is deterministic. In this subsection, however, in order to test the limits of applicability of our method and possibilities for future extensions, we applied it to a case where the volatility function was stochastic. The study in [mykland12] lends support to this approach, but here we concentrate on practical aspects and defer the corresponding theoretical investigation until another occasion.
Specifically, we considered the CoxRossIngersoll (CIR) model or the square root process,
(10) 
Here are parameters of the model. This diffusion process was introduced in [feller51a] and [feller51b], and gained popularity in finance as a model for shortterm interest rates, see [cox85]. The condition ensures strict positivity and ergodicity of . The volatility function from (1) now corresponds to a realisation of a stochastic process , where solves the CIR equation (10).
We took arbitrary parameter values
(11) 
A sample path of is plotted in the left panel of Figure 5, whereas the corresponding volatility is given in the right panel of the same figure. In Figure 6 we display estimation results obtained with the IGMC prior, using and bins and hyperparameter specifications and . A conclusion that emerges from this figure is that our Bayesian method captures the overall shape of the realised volatility in a rather satisfactory manner.
4. DowJones industrial averages
In this section we provide a reanalysis of a classical dataset in changepoint detection in time series; see, e.g., [chen12], [diaz82], [hsu77], [hsu79] and [iacus08]. Specifically, we consider weekly closing values of the DowJones industrial averages in the period 2 July 1971 – 2 August 1974. In total there are 162 observations available, which constitute a relatively small sample, and thus the inference problem is rather nontrivial. The data can be accessed as the dataset DWJ in the sde package (see [iacus16]) in R (see [r]). See the left panel of Figure 7 for a visualisation. In [iacus08] the weekly data , , are transformed into returns , and the least squares changepoint estimation procedure from [degregorio08] has been performed. Reproducing the corresponding computer code in R results in a changepoint estimate of 16 March 1973. That author speculates that this changepoint is related to the Watergate scandal.
Similar to [iacus08], parametric changepoint analyses in [chen12], [diaz82] and [hsu79] give a changepoint in the third week of March 1973. However, as noted in [iacus08], examination of the plot of the time series (see Figure 7, the right panel) indicates that another changepoint may be present in the data. Then dropping observations after 16 March 1973 and analysing the data for existence of a changepoint using only the initial segment of the time series, the author discovers another changepoint on 17 December 1971, which he associates with suspending the convertibility of the US dollar into gold under President Richard Nixon’s administration.
From the above discussion it should be clear that nonparametric modelling of the volatility may provide additional insights for this dataset. We first informally investigated the fact whether an SDE driven by the Wiener process is a suitable model for the data at hand. Many of such models, e.g. the geometric Brownian motion, a classical model for evolution of asset prices over time (also referred to as the Samuelson or BlackScholes model), rely on an old tenet that returns of asset prices follow a normal distribution. Although the assumption has been empirically disproved for highfrequency financial data (daily or intraday data; see, e.g.,
[carr02], [eberlein95] and [kuchler91]), its violation is less severe for widely spaced data in time (e.g. weekly data, as in our case). In fact, the ShapiroWilk test that we performed in Ron the returns past the changepoint 16 March 1973 did not reject the null hypothesis of normality (
value 0.4). On the other hand, the quantilequantile (QQ) plot of the same data does perhaps give an indication of a certain mild deviation from normality, see Figure 8, where we also plotted a kernel density estimate of the data (obtained via the command
density in R, with bandwidth determined automatically through crossvalidation).In Figure 9 we plot the sample autocorrelation and partial autocorrelation functions based on returns ’s past the changepoint 16 March 1973. These do not give decisive evidence against the assumption of independence of ’s. Neither does the LjungBox test (the test is implemented in R via the command Box.test), which yields a value when applied with lags (the value is certainly small, but not overwhelmingly so).
Summarising our findings, we detected only a mild evidence against the assumption that the returns of the DowJones weekly closings of industrial averages (over the period 16 March 1973 – 2 August 1974, but similar conclusions can be reached also over the other subperiods covered by the DWJ dataset) are approximately independent and follow a normal distribution. Thus there is no strong a priori reason to believe that a geometric Brownian motion is an outright unsuitable model in this setting: it can be used as a first approximation. To account for timevariability of volatility (as suggested by the changepoint analysis), we incorporate a timedependent volatility function in the model, and for additional modelling flexibility we also allow a statedependent drift. Setting , our model is thus given by
(12) 
An alternative here is to directly (i.e. without any preliminary transformation) model the DowJones data using equation (1). We consider both possibilities, starting with the model (12).
As in our simulation examples in Section 3, we used a vague prior on corresponding to the limit , whereas for the other two hyperparameters we assumed . The scaling in the independent Gaussian random walk proposal in the MetropolisHastings step was chosen in such a way so as to yield an acceptance rate of ca. . The Gibbs sampler was run for iterations, and the first samples were dropped as a burnin. We present the estimation results we obtained using and bins, see Figure 10. Although the sample size is quite small in this example, the data are informative enough to yield nontrivial inferential conclusions even with diffuse priors. Both plots in Figure 10 are qualitatively similar and suggest:

A decrease in volatility at the end of 1971, which can be taken as corresponding to the changepoint in December 1971 identified in [iacus08]. Unlike that author, we do not directly associate it with suspending the convertibility of the US dollar into gold (that took place in August 1971 rather than December 1971).

A gradual increase in volatility over the subsequent period stretching until the end of 1973. Rather than only the Watergate scandal (and a changepoint in March 1973 as in [iacus08]), there could be further economic causes for that, such as the 1973 oil crisis and the 1973–1974 stock market crash.

A decrease in volatility starting in early 1974, compared to the immediately preceding period.
In general, in this work we do not aim at identifying causes for changes in volatility regimes, but prefer to present our inference results, that may subsequently be used in econometric analyses.
Now we move to the Bayesian analysis of the data using model (1). The prior settings were as in the previous case, and we display the results in Figure 11. The overall shapes of the inferred volatility functions are the same in both Figure 10 and Figure 11, and hence similar conclusions apply.
Finally, we stress the fact that our nonparametric Bayesian approach and changepoint estimation are different in their scope: whereas our method aims at estimation of the entire volatility function, changepoint estimation (as its name actually suggests) concentrates on identifying changepoints in the variance of the observed time series, which is a particular feature of the volatility. To that end it assumes the (true) volatility function is piecewise constant, which on the other hand is not an assumption required in our method. Both techniques are useful, and each can provide insights that may be difficult to obtain from another.
5. Conclusions
Bayesian inference for SDEs from discretetime observations is a difficult task, owing to intractability of the likelihood and the fact that the posterior is not available in closed form. Posterior inference therefore typically requires the use of intricate MCMC samplers. Designing algorithms that result in Markov chains that mix well and explore efficiently the posterior surface is a highly nontrivial problem. Inspired by some ideas from the audio signal processing literature and our earlier work [gugu17], in this paper we introduced a novel nonparametric Bayesian approach to estimation of the volatility coefficient of an SDE. Our method is easy to understand and straightforward to implement via Gibbs sampling, and performs well in practice. Thereby our hope is that our work will contribute to further dissemination and popularisation of a nonparametric Bayesian approach to inference in SDEs, specifically with financial applications in mind. Our work can also be viewed as a partial fulfillment of anticipation in [godsill07] that some ideas developed originally in the context of audio and music processing “will also find use in other areas of science and engineering, such as financial or biomedical data analysis”.
As a final remark, we do not attempt to provide a theoretical, i.e. asymptotic frequentist analysis of our new approach here (see, e.g., the recent monograph [ghosal17], and specifically [gugu17] for such an analysis in the SDE setting), but leave this as a topic of future research.
6. Formulae for parameter updates
In this section we present additional details on the derivation of the update formulae for the Gibbs sampler from Section 2. The starting point is to employ the Markov property from (4), and using the standard Bayesian notation, to write the joint density of and as
(13) 
6.1. Full conditional distributions
We first indicate how (5) was derived. Insert expressions for the individual terms in (13) from (4) and collect separately terms that depend on only, to see that the density of the full conditional distribution of , , is proportional to
Upon normalisation, this expression is the density of the distribution, which proves fromula (5). Formula (7) follows directly from the last expression in (4). Formula (8) is proved analogously to (5). Finally, (6) follows from (4) and Bayes’ formula. Cf. also [dikmen08], Appendix B.6.
6.2. MetropoliswithinGibbs step
Now we describe the MetropolisHastings step within the Gibbs sampler, that is used to update the hyperparameters of our algorithm, in case the latter are equipped with a prior. For simplicity, assume (we note that such a choice is used in practical examples in [cemgil:proc:08]), and suppose is equipped with a prior, . Let the hyperparameter be fixed. Obviously, could have been equipped with a prior as well, but this would have further slowed down our estimation procedure, whereas the practical results in Sections 3 and 4 we obtained are already satisfactory with fixed. Using (4) and (13), one sees that the joint density of , and is proportional to
This in turn is proportional to
The latter expression is an unnormalised full conditional density of , and can be used in the MetropoliswithinGibbs step to update .
The rest of the MetropolisHastings step is standard, and the following approach was used in our practical examples: pick a proposal kernel , for instance a Gaussian random walk proposal , where is the density of a normal random variable with mean zero and variance . Note that this specific choice may result in proposing a negative value , which needs to be rejected straightaway as invalid. Then, for computational efficiency, instead of moving to another step within the Gibbs sampler, one keeps on proposing new values until a positive one is proposed. This is then accepted with probability
where
is the cumulative distribution function of a normal random variable with mean zero and variance
; otherwise the current value is retained. See [wilkinson12] for additional details and derivations. Finally, one moves to other steps in the Gibbs sampler, namely to updating ’s and ’s, and iterates the procedure. The acceptance rate in the MetropolisHastings step can be controlled through the scale parameter of the proposal density . Some practical rules for determination of an optimal acceptance rate in the MetropolisHastings algorithm are given in [gelman:proc:96], and those for the MetropoliswithinGibbs algorithm in [sherlock10].
Comments
There are no comments yet.