Buckets:
Adaptive sequential Monte Carlo by means of mixture of experts
Julien Cornebise* E. Moulines† J. Olsson‡
September 13thth, 2012
Abstract: Appropriately designing the proposal kernel of particle filters is an issue of significant importance, since a bad choice may lead to deterioration of the particle sample and, consequently, waste of computational power. In this paper we introduce a novel algorithm adaptively approximating the so-called optimal proposal kernel by a mixture of integrated curved exponential distributions with logistic weights. This family of distributions, referred to as mixtures of experts, is broad enough to be used in the presence of multi-modality or strongly skewed distributions. The mixtures are fitted, via online-EM methods, to the optimal kernel through minimisation of the Kullback-Leibler divergence between the auxiliary target and instrumental distributions of the particle filter. At each iteration of the particle filter, the algorithm is required to solve only a single optimisation problem for the whole particle sample, yielding an algorithm with only linear complexity. In addition, we illustrate in a simulation study how the method can be successfully applied to optimal filtering in nonlinear state-space models.
Keywords: Optimal proposal kernel, Adaptive algorithms, Kullback-Leibler divergence, Coefficient of variation, Expectation-maximisation, Particle filter, Sequential Monte Carlo, Shannon entropy
1 Introduction
During the last decade, sequential Monte Carlo (SMC) methods have developed from being a tool for solving specific problems, such as the optimal filtering problem in general state-space models or simulation of rare events
(these topics are abundantly covered in Liu, 2001; Doucet et al., 2001; Ristic et al., 2004; Cappé et al., 2005; Del Moral et al., 2006, and the references therein). The high flexibility of these methods makes it possible to use them for efficiently executing key steps of composite algorithms tackling very complex problems, e.g., as in particle Markov chain Monte Carlo (MCMC) methods for joint inference on static and dynamic variables (see Andrieu et al., 2010).
However, while there has been abundant work on adaptation of MCMC methods, from the early algorithms of Haario et al. (2001) to the mathematical study of diminishing adaptation established in Andrieu and Moulines (2006) and Roberts and Rosenthal (2009), adaptive methods for SMC are still in their infancy. This article, which sets out from a theoretical framework for adaptive SMC presented in Cornebise et al. (2008), is part of the effort to fill this gap—which is required to achieve maximum efficiency and user-friendliness. A mathematically less detailed pre-version of the present paper appeared as part of the Ph. D. dissertation by Cornebise (2009).
Such adaptive methods are especially suited to cope with the rising complexity of state-space models (SSMs) in recent applications. While simple unidimensional filtering problems will rarely justify even a slight computational overhead of adaptive methods over plain SMC methods (such as the bootstrap particle filter; see Section 3.1), the situation is reversed for many state-of-the-art SSMs, where forward simulation may require costly physical emulators or the evaluation of computationally expensive transition densities. This occurs, e.g., as soon as the model involves differential equations and numeric solutions of
†Département Traitement du Signal et de l'Image, Telecom ParisTech
‡Centre for Mathematical Sciences, Lund Universitythe same. Recent examples range from health sciences (see e.g. Korotsil et al., 2012, for a study of the epidemiology of human papillomavirus infections) to environmental sciences (see e.g. Robins et al., 2009, who studied data assimilation for atmospheric dispersion of hazardous airborne biological material).
In the SMC framework, the objective is to approximate a sequence of target densities, defined on a sequence of state spaces as follows. Given a density $\mu$ on some state space $\Xi$ , the next density $\tilde{\mu}$ in the sequence is defined by
where $\tilde{\Xi}$ is some other state space and $l : \Xi \times \tilde{\Xi} \rightarrow \mathbb{R}_+$ is an un-normalised transition density. Here the dimensions of $\tilde{\Xi}$ and $\Xi$ typically differ. This framework encompasses, inter alia, the classical problem of optimal filtering in SSMs recalled in Section 6 (see in particular Equation (6.4)).
SMC methods approximate online the densities generated by (1.1) by a recursively updated set of random particles ${\mathbf{X}i}{i=1}^N$ and associated importance weights ${\omega_i}{i=1}^N$ . At each iteration of the algorithm, the weighted particle sample ${(\mathbf{X}i, \omega_i)}{i=1}^N$ approximates the corresponding target density $\mu$ in the sense that $\sum{i=1}^N \omega_i f(\mathbf{X}i)/\Omega$ , with $\Omega := \sum{\ell=1}^N \omega_\ell$ , approximates $\mathbb{E}\mu[f(\mathbf{X})] := \int f(\mathbf{x})\mu(\mathbf{x}) d\mathbf{x}$ (this will be our generic notation for expectations) for all bounded measurable functions $f$ on the corresponding state space $\Xi$ . All SMC methods have two main operations in common: a mutation step and a selection step. In the former, the particles are randomly scattered in the state space by means of a proposal transition density $r : \Xi \times \tilde{\Xi} \rightarrow \mathbb{R}+$ ; more specifically, the mutation operation generates a set of new, $\tilde{\Xi}$ -valued particles by moving the particles according to the transition density $r$ and assigning each mutated particle an updated importance weight proportional to $l/r$ ; consequently, each weight reflects the relevance of the corresponding particle as measured by the likelihood ratio. In general, the mutated particles are conditionally independent given the ancestor particles. In the selection step, the particles are duplicated or eliminated with probabilities depending on the importance weights. This is typically done by multinomially resampling the particles with probabilities proportional to the corresponding weights. In the auxiliary particle filter
(proposed by Pitt and Shephard, 1999) the particles are resampled according to weights that are modified by a factor determined by an adjustment multiplier function $a : \Xi \rightarrow \mathbb{R}_+$ . Introducing adjustment multipliers makes it possible to, by using some convenient lookahead strategy, increase the weight of those particles that will most likely contribute significantly to the approximation of the subsequent target distribution after having been selected. In this way computational power is directed toward zones of the state space where the mass of the subsequent target density is located.
The choice of the adjustment weight function $a$ and the proposal density $r$ affects significantly the quality of the generated sample, and in this paper we focus on adaptive design of the latter. Letting $r(\mathbf{x}, \tilde{\mathbf{x}}) = l(\mathbf{x}, \tilde{\mathbf{x}})/\int l(\mathbf{x}, \tilde{\mathbf{x}}) d\tilde{\mathbf{x}}$ is appealing as the particles in this case are propagated according to a dynamics that is closely related to that of the recursion (1.1). Indeed, by combining this choice of proposal kernel with the adjustment multiplier function $a(\mathbf{x}) = \int l(\mathbf{x}, \tilde{\mathbf{x}}) d\tilde{\mathbf{x}}$ yields perfectly uniform importance weights. In this case the SMC algorithm is referred to as fully adapted. However, this so-called optimal proposal kernel and adjustment multiplier weight function can be expressed on closed-form only in a few cases and one is in general referred to approximations of these quantities.
There is a large literature dealing with the problem of approximating the optimal proposal kernel. When using SMC for optimal filtering in SSMs, Doucet et al. (2000) suggest approximating each function $l(\mathbf{x}, \cdot)$ by a Gaussian density whose mean and covariance are obtained using the extended Kalman filter. A similar approach based on the unscented Kalman filter was proposed by Van der Merwe et al. (2000) (see also Van der Merwe and Wan, 2003). This technique presupposes that the model under consideration can be well approximated by a nonlinear Gaussian SSM with linear measurement equation; however, this far from always the case, and careless linearisation (e.g. in cases where the local likelihood is multimodal) may boomerang and cause severe deterioration of the particle sample. Cappé et al. (2005, Section 7.2.2.4) make a form of Laplace approximation of $l(\mathbf{x}, \cdot)$ by simply locating the mode of the function in question and centering a Gaussian density or the density of a student's $t$ -distribution around the same. This technique, which goes back to Pitt and Shephard (1999), is appro-priate when the function is log-concave (or strongly unimodal); nevertheless, as the mode of each $l(\mathbf{x}, \cdot)$ is a function of $\mathbf{x}$ , this approach involves solving one optimisation problem per particle. Thus, in spite of this intense recent activity in the field, the state-of-the-art algorithms have met only mitigated success as they implicitly assume that the functions $l(\mathbf{x}, \cdot)$ have a single mode.
A common practice (see e.g. Oh and Berger, 1993) for designing proposal distributions in standard (non-sequential) importance sampling is to consider a parametric family of proposal distributions and then identify a parameter that minimizes some measure of discrepancy between the target and the proposal distributions. Common choices are the widely used squared coefficient of variation of the importance weights or the negated Shannon entropy (see e.g. Cappé et al., 2005, p. 235, for definitions of these quantities) of the same. Both are maximal when the sample is completely degenerated, i.e. when one single particle carries all the weight, and minimal when all the importance weights are equal. The coefficient of variation often appears in a transformed form as the effective sample size (we refer again to Cappé et al., 2005, for a definition). The same quantities have been widely used also within the framework of SMC, but until recently only for triggering selection and not as a tool for adaptive design of the instrumental distribution. The key result of Cornebise et al. (2008, Theorems 1 and 2) was to relate the Shannon entropy to the Kullback-Leibler divergence (KLD; see Equation (3.6)) between the instrumental and proposal distributions of the particle filter. More specifically, when selection is carried through using multinomial resampling, each particle cloud update (selection followed by resampling) can be viewed as standard importance sampling where a weighted mixture of $N$ stratas—the $i$ th stratum being proportional to $\omega_i l(\mathbf{X}_i, \cdot)$ —approximating $\tilde{\mu}$ is targeted using a mixture proposal $\pi$ comprising $N$ stratas proportional to ${\omega_i r(\mathbf{X}i, \cdot)}{i=1}^N$ . As the number of particles tends to infinity, one may prove that the KLD between these two mixtures tends to the KLD between two distributions having densities $\mu^*(\mathbf{x}, \tilde{\mathbf{x}}) \propto l(\mathbf{x}, \tilde{\mathbf{x}})\mu(\mathbf{x})$ and $\pi^*(\mathbf{x}, \tilde{\mathbf{x}}) \propto a(\mathbf{x})r(\mathbf{x}, \tilde{\mathbf{x}})\mu(\mathbf{x})$ , which can be viewed as “asymptotic” target and proposal distributions on the product space $\Xi \times \tilde{\Xi}$ . In addition, one may prove that the Shannon entropy has the same limit. (Cornebise et al. (2008) established a similar relation be-
tween the coefficient of variation of the particle weights and the chi-square divergence between the same distributions.) This gives a sound theoretical support for using the Shannon entropy of the importance weights for measuring the quality of the particle sample. In addition, it suggests that the KLD between the target mixture and $\pi$ could be used in lieu of the Shannon entropy for all purposes, especially adaptation. As a matter of fact, in the context of adaptive design of SMC methods, the KLD is (as pointed out by Cornebise et al., 2008, Section 2.3) highly practical since it decouples the adaptation of the adjustment weight function $a$ and that of the proposal kernel $r$ ; see the next section.
Henceforth, in the present article we extend and implement fully the methodology indicated and sketched by Cornebise et al. (2008) and select the auxiliary proposal distribution $\pi$ from a family ${\pi_{\boldsymbol{\theta}}; \boldsymbol{\theta} \in \Theta}$ of candidate distributions by picking $\boldsymbol{\theta}^*$ such that the KLD between the target mixture and $\pi_{\boldsymbol{\theta}^*}$ is minimal and letting $\pi = \pi_{\boldsymbol{\theta}^*}$ . On the one hand, the chosen parametric family should be flexible enough to approximate complex transition kernels; on the other hand, sampling from $\pi_{\boldsymbol{\theta}}$ should be easy. Finally, the parameterisation should be done in such a way that the problem of estimating the parameters is as simple as possible. In this article, we suggest modeling the proposal $\pi_{\boldsymbol{\theta}}$ as a mixture of integrated curved exponential distributions. By letting the mixture weights of the chosen proposal depend on the ancestor particles ${\mathbf{X}i}{i=1}^N$ , we allow for partitioning of the input space into regions corresponding to a specialised kernel. Each component of the mixture belongs to a family of integrated curved exponential distributions, whose two most known members are the multivariate Gaussian distribution and the Student’s $t$ -distribution. Also the parameters of the mixture depend on the ancestors ${\mathbf{X}i}{i=1}^N$ . This parameterisation of the proposal distribution is closely related to the (hierarchical) mixture of experts appearing in the machine learning community and described in (Jordan and Jacobs, 1994; Jordan and Xu, 1995). The flexibility of our approach allows for efficient approximation of the optimal kernel for a large class of intricate (nonlinear non-Gaussian) models. Unlike typical EM-type parameter estimation procedures, which are in general run iteratively until stabilisation, our adaptive SMC algorithm only requires a decrease of the KLD, not an exact minimum.As illustrated in our examples in Section 6, such a gain typically occurs already at early iterations of the algorithm (a characteristic of EM-type algorithms), implying the need of only very few EM iterations and therefore a minimal computational overhead.
The paper is organized as follows. In the next section we introduce some matrix notation and list the most common notation used throughout the paper. Section 3 describes, in a nutshell, auxiliary SMC methods as well as a KLD-based approach to adaptation of these methods that is used in the paper. Section 4 recalls mixtures of experts and Section 5 treats optimisation of the mixture parameters by means of stochastic approximation methods; the latter section is the core of the paper and contains our main results. In Section 6 we illustrate the efficiency of the method on several simulation-based examples and Section A contains some proofs.
2 Notation
2.1 Matrix notation
All vectors and matrices are typeset in boldface. Vectors are column vectors unless specified differently. The $i$ th column vector of a matrix $\mathbf{A}$ is denoted by $\mathbf{A}_{|i}$ . We denote by $\text{Tr}(\mathbf{A})$ the trace of a matrix $\mathbf{A}$ and for any matrices $\mathbf{A}$ and $\mathbf{B}$ of dimensions $m \times n$ and $p \times q$ , respectively, we denote their direct sum by
which is such that $(\mathbf{A} \oplus \mathbf{B})^\top = \mathbf{A}^\top \oplus \mathbf{B}^\top$ and, for any two matrices $\mathbf{C}$ and $\mathbf{D}$ of compatible dimensions, $(\mathbf{A} \oplus \mathbf{B})(\mathbf{C} \oplus \mathbf{D})^\top = (\mathbf{A}\mathbf{C}^\top) \oplus (\mathbf{B}\mathbf{D}^\top)$ , where $\top$ denotes the transpose.
2.2 List of notation
The following quantities are defined in the corresponding equations.
| Quantity | Equation |
|---|---|
| (3.3) | |
| (4.5) | |
| (3.6) | |
| (4.2) | |
| (6.3) | |
| (6.2) | |
| (6.2) |
| Quantity | Equation |
|---|---|
| (5.15) | |
| (5.23) | |
| (5.8) | |
| (3.2) | |
| (1.1) | |
| (5.7) | |
| (5.21) | |
| (4.3) | |
| (5.1) | |
| (5.5) | |
| (4.1) | |
| (2.1) | |
| (6.1) | |
| (6.1) | |
| (4.1) | |
| (4.6) | |
| (5.2) | |
| (5.6) | |
| (5.20) | |
| (5.27) | |
| (5.28) | |
| (4.2) | |
| (5.9) | |
| (5.27) | |
| (5.28) | |
| (4.4) |
3 Preliminaries
3.1 Auxiliary SMC methods
In order to precisely describe one iteration of the SMC algorithm, let $\mu$ be a target probability density function over a space $\Xi$ and suppose that we have at hand a weighted sample ${(\mathbf{X}i, \omega_i)}{i=1}^N$ such that $\sum_{i=1}^N \omega_i f(\mathbf{X}i)/\Omega$ , with $\Omega = \sum{i=1}^N \omega_i$ , estimates $\int f(\mathbf{x})\mu(\mathbf{x}) d\mathbf{x}$ for any $\mu$ -integrable function $f$ . As a notational convention, we use capitals to denote random variables and lower case for function arguments.
We now wish to transform (by moving the particles and adjusting accordingly the importance weights) the sample ${(\mathbf{X}i, \omega_i)}{i=1}^N$ into a new weighted particle sample approximating the probability density $\tilde{\mu}$ defined on $\tilde{\Xi}$ through (1.1). Having access to ${(\mathbf{X}i, \omega_i)}{i=1}^N$ , an approximation of $\tilde{\mu}$ is naturally obtained by plugging the weighted empirical measure associated with this weighted sample into (1.1),yielding the mixture density
where we have introduced the normalised transition density
and the partition function
of $l(\mathbf{x}, \cdot)$ , and ideally an updated particle sample would be obtained by drawing new particles ${\tilde{\mathbf{X}}i}{i=1}^N$ independently from (3.1). There are however two problems with this approach: firstly, in general, the integral $\int l(\cdot, \tilde{\mathbf{x}}) d\tilde{\mathbf{x}}$ lacks closed-form expression, ruling out direct computation of the mixture weights; secondly, even if the integral was known in closed form, the resulting algorithm would have an $\mathcal{O}(N^2)$ computational complexity due to the normalization of the mixture weights. To cope with these issues, we will, as proposed by Pitt and Shephard (1999), aim instead at sampling from an auxiliary target distribution having density
over the product space ${1, \dots, N} \times \tilde{\Xi}$ of indices and particle positions. To sample $\bar{\mu}$ we take an importance sampling approach consisting in drawing independent pairs ${(I_i, \tilde{\mathbf{X}}i)}{i=1}^N$ of indices and positions from the proposal distribution
over the same extended space and assigning each draw $(I_i, \tilde{\mathbf{X}}_i)$ the importance weight $\tilde{\omega}_i := w(I_i, \tilde{\mathbf{X}}_i)$ , where
Here $r$ and $a$ are the proposal transition density resp. adjustment multiplier weight function mentioned in the introduction. As for
all importance sampling, it is crucial that the target distribution $\bar{\mu}$ is absolutely continuous with respect to the proposal distribution $\pi$ . We thus require that the adjustment multiplier weight function is positive and that the proposal transition density is such that for all $\mathbf{x} \in \Xi$ ,
i.e. the support of $l(\mathbf{x}, \tilde{\mathbf{x}})$ is included in that of $r(\mathbf{x}, \tilde{\mathbf{x}})$ . Finally, we discard the indices $I_i$ and keep ${(\tilde{\mathbf{X}}_i, \tilde{\omega}i)}{i=1}^N$ as an approximation of $\tilde{\mu}$ .
This scheme, which is traditionally referred to as the auxiliary particle filter (Pitt and Shephard, 1999), encompasses the simpler framework of the bootstrap particle filter proposed originally by Gordon et al. (1993), which simply amounts to setting $a = \mathbb{1}_{\Xi}$ (implying a gain of simplicity at the price of flexibility). Moreover, SMC schemes where resampling is performed only at random times can similarly be cast into the setting of the auxiliary particle filter by composing the kernels involved in several consecutive steps of the of recursion (1.1) (see e.g. Cornebise, 2009, Chapter 4, for details). A theoretical analysis of bootstrap-type SMC methods with random resampling schedule is also given by Del Moral et al. (2012). Therefore, any methodology built for the auxiliary particle filter also applies to these simpler frameworks without modification.
3.2 A foundation of SMC adaptation
We now describe a foundation of SMC adaptation underpinning the coming developments. Recall that the KLD between two distributions $\mu$ and $\nu$ , defined on the same state space $\Xi$ , is defined as
provided that $\mu$ is absolutely continuous with respect to $\nu$ . Here $\mathbb{E}_{\mu}$ denotes the expectation under $\mu$ . Casting the KLD into the framework of SMC methods and replacing $\mu$ and $\nu$ by the auxiliary target distribution $\bar{\mu}$ and the proposal $\pi$ , respectively, yields
which decouples as
As clear from the definition (3.4), the measure $\bar{\mu}$ is random as it depends on the ancestor particles ${\mathbf{X}i}{i=1}^N$ ; the locations of the latter should thus be viewed as fixed within the brackets of the expectations in (3.7) and (3.8). The first term in (3.8) corresponds to the discrepancy induced by mutating the particles ${\tilde{\mathbf{X}}i}{i=1}^N$ using a suboptimal proposal kernel, and the second term corresponds to the discrepancy induced by sampling the ancestor indices ${I_i}_{i=1}^N$ according to suboptimal adjustment weights. Moreover,
where the additive term $c$ involves the optimal quantities $a^*$ and $l^*$ only and is irrelevant for the optimisation problem. Equality up to a constant will in the following be denoted by $\equiv$ . Restricting ourselves to adaptation of the proposal kernel, we obtain the simple expression
In the next section we present a method for minimising the KLD (3.9) over a family ${\pi_{\theta}; \theta \in \Theta}$ of instrumental distributions where the proposal transition density of each member belongs to parametrised family of mixtures of experts. Formally, we will solve
and take $\pi = \pi_{\theta^*}$ . Although the right hand side of (3.9) is most often not directly computable, such a quantity can be approximated on-the-fly using the weighted sample already generated; the resulting algorithms (Algorithms 1 and 2; see Section 5) closely resembles the cross-entropy method (see Rubinstein and Kroese, 2004) method.
4 Mixture of experts
In this contribution, we consider $\Xi = \mathbb{R}^p$ and $\tilde{\Xi} = \mathbb{R}^{\bar{p}}$ and let the proposal kernel have density (with respect to Lebesgue measure)
where ${\alpha_j}_{j=1}^d$ are nonnegative weighting functions summing into unity and $\rho$ is a Markov transition kernel from $\Xi$ to $\tilde{\Xi}$ . The weighting functions and the Markov transition density are parametrised by parameters $\underline{\beta}$ and $\boldsymbol{\eta}$ , respectively, and the $j$ th stratum of the mixture is characterised by a certain choice $\boldsymbol{\eta}_j$ of the latter. The integer $d$ is fixed. Denote
Putting (4.1) into the auxiliary particle filter framework in Section 3.1, the associated proposal distribution is given by
We then assign the importance weight $\tilde{\omega}i = w{\theta}(I_i, \tilde{\mathbf{X}}_i)$ to each draw $(I_i, \tilde{\mathbf{X}}i)$ from $\pi{\theta}$ , where
The set ${\alpha_j}{j=1}^d$ of weight functions typically partitions the input space $\Xi$ into sub-regions with smooth transitions by assigning a vector of mixture weights to each point of the input space. In each sub-region, one proposal kernel will eventually dominate the mixture, specialising the whole proposal $r{\theta}$ on a per-region basis. As in Jordan and Xu (1995), we consider logistic weight functions, i.e. $\underline{\beta} = {\beta_1, \dots, \beta_{d-1}}$ and
where $\bar{\mathbf{x}} := (\mathbf{x}^\top \ 1)^\top$ and the $\beta_j$ 's are vectors in $\mathbb{R}^{p+1}$ . It is sometimes of interest to resort to simpler mixtures whose weights do not depend on $\mathbf{x}$ by letting, for a set $\underline{\beta} = (\beta_j)_{j=1}^d$ nonnegative scalars summing into unity, $\alpha_j(\mathbf{x}, \underline{\beta}) := \beta_j$ , $1 \leq j \leq d$ , independently of $\mathbf{x}$ and hence without partitioning the input space. This model for the transition kernel is then similar to the switching regression model (Quandt and Ramsey, 1972).
The original mixture of experts proposed by Jordan and Jacobs (1994) was based on Gaussian strata $\rho$ , which is indeed is the most straightforward choice. However the same algorithm can, at no additional cost, be extended to the broader class of densities of integrated curved exponential form, described in Assumption 4.1 below, which still allows for the same trade-off between flexibility, ease of sampling, and convenient estimation.
Assumption 4.1 There exist a state space $\Upsilon \subseteq \mathbb{R}^u$ and a Markov transition kernel from $\Xi$ to $\tilde{\Xi} \times \Upsilon$ having a density of form
where the functions $\gamma$ , $h$ , $a$ and $\mathbf{s}$ , $\mathbf{b}$ are real-valued resp. $\mathbb{R}^{s \times s'}$ -valued, such that
In other words, $\rho(\mathbf{x}, \tilde{\mathbf{x}}; \boldsymbol{\eta})$ is the density of the marginal of a curved exponential distribution. In (4.6), $\mathbf{s}$ is a vector of sufficient statistics.
In the next section we present a method fitting efficiently the proposal $\pi_\theta$ to the target $\bar{\mu}$ by solving (3.10). We will illustrate the details for the specific case of multidimensional Gaussian distributions and multidimensional student's $t$ -distributions. It should however be kept in mind that the algorithm is valid for any member of the family of integrated curved exponential distributions as long as Assumption 5.1 below is fulfilled.
5 Main results and algorithms
In the following we discuss how the intricate optimisation problem (3.10) can be cast into the framework of latent data problems. Parameter estimation for mixtures of experts is most often carried through using the EM algorithm (see McLachlan and Krishnan, 2008,
Section 8.4.6), and also in this paper a recursive EM-type algorithm will be used for finding close to optimal mixture parameters. More precisely, we develop an algorithm that is closely related to the online EM algorithm for latent data models proposed by Cappé and Moulines (2009). From a practical point of view, a key feature of the adaptive SMC approach described in Section 3.2 is that it aims at decreasing the KLD rather than obtaining a fully converged parameter estimate: very few EM iterations are therefore needed for obtaining a significant computational gain. This will be illustrated further in the simulation study of Section 6.
5.1 Main assumptions and definitions
In order to describe clearly the algorithm, we limit ourselves initially to the case of constant mixture weights, i.e. we let $\alpha_j(\mathbf{x}, \underline{\beta}) := \beta_j$ , $1 \leq j \leq d$ , for a set $\underline{\beta} = {\beta_j}_{j=1}^d$ of nonnegative scalars summing into unity. The more general choice (4.5) will be treated in Section 5.3.
We now augment the pair $(I, \tilde{\mathbf{X}})$ with the index $J$ of the mixture component as well as the auxiliary, $\Upsilon$ -valued variable $\mathbf{U}$ of the curved exponential family. More specifically, we introduce an extended random variable $(I, J, \tilde{\mathbf{X}}, \mathbf{U})$ having mixed-type distribution with probability function
on the product space ${1, \dots, N} \times {1, \dots, d} \times \tilde{\Xi} \times \Upsilon$ . It is easily checked that $\pi_\theta$ is the marginal of $\bar{\pi}_\theta$ in $I$ and $\tilde{\mathbf{X}}$ . The following assumption is critical for convenient estimation. Although surprising at first glance, it simply requires that the state space extension used for handling integrated curved exponential distributions can be easily reverted for the sufficient statistics: the expectation of the sufficient statistics under the conditional distribution of $\mathbf{U}$ given $\mathbf{X}_I = \mathbf{x}$ and $\tilde{\mathbf{X}} = \tilde{\mathbf{x}}$ , having density
should be available in closed form. Since the sufficient statistics of exponential families often have a simple form, this assumption ismost often satisfied; e.g. it is satisfied for the Student's $t$ -distribution in Example 5.2.
Assumption 5.1 For all $\boldsymbol{\eta}$ and $(\mathbf{x}, \tilde{\mathbf{x}}) \in \Xi \times \tilde{\Xi}$ , the expectation
is available in closed form.
Under Assumption 5.1, let for $j \in {1, \dots, d}$ ,
where we have defined the responsibilities
Note that the quantities defined in (5.4) depend implicitly on the ancestor sample. In addition, we collect these statistics in the matrix
and the vector
We impose the following additional assumption.
Assumption 5.2 There are subsets $\Pi \subseteq \mathbb{R}_+^d$ and $\Sigma \subseteq \mathbb{R}^{s \times s'}$ such that for all ${\mathbf{s}j}{j=1}^d \in \Sigma^d$ and $\mathbf{p} \in \Pi$ , the mapping
has unique global maximum denoted by $\bar{\boldsymbol{\theta}}(\mathbf{s}, \mathbf{p})$ . In particular,
Remark 5.1 (Decoupling, optimal strata weights) It should be noted that due to the additive form of $l$ , the optimisation problem (5.9) can be split into the two separate subproblems
corresponding to maximisation over mixture weights and strata parameters, respectively. In the framework considered so far, where the mixture weights are assumed to be constant, the first of these problems (1) has, for all $\mathbf{p} \in \mathbb{R}_+^d$ the solution
for all $j$ .
In the following two examples we specify the solution to the second problem (2) in Remark 5.1 for the two—fundamental—special cases of multivariate Gaussian distributions and multivariate Student's $t$ -distributions; it should however be kept in mind that the method is valid for any integrated curved exponential distribution satisfying Assumption 5.1.
Example 5.1 (Updating formulas for Gaussian strata) In a first example we let each stratum be the density of a linear Gaussian regression parameterised by $\boldsymbol{\eta} = (\boldsymbol{\mu}, \boldsymbol{\Sigma})$ . The original hierarchical mixture of experts was based on plain Gaussian distributions and a possibly deep hierarchy of experts, incurring a computational overhead possibly larger than affordable for an algorithm used for adaptation. Dropping the hierarchical approach is a possible way to reduce this overhead at the price of reduced flexibility. Here we compensate somewhat for this loss by using a linear regression within each expert. Thus, given $J = j$ , the new particle location $\tilde{\mathbf{X}}$ has $p'$ -dimensional Gaussian distribution with mean $\boldsymbol{\mu}_j \bar{\mathbf{X}}_I$ , where the regression matrix $\boldsymbol{\mu}_j$ is of size $p' \times (p + 1)$ , and symmetric, positive definite covariance matrix $\boldsymbol{\Sigma}_j$ of size$p' \times p'$ . In this case, the sufficient statistics are $\mathbf{s}(\mathbf{x}, \tilde{\mathbf{x}}, \mathbf{u}) = (\tilde{\mathbf{x}}\tilde{\mathbf{x}}^\top) \oplus (\overline{\mathbf{x}\mathbf{x}^\top}) \oplus (\tilde{\mathbf{x}}\bar{\mathbf{x}}^\top)$ , leading to the expected responsibilities
and the expected sufficient statistics $\bar{\mathbf{s}}j(\boldsymbol{\theta}) := \bar{\mathbf{s}}{j,1}(\boldsymbol{\theta}) \oplus \bar{\mathbf{s}}{j,2}(\boldsymbol{\theta}) \oplus \bar{\mathbf{s}}{j,3}(\boldsymbol{\theta})$ with
In addition, the functions $\mathbf{a}$ and $\mathbf{b}$ in (4.6) are in this case given by
Consequently, the argument $\boldsymbol{\eta}_j$ that maximizes $l(\mathbf{s}, \mathbf{p}; \boldsymbol{\theta})$ for a given sufficient statistics $\mathbf{s}_j$ is given by
Remark 5.2 (Pooling the covariances) In practice, a robustified version of the algorithm above can be obtained by pooling the covariances (see e.g. Rayens and Greene, 1991). This means that a common covariance matrix is used for all the components of the mixture, i.e. $\boldsymbol{\Sigma}_j = \boldsymbol{\Sigma}$ for all $j$ . By doing this, the well-known problem of mixture models with strata components degenerating to Dirac masses is avoided. It is straightforward to enforce such a restriction in the optimisation procedure above, leading to
for all $j$ .
Example 5.2 (Updating formulas for Student's t-distribution-based strata)
A common grip in importance sampling (see, for instance, Oh and Berger, 1993) is to replace, in order to allow for more efficient exploration of the state space, Gaussian distributions by Student's t-distributions. Therefore,
as a more robust alternative to the approach taken in Example 5.1, one may use instead a $p'$ -dimensional Student's t-distribution with $\nu$ degrees of freedom, i.e.
Remark 5.3 (Fixing the degrees of freedom) The number $\nu$ of degrees of freedom of the Student's t-distributions is fixed, typically to $\nu \in {3, 4}$ , beforehand and is common to all strata. A similar choice has been made by, among others, Peel and McLachlan (2000, Section 7) and Cappé et al. (2008) as it allows for closed form optimisation in the M-step.
The choice (5.12) can be cast into the framework of Section 4, with $\Upsilon = \mathbb{R}$ , thanks to the Gaussian-Gamma decomposition of multivariate Student's t-distributions used by Liu and Rubin (1995, Section 2) and Peel and McLachlan (2000, Section 3):
where $\gamma(u; a, b) := b^a u^{a-1} \exp(-bu)/\Gamma(a)$ is the density of the Gamma distribution with shape parameter $a$ and scale parameter $b$ . Hence, the multivariate Student's t-distribution is an integrated curved exponential distribution (4.6) with $\gamma(u) = \gamma(u; \nu/2, \nu/2)$ , $h(\mathbf{x}, \tilde{\mathbf{x}}, u) = (2\pi)^{-p/2}$ , sufficient statistics $\mathbf{s}(\mathbf{x}, \tilde{\mathbf{x}}, u) = (u\tilde{\mathbf{x}}\tilde{\mathbf{x}}^\top) \oplus (u\overline{\mathbf{x}\mathbf{x}^\top}) \oplus (u\tilde{\mathbf{x}}\bar{\mathbf{x}}^\top)$ , and the same $\mathbf{a}$ and $\mathbf{b}$ as in the Gaussian case (5.10). Since the Gamma distribution is conjugate prior with respect to precision for the Gaussian distribution with known mean, the expectation $\eta$ of $U$ under $\bar{\rho}(u|\mathbf{x}, \tilde{\mathbf{x}}; \boldsymbol{\eta})$ can in this case be expressed as
where $\delta(\tilde{\mathbf{x}}, \boldsymbol{\mu} \bar{\mathbf{x}}, \boldsymbol{\Sigma}) = (\tilde{\mathbf{x}} - \boldsymbol{\mu} \bar{\mathbf{x}})^\top \boldsymbol{\Sigma}^{-1} (\tilde{\mathbf{x}} - \boldsymbol{\mu} \bar{\mathbf{x}})$ is the Mahalanobis distance with covariance matrix $\boldsymbol{\Sigma}$ . This leads to the expected responsibilities
and the expected sufficient statistics $\bar{\mathbf{s}}j(\boldsymbol{\theta}) =$$\bar{s}{j,1}(\boldsymbol{\theta}) \oplus \bar{s}{j,2}(\boldsymbol{\theta}) \oplus \bar{s}{j,3}(\boldsymbol{\theta})$ with
Since the functions $\mathbf{a}$ and $\mathbf{b}$ are unchanged from the Gaussian case, the same equations (5.11) can be applied for updating the parameter $\boldsymbol{\eta}$ based on the expected sufficient statistics (5.14). Moreover, covariance pooling can be achieved in exactly same way as in Remark 5.2.
5.2 Main algorithm
Under the assumptions above we define the mean field
We now have the following result (proved in Section A), which serves as a basis for the method proposed in the following.
Proposition 5.1 Under Assumptions 4.1–5.2, the following holds true. If $(\mathbf{s}^, \mathbf{p}^*)$ is a root of the mean field $\mathbf{h}$ , defined in (5.15), then $\bar{\boldsymbol{\theta}}(\mathbf{s}^*, \mathbf{p}^*)$ is a stationary point of $\boldsymbol{\theta} \mapsto \nabla_{\boldsymbol{\theta}} d_{\text{KL}}(\bar{\mu} || \pi_{\boldsymbol{\theta}})$ . Conversely, if $\boldsymbol{\theta}^*$ is a stationary point of the latter mapping, then $(\mathbf{s}^*, \mathbf{p}^*)$ , with*
is a root of $\mathbf{h}$ .
From Proposition 5.1 it is clear that the optimal parameter $\boldsymbol{\theta}^*$ solving the optimisation problem (3.10) over the parameter space $\Theta$ can, under the assumptions stated above, be obtained by solving the equation $\mathbf{h}(\mathbf{s}, \mathbf{p}) = \mathbf{0}$ in the space $\boldsymbol{\Sigma}^d \times \boldsymbol{\Pi}$ of sufficient statistics. Nevertheless, solving the latter equation obstructed by the fact that the expected sufficient statistics (5.4), and consequently the mean field itself, cannot be expressed on closed form (as we do not know the target distribution $\bar{\mu}$ ). Thus, following the online EM approach taken by Cappé and Moulines (2009), we aim at finding this root using stochastic approximation. More specifically, we apply the classical Robbins-Monro procedure
where ${\lambda_{\ell}}_{\ell \geq 1}$ is a decreasing sequence of positive step sizes such that
and $\tilde{\mathbf{h}}(\hat{\mathbf{s}}{\ell}, \hat{\mathbf{p}}{\ell})$ is a noisy observation of $\mathbf{h}(\hat{\mathbf{s}}{\ell}, \hat{\mathbf{p}}{\ell})$ . We form these noisy observations by means importance sampling in the following way: at iteration $\ell + 1$ (i.e. when computing $\hat{\mathbf{s}}{\ell+1}$ and $\hat{\mathbf{p}}{\ell+1}$ given $\hat{\mathbf{s}}{\ell}$ and $\hat{\mathbf{p}}{\ell}$ ) we estimate
by, first, setting $\boldsymbol{\theta}{\ell} := \bar{\boldsymbol{\theta}}(\hat{\mathbf{s}}{\ell}, \hat{\mathbf{p}}{\ell})$ , second, estimating $\bar{\mathbf{s}}(\boldsymbol{\theta}{\ell})$ and $\bar{\mathbf{p}}(\boldsymbol{\theta}{\ell})$ by drawing particles and indices ${(I_i^{\ell}, \tilde{\mathbf{X}}i^{\ell})}{i=1}^{N{\ell}}$ from $\pi_{\boldsymbol{\theta}{\ell}}$ (the currently best fitted proposal distribution), computing the associated weights ${\tilde{\omega}i^{\ell}}{i=1}^{N{\ell}}$ , and forming the importance sampling estimates
Note that we let the Monte Carlo sample size $N_{\ell}$ , which can be relatively small compared to the sample size $N$ of the particle filter, increase with the iteration index $\ell$ . As before, we collect the statistics defined in (5.19) in a matrix
and a vector
Note that the importance sampling estimates defined in (5.19) lack normalisation $\sum_{i=1}^{N_{\ell}} \tilde{\omega}_i^{\ell}$ . The role of the latter is to estimate the normalising constant $\iint \mu(\mathbf{x}) l(\mathbf{x}, \tilde{\mathbf{x}}) d\mathbf{x} d\tilde{\mathbf{x}}$ of the target distribution defined in (1.1), as it holds that
as $N \rightarrow \infty$ (see e.g. Douc et al., 2008). Here $\xrightarrow{\mathbb{P}}$ denotes convergence in probability. However, as $N_{\ell}$ is supposed to be considerably smaller than $N$ , which is necessary for obtaining a computationally efficient algorithm, thisestimator suffers from large variance in our case. Thus, in order to robustify the estimator we combine the Robbins-Monro procedure (5.17) with a similar procedure
for the normalising constant. The recursion (5.22) is typically initialised with
Obviously, (5.22) does not guarantee that the normalised weights ${\tilde{\omega}i^\ell / \hat{c}{\ell+1} N_\ell}_{i=1}^{N_\ell}$ sum to unity, and our approach thus leads to approximation of a probability measure with a finite measure. However, this does not impede the convergence of the stochastic approximation algorithm. Needless to say, this does not effect the properties of the particle approximation as the weights of the final particle sample obtained using the adapted kernel are renormalised in the usual manner and thus sum to unity. The final weighted empirical measure associated with the particle cloud is therefore still a probability distribution.
So, the Robbins-Monro recursion (5.17) is executed for, say, $L$ iterations before every updating step of the SMC algorithm or only at time steps decided by the user, and the final parameter fit $\theta_L$ defines the instrumental distribution $\pi_{\theta^*}$ used for propagating the particles. Before formulating the final algorithm, working in practice, the following assumption is needed to formulate a practically useful algorithm.
Assumption 5.3 There is a set $\Gamma \subseteq \mathbb{R}+$ such that for all $N \geq 1$ , ${\tilde{\mathbf{X}}i}{i=1}^N \in \tilde{\Xi}^N$ , ${I_i}{i=1}^N \in {1, \dots, N}^N$ , $\theta \in \Theta$ , $c \in \Gamma$ , $\mathbf{s} \in \Sigma^d$ , $\mathbf{p} \in \Pi$ , and $\lambda \in (0, 1)$ , it holds that
where the quantities $\tilde{\mathbf{s}}(\theta)$ and $\tilde{\mathbf{p}}(\theta)$ are defined through (5.20) and (5.21), respectively.
Under Assumption 5.3 our main algorithm, combining the two procedures (5.17) and (5.22), is well defined and goes as follows.
Algorithm 1
Require: ${(\mathbf{X}i, \omega_i)}{i=1}^N$ , $\theta_0$ , $\hat{c}_0$ , $\hat{\mathbf{s}}_0$ , $\hat{\mathbf{p}}_0$
1: for $\ell = 0 \rightarrow L - 1$ do
2: for $i = 1 \rightarrow N_\ell$ do
3: draw $(I_i^\ell, \tilde{\mathbf{X}}i^\ell) \sim \pi{\theta_\ell}$
4: set $\tilde{\omega}i^\ell \leftarrow w{\theta_\ell}(I_i^\ell, \tilde{\mathbf{X}}_i^\ell)$
5: end for
6: compute $\tilde{\mathbf{s}}(\theta_\ell)$ and $\tilde{\mathbf{p}}(\theta_\ell)$ as in (5.19)
7: set
8: set
9: set $\theta_{\ell+1} \leftarrow \bar{\theta}(\hat{\mathbf{s}}{\ell+1}, \hat{\mathbf{p}}{\ell+1})$
10: end for
5.3 Algorithm formulation for logistic weights
We now extend the case of constant mixture weights to the considerably more complicated case of logistic weights (4.5). In this case the gradient $\nabla_{\theta} d_{\text{KL}}(\bar{\mu} \parallel \pi_{\theta})$ , the crucial quantity in the proof of Proposition 5.1 (given in Section A), is given by
where
with responsibilities $\bar{\pi}_{\theta}(j|i, \tilde{\mathbf{x}})$ given by, in this case,
and the vectors $\bar{\mathbf{p}}(\theta)$ and $\bar{\mathbf{s}}(\theta)$ being defined as in (5.7). However, since we are no longer within the framework of exponential families, we will not be able to find a closed-form zeroof $\nabla_{\underline{\theta}} \kappa(\underline{\theta})$ (e.g. a closed-form global maximum of $\kappa(\underline{\theta})$ ) when $\bar{\mathbf{p}}(\underline{\theta})$ and $\bar{\mathbf{s}}(\underline{\theta})$ are replaced by stochastic approximation iterates $\hat{\mathbf{p}}_\ell$ and $\hat{\mathbf{s}}_\ell$ , respectively. On the other hand, for all $\underline{\theta}_\ell \in \Theta$ , the mapping
is concave (this stems directly from the multinomial logistic regression by computing conditional expectations) and can thus be approximated well by a second degree polynomial (in $\underline{\beta}$ ). More specifically, consider the gradient $\nabla_{\underline{\beta}} \bar{\kappa}(\underline{\beta}; \underline{\theta}_\ell)$ , whose $j$ th block ( $j \in {1, \dots, d-1}$ ) is given by
and the Hessian $\nabla_{\underline{\beta}} \nabla_{\underline{\beta}} \bar{\kappa}(\underline{\beta}; \underline{\theta}_\ell)$ , having the matrix
as block $(j, j') \in {1, \dots, d-1}^2$ . Note that the Hessian above does not depend on $\underline{\theta}_\ell$ .
Using these definitions, the function $\bar{\kappa}(\underline{\beta}; \underline{\theta}_\ell)$ can be approximated by means of a second order Taylor expansion around the previous parameter estimate $\underline{\beta}^\ell$ according to
In the light of (5.26), we approximate the gradient and Hessian quantities (which again lack closed-form expressions due to the expectation over $\bar{\mu}$ ) using stochastic approximation. For this purpose, we set
and define
and find, using again the Robbins-Monro procedure, a root to the extended mean field
In this case, the mapping $\bar{\underline{\theta}}$ updates the strata parameters in analogy with the case of constant mixture weights (Section 5.1) while the $\underline{\beta}$ parameter gets updated according to the Newton-Raphson-type formula
The full procedure is summarised below, where we have defined $\tilde{\mathbf{p}}(\underline{\theta}_\ell)$ as in (5.21) and
Algorithm 2
Require: ${(\mathbf{X}i, \omega_i)}{i=1}^N, \underline{\theta}_0, \hat{c}_0, \hat{s}_0, \hat{\mathbf{p}}_0, \hat{\mathbf{t}}_0, \hat{\mathbf{v}}_0$
1: for $\ell = 0 \rightarrow L-1$ do
2: for $i = 1 \rightarrow N_\ell$ do
3: draw $(I_i^\ell, \tilde{\mathbf{X}}i^\ell) \sim \pi{\underline{\theta}_\ell}$
4: set $\tilde{\omega}i^\ell \leftarrow w{\underline{\theta}_\ell}(I_i^\ell, \tilde{\mathbf{X}}_i^\ell)$
5: end for
6: compute $\tilde{\mathbf{s}}(\underline{\theta}_\ell)$ and $\tilde{\mathbf{p}}(\underline{\theta}_\ell)$ through (5.19)
7: compute $\tilde{\mathbf{t}}(\underline{\theta}_\ell)$ and $\tilde{\mathbf{v}}(\underline{\theta}_\ell)$ through (5.28)
8: set
$$\hat{c}{\ell+1} \leftarrow (1 - \lambda{\ell+1}) \hat{c}\ell + \frac{\lambda{\ell+1}}{N_\ell} \sum_{i=1}^{N_\ell} \tilde{\omega}_i^\ell$$
9: set
$$\hat{\mathbf{s}}{\ell+1} \oplus \hat{\mathbf{p}}{\ell+1} \oplus \hat{\mathbf{t}}{\ell+1} \oplus \hat{\mathbf{v}}{\ell+1} \ \leftarrow (1 - \lambda_{\ell+1}) \hat{\mathbf{s}}_\ell \oplus \hat{\mathbf{p}}_\ell \oplus \hat{\mathbf{t}}\ell \oplus \hat{\mathbf{v}}\ell \ + \frac{\lambda{\ell+1}}{\hat{c}{\ell+1} N_\ell} \tilde{\mathbf{s}}(\underline{\theta}_\ell) \oplus \tilde{\mathbf{p}}(\underline{\theta}_\ell) \oplus \tilde{\mathbf{t}}(\underline{\theta}_\ell) \oplus \tilde{\mathbf{v}}(\underline{\theta}_\ell)$$```
10: set $\theta_{\ell+1} \leftarrow \bar{\theta}(\hat{s}{\ell+1}, \hat{p}{\ell+1})$ $\triangleright$ Here $\underline{\beta}$ is updated according to $\underline{\beta}^{\ell+1} :=$ $\underline{\beta}^{\ell} - \hat{v}{\ell+1}^{-1} \hat{t}{\ell+1}$ . 11: end for
## 5.4 Some practical guidelines to choose the algorithm's parameters
The algorithm outlined above involves parameters, such as the number $d$ of mixture components and Monte Carlo sample sizes $\{N_{\ell}\}_{\ell=0}^{L-1}$ , that need to be set. Fortunately, these parameters are easy to tune. Moreover, we may expect a significant improvement of the SMC proposal distribution also with a suboptimal choice of the same. Without constituting a definite answer on how to tune the algorithmic parameters, the following guidelines emerge from the simulations in Section 6.
- • Concerning the number $d$ of experts, a too small number of strata averts good approximation of strongly nonlinear kernels while a too large $d$ leads to computational overhead. For instance, in the simulations in Section 6 we use up to $d = 8$ strata to approximate well a kernel with a ring-shaped, 2-dimensional support.
- • The initial values of the mixture parameters could be chosen based on prior knowledge of the model by, e.g., fitting, in the case of optimal filtering with informative or non-informative observations, the local likelihood or the prior kernel, respectively. In our simulations, we initialise the algorithm with uniform logistic weights (i.e. with $\underline{\beta}$ being a null matrix), regressions being null except for the intercept (implying independence of the ancestors), and covariances making the strata widely overspread on the support of the target.
- • For simplicity, the importance sampling sizes can be taken constant $N_0 = N_1 = \dots = N_{L-1}$ over all iterations. This is well in line with the theory of stochastic approximation (see e.g. Duflo, 1997; Benveniste et al., 1990), which in general guarantees, if the step size sequence is chosen correctly, convergence of the algorithm as long as the variance of the observation noise is not increasing with the iteration index. Moreover, the importance sampling size can be chosen relatively to the computational budget, by using, say, a proportion $\alpha = 20\%$ or
$\alpha = 50\%$ of the total number $N$ of particles that would be used for a plain SMC algorithm for the adaptation step. More specifically, this is achieved by using $N_{\ell} = \alpha N / L$ samples per iteration in the adaptation step, and propagating $(1 - \alpha)N$ particles through the SMC updating step (based on the adapted proposal). In order assure reasonably fast convergence, the proportion $\alpha$ should be large enough to assure a decent sampling size $N_{\ell}$ , typically not less than a hundred particles. It can be helpful to let the sample size $N_0$ be twice or three times larger than the size used at a typical iteration, to recover from a potentially bad initial choice.
- • The number $L$ of iterations can typically be very small compared to typical stochastic approximation EM runs: as mentioned at the beginning of Section 5, the main gain in KLD typically occurs in the first dozen iterations.
- • The least precise guideline concerns the step sizes $\{\lambda_{\ell}\}_{\ell=0}^{L-1}$ . Since we not search an exact optimum but rather aim at fast convergence towards the region of interest, we use a slowly decreasing step size. A rate matching this could be, say, letting $\lambda_{\ell} = \ell^{-0.6}$ (which was used by e.g. Cappé et al., 2005, Section 11.1.5, within the framework of stochastic approximation EM methods), which satisfies (5.18). However, in our simulations we use, for simplicity, a constant step size. This is well motivated by the fact that the adaptation algorithm is run for only a few iterations.
## 6 Applications to nonlinear SSMs
As mentioned, SMC methods can be successfully applied to optimal filtering in SSMs. An SSM is a bivariate process $\{(\mathbf{X}_k, \mathbf{Y}_k)\}_{k \geq 0}$ , where $\mathbf{X} := \{\mathbf{X}_k\}_{k \geq 0}$ is an unobserved Markov chain on some state space $\mathbf{X} \subseteq \mathbb{R}^x$ and $\{\mathbf{Y}_k\}_{k \geq 0}$ is an observation process taking values in some space $\mathbf{Y} \subseteq \mathbb{R}^y$ . We denote by $Q$ and $\pi_0$ the transition density (in the following sometimes referred to as the *prior kernel*) and initial distribution of $\mathbf{X}$ , respectively. Conditionally on $\mathbf{X}$ , the observations are assumed to be conditionally independent with the conditional distribution $G$ of a particular $\mathbf{Y}_k$ depending on the corresponding $\mathbf{X}_k$ only. We will assume that $Q$ and $G$ have densities $q$and $g$ , respectively, with respect to Lebesgue measure, i.e.
and
For simplicity, we assume that the Markov chain $\mathbf{X}$ is time-homogeneous and that the distribution $G$ does not depend on $k$ ; however, all developments that follow may be extended straightforwardly to the time-inhomogeneous case. The optimal filtering problem consists of computing, given a fixed record $\mathbf{Y}_{0:n} := (\mathbf{Y}_0, \dots, \mathbf{Y}_n)$ of observations, the *filter distributions*
for $k = 0, 1, \dots, n$ . As the sequence of filter distributions satisfies the nonlinear recursion
the optimal filtering problem can be perfectly cast into the sequential importance sampling framework (1.1) with $\Xi = \tilde{\Xi} \equiv \mathbf{X}$ , $\mu \equiv \phi_k$ , $\tilde{\mu} \equiv \phi_{k+1}$ , and $l(\mathbf{x}, \tilde{\mathbf{x}}) \equiv g(\tilde{\mathbf{x}}, \mathbf{Y}_{k+1}) q(\mathbf{x}, \tilde{\mathbf{x}})$ . Consequently, a particle sample $\{(\mathbf{X}_i, \omega_i)\}_{i=1}^N$ (here and in the following we have omitted, for brevity, the time index from the particles and the associated weights) approximating $\phi_k$ can be transformed into a sample $\{(\tilde{\mathbf{X}}_i, \tilde{\omega}_i)\}_{i=1}^N$ approximating the filter $\phi_{k+1}$ at the subsequent time step by executing the following steps.
---
### Algorithm 3
**Require:** $\{(\mathbf{X}_i, \omega_i)\}_{i=1}^N$
1: **for** $i = 1 \rightarrow N$ **do**
2: draw $I_i \sim \{a_\ell \omega_\ell\}_{\ell=1}^N$
3: draw $\tilde{\mathbf{X}}_i \sim r(\mathbf{X}_{I_i}, \cdot)$
4: set $\tilde{\omega}_i \leftarrow \frac{g(\tilde{\mathbf{X}}_i, \mathbf{Y}_{k+1}) q(\mathbf{X}_{I_i}, \tilde{\mathbf{X}}_i)}{a_{I_i} r(\mathbf{X}_{I_i}, \tilde{\mathbf{X}}_i)}$
5: **end for**
6: **return** $\{(\tilde{\mathbf{X}}_i, \tilde{\omega}_i)\}_{i=1}^N$
---
Here $\{a_i\}_{i=1}^N$ is a set of nonnegative adjustment multipliers and the proposal $r$ is a Markov transition density. In the following examples we illustrate how the proposal $r$
can be designed adaptively using Algorithm 2 described in the previous section. In Sections 6.1 and 6.2 we adapt mixtures of Gaussian strata (following Example 5.1) only, while Section 6.3 provides also a comparison with Student's $t$ -distributed strata.
## 6.1 Multivariate linear Gaussian model
We start by considering a simple multivariate linear Gaussian model. In this toy example, the optimal kernel $l^*$ and the optimal adjustment weight function $a^*$ (defined in (3.2) and (3.3), respectively) are available in closed form; here we use the term “optimal”—which is standard in the literature—as a proposal distribution $\pi$ based these quantities minimise the KLD (3.7) (which of course then vanishes as $\pi = \bar{\mu}$ in that case). This makes it possible to compare our algorithm to an exact reference. The optimal kernel does not belong to our family of mixture of experts. In the model under consideration:
- • Each $\mathbf{Y}_k$ , taking values in $\mathbf{Y} = \mathbb{R}^2$ , is a noisy observation of the corresponding hidden state $\mathbf{X}_k$ with local likelihood $g(\tilde{\mathbf{x}}, \mathbf{y}) = \mathcal{N}_2(\mathbf{y}; \tilde{\mathbf{x}}, \Sigma_{\mathbf{Y}})$ , where $\Sigma_{\mathbf{Y}} = 0.1 \times \mathbf{I}_2$ .
- • The prior kernel density is a mixture of multivariate Gaussian distributions $q(\mathbf{x}, \tilde{\mathbf{x}}) = (\mathcal{N}_2(\tilde{\mathbf{x}}; \Lambda_1 \bar{\mathbf{x}}, \Sigma) + \mathcal{N}_2(\tilde{\mathbf{x}}; \Lambda_2 \bar{\mathbf{x}}, \Sigma))/2$ with regression matrices
- • The filter $\phi_k$ at time $k$ is assumed to be a bimodal distribution $\mathbf{X} = \mathbb{R}^2$ with density $(\mathcal{N}_2(\mathbf{x}; (0, 1)^\top, \Sigma) + \mathcal{N}_2(\mathbf{x}; (0, -1)^\top, \Sigma))/2$ , where $\Sigma = 0.1 \times \mathbf{I}_2$ . The two modes are hence well separated.
In this first example, we consider *one single updating step of the particle filter*, with initial particles $\{\mathbf{X}_i\}_{i=1}^N$ sampled exactly from the filter $\phi_k$ . For this single step we study the performance of Algorithm 2. The observation $\mathbf{Y}_{k+1}$ is chosen to be $\mathbf{Y}_{k+1} = (1, 0)^\top = \Lambda_2(0, 1)^\top = \Lambda_1(0, -1)^\top$ . The likelihood of $\mathbf{X}_k$ given $\mathbf{Y}_{k+1}$ is thus symmetric around the $x$ -axis, giving equal weight to each of the two components of $\phi_k$ . Even though the optimal kernel $l^*$ does not belong to our family of experts, it is still a mixture whose weights are highly dependent of location of the ancestor, and we can expect our algorithm to adapt accordingly.The histogram in Figure 1a of the importance weights of the particle swarm produced using the prior kernel shows that 50% of these weights are nearly equal to zero, and the remaining ones are spread over a large range of values. This is confirmed by looking at the corresponding curve of proportions in Figure 1b, showing that 80% of the total mass is carried by only 25% of the particles and 99% of the total mass by 40% of the particles.
(a) Histogram of the importance weights obtained using the prior kernel and the final adapted kernel in the linear Gaussian model. Weights are renormalised and multiplied by the size of the sample. In the fully adapted case, all weights would be equal to unity.
(b) *Curve of proportions*: proportion of particles, sorted by decreasing order of importance weight, against proportion of the total mass, for the prior kernel and for different numbers of iterations of the adaptation algorithm.
Figure 1: Evolution of the distribution of the importance weights before and after adaptation.
One single iteration of the adaptation scheme described in Section 5.3 is sufficient to improve those proportions to 80% of the mass for 40% of the particles and 99% of the mass for 55% of the particles; see the correspond-
ing curve of proportions in Figure 1b. After 10 such iterations, the corresponding curve of proportions (again in Figure 1b) shows that close to maximum improvement has been reached: the first few steps of the optimisation are enough to bring significant improvement. The histogram in Figure 1a of the weights obtained at the final iteration shows how the weights are concentrated around unity, the value that corresponds to sampling with the optimal kernel. We display in Figure 2a the KLD (3.8) between the fit and the target (estimated by means of a Monte Carlo approximation with a large sample size), and the same KLD for the proposal based on the prior kernel as well as the proposal based on the optimal kernel with uniform adjustment weights—i.e. all ancestor particles have practically the same optimal adjustment weight: choosing $a \equiv \mathbb{1}_X$ makes the second term in (3.8) negligible. As mentioned earlier, the KLD decreases very fast, and most of the improvement is obtained after only a few iterations. Figure 2b compares the evolution of the KLD for several step sizes covering four orders of magnitude. As the step size is increased, the algorithm converges faster, although less smoothly: this has no impact from a practical point of view as we are only looking for a good proposal kernel in an importance sampling setting, not for the exact optimal one.
## 6.2 Bessel process observed in noise
As a more challenging example we consider optimal filtering of the Brownian motion underlying a Bessel process observed in noise, also known as *range-only filtering* of a Gaussian random walk. We let the state process evolve in the plane for visualisation purposes. More specifically, the SSM is given by
where $\{\mathbf{V}_k\}_{k \geq 0}$ and $\{W_k\}_{k \geq 0}$ are sequences of mutually independent $\mathcal{N}_2(\mathbf{0}, \Sigma_{\mathbf{X}})$ -distributed and $\mathcal{N}(0, \sigma_Y^2)$ -distributed, respectively, noise variables and $\|\mathbf{X}\|_2$ is the Euclidean $\mathcal{L}_2$ norm on the state space $\mathbf{X} := \mathbb{R}^2$ . Note that the observations $\{Y_k\}_{k \geq 0}$ are real-valued in this case. With the notation in (6.1) and (6.2), it holds that $q(\mathbf{x}, \tilde{\mathbf{x}}) := \mathcal{N}_2(\tilde{\mathbf{x}}; \mathbf{x}, \Sigma_{\mathbf{X}})$ the density of the prior kernel and by $g(\tilde{\mathbf{x}}, y) := \mathcal{N}(y, \|\tilde{\mathbf{x}}\|_2, \sigma_Y^2)$ the local likelihood. We ensure a diffuse prior kernel and informative observations by setting $\Sigma_{\mathbf{X}} = \mathbf{I}_2$ and $\sigma_Y^2 = 0.01$ .(a) Evolution of the KLD for the step size $0.1/\sqrt{L}$ .
(b) Comparison of evolution of KLDs for step sizes $\{0.001, 0.01, 0.1, 1, 10, 15\}/\sqrt{L}$ .
Figure 2: Evolution of the KLD over $L = 20$ iterations of adaptation for the linear Gaussian model.
Figure 3: Comparison of 1,000 and 200 particles from the prior and final adapted kernels, respectively, with the 10% largest weights plotted in red, along with the corresponding ancestors and the distribution of the corresponding importance weights. For the prior kernel, the support of the proposal distribution is over-spread and as much as 80% of particles have null weight. On the other hand, the adaptation algorithm concentrates the particles on the support of the target distribution, thus narrowing the span of the importance weights and balancing the contributions of all particles.Figure 4: Evolution of the optimal and adapted kernels for the Bessel model for 3 distinct ancestors in different regions and for the observation $Y_{k+1} = 1.0$ . The last three rows show the evolution of the adapted proposal kernel after the initial fit, one iteration, and 30 iterations, respectively. The adaptation algorithm used a constant sample size $N_\ell = 200$ for all $\ell \in \{0, 2, \dots, 30\}$ . After 30 iterations (the last row), the adapted kernel is visually very close to the optimal kernel.As the hidden state is observed range-only, the state equation provides most of the information concerning the bearing while the local likelihood is invariant under rotation of the state around the origin. This induces a variety of nonlinear shapes of the optimal kernel depending on the location of the ancestor—see the three top rows of Figure 4.
To start with, we consider again a *single step* of the particle filter, updating a weighted particle sample $\{(\mathbf{X}_i, \omega_i)\}_{i=1}^N$ approximating the filter $\phi_k$ at some time index $k$ to a weighted sample $\{(\tilde{\mathbf{X}}_i, \tilde{\omega}_i)\}_{i=1}^N$ approximating the filter $\phi_{k+1}$ at the next time step. We assume that $\phi_k = \mathcal{N}_2((0.7, 0.7)^\top, 0.5 \times \mathbf{I}_2)$ . In our experiment, the original weighted sample $\{(\mathbf{X}_i, \omega_i)\}_{i=1}^N$ consists of a sample of $N = 20,000$ i.i.d. particles drawn exactly from the filter distribution $\phi_k$ (hence, the particles have uniform weights). The observation at time $k + 1$ is $Y_{k+1} = 1.0$ . We initialise the algorithm with a sample $\{(\tilde{\mathbf{X}}_i, \tilde{\omega}_i)\}_{i=1}^{N_0}$ of size $N_0 = 1,000$ using the prior kernel. The resulting cloud, along with the 100 particles with largest weights and the corresponding ancestors, is plotted in Figure 3 (top row). The support of the proposal distribution is overspread: most of the particles have negligible weights, and only a few particles have comparatively large weights; indeed, 80% of particles have null weight. This is confirmed by the curve of proportions in Figure 5 (left): only 20% of the proposed particles carry the total mass. Adaptation of the proposal kernel is thus highly relevant. Here again, adaptation of the adjustment weights is not required, as the ancestors of the particles with highest importance weights are not located in any specific region of the state space.
Based on these 1,000 particles, the first iteration of the adaptation is carried through using conditional probabilities from the initial fit displayed in Figure 4 (fourth row), whose components are chosen to be independent of the ancestor, i.e. only the constant term is non-zero. The resulting kernel is plotted in Figure 4 (fifth row). We use it to propose 200 new particles, serving as an importance sampling approximation of $\bar{\mu}$ at the second iteration. After 30 such iterations, the adapted kernel visible in Figure 4 (last row) is visually very close to the optimal kernel. Note the impact of the location of the ancestor on the un-normalised transition kernel, whose mass shifts on a circle, and how the adapted kernel shifts accordingly. Figure 3 (bottom row) shows that the adaptation algo-
rithm concentrates the particles on the support of the target distribution, thus narrowing the span of the importance weights and balancing the contributions of all particles.
Figure 5: Improvement of the proposal via adaptation over 30 iterations. The plot of proportions (left) and the KLD (right) improve dramatically over the very first iterations. While the prior kernel puts 90% of the mass on 15% of the particles, adaptation increases it to 70% of the particles in one step and stabilises to 80% after a very few iterations.
Most importantly, a very small number of iterations suffices to achieve significantly more uniformly distributed importance weights, i.e. to significantly lower the KLD: Figure 5 (right) shows how the KLD drops after the first 2 iterations and then stabilises near null, while the curve of proportions in Figure 5 (left) shows that the distribution of the weights is essentially unchanged past the first few iterations.
As a final look at convergence, Figure 6 displays the evolution of all the estimated parameters over the 30 iterations, confirming that the fit stabilises after a few steps: the result of the very first couple of iterationscould serve as a more efficient proposal than the prior kernel. The parameters of the logistic weights $\beta^\ell$ are the slowest to stabilise due to the stochastic gradient used to palliate for the lack of closed-form update as mentioned in Section 5.3.
We now turn focus away from a single time step and run the adaptive particle filter in parallel with a plain bootstrap filter for 50 time steps with a simulated observation record $\mathbf{Y}_{0:50}$ as input. At each iteration we let the adaptive filter optimise the proposal kernel using Algorithm 2. The forgetting of the initial distribution of the random walk entails that the filter distribution flow converges to a distribution symmetric under rotation and supported on a ring centered at $\mathbf{0}$ . To connect with widespread measures of efficiency discussed in Cornebise et al. (2008), we compare in Figure 7 the negated Shannon entropy (lower is better) of the importance weights and relative effective sample size (in percentage of the total sample; higher is better) of the two filters. The negated Shannon entropy shows improvement by our adaptive filter, which is not surprising in the light of the convergence results of Cornebise et al. (2008, Theorems 1–2): it is a consistent estimate of the same KLD when the number of particles tends to infinity. In this example also the relative effective sample size also exhibits an improvement, even though it is a consistent estimate of the chi-square divergence between the same distributions rather than the KLD (see again Cornebise et al., 2008, Theorem 1–2, for an exact formulation of this result). Even though this makes sense to some extent, one should not expect this to hold systematically.
### 6.3 Multivariate tobit model
We now briefly illustrate how far adaptation of the proposal kernel leads—and where it stops. Consider a partially observed multivariate dynamic *tobit* model
where $\mathbf{X} = \mathbb{R}^2$ , $\mathbf{A} = 0.8 \times \mathbf{I}_2$ , and $\mathbf{B} = (1, 1)^\top \in \mathbb{R}^2$ , so that $Y_k$ takes values in $\mathbb{R}$ . Here $\{\mathbf{U}_k\}_{k \geq 1}$ and $\{V_k\}_{k \geq 0}$ are independent sequences of mutually independent and identically distributed Gaussian variables with variances $\Sigma_{\mathbf{U}} = 2 \times \mathbf{I}_2$ and $\sigma_v^2 = 0.1$ , respectively. The observation process consists of noisy observations of the sum of its components of the hidden states. In addition, the
observations are left-censored. We consider again a single update of the particle swarm for a given time step $k$ , where we have $N = 20,000$ ancestor particles distributed according to $\mathcal{N}_2((1, 1)^\top, 10 \times \mathbf{I}_2)$ and set $Y_{k+1} = 0$ . The local likelihood is hence null above the line $\Delta = \{\mathbf{x} \in \mathbb{R}^2 : \mathbf{B}^\top \mathbf{x} = 0\}$ and constant below with a narrow transition region, as displayed in Figure 8 (second row). The prior kernel displayed in Figure 8 (first row) can have most of its mass out of the high-likelihood regions, depending on the ancestor. Figure 8 (third row) illustrates the unnormalised optimal transition kernel for three reference ancestors, showing how the match between the supports of the prior kernel and the local likelihood varies depending on the position of the ancestor particle relatively to the line $\mathbf{A}^{-1}\Delta$ . Hence, half of the original particles have very small adjustment multiplier weights.
Adapting the proposal kernel will hence require at least two components, and will not lead to perfect adaptation, as can be seen on the fit after 10 iterations displayed in Figure 8 (last row). The ancestor $(1, 1)^\top$ is the center of the ancestor sample, and the un-normalised optimal kernel is the prior kernel truncated in its middle; $(-3, -1)^\top$ is the bottom left of the ancestor sample, and the un-normalised optimal kernel almost matches the prior, save for a truncation in the upper right tail; $(9, 5)^\top$ is the top right of the ancestor sample, and the un-normalised optimal kernel differs widely from the prior kernel, as only the very far tails of the latter match non-null local likelihood.
Finally, without getting into details, we make our point through Figure 9 showing again how the KLD drops in the first iteration for families of both Gaussian distributions and Student’s $t$ -distributions. We nevertheless purposefully keep the algorithm running for a large number of iterations, to illustrate the difference between the Gaussian and the Student’s $t$ parametrisations. The Gaussian parametrisation stabilises close to the minimum achieved by the optimal kernel with uniform adjustment weights, which is not negligible. In addition, the Student’s $t$ -distribution allows for heavier tails at the price of a higher attainable lower bound on the KLD.Figure 6: Parameters $\theta_\ell$ of the adapted kernel over 30 iterations of the algorithm: practically, convergence is achieved after a few steps only.
Figure 7: Evolution of the effective sample size and the entropy between target and proposal for the bootstrap filter and the adaptive filter. As intended, the adaptive filter leads to higher effective sample size and lower negated Shannon entropy.Figure 8: Evolution of the adapted kernel, for three ancestors in different regions for the tobit model. In this example the observation brings, thanks to the censorship, a lot of information concerning the location of the ancestor. Thus, adapting only the proposal kernel will not lead to perfect adaptation, as can be seen on the fit after 10 iterations. In this case it is thus highly relevant to consider adaptation also of the adjustment multiplier weights.Figure 9: Comparison of the evolution of the KLD for the Gaussian experts and the Student’s $t$ -distributed expert over $L = 5,000$ iterations of the algorithm.
Finally, Figure 9 illustrates a limit case for kernel-only adaptation: replacing the prior kernel by the optimal one reduces only the KLD, which is still far from zero, by half. In this example, chosen for that purpose, the observation brings, thanks to the censorship, a lot of information concerning the location of the ancestor. In this case it is thus highly relevant to consider adaptation also of the adjustment multiplier weights. Naturally, this could, as a logical next step in development of adaptive SMC algorithms, be done by designing a minimisation algorithm for the second term of (3.8), which we leave as an open problem.
## 7 Future work and Conclusion
Relying on the results of Cornebise et al. (2008), we have built new algorithms approximating the so-called optimal proposal kernel at a given time step of an auxiliary particle filter by means of minimisation of the KLD between the auxiliary target and instrumental distributions of the particle filter. More specifically, the algorithm fits a weighted mixture of integrated curved exponential distributions with logistic weights to the auxiliary target distribution by minimising the KLD between the two using a Monte Carlo version of the online EM method proposed by Cappé and Moulines (2009).
In addition, we have applied successfully this relatively simple algorithm to optimal filtering in SSMs; indeed, running the stochastic approximation-based adaptation procedure for only a few iterations at every time
step as the particles evolve leveled off significantly the distribution of the total weight mass among the particles also for models exhibiting very strong nonlinearity. Thus, adding the adaptation step to an existing particle filter implementation implies only limited computational demands.
A proof of convergence, combining the theory developed in Section 5 with existing theory of stochastic approximation with Markovian perturbations (treated by, e.g., Duflo, 1997; Benveniste et al., 1990) of the algorithm is currently in progress. In addition, we investigate at present the possibility of extending the approach to comprise adaptation also of the adjustment multipliers.
## Acknowledgement
We thank the anonymous referees for insightful comments that significantly improved the presentation of the paper.
## A Proof of Proposition 5.1
The proof follows the lines the proof of Cappé and Moulines (2009, Proposition 1). For vector-valued differentiable functions $h = (h_1, \dots, h_m)^\top$ from $\Theta$ to $\mathbb{R}^m$ we denote by $\nabla_\theta h^\top$ the $|\Theta| \times m$ matrix having the gradient $\nabla_\theta h_j$ as $j$ th column, this is, the inverse of the Jacobian of the same mapping.
Let $(\mathbf{s}^*, \mathbf{p}^*)$ be a zero of $\mathbf{h}$ and set $\boldsymbol{\theta}^* := \bar{\boldsymbol{\theta}}(\mathbf{s}^*, \mathbf{p}^*)$ . Since for all $\boldsymbol{\eta}$ and $\mathbf{s} \in \mathbb{R}^{s \times s'}$ ,
it holds, by definition,
On the other hand, by Fisher's identity,
and, by (A.1),
Consequently,
Thus, as
we obtain
Conversely, let $\boldsymbol{\theta}^*$ be a stationary point of $\boldsymbol{\theta} \mapsto d_{\text{KL}}(\bar{\mu} || \pi_{\boldsymbol{\theta}})$ (i.e. (A.4) holds) and let $(\mathbf{s}^*, \mathbf{p})$ be given by (5.16). Then we conclude, via (A.2) and (A.3), that $\boldsymbol{\theta}^*$ is a stationary point of $\boldsymbol{\theta} \mapsto l(\mathbf{s}^*, \mathbf{p}^*; \boldsymbol{\theta})$ as well. However, by Assumption 5.2, this point is unique and
equal to $\bar{\boldsymbol{\theta}}(\mathbf{s}^*, \mathbf{p}^*)$ ; thus,
which completes the proof.
## References
C. Andrieu and É. Moulines. On the ergodicity properties of some adaptive MCMC algorithms. *The Annals of Applied Probability*, 16:1462–1505, 2006.
C. Andrieu, A. Doucet, and R. Holenstein. Particle Markov chain Monte Carlo methods. *Journal of the Royal Statistical Society: Series B (Statistical Methodology)*, 72(3):269–342, 2010.
A. Benveniste, M. Métivier, and P. Priouret. *Adaptive Algorithms and Stochastic Approximations*, volume 22. Springer, Berlin, 1990. Translated from the French by Stephen S. S. Wilson.
O. Cappé and É. Moulines. On-line expectation–maximization algorithm for latent data models. *J. Roy. Statist. Soc. Ser. B*, 71(3):593–613, 2009.
O. Cappé, E. Moulines, and T. Rydén. *Inference in Hidden Markov Models*. Springer, 2005.
O. Cappé, R. Douc, A. Guillin, J. M. Marin, and C. P. Robert. Adaptive importance sampling in general mixture classes. *Statistics and Computing*, 18(4):447–459, 2008.
J. Cornebise. *Adaptive Sequential Monte Carlo Methods*. PhD thesis, Université Pierre et Marie Curie – Paris 6, June 2009.
J. Cornebise, E. Moulines, and J. Olsson. Adaptive methods for sequential importance sampling with application to state space models. *Statistics and Computing*, 18(4):461–480, 2008.
P. Del Moral, Arnaud Doucet, and Ajay Jasra. Sequential Monte Carlo samplers. *J. Roy. Statist. Soc. Ser. B*, 68(3):411, 2006.
P. Del Moral, Arnaud Doucet, and Ajay Jasra. On adaptive resampling strategies for sequential Monte Carlo methods. *Bernoulli*, 18(1):252–278, 2012.R. Douc, É. Moulines, and J. Olsson. Optimality of the auxiliary particle filter. *Probability and Mathematical Statistics*, 29(1): 1–28, 2008.
A. Doucet, S. Godsill, and C. Andrieu. On sequential Monte Carlo sampling methods for Bayesian filtering. *Statistics and Computing*, 10(3):197–208, 2000.
A. Doucet, N. De Freitas, and N. Gordon, editors. *Sequential Monte Carlo Methods in Practice*. Springer, New York, 2001.
M. Duflo. *Random Iterative Models*, volume 34. Springer, Berlin, 1997. Translated from the 1990 French original by S. S. Wilson and revised by the author.
N. Gordon, D. Salmond, and A. F. Smith. Novel approach to nonlinear/non-Gaussian Bayesian state estimation. *IEE Proc. F, Radar Signal Process.*, 140:107–113, 1993.
H. Haario, E. Saksman, and J. Tamminen. An adaptive Metropolis algorithm. *Bernoulli*, 7(2):223–242, 2001.
M. Jordan and R. Jacobs. Hierarchical mixtures of experts and the EM algorithm. *Neural computation*, 6:181–214, 1994.
M. Jordan and L. Xu. Convergence results for the EM approach to mixtures of experts architectures. *Neural Networks*, 8(9):1409–1431, 1995. ISSN 0893-6080.
I. Korotsil, G.W. Peters, J. Cornebise, and D. Regan. Adaptive Markov chain Monte Carlo forward simulation for statistical analysis in epidemic modeling of human papilloma virus. *To appear in Statistics in Medicine.*, 2012.
C. Liu and D.B. Rubin. ML estimation of the t distribution using EM and its extensions, ECM and ECME. *Statistica Sinica*, 5(1): 19–39, 1995.
J.S. Liu. *Monte Carlo Strategies in Scientific Computing*. Springer, New York, 2001.
G.J. McLachlan and T. Krishnan. *The EM algorithm and extensions, second edition*. Wiley New York, 2008.
M.-S. Oh and J. O. Berger. Integration of multimodal functions by Monte Carlo importance sampling. *J. Amer. Statist. Assoc.*, 88(422):450–456, 1993. ISSN 0162-1459.
D. Peel and G.J. McLachlan. Robust mixture modelling using the t distribution. *Statistics and Computing*, 10(4):339–348, 2000.
M. K. Pitt and N. Shephard. Filtering via simulation: Auxiliary particle filters. *J. Am. Statist. Assoc.*, 94(446):590–599, 1999.
R.E. Quandt and J.B. Ramsey. A new approach to estimating switching regressions. *J. Am. Statist. Assoc.*, 67:306–310, 1972.
W Rayens and T. Greene. Covariance pooling and stabilization for classification. *Computational Statistics & Data Analysis*, 11(1): 17–42, 1991.
B. Ristic, M. Arulampalam, and A. Gordon. *Beyond Kalman Filters: Particle Filters for Target Tracking*. Artech House, 2004.
G. O. Roberts and J. S. Rosenthal. Examples of adaptive MCMC. *Journal of Computational and Graphical Statistics*, 18(2): 349–367, 2009.
P. Robins, V. E. Rapley, and N. Green. Real-time sequential inference of static parameters with expensive likelihood calculations. *J. Roy. Statist. Soc. Ser. B*, 58:641–662, 2009.
R. Y. Rubinstein and D. P. Kroese. *The Cross-Entropy Method*. Springer, 2004.
R. Van der Merwe and E. Wan. Sigma-point Kalman filters for probabilistic inference in dynamic state-space models. In *Proceedings of the Workshop on Advances in Machine Learning*, Montreal, Canada, June 2003.
R. Van der Merwe, A. Doucet, N. De Freitas, and E. Wan. The unscented particle filter. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, *Adv. Neural Inf. Process. Syst.*, volume 13. MIT Press, 2000.
Xet Storage Details
- Size:
- 106 kB
- Xet hash:
- 40948062b1ce9eb3e98545e508a2ab04b36e039e3deded21e5cdc2e46af4a846
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.