List of abstracts
-
Christophe Andrieu: Particle MCMC
-
Yves Atchade:
-
Witold Bednorz: Exponential type Inequalities for Markov chains
-
Kasper Berthelsen: Perfect posterior inference for mixture models.
-
Alex Beskos: Diffusion limits for MCMC paths
-
Roberto Casarin: Sequential Monte Carlo for complex sampling problems
-
Siddhartha Chib: Tailored Multiple-block MCMC Methods for Analysis of DSGE Models
-
Nicolas Chopin: Bayesian nonparametric estimation of a long-memory Gaussian process: computational aspects
-
Dan Crisan: Monte-Carlo approximations of Feynman-Kac representations
-
Andreas Eberle: Quantitative approximations of evolving probability measures and sequential Markov Chain Monte Carlo methods
-
Yalchin Efendiev: Uncertainty quantification with multi-scale models in porous media flow applications
-
Jo Eidsvik: Bayesian inference in spatial models with skew Normal latent variables
-
Chris L. Farmer: Optimal control and data assimilation: principles and approximations
-
Gersende Fort:
-
Flavio Goncalves: Monte Carlo Inference for Jump-diffusion Processes
-
Nial Friel: Bayesian inference for social network models
-
Radu Herbei: Hybrid samplers for ill-posed inverse problems
-
Mark Huber: Perfect simulation of repulsive point processes
-
Ajay Jasra: On Non-linear MCMC
-
Adam Johansen:
-
Galin Jones:
-
Ruth King:
-
Samuel Kou:
-
Theodore Kypraios: Bayesian Inference and Model Choice for Nonlinear Stochastic Processes: Applications to stochastic epidemic modelling.
-
krzysztof Latuszynski: Characterization of CLTs for Ergodic Markov Chains via Regeneration
-
Faming Liang: Trajectory Averaging for Stochastic Approximation MCMC Algorithm
-
Robert McKay: Why Markov Monte Carlo Methods give such good answers for high-dimensional integrals.
-
Xiao-Li Meng:
-
Antonietta Mira: Zero Variance MCMC
-
Jesper Moller: Likelihood inference and computational problems for a random set model for the union of interacting discs.
-
Eric Moulines: On interacting particle approximations of the smoothing distribution in general state spaces models
-
Wojciech Niemiro: Nonasymptotic bounds on the estimation error for regenerative MCMC
-
Omiros Papaspiliopoulos: Posterior Simulation for nonparametric hidden Markov models
-
Sylvia Richardson:
-
Christian Robert: Computational approaches to Bayesian model choice, Recent Advances in ABC Methodology
-
Judith Rousseau: The Use of Importance sampling for repeated MCMC
-
Harvard Rue: Some alternatives to MCMC
-
Sujit Sahu: Fusing point and areal level space-time data with application to wet deposition
-
Giovanni Sebastiani: Markov chain analysis of ant colony optimization
-
Giorgos Sermaidis: Exact inference for discretely observed diffusions
-
Christopher Sherlock: The random walk Metropolis: linking theory and practice through a case study
-
Geir Storvik: On the flexibility of acceptance probabilities in auxiliary variable Metropolis-Hastings algorithms
-
Andrew Stuart: Bayesian Approach to Inverse Problems in PDEs
-
Haakon Tjelmeland and Haakon Austad: On approximate calculations for binary Markov random fields
-
Simon Wilson: Factor Analysis on multi-channel image data with gaussian mixture priors
-
Yaming Yu:
- Christophe Andrieu: Particle MCMC
Markov chain Monte Carlo (MCMC) and Sequential Monte Carlo (SMC) methods have emerged as the two main tools to sample from high-dimensional probability distributions. Although asymptotic convergence of MCMC algorithms is ensured under weak assumptions, the performance of these latters is unreliable when the proposal distributions used to explore the space are poorly chosen and/or if highly correlated variables are updated independently. We show here how it is possible to build efficient high-dimensional proposal distributions using SMC methods. This allows us not only to improve over standard MCMC schemes but also to make Bayesian inference feasible for a large class of statistical models where this was not previously the case. We demonstrate these algorithms on various non-linear non-Gaussian state-space models, a stochastic kinetic model and Dirichlet process mixtures. - Witold Bednorz: Exponential type inequalities for Markov chains.
There are several known results on the exponential type inequalities for functionals of independent random variables. Such inequalities are used to describe the concentration property i.e. that sums of r.v.'s cannot deviate much from their expectations. The natural question arises which of the results can be reproved for functionals of Markovian variables. Obviously to have a chance for showing exponentially fast concentration we have to assume that the Markov chain is geometrically ergodic. In this talk we discuss more general case and show that whenever geometric regularity condition holds some form of the concentration behavior can be observed.
- Kasper Berthelsen: Perfect posterior inference for mixture models.
We consider a standard Bayesian analysis of the unknown mixture weights for a mixture of known densities. It is well known how to approximately sample the posterior weights by Gibbs sampling using the duality principle, i.e. introducing auxiliary allocation variables In this talk we formulate an algorithm for perfect simulation of the posterior weights. The algorithm is an example of Wilson's read-once coupling from the past algorithm (read-once CFTP). The key ingredient of this algorithm is constructing compounds of random maps so that stationarity is preserved and there is a positive probability of detecting coalescence, i.e. concluding that the compound map maps the entire state space into a single point. - Alex Beskos: Diffusion limits for MCMC paths.
MCMC trajectories resemble diffusion paths. Such an interpretation has been substantiated in earlier works in the literature when it has been proven that, when the dimensionality of the state space increases, the MCMC trajectory converges to a particular diffusion process. Such a result, though proven in simplified scenaria of iid targets, provides insight in the behavior of MCMC algorithms in high dimensions. We examine the case of the so-called "hybrid Monte-Carlo" MCMC algorithm, invoking Hamiltonian dynamics, employed by physicists in molecular dynamics applications and elsewhere. Bridging the machinery employed above with tools from numerical analysis we show that the MCMC trajectory of the hybrid algorithm converges (when appropriately rescaled) to a hypoelliptic SDE. Such a result provides a complete characterization of the efficiency of the algorithm: we conclude that the hybrid algorithm should be scaled as 1/n^{1/6} (n being the dimensionality) , with optimal acceptance probability 0.743.
- Roberto Casarin: Sequential Monte Carlo for complex sampling problems (joint with Christophe Andrieu)
Sequential Monte Carlo (SMC) represents a class of general sampling methods which allow us to design good samplers for difficult sampling problems. With respect to the classical Markov Chain Monte Carlo (MCMC) algorithms the kernels of the SMC need not to be reversible or even Markov. Moreover for inference on high-dimensional static models a strategy based on the combination of a long tempering sequence with a SMC sampling strategy could overcome the degeneracy problem of the SMC sampling and performs as well as a MCMC algorithms or even better when the last exhibits the traditional problem of slow mixing. Note however that the recent literature on population of MCMC chains provides an alternative and promising way to solve the problem of exploration of the space and to improve on the classical MCMC. In this work we follow the SMC framework, but suggest some news algorithms which result from the combination of SMC with MCMC and could be particularly useful in the context of simulation from multimodal distributions and from high-dimensional distributions. The performance of the proposed algorithms will be demonstrated on some challenging sampling problems.
- Nicolas Chopin: Bayesian nonparametric estimation of a long-memory Gaussian process: computational aspects (joint work with Judith Rousseau and Brunero Liseo)
In Rousseau et al. (2008), we proposed a novel method for the Bayesian nonparametric estimation of the spectral density of a Gaussian long-memory process, based on the true likelihood rather than on Whittle's approximation, which is not valid in long memory settings. We also established the convergence properties (consistency, rates of convergence) of the corresponding estimates. In this paper, we discuss the computational implementation of this procedure. The two main computational challgenges are(a) the likelihood function involves the inverse of a possibly big Toeplitz matrix, and
(b) the posterior distribution is trans-dimensional.
With respect to (a), we consider a simple approximation of the likelihood, which may be used either directly or as a tool for building an importance sampling proposal. With respect to (b), we compare different methods, focussing on reversible jump MCMC, and population/sequential monte carlo methods. In particular, we explain how the likelihood function may be computed recursively, which makes the use of sequential Monte Carlo particularly interesting, especially if the dataset is large.
- Dan Crisan: Monte-Carlo approximations of Feynman-Kac representations
Feynman-Kac representations are probabilistic expressions of solutions of linear/nonlinear PDEs and stochastic PDEs. Over the last ten years, these type of representations spawned
a variety of Monte Carlo methods for approximating numerically solutions of certain classes of linear/nonlinear PDEs and SPDEs. The aim of the talk is to present a survey of Monte-Carlo
approximations of Feynman-Kac representations and their applications to filtering and finance. The talk is based on joint work with Manolarakis and Ghazali.
- Andreas Eberle: Quantitative approximations of evolving probability measures and sequential Markov Chain Monte Carlo methods
We study approximations of evolving probability measures by an interacting particle system. The particle system dynamics is a combination of independent Markov chain moves and importance sampling/resampling steps. Under global regularity conditions, we derive non-asymptotic error bounds for the particle system approximation. The main motivation are applications to sequential MCMC methods for Monte Carlo integral estimation.
- Yalchin Efendiev: Uncertainty quantification with multi-scale models in porous media flow applications
In this talk, I will describe the use of coarse-scale models in uncertainty quantification. The problem under consideration is posed as a sampling problem from a posterior distribution. The main goal is to develop an efficient sampling technique within the framework of Markov chain Monte Carlo methods that uses coarse-scale models and gradients of the target distribution. The purpose is to reduce the computational cost of Langevin algorithms for dynamic data integration problems. We propose to use inexpensive coarse-scale solutions in calculating the proposals of Langevin algorithms.To guarantee the correct and efficient sampling of the proposed algorithm, we also intend to test the proposals with coarse-scale solutions. Comparing with the direct Langevin algorithm based on fine-scale solutions the proposed method generates a modified Markov chain by incorporating the coarse-scale information of the problem. Under some mild technical conditions we show that the modified Markov chain converges to the correct posterior distribution. Our numerical examples show that the proposed coarse-gradient Langevin algorithms are much faster than the direct Langevin algorithms bt have similar acceptance rates. Numerical results for multi-phase flow and transport are presented. Both Gaussian and non-Gaussian prior models are studied.
- Chris L. Farmer: Optimal control and data assimilation: principles and approximations
There are four basic types of activity involving combinations of uncertainty and optimisation:- Uncertainty propagation where the problem is to predict the probabilistic behaviour of an uncertain system, with uncertainty in the initial or boundary conditions or in the static properties.
- Data assimilation, also known as "history matching", "system identification" or inverse problems.
- Decision making. Here a choice must be made between competing courses of action. For each choice of action the outcome is uncertain.
- Optimal control of an uncertain system. A system is only known in a probabilistic way. One has to design a control policy that optimises the system. The problem is particularly difficult when optimisation of the measurement system is included in the problem.
- Nial Friel: Bayesian inference for social network models
The exponential random graph model (or p* model) is very popular in social network analysis. However it is complicated to handle statistically, since the corresponding likelihood can only be calculated exactly for all but trivially small networks. This is due to the fact that the p* model is specified up to an intractable normalising constant. This talk will present various approaches to overcome this problem, and has some connections with approximate Bayesian computation methods.
- Flavio Goncalves: Monte Carlo inference for Jump-Diffusion Proesses
This work proposes a Monte Carlo Expectation-Maximisation algorithm for inference in discretely observed jump-diffusion processes. The most challenging step of the algorithm is to sample from the jump-diffusion conditional on the observations. Such step is performed using the Conditional Jump Exact Algorithm which is also proposed in this work. The algorithm draws exact samples from the conditional jump-diffusion via retrospective rejection sampling.
- Radu Herbei: Hybrid samplers for ill-posed inverse problems
In the Bayesian approach to ill-posed inverse problems, regularization is imposed by specifying a prior distribution on the parameters of interest and MCMC samplers are used to extract information about its
posterior distribution. The aim of this work is to investigate the convergence properties of the random-scan random walk Metropolis (RSM) algorithm for posterior distributions in ill-posed inverse problems. We provide an accessible set of sufficient conditions, in terms of the observational model and the prior, to ensure geometric ergodicity of RSM samplers of the posterior distribution. We illustrate how these conditions can be checked in an application to the inversion of oceanographic tracer data. - Mark Huber: Perfect simulation of repulsive point processes
Repulsive spatial data arise whenever objects such as trees or towns are competing for scarce resources. Many different models have been created to deal with such data. In this talk I will explain improved perfect simulation methods for two of these models that generate Monte Carlo variates exactly from the desired distributions. For density based models such as the Strauss process, simple birth-death chains can be augmented with a swapping move to speed up convergence. Then continuous time bounding chains can be employed to obtain
perfect samples.
Another class of models from Matérn requires a different approach. Here construction of a Metropolis Markov chain can be accomplished with auxiliary Poisson processes. Fortunately, bounding chains can also be used with these new Metropolis chains to also obtain perfect samples. Finally, I will show how a product estimator method for these models can be created for precise approximation without the need to know sample variance.
for simulating from a probability measure B&P 2 P(E). Non-linear Markov kernels (e.g. [10, 11]) K : P(E) B!_ E B"* P(E) can be constructed to, in some sense, improve over MCMC methods. However, such non-linear kernels cannot be simulated exactly, so approximations of the nonlinear kernels are constructed using auxiliary or self-interacting chains. Several non-linear kernels are presented and it is demonstrated that, under some conditions, the associated selfinteracting approximations exhibit a strong law of large numbers; our proof technique is via the Poisson equation and Foster-Lyapunov conditions. We investigate the performance of our approximations with some simulations.
- Theodore Kypraios: Bayesian Inference and Model Choice for Nonlinear Stochastic Processes: Applications to stochastic epidemic modelling.
Analysing infectious disease data is a non-standard problem. In general,inference problems for disease outbreak data are complicated by the facts that (i) the data are inherently dependent and (ii) the data are usually incomplete in the sense that the actual process of infection is not observed. However,it is often possible to formulate
simple stochastic models which describe the key features of epidemic spread. Markov Chain Monte Carlo methods play a vital role in drawing efficiently Bayesian inference for the parameters of interest which govern transmission. Although standard data augmentation MCMC algorithms perform well in specific circumstances, it is often the case that their
performance deteriorates due the high dimension of the missing data and the dependence between them and the model parameters.
In this talk we will discuss a class of non-centered MCMC algorithm for stochastic epidemic models and an application of these to model the 2001 UK FMD outbreak. Furthermore, we will present a range of different models to model the spread of hospital infections such as MRSA in wards and how MCMC allow us to estimate the parameters of interest. Finally, we show how Reversible-Jump MCMC algorithms and algorithms based on the developed methodology of power posteriors (Friel and Pettit, 2008) can be used for Bayesian model choice in stochastic epidemics.
I suggest that when a Markov Monte Carlo method gives great accuracy for a multivariate integral it is because the dynamics of the variables are "weakly dependent" and the observables are of "Dobrushin class". Then a large deviation estimate gives exponentially small probability for the time average of the observable to fail to be within a prescribed tolerance of its integral, with respect to number of iterations. Comments from people more expert than me are welcome!
We propose a general purpose variance reduction technique for Markov Chain Monte Carlo estimators based on the zero-variance principle introduced in the physics literature. We show how it is possible, for a large class of interesting Bayesian statistical model , to generate moments estimators with ideally null variance. The variance reduction achieved in practice is substantial as we show with some toy examples and a real application to Bayesian inference for credit risk estimation.
- Jesper Moller: Likelihood inference and computational problems for a random set model for the union of interacting discs.
This talk is based on the papers:
- J. Møller and K. Helisova. Power diagrams and interaction processes for unions of discs. Advances in Applied Probility, 40, 321-347.
- J. Møller and K. Helisova. Likelihood inference for unions of interacting discs. Research Report R-2008-18, Department of Mathematical Sciences, Aalborg University. (Submitted for publication.)
The first paper studies a flexible class of finite disc process models with interaction between the discs. We let U denote the random set given by the union of discs, and use for the disc process an exponential family density with the canonical sufficient statistic only depending on geometric properties of U such as the area, perimeter, Euler-Poincar'echaracteristic, and number of holes. This includes the quermass-interaction process and the continuum random cluster model as special cases. Viewing our model as a connected component Markov point process, and thereby establish local and spatial Markov properties, becomes useful for handling the problem of edge effects when only U is observed within a bounded observation window.The power tessellation and its dual graph become major tools when establishing inclusion-exclusion formulae, formulae for computing geometric characteristics of U, and stability properties of the underlying disc process density.
Algorithms for constructing the power tessellation of U and for simulating the disc process are discussed.
- Eric Moulines: On interacting particle approximations of the smoothing distribution in general state spaces models
A long-lasting problem in general state-space models is the approximation of the smoothing distribution of a given state, or a sequence of states, conditional on the observations from the past, the present, and the future. The aim of this talk is to provide a rigorous foundation for the calculation, or approximation, of such smoothed distributions, and to analyze in a common unifying framework different schemes to reach this goal. Through a cohesive and generic exposition of the scientific literature we offer several novel extensions allowing to approximate and sample from the joint smoothing distribution.
Dirichlet process mixtures (MDPs), are now standard in semiparametric modelling. Posterior inference for such models is typically performed using Markov chain Monte Carlo methods, which can be roughly categorised into marginal and conditional methods. The former integrate out analytically the infinite-dimensional component of the hierarchical model and sample from the marginal distribution of the remaining variables using the Gibbs sampler. Conditional methods impute the Dirichlet process and update it as a component of the Gibbs sampler. Since this requires imputation of an infinite-dimensional process, implementation of the conditional method has relied on finite approximations.In the first part of the talk we show how to avoid such approximations by novel Gibbs sampling algorithms which sample from the exact posterior distribution of quantities of interest. The approximations are avoided by the technique of retrospective sampling. Motivated by the modelling of copy number variation (CNVs) in the human genome we have developed
hidden Markov models where the likelihood is given by an MDP. We term the resulting model an HMM-MDP model. Thus, we deal with a model with two levels of clustering for the observed data, a temporally persisting (local) clustering induced by the HMM and a global clustering induced by the Dirichlet process. The second part of the talk shows how to design efficient conditional methods for fitting these models elaborating on the methods developed on the first part of the talk and on dynamic programming techniques.
- Christian Robert: Computational approaches to Bayesian model choice (Joint work with Jean-Michel MARIN, INRIA Futurs, Universite Paris Sud Orsay, and Nicolas CHOPIN, CREST.)
In this talk, we will survey the flurry of recent results on the ABC algorithm that have appeared in the literature, including our own on the ABC-PMC version of the ABC algorithm and on the use of "exact" ABC algorithms for the selection of Ising models.
- Judith Rousseau: Use of Importance Sampling for Repeated MCMC
The idea of using importance sampling (IS) with Markov Chain Monte Carlo (MCMC) has been around for over a decade. It has been used in Bayesian statistics to assess prior sensitivity and to carry out cross-validation
and in maximum likelihood estimation when MCMC is required to evaluate the likelihood function. While there have been a number of successful applications, some properties of IS when used with MCMC have had relatively little investigation. It is the purpose of this paper to further investigate the efficency of IS with MCMC.
- Giovanni Sebastiani: Markov chain analysis of Ant Colony Optimization
We describe some theoretical results of a study on the expected time needed by a class of Ant Colony Optimization algorithms to solve combinatorial optimization problems. The algorithm is described by means of a suitable Markov chain with absorbing states in set of maximizers. First, we present some general results on the expected runtime of the considered class of algorithms. These results are then specialized to the case of some pseudo-Boolean functions. The results obtained for these functions are also compared to those from the well-investigated (1+1)-Evolutionary Algorithm.
- Giorgos Sermaidis: Exact inference for discretely observed diffusions
The aim of this work is to make Bayesian inference for discretely observed diffusions, the challenge being that the transition density of the process is typically unavailable. Competing methods rely on augmenting the data with the missing paths since there exists an analytical expression for the complete-data likelihood. Such implementations require a rather fine discretization of the imputed path leading to convergence issues and computationally expensive algorithms.
Our method is based on exact simulation of diffusions (Beskos et al 2006) and has the advantage that there is no discretization error. We present a Gibbs sampler for sampling from the posterior distribution of the parameters and discuss how to increase its efficiency using reparametrizations of the augmentation scheme (Papaspiliopoulos et al 2007 ).
- Geir Storvik: On the flexibility of acceptance probabilities in auxiliary variable Metropolis-Hastings algorithms
Use of auxiliary variables for generating proposal variables within a Metropolis-Hastings setting has been suggested in many different settings. This has in particular been of interest for simulation from complex distributions such as multimodal distributions or in transdimensional approaches. For many of these approaches, the acceptance probabilities that are used turn up somewhat magic and different proofs for their validity have been given in each case.
In this talk I will present a general framework for construction of acceptance probabilities in auxiliary variable proposal generation. In addition to demonstrate the similarities between many of the proposed algorithms in the literature, the framework also demonstrate that there is a great flexibility in how to construct such acceptance probabilities, in addition to the flexibility in how to construct the proposals. With this flexibility, alternative acceptance probabilities are suggested. Some numerical experiments will also be reported.
A wide variety of inverse problems in PDEs can be formulated in a common mathematical framework, by adopting the perspective of Bayesian statistics on a function space. This common mathematical structure can then be exploited to construct robust MCMC algorithms to gain insight into the structure of the inverse problem.
The common mathematical structure is that the posterior measure has density with respect to a Gaussian reference measure, with log density satisfying certain natural bounds and Lipschitz continuity properties. The ideas will be illustrated on a number of applications including oceanography and nuclear waste management.
In this presentation we limit the attention to binary MRFs defined on a regular lattice and propose an approximate forward-backward algorithm for such models. The forward part of the (exact) forward-backward algorithm computes a series of joint marginal distributions by summing out each of the variables in turn. We represent these joint marginal distributions by interactions of different orders. We develop recursive formulas for these interaction parameters, first computing first order interactions, then the second order interactions and so on. The approximation is defined by approximating to zero all interaction parameters that are sufficiently close to zero. In addition, an interaction parameter is set to zero whenever all associated lower level interactions are (approximated to) zero. For the approximative algorithm to be computationally feasible for large lattices the number of interactions that are approximated to zero must be sufficiently large. We illustrate the performance of the resulting algorithm by a number of examples. For many models the algorithm is computationally feasible even for large lattices. This includes the Ising model, for which the computational cost of the approximative algorithm is linear in the number of nodes. We also present examples of models where the number of remaining non-zero interaction parameters is too high, so that also the approximative algorithm is computationally infeasible.
Authors: Simon Wilson (Trinity College Dublin) - speaker; Ercan Kuruoglu (CNR Pisa, Italy); Alicia Quiros Carretero (University Rey Juan Carlos, Madrid)
We consider a factor analysis problem where the data are a set of images of the same scene observed at different frequencies. The goal is to separate out the different factors (or sources) of light that contributed to the scene. The motivating example for this work is data on extra-terrestrial microwaves observed at 5 channels, for which it is desired to separate the observed intensities into their sources, such as galactic dust, cosmic microwave background, etc.
There is a lot of prior information on these sources, which can be reasonably modelled as Gaussian mixtures. Within and between pixel correlations between sources suggest mixtures of multivariate Gaussians or even mixtures of multivariate Gaussian Markov random ields as desirable factor priors. We discuss MCMC schemes to implement such an inference and in particular the computational difficulties that arise when these prior dependencies are incorporated.