Jump to Main Content
Convergence rates and asymptotic standard errors for Markov chain Monte Carlo algorithms for Bayesian probit regression Series B Statistical methodology
- Roy, Vivekananda, Hobert, James P.
- Journal of the Royal Statistical Society 2007 v.69 no.4 pp. 607-623
- Markov chain, algorithms, normal distribution, variance
- Consider a probit regression problem in which Y₁, [ellipsis (horizontal)], Yn are independent Bernoulli random variables such that [graphic removed] where xi is a p-dimensional vector of known covariates that are associated with Yi, β is a p-dimensional vector of unknown regression coefficients and Φ(·) denotes the standard normal distribution function. We study Markov chain Monte Carlo algorithms for exploring the intractable posterior density that results when the probit regression likelihood is combined with a flat prior on β. We prove that Albert and Chib's data augmentation algorithm and Liu and Wu's PX-DA algorithm both converge at a geometric rate, which ensures the existence of central limit theorems for ergodic averages under a second-moment condition. Although these two algorithms are essentially equivalent in terms of computational complexity, results of Hobert and Marchev imply that the PX-DA algorithm is theoretically more efficient in the sense that the asymptotic variance in the central limit theorem under the PX-DA algorithm is no larger than that under Albert and Chib's algorithm. We also construct minorization conditions that allow us to exploit regenerative simulation techniques for the consistent estimation of asymptotic variances. As an illustration, we apply our results to van Dyk and Meng's lupus data. This example demonstrates that huge gains in efficiency are possible by using the PX-DA algorithm instead of Albert and Chib's algorithm.