Группа авторов

Computational Statistics in Data Science


Скачать книгу

p"/>s toward zero, while the local scale lamda Subscript ps, with its heavy‐tailed prior pi Subscript l o c a l Baseline left-parenthesis dot right-parenthesis, allow a small number of tau lamda Subscript p and hence theta Subscript ps to be estimated away from zero. While motivated by two different conceptual frameworks, the spike‐and‐slab can be viewed as a subset of global–local priors in which pi Subscript l o c a l Baseline left-parenthesis dot right-parenthesis is chosen as a mixture of delta masses placed at lamda Subscript p Baseline equals 0 and lamda Subscript p Baseline equals sigma slash tau. Continuous shrinkage mitigates the multimodality of spike‐and‐slab by smoothly bridging small and large values of lamda Subscript p.

      On the other hand, the use of continuous shrinkage priors does not address the increasing computational burden from growing upper N and upper P in modern applications. Sparse regression posteriors under global–local priors are amenable to an effective Gibbs sampler, a popular class of MCMC we describe further in Section 4.1. Under the linear and logistic models, the computational bottleneck of this Gibbs sampler stems from the need for repeated updates of bold-italic theta from its conditional distribution

      3.1.2 Conjugate gradient sampler for structured high‐dimensional Gaussians

      The conjugate gradient (CG) sampler of Nishimura and Suchard [57] combined with their prior‐preconditioning technique overcomes this seemingly inevitable script í’ª left-parenthesis min left-brace right-brace comma times times upper N 2 upper P comma times times NP 2 right-parenthesis growth of the computational cost. Their algorithm is based on a novel application of the CG method [59, 60], which belongs to a family of iterative methods in numerical linear algebra. Despite its first appearance in 1952, CG received little attention for the next few decades, only making its way into major software packages such as MATLAB in the 1990s [61]. With its ability to solve a large and structured linear system bold upper Phi bold-italic theta equals bold-italic b via a small number of matrix–vector multiplications bold-italic v right-arrow bold upper Phi bold-italic v without ever explicitly inverting bold upper Phi, however, CG has since emerged as an essential and prototypical algorithm for modern scientific computing [62, 63].