Uh-oh, it looks like your Internet Explorer is out of date.

For a better shopping experience, please upgrade now.

Applied Bayesian Modeling and Causal Inference from Incomplete-Data Perspectives / Edition 1

Applied Bayesian Modeling and Causal Inference from Incomplete-Data Perspectives / Edition 1

by Xiao-Li Meng, Andrew Gelman

ISBN-10: 047009043X

ISBN-13: 9780470090435

Pub. Date: 09/20/2004

Publisher: Wiley

This book brings together a collection of articles on statistical methods relating to missing data analysis, including multiple imputation, propensity scores, instrumental variables, and Bayesian inference. Covering new research topics and real-world examples which do not feature in many standard texts. The book is dedicated to


This book brings together a collection of articles on statistical methods relating to missing data analysis, including multiple imputation, propensity scores, instrumental variables, and Bayesian inference. Covering new research topics and real-world examples which do not feature in many standard texts. The book is dedicated to Professor Don Rubin (Harvard). Don Rubin  has made fundamental contributions to the study of missing data.

Key features of the book include:

  • Comprehensive coverage of an imporant area for both research and applications.
  • Adopts a pragmatic approach to describing a wide range of intermediate and advanced statistical techniques.
  • Covers key topics such as multiple imputation, propensity scores, instrumental variables and Bayesian inference.
  • Includes a number of applications from the social and health sciences.
  • Edited and authored by highly respected researchers in the area.

Product Details

Publication date:
Wiley Series in Probability and Statistics Series , #561
Product dimensions:
6.40(w) x 9.30(h) x 1.30(d)

Related Subjects

Table of Contents


I Casual inference and observational studies.

1 An overview of methods for causal inference from observational studies, by Sander Greenland.

1.1 Introduction.

1.2 Approaches based on causal models.

1.3 Canonical inference.

1.4 Methodologic modeling.

1.5 Conclusion.

2 Matching in observational studies, by Paul R. Rosenbaum.

2.1 The role of matching in observational studies.

2.2 Why match?

2.3 Two key issues: balance and structure.

2.4 Additional issues.

3 Estimating causal effects in nonexperimental studies, by Rajeev Dehejia.

3.1 Introduction.

3.2 Identifying and estimating the average treatment effect.

3.3 The NSWdata.

3.4 Propensity score estimates.

3.5 Conclusions.

4 Medication cost sharing and drug spending in Medicare, by Alyce S. Adams.

4.1 Methods.

4.2 Results.

4.3 Study limitations.

4.4 Conclusions and policy implications.

5 A comparison of experimental and observational data analyses, by Jennifer L. Hill, Jerome P. Reiter, and Elaine L. Zanutto.

5.1 Experimental sample.

5.2 Constructed observational study.

5.3 Concluding remarks.

6 Fixing broken experiments using the propensity score, by Bruce Sacerdote.

6.1 Introduction.

6.2 The lottery data.

6.3 Estimating the propensity scores.

6.4 Results.

6.5 Concluding remarks.

7 The propensity score with continuous treatments, by Keisuke Hirano and Guido W. Imbens.

7.1 Introduction.

7.2 The basic framework.

7.3 Bias removal using the GPS.

7.4 Estimation and inference.

7.5 Application: the Imbens–Rubin–Sacerdote lottery sample.

7.6 Conclusion.

8 Causal inference with instrumental variables, by Junni L. Zhang.

8.1 Introduction.

8.2 Key assumptions for the LATE interpretation of the IV estimand.

8.3 Estimating causal effects with IV.

8.4 Some recent applications.

8.5 Discussion.

9 Principal stratification, by Constantine E. Frangakis.

9.1 Introduction: partially controlled studies.

9.2 Examples of partially controlled studies.

9.3 Principal stratification.

9.4 Estimands.

9.5 Assumptions.

9.6 Designs and polydesigns.

II Missing data modeling.

10 Nonresponse adjustment in government statistical agencies: constraints, inferential goals, and robustness issues, by John L. Eltinge.

10.1 Introduction: a wide spectrum of nonresponse adjustment efforts in government statistical agencies.

10.2 Constraints.

10.3 Complex estimand structures, inferential goals, and utility functions.

10.4 Robustness.

10.5 Closing remarks.

11 Bridging across changes in classification systems, by Nathaniel Schenker.

11.1 Introduction.

11.2 Multiple imputation to achieve comparability of industry and occupation codes.

11.3 Bridging the transition from single-race reporting to multiple-race reporting.

11.4 Conclusion.

12 Representing the Census undercount by multiple imputation of households, by Alan M. Zaslavsky.

12.1 Introduction.

12.2 Models.

12.3 Inference.

12.4 Simulation evaluations.

12.5 Conclusion.

13 Statistical disclosure techniques based on multiple imputation, by Roderick J. A. Little, Fang Liu, and Trivellore E. Raghunathan.

13.1 Introduction.

13.2 Full synthesis.

13.3 SMIKe andMIKe.

13.4 Analysis of synthetic samples.

13.5 An application.

13.6 Conclusions.

14 Designs producing balanced missing data: examples from the National Assessment of Educational Progress, by Neal Thomas.

14.1 Introduction.

14.2 Statistical methods in NAEP.

14.3 Split and balanced designs for estimating population parameters.

14.4 Maximum likelihood estimation.

14.5 The role of secondary covariates.

14.6 Conclusions.

15 Propensity score estimation with missing data, by Ralph B. D’Agostino Jr.

15.1 Introduction.

15.2 Notation.

15.3 Applied example:March of Dimes data.

15.4 Conclusion and future directions.

16 Sensitivity to nonignorability in frequentist inference, by Guoguang Ma and Daniel F. Heitjan.

16.1 Missing data in clinical trials.

16.2 Ignorability and bias.

16.3 A nonignorable selection model.

16.4 Sensitivity of the mean and variance.

16.5 Sensitivity of the power.

16.6 Sensitivity of the coverage probability.

16.7 An example.

16.8 Discussion.

III Statistical modeling and computation.

17 Statistical modeling and computation, by D. Michael Titterington.

17.1 Regression models.

17.2 Latent-variable problems.

17.3 Computation: non-Bayesian.

17.4 Computation: Bayesian.

17.5 Prospects for the future.

18 Treatment effects in before-after data, by Andrew Gelman.

18.1 Default statistical models of treatment effects.

18.2 Before-after correlation is typically larger for controls than for treated units.

18.3 A class of models for varying treatment effects.

18.4 Discussion.

19 Multimodality in mixture models and factor models, by Eric Loken.

19.1 Multimodality in mixture models.

19.2 Multimodal posterior distributions in continuous latent variable models.

19.3 Summary.

20 Modeling the covariance and correlation matrix of repeated measures, by W. John Boscardin and Xiao Zhang.

20.1 Introduction.

20.2 Modeling the covariance matrix.

20.3 Modeling the correlation matrix.

20.4 Modeling a mixed covariance-correlation matrix.

20.5 Nonzero means and unbalanced data.

20.6 Multivariate probit model.

20.7 Example: covariance modeling.

20.8 Example: mixed data.

21 Robit regression: a simple robust alternative to logistic and probit regression, by Chuanhai Liu.

21.1 Introduction.

21.2 The robit model.

21.3 Robustness of likelihood-based inference using logistic, probit, and robit regression models.

21.4 Complete data for simple maximum likelihood estimation.

21.5 Maximum likelihood estimation using EM-type algorithms.

21.6 A numerical example.

21.7 Conclusion.

22 Using EM and data augmentation for the competing risks model, by Radu V. Craiu and Thierry Duchesne.

22.1 Introduction.

22.2 The model.

22.3 EM-based analysis.

22.4 Bayesian analysis.

22.5 Example.

22.6 Discussion and further work.

23 Mixed effects models and the EM algorithm, by Florin Vaida, Xiao-Li Meng, and Ronghui Xu.

23.1 Introduction.

23.2 Binary regression with random effects.

23.3 Proportional hazards mixed-effects models.

24 The sampling/importance resampling algorithm, by Kim-Hung Li.

24.1 Introduction.

24.2 SIR algorithm.

24.3 Selection of the pool size.

24.4 Selection criterion of the importance sampling distribution.

24.5 The resampling algorithms.

24.6 Discussion.

IV Applied Bayesian inference.

25 Whither applied Bayesian inference?, by Bradley P. Carlin.

25.1 Where we’ve been.

25.2 Where we are.

25.3 Where we’re going.

26 Efficient EM-type algorithms for fitting spectral lines in high-energy astrophysics, by David A. van Dyk and Taeyoung Park.

26.1 Application-specific statistical methods .

26.2 The Chandra X-ray observatory.

26.3 Fitting narrow emission lines.

26.4 Model checking and model selection.

27 Improved predictions of lynx trappings using a biological model, by Cavan Reilly and Angelique Zeringue.

27.1 Introduction.

27.2 The current best model.

27.3 Biological models for predator prey systems.

27.4 Some statistical models based on the Lotka-Volterra system.

27.5 Computational aspects of posterior inference.

27.6 Posterior predictive checks and model expansion.

27.7 Prediction with the posterior mode.

27.8 Discussion.

28 Record linkage using finite mixture models, by Michael D. Larsen.

28.1 Introduction to record linkage.

28.2 Record linkage.

28.3 Mixture models.

28.4 Application.

28.5 Analysis of linked files.

28.6 Bayesian hierarchical record linkage.

28.7 Summary.

29 Identifying likely duplicates by record linkage in a survey of prostitutes, by Thomas R. Belin, Hemant Ishwaran, Naihua Duan, Sandra H. Berry, and David E. Kanouse.

29.1 Concern about duplicates in an anonymous survey.

29.2 General frameworks for record linkage.

29.3 Estimating probabilities of duplication in the Los Angeles Women’s Health Risk Study.

29.4 Discussion.

30 Applying structural equation models with incomplete data, by Hal S. Stern and Yoonsook Jeon.

30.1 Structural equation models.

30.2 Bayesian inference for structural equation models.

30.3 Iowa Youth and Families Project example.

30.4 Summary and discussion.

31 Perceptual scaling, by Ying Nian Wu, Cheng-En Guo, and Song Chun Zhu.

31.1 Introduction.

31.2 Sparsity and minimax entropy.

31.3 Complexity scaling law.

31.4 Perceptibility scaling law.

31.5 Texture = imperceptible structures.

31.6 Perceptibility and sparsity.



Customer Reviews

Average Review:

Post to your social network


Most Helpful Customer Reviews

See all customer reviews