An ‘Og reader. Emmanuel Charpentier, sent me the following email about model choice:
I read with great interest your critique of Peter Congdon’s 2006 paper (CSDA, 50(2):346-357) proposing a method of estimation of posterior model probabilities based on improper distributions for parameters not present in the model inder examination, as well as a more general critique in your recent review of M. Aitkin’s recent book.
However, Peter Congdon’s 2007 proposal (Statistical Methodology. 4(2):143-157.) of another method for model weighting seems to have flown under your radar ; more generally, while the 2006 proposal seems to have been somewhat quoted and used in at least one biological application and two financial applications, ihis 2007 proposal seems to have been largely ignored (as far as a naïve Google Scholar’s user can tell) ; I found no allusion to this technique neither in your blog nor on Andrew Gelman’s blog.
This proposal, which uses a full probability model with proper priors and pseudo-priors, seems, however, to answer your critiques, and offers a number of technical advantages over other proposal :
- it can be computed from separate MCMC samples, with no regard to the MCMC sapling technique used to obtain them, therefore allowing the use of the « canned expertise » existing in WinBUGS, OpenBUGS or JAGS (which entails the impossibility of controlling the exact sampling methods used to solve a given problem) ;
- it avoids the needs of very long runs to sufficiently explore unlikely models (which is the curse of Carlin & Chib (1995) method) ;
- it seems relatively easy to compute in most situations.
I’d be quite interested by any writings, thoughts or reactions to this proposal.
As I had indeed missed this paper, I went and took a look at it.
“Profiles can be obtained of differences in parameters or model fit measures between models.” (page 144)
Concerning the motivations of the approach given in the introduction, I am rather surprised at the goal of getting posterior distributions of differences like and
Now, looking at the method itself, it boils down to an importance sampling solution, namely that is estimated by (page 147)
where is the importance function approximating the true posterior for model m and where is one generation from this importance function. (I find the paper confusing at times as those simulations could also be understood as generated from the true posterior.) Obviously the quality of the approximation of the posterior density will impact the quality of the approximation of the posterior probability. My only technical objection to the paper is about the approximation of the “posterior” distributions of differences like or of “posterior” probabilities
rather than which seems to be implied in the paper (page 150). (But this may be an over-interpretation of mine…)