Closed

Description
While the inverse Wishart distribution has significant problems, it also has advantages:
- Conjugacy.
- A strong maximum entropy argument (the inverse Wishart is the least informative prior for a given harmonic and geometric mean).
- The Wishart and inverse-Wishart distributions can be used as informative priors, unlike the LKJ prior on correlations, which is a purely regularizing prior. While a purely regularizing prior that shrinks correlations towards 0 is a good idea in cases where the likelihood swamps the prior, in cases where there is important prior information that should be included in the model about the sign and size of correlations, this is not optimal.
In general, there are two strong arguments against using the inverse Wishart: - The inverse Wishart is highly informative, even when its parameters are set to very small values. This is correct, but doesn't imply they can't be useful when prior information suggests an inverse-Wishart is reasonable, e.g. when using the sampling distribution from a past study.
- The inverse Wishart is difficult to interpret and unsuited to HMC. While this is correct in its usual form, it is possible to reparametrize the distribution into the restricted inverse-Wishart distribution using the separation strategy. Such distributions are easier to sample from and to interpret -- see here and here.
Given the restricted inverse Wishart retains some of the advantages of its more common Wishart counterpart while avoiding some of its pitfalls, I think it makes sense to include it in PyMC3 for modeling informative priors, as long as a note is included specifying that it is not a good uninformative prior.