George, Edward ILiang, FengXu, Xinyi2023-05-232023-05-232006-05-012017-08-24https://repository.upenn.edu/handle/20.500.14332/47748Let X|μ∼Np(μ,vxI) and Y|μ∼Np(μ,vyI) be independent p-dimensional multivariate normal vectors with common unknown mean μ. Based on only observing X=x, we consider the problem of obtaining a predictive density p̂(y|x) for Y that is close to p(y|μ) as measured by expected Kullback–Leibler loss. A natural procedure for this problem is the (formal) Bayes predictive density p̂U(y|x) under the uniform prior πU(μ)≡1, which is best invariant and minimax. We show that any Bayes predictive density will be minimax if it is obtained by a prior yielding a marginal that is superharmonic or whose square root is superharmonic. This yields wide classes of minimax procedures that dominate p̂U(y|x), including Bayes predictive densities under superharmonic priors. Fundamental similarities and differences with the parallel theory of estimating a multivariate normal mean under quadratic loss are described.The original and published work is available at: https://projecteuclid.org/euclid.aos/1146576256#abstractBayes ruleshear equationinadmissibilitymultiple shrinkagemultivariate normalprior distributionsshrinkage estimationsuperharmonic marginalssuperharmonic priorsunbiased estimate of riskPhysical Sciences and MathematicsImproved Minimax Predictive Densities Under Kullback-Leibler LossArticle