Information Geometry and Asymptotic Theory for SMML Estimators
We develop an asymptotic theory for strict minimum message length (SMML) estimators in regular parametric models with countable data spaces. We show that, asymptotically, the optimal SMML partition is induced by a weighted Fisher--Voronoi tessellation in parameter space, pulled back through the maximum likelihood estimator. We further show that each SMML codepoint is asymptotically a weighted average of the maximum likelihood estimates associated with observations in its cell. These results imply that the SMML estimator is consistent and converges at the usual parametric $n^{-1/2}$ rate under standard regularity conditions. We also give a Kullback--Leibler projection interpretation of SMML codepoints and a decomposition of the expected SMML codelength into an assertion entropy and an expected conditional cross-entropy. In exponential families, the theory simplifies further: SMML codepoints satisfy a moment-matching condition, and optimal SMML cells are induced by a polyhedral partition of the sufficient-statistic space.