Point distribution model

The point distribution model is a model for representing the mean geometry of a shape and some statistical modes of geometric variation inferred from a training set of shapes.

The point distribution model concept has been developed by Cootes,[1] Taylor et al.[2] and became a standard in computer vision for the statistical study of shape[3] and for segmentation of medical images[2] where shape priors really help interpretation of noisy and low-contrasted pixels/voxels.

A landmark is an annotating point posed by an anatomist onto a given locus for every shape instance across the training set population.

For instance, the same landmark will designate the tip of the index finger in a training set of 2D hands outlines.

Principal component analysis (PCA), for instance, is a relevant tool for studying correlations of movement between groups of landmarks among the training set population.

Typically, it might detect that all the landmarks located along the same finger move exactly together across the training set examples showing different finger spacing for a flat-posed hands collection.

These landmarks are aligned using the generalized procrustes analysis, which minimizes the least squared error between the points.

Assuming the scattering is gaussian in this space, PCA is used to compute normalized eigenvectors and eigenvalues of the covariance matrix across all training shapes.

To ensure that the new shapes are all within the variation seen in the training set, it is common to only allow each element of

PDM's can be extended to any arbitrary number of dimensions, but are typically used in 2D image and 3D volume applications (where each landmark point is

euclidean vectors associated to corresponding landmark and designating a compound move for the whole shape.

Typically, a twisting nematode worm is used as an example in the teaching of kernel PCA-based methods.

Due to the PCA properties: eigenvectors are mutually orthogonal, form a basis of the training set cloud in the shape space, and cross at the 0 in this space, which represents the mean shape.

Also, PCA is a traditional way of fitting a closed ellipsoid to a Gaussian cloud of points (whatever their dimension): this suggests the concept of bounded variation.

The idea behind PDMs is that eigenvectors can be linearly combined to create an infinity of new shape instances that will 'look like' the one in the training set.

The coefficients are bounded alike the values of the corresponding eigenvalues, so as to ensure the generated 2n/3n-dimensional dot will remain into the hyper-ellipsoidal allowed domain—allowable shape domain (ASD).