In data mining, cluster-weighted modeling (CWM) is an algorithm-based approach to non-linear prediction of outputs (dependent variables) from inputs (independent variables) based on density estimation using a set of models (clusters) that are each notionally appropriate in a sub-region of the input space. The overall approach works in jointly input-output space and an initial version was proposed by Neil Gershenfeld.[1][2]
Basic form of model
The procedure for cluster-weighted modeling of an input-output problem can be outlined as follows.[2] In order to construct predicted values for an output variable y from an input variable x, the modeling and calibration procedure arrives at a joint probability density function, p(y,x). Here the "variables" might be uni-variate, multivariate or time-series. For convenience, any model parameters are not indicated in the notation here and several different treatments of these are possible, including setting them to fixed values as a step in the calibration or treating them using a Bayesian analysis. The required predicted values are obtained by constructing the conditional probability densityp(y|x) from which the prediction using the conditional expected value can be obtained, with the conditional variance providing an indication of uncertainty.
The important step of the modeling is that p(y|x) is assumed to take the following form, as a mixture model:
where n is the number of clusters and {wj} are weights that sum to one. The functions pj(y,x) are joint probability density functions that relate to each of the n clusters. These functions are modeled using a decomposition into a conditional and a marginal density:
where:
pj(y|x) is a model for predicting y given x, and given that the input-output pair should be associated with cluster j on the basis of the value of x. This model might be a regression model in the simplest cases.
pj(x) is formally a density for values of x, given that the input-output pair should be associated with cluster j. The relative sizes of these functions between the clusters determines whether a particular value of x is associated with any given cluster-center. This density might be a Gaussian function centered at a parameter representing the cluster-center.
In the same way as for regression analysis, it will be important to consider preliminary data transformations as part of the overall modeling strategy if the core components of the model are to be simple regression models for the cluster-wise condition densities, and normal distributions for the cluster-weighting densities pj(x).
General versions
The basic CWM algorithm gives a single output cluster for each input cluster. However, CWM can be extended to multiple clusters which are still associated with the same input cluster.[3] Each cluster in CWM is localized to a Gaussian input region, and this contains its own trainable local model.[4] It is recognized as a versatile inference algorithm which provides simplicity, generality, and flexibility; even when a feedforward layered network might be preferred, it is sometimes used as a "second opinion" on the nature of the training problem.[5]
The original form proposed by Gershenfeld describes two innovations:
Enabling CWM to work with continuous streams of data
Addressing the problem of local minima encountered by the CWM parameter adjustment process[5]
CWM can be used to classify media in printer applications, using at least two parameters to generate an output that has a joint dependency on the input parameters.[6]
^ abProkhorov, A New Approach to Cluster-Weighted Modeling Danil V.; Lee A. Feldkamp; Timothy M. Feldkamp. "A New Approach to Cluster-Weighted Modeling"(PDF). Dearborn, MI: Ford Research Laboratory.