The concept of stacking (short for Stacked Generalization) applies to the area of predictive data mining to combine the predictions from multiple models. It is particularly useful when the types of models included in the project are very different.

Suppose your data mining project includes tree classifiers, such as
C&RT and CHAID,
linear discriminant analysis (e.g., see GDA),
and STATISTICA
Automated Neural Networks. Each computes predicted classifications
for a cross-validation sample,
from which overall goodness-of-fit statistics (e.g., misclassification
rates) can be computed. Experience has shown that combining the predictions
from multiple methods often yields more accurate predictions than can
be derived from any one method (e.g., see Witten and Frank, 2000). In
stacking, the predictions from different classifiers are used as input
into a meta-learner, which attempts
to combine the predictions to create a final best predicted classification.
So, for example, the predicted classifications from the tree classifiers,
linear model, and the neural network classifier(s) can be used as input
variables into a neural network

Other methods for combining the prediction from multiple models or methods (e.g., from multiple data sets used for learning) are Boosting and Bagging (Voting). See also, Data Mining.