The 2006 paper defined model ensembles. Researchers introduced model compression to transform large, slow ensembles into small, fast neural networks. Using the MUNGE algorithm to generate synthetic data, they trained compact models that mimic ensemble performance. This achieves a 1000x reduction in size and latency. Source: Title: Model Compression Institution: Cornell University URL: https://www.cs.cornell.edu/~caruana/compression.kdd06.pdf