Core Concepts
Ensemble models, particularly stacking ensembles with non-negative constraints, consistently outperformed individual candidate models in predicting milk quality traits and animal diet from mid-infrared spectroscopic data.
Abstract
The study evaluated the performance of various statistical machine learning models, including dimension reduction methods, regularized regression, kernel methods, neural networks, and tree-based ensembles, in two chemometric data analysis challenges:
Regression challenge: Predicting 14 milk quality traits from mid-infrared (MIR) milk spectra. The datasets contained 622 samples.
Classification challenge: Predicting animal diet (grass-only, grass-white clover, or total mixed ration) from MIR milk spectra. The dataset contained 3,275 samples.
The models were trained and evaluated using random splits of the data, with further cross-validation for tuning hyperparameters. A linear mixed effects model was used to statistically analyze the prediction performance metrics (RMSE for regression, accuracy for classification).
The key findings are:
Stacking ensembles, particularly those with non-negative constraints on the meta-learner coefficients, consistently outperformed the best individual candidate models across both datasets.
In the regression challenge, the stacking ensemble reduced the average RMSE from 0.85 to 0.84 compared to the best candidate model (PLS).
In the classification challenge, the stacking ensemble increased the average accuracy from 0.78 to 0.81 compared to the best candidate model (LDA).
The improvement in performance, while modest, highlights the value of ensemble methods in leveraging the strengths of diverse candidate models to improve prediction accuracy.
The statistical analysis showed that the variability in prediction performance across random data splits was much larger than the variability across different algorithms in the regression dataset, emphasizing the importance of robust experimental design.
Overall, the results demonstrate that ensemble models, particularly stacking ensembles, can be a valuable tool for improving prediction from spectroscopic data compared to relying on a single candidate model.
Stats
The regression dataset contained 14 milk quality traits, including:
Rennet coagulation time (RCT)
Curd-firming time (k20)
Curd firmness at 30 and 60 min (a30, a60)
Casein micelle size (CMS)
pH
Heat stability
Casein composition (αS1-CN, αS2-CN, β-CN, κ-CN)
Whey protein composition (α-LA, β-LG A, β-LG B)
The classification dataset contained 3,275 milk samples from cows fed one of three diet regimens: grass-only (GRS), grass-white clover (CLV), and total mixed ration (TMR).
Quotes
"Stacking ensembles offer an elegant way of combining predictions of different candidate models."
"While there was some variability in algorithm performance for different traits, the LME model showed the stacking ensembles significantly outperformed model averaging (Ens_MA) and majority voting (Ens_maj_vote) ensembles in our application."
"Stacking ensemble model implementations can increase diversity of predictions by using different hyper-parameter settings, however we chose to blend the predictions of tuned models."