Accuracy Analysis

Accuracy of TRUST vs Alternative Models

The accuracy of TRUST relative to other machine learning models is tested in 60 datasets: 20 synthetic (16 specially designed to challenge TRUST) and 40 real-world benchmark datasets (e.g. Abalone, Boston, Diamonds, Mpg, or Wine).

The alternative models are:

The first 4 are usually regarded as interpretable models, while the 5th one is a black box.

Global accuracy results across 60 datasets Accuracy results by dataset group

The first plot above shows that overall TRUST has comparable accuracy to a top-performing model like Random Forest (or slightly better), and tends to outperform other interpretable models.

The second plot shows a breakdown by dataset group, illustrating that TRUST is the only model that performs well under all conditions.

Lastly, other models have been included in a more comprehensive analysis (e.g. off-the-shelf XGBoost, deep neural networks or splines) and a similar conclusion holds.