Neural Additive Models (NAMs) combine some of the expressivity of DNNs with the inherent intelligibility of generalized additive models. NAMs learn a linear combination of neural networks that each attend to a single input feature. These networks are trained jointly and can learn arbitrarily complex relationships between their input feature and the output. Our experiments on regression and classification datasets show that NAMs are more accurate than widely used intelligible models such as logistic regression and shallow decision trees. They perform similarly to existing state-of-the-art generalized additive models in accuracy, but can be more easily applied to real-world problems.
EBM is a glassbox model indented to have comparable accuracy to machine learning models such as Random Forest and Boosted Trees as well as interpretability capabilities. EBM is a Generalized Additive Model (GAM) which is very similar to linear models.EBM offers some improvements to traditional GAMs: each feature function is learned using modern machine learning technique such as bagging and boosting. Training is carried out in many iterations where each iteration consists of constructing a boosting procedure separately for each feature. As this is done with low learning rates, the order of the features does not matter. The high number of iterations aims to mitigate the effects of co-linearity in order to learn best what is the contribution of each feature to the model’s prediction. Furthermore, EBM can automatically detect and include pairwise interaction terms, this increases the accuracy of the model while preserving its explainability. Since EBM is an additive model, the contribution of each feature can be captured and visualized, and therefore enhance explainability.
name | value |
---|---|
random_state | 1377 |
test_size | 0.15 |
data_n_episodes | 30000 |
n_estimators | 1 |
feature_step_n_inner_bags | 0 |
n_jobs | -1 |
min_cases_for_splits | 2 |
binning_strategy | uniform |
n_splits | 5 |
bf | best first |
o50 | (n_estimators= 50) |
i50 | (feature_step_n_inner_bags = 50) |
q | (binning_strategy= quantile) |
XGB is a scalable and highly accurate implementation of gradient boosting that pushes the limits of computing power for boosted tree algorithms, being built largely for energizing machine learning model performance and computational speed. With XGB, trees are built in parallel, instead of sequentially like GBDT. It follows a level-wise strategy, scanning across gradient values and using these partial sums to evaluate the quality of splits at every possible split in the training set.
name | value |
---|---|
random_state | 1377 |
n_estimators | 30000 |
min_child_weight | 1 |
tree_method | exact |
reg_lambda | 0 |
n_jobs | -1 |
colsample_bytree | 1 |
subsample | 1 |
holdout_split | 0.176 |
early_stopping_rounds | 50 |
d1 | (max_depth=1 abductiu) |
d3 | (max_depth=3 teleologic) |
o20 | (n_estimators=20) |
cols0.00001 | (colsample_bytree=0.00001) |
cols0.5 | (colsample_bytree=0.5) |
cols1 | (colsample_bytree=1) |
Spline Regression is a non-parametric regression technique. This regression technique divides the datasets into bins at intervals or points called knots and each bin has its separate fit.
rSpline:
name | value |
---|---|
random_state | 1377 |
maxk | 100 |
nthreads | 30 |
model_to_use | bam |
Spline:
name | value |
---|---|
random_state | 1377 |
search | True |
max_iter | 500 |
n_splines | 50 |
fit_binary_feat_as_factor_term | False |
b | (fit_binary_feat_as_factor_term=True) |
Agarwal, R., Melnick, L., Frosst, N., Zhang, X., Lengerich, B., Caruana, R., & Hinton, G. E. (2021). Neural additive models: Interpretable machine > learning with neural nets. Advances in Neural Information Processing Systems, 34.
Chang, C. H., Tan, S., Lengerich, B., Goldenberg, A., & Caruana, R. (2021, August). How interpretable and trustworthy are gams?. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (pp. 95-105).