Skip to content

Latest commit

 

History

History

Modelatge

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

NAM

Neural Additive Models (NAMs) combine some of the expressivity of DNNs with the inherent intelligibility of generalized additive models. NAMs learn a linear combination of neural networks that each attend to a single input feature. These networks are trained jointly and can learn arbitrarily complex relationships between their input feature and the output. Our experiments on regression and classification datasets show that NAMs are more accurate than widely used intelligible models such as logistic regression and shallow decision trees. They perform similarly to existing state-of-the-art generalized additive models in accuracy, but can be more easily applied to real-world problems.

EBM

EBM is a glassbox model indented to have comparable accuracy to machine learning models such as Random Forest and Boosted Trees as well as interpretability capabilities. EBM is a Generalized Additive Model (GAM) which is very similar to linear models.EBM offers some improvements to traditional GAMs: each feature function is learned using modern machine learning technique such as bagging and boosting. Training is carried out in many iterations where each iteration consists of constructing a boosting procedure separately for each feature. As this is done with low learning rates, the order of the features does not matter. The high number of iterations aims to mitigate the effects of co-linearity in order to learn best what is the contribution of each feature to the model’s prediction. Furthermore, EBM can automatically detect and include pairwise interaction terms, this increases the accuracy of the model while preserving its explainability. Since EBM is an additive model, the contribution of each feature can be captured and visualized, and therefore enhance explainability.

name value
random_state 1377
test_size 0.15
data_n_episodes 30000
n_estimators 1
feature_step_n_inner_bags 0
n_jobs -1
min_cases_for_splits 2
binning_strategy uniform
n_splits 5
bf best first
o50 (n_estimators= 50)
i50 (feature_step_n_inner_bags = 50)
q (binning_strategy= quantile)

XGB

XGB is a scalable and highly accurate implementation of gradient boosting that pushes the limits of computing power for boosted tree algorithms, being built largely for energizing machine learning model performance and computational speed. With XGB, trees are built in parallel, instead of sequentially like GBDT. It follows a level-wise strategy, scanning across gradient values and using these partial sums to evaluate the quality of splits at every possible split in the training set.

name value
random_state 1377
n_estimators 30000
min_child_weight 1
tree_method exact
reg_lambda 0
n_jobs -1
colsample_bytree 1
subsample 1
holdout_split 0.176
early_stopping_rounds 50
d1 (max_depth=1 abductiu)
d3 (max_depth=3 teleologic)
o20 (n_estimators=20)
cols0.00001 (colsample_bytree=0.00001)
cols0.5 (colsample_bytree=0.5)
cols1 (colsample_bytree=1)

Spline

Spline Regression is a non-parametric regression technique. This regression technique divides the datasets into bins at intervals or points called knots and each bin has its separate fit.

rSpline:

name value
random_state 1377
maxk 100
nthreads 30
model_to_use bam

Spline:

name value
random_state 1377
search True
max_iter 500
n_splines 50
fit_binary_feat_as_factor_term False
b (fit_binary_feat_as_factor_term=True)

Reference

Agarwal, R., Melnick, L., Frosst, N., Zhang, X., Lengerich, B., Caruana, R., & Hinton, G. E. (2021). Neural additive models: Interpretable machine > learning with neural nets. Advances in Neural Information Processing Systems, 34.

Chang, C. H., Tan, S., Lengerich, B., Goldenberg, A., & Caruana, R. (2021, August). How interpretable and trustworthy are gams?. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (pp. 95-105).