PyCaret Model Training Report

Setup & Best Model
Best Model Plots
Feature Importance
Explainer

Setup Parameters

ParameterValue
target PCR
session_id 42
index False
If you want to know all the experiment setup parameters, please check the PyCaret documentation for the classification/regression exp function.

Best Model: LGBMClassifier

ParameterValue
boosting_type gbdt
class_weight None
colsample_bytree 1.0
importance_type split
learning_rate 0.1
max_depth -1
min_child_samples 20
min_child_weight 0.001
min_split_gain 0.0
n_estimators 100
n_jobs -1
num_leaves 31
objective None
random_state 42
reg_alpha 0.0
reg_lambda 0.0
subsample 1.0
subsample_for_bin 200000
subsample_freq 0

Comparison Results on the Cross-Validation Set

Model Accuracy ROC-AUC Recall Prec. F1 Kappa MCC PR-AUC-Weighted TT (Sec)
Light Gradient Boosting Machine 0.715 0.6000 0.6500 0.6917 0.6357 0.4380 0.4748 0.6822 0.228
Logistic Regression 0.670 0.6500 0.7500 0.6167 0.6633 0.3478 0.3742 0.7144 0.331
Ridge Classifier 0.670 0.6167 0.7500 0.6167 0.6633 0.3478 0.3742 0.0000 0.180
Naive Bayes 0.650 0.6333 0.6833 0.5917 0.6257 0.2969 0.3112 0.6978 2.694
Quadratic Discriminant Analysis 0.610 0.6333 0.4667 0.5333 0.4733 0.2256 0.2488 0.7033 0.158
Linear Discriminant Analysis 0.605 0.6000 0.7000 0.5900 0.6105 0.2372 0.2577 0.6594 0.110
CatBoost Classifier 0.595 0.6167 0.6167 0.5500 0.5600 0.2165 0.2207 0.6861 12.075
Extra Trees Classifier 0.590 0.6000 0.5833 0.5000 0.5300 0.2103 0.2167 0.6811 0.775
SVM - Linear Kernel 0.585 0.6500 0.5333 0.4667 0.4521 0.1429 0.1690 0.0000 0.217
K Neighbors Classifier 0.565 0.6292 0.5000 0.5750 0.5057 0.1413 0.1469 0.6717 0.685
Random Forest Classifier 0.555 0.5667 0.5833 0.5167 0.5233 0.1524 0.1540 0.6211 0.847
Dummy Classifier 0.540 0.5000 0.0000 0.0000 0.0000 0.0000 0.0000 0.4600 0.165
Ada Boost Classifier 0.510 0.4417 0.5667 0.4650 0.4971 0.0656 0.0275 0.5819 0.645
Decision Tree Classifier 0.495 0.5000 0.4333 0.4333 0.4133 0.0049 0.0040 0.5483 0.329
Gradient Boosting Classifier 0.475 0.4333 0.4500 0.4000 0.4033 -0.0033 -0.0239 0.5800 0.643
Extreme Gradient Boosting 0.460 0.4833 0.4333 0.3333 0.3667 -0.0489 -0.0537 0.6281 0.422

Results on the Test Set for the best model

Model Accuracy ROC-AUC Recall Prec. F1 Kappa MCC PR-AUC-Weighted
Light Gradient Boosting Machine 0.8095 0.7454 0.6667 0.8571 0.75 0.6 0.6124 0.6799

Best Model Plots on the testing set

Confusion_matrix

confusion_matrix

Auc

auc

Threshold

threshold

Pr

pr

Error

error

Class_report

class_report

Learning

learning

Calibration

calibration

Vc

vc

Dimension

dimension

Manifold

manifold

Rfe

rfe

Feature

feature

Feature_all

feature_all

PyCaret Feature Importance Report

Feature importance analysis from atrained Random Forest

Use gini impurity forcalculating feature importance for classificationand Variance Reduction for regression

tree_importance

SHAP Summary from a trained lightgbm

shap_summary