Predictive Analytics (with R)
Input File (.csv only):
Condition(s) for Subset:
Categorical Variable
Categorical Variable
Categorical Variable
Categorical Variable
Dependent Variable:
Change to Categorical Variable
Independent Variable(s):
From Top From Bottom
% of Training Data
Include Prediction for Validation Data with Testing Data
Linear Regression
Lasso Regression (i.e., L1 Regularization)
Ridge Regression (i.e., L2 Regularization)
Elastic Net Regression with alpha =
Decision Tree (with Levels)
Random Forest with trees with maximum nodes
Gradient Boosting with maximum depth = , number of trees =
XGBoost with maximum depth = , number of trees =
LightGBM with maximum depth = , number of trees =
Support Vector Machine with kernel = , cost = , gamma =
Linear Discriminant Analysis
Logistic Regression for binary classification with cut-off =
Decision Tree with Levels
Random Forest with trees with maximum nodes
Adaptive Boosting (AdaBoost) with number of trees =
Gradient Boosting for binary classification with cut-off = , maximum depth = , number of trees =
XGBoost for binary classification with cut-off = , maximum depth = , number of trees =
LightGBM for binary classification with cut-off = , maximum depth = , number of trees =
Support Vector Machine with kernel = , cost = , gamma =
Naive Bayes
k-Nearest Neighbor with k = (Note: numeric independent variables only)
Neural Network (with hidden layers)
Mean Squared Error (MSE)
Mean Absolute Error (MAE)
Precision
Recall or Sensitivity
Specificity
F1 Score
Condition(s) for Subset:
Categorical Variable
Categorical Variable
Categorical Variable
Categorical Variable
Dependent Variable:
Change to Categorical Variable
Independent Variable(s):
From Top From Bottom
% of Training Data
Include Prediction for Validation Data with Testing Data
Linear Regression
Lasso Regression (i.e., L1 Regularization)
Ridge Regression (i.e., L2 Regularization)
Elastic Net Regression with alpha =
Decision Tree (with Levels)
Random Forest with trees with maximum nodes
Gradient Boosting with maximum depth = , number of trees =
XGBoost with maximum depth = , number of trees =
LightGBM with maximum depth = , number of trees =
Support Vector Machine with kernel = , cost = , gamma =
Linear Discriminant Analysis
Logistic Regression for binary classification with cut-off =
Decision Tree with Levels
Random Forest with trees with maximum nodes
Adaptive Boosting (AdaBoost) with number of trees =
Gradient Boosting for binary classification with cut-off = , maximum depth = , number of trees =
XGBoost for binary classification with cut-off = , maximum depth = , number of trees =
LightGBM for binary classification with cut-off = , maximum depth = , number of trees =
Support Vector Machine with kernel = , cost = , gamma =
Naive Bayes
k-Nearest Neighbor with k = (Note: numeric independent variables only)
Neural Network (with hidden layers)
Mean Squared Error (MSE)
Mean Absolute Error (MAE)
Precision
Recall or Sensitivity
Specificity
F1 Score
Copyright © 2019 Abhimanyu Gupta. All rights reserved.