Skip to content

Latest commit

 

History

History
101 lines (72 loc) · 4.88 KB

README.md

File metadata and controls

101 lines (72 loc) · 4.88 KB

Credit_Risk_Analysis

8-Credit Risk

Challenge Overview

Purpose:

The purpose of this analysis is to predict credit risk with machine learning models by using different techniques to train and evaluate models with unbalanced classes.

  • Resampling Models

    • Over-sampling method: using the RandomOverSampler & SMOTE algorithms
    • Under-sampling method: using the ClusterCentroids algorithm
    • Combination Sampling (a combinatorial approach of oversampling and undersampling): using the SMOTEENN algorithm
  • Ensemble Classifiers method

    • using BalancedRandomForestClassifier & EasyEnsembleClassifier algorithms

Resources

  • Software:

    • Jupyter Notebook 6.4.6
    • Machine Learning
      • Python
        • scikit-learn library
        • imbalanced-learn library
  • Data source:

Results

1 Random_Oversampling

  • Random Oversampling
    • the balanced accuracy score: 66%
    • the precision and recall scores (high_risk):
      • The sensitivity/recall (71%) is more than the precision (1%)
      • there are many false positives (predicted high risk but actually low risk)
      • making this a poor algorithm for this dataset

2 SMOTE

  • Synthetic Minority Oversampling Technique (SMOTE)
    • the balanced accuracy score: 66%
    • the precision and recall scores (high_risk):
      • The sensitivity/recall (63%) is more than the precision (1%)
      • there are many false positives (predicted high risk but actually low risk)
      • making this a poor algorithm for this dataset

3 Undersampling_Cluster_Centroid

  • Cluster Centroid Undersampling
    • the balanced accuracy score: 54%
    • the precision and recall scores (high_risk):
      • The sensitivity/recall (69%) is more than the precision (1%)
      • there are many false positives (predicted high risk but actually low risk)
      • making this a poor algorithm for this dataset

4 Combination

  • Combination Sampling With SMOTEENN
    • The balanced accuracy score: 64%
    • the precision and recall scores (high_risk):
      • The sensitivity/recall (72%) is more than the precision (1%)
      • there are many false positives (predicted high risk but actually low risk)
      • making this a poor algorithm for this dataset

5 Balanced_Random_Forest_Classifier

  • BalancedRandomForestClassifier
    • The balanced accuracy score: 79%
    • the precision and recall scores (high_risk):
      • The sensitivity/recall (70%) is more than the precision (3%)
      • there are many false positives (predicted high risk but actually low risk)
      • making this a poor algorithm for this dataset
    • The total_rec_prncp and total_pymnt of the credit dataset are the more relevant features or columns

6 Easy_Ensemble_AdaBoost_Classifier

  • EasyEnsembleClassifier
    • The balanced accuracy score: 93%
    • the precision and recall scores (high_risk):
      • The sensitivity/recall (92%) is more than the precision (9%)
      • there are many false positives (predicted high risk but actually low risk)
      • making this a poor algorithm for this dataset

Summary:

Eventhough the EasyEnsembleClassifier algorithm has the highest balanced accuracy score, 93%, this algorithm and the other algorithms still are not good enough to determine if a credit is high risk because the sensitivity/recall is very high, while the precision is very low. It indicates that there are many false positives (predicted high risk but actually low risk). Clearly, they are not useful algorithms for this dataset. Therefore I would not recommend that they be used to predict credit risk. Maybe a dataset with more obsevations would produce a better result.