Performance metrics (classification report) for each model on the balanced dataset:
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Not Fraud | 0.93 | 0.99 | 0.96 | 69 |
Fraud | 0.99 | 0.93 | 0.96 | 73 |
Accuracy | 0.96 | 142 | ||
Macro Avg | 0.96 | 0.96 | 0.96 | 142 |
Weighted Avg | 0.96 | 0.96 | 0.96 | 142 |
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Not Fraud | 0.86 | 1.00 | 0.93 | 69 |
Fraud | 1.00 | 0.85 | 0.92 | 73 |
Accuracy | 0.92 | 142 | ||
Macro Avg | 0.93 | 0.92 | 0.92 | 142 |
Weighted Avg | 0.93 | 0.92 | 0.92 | 142 |
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Not Fraud | 0.91 | 1.00 | 0.95 | 69 |
Fraud | 1.00 | 0.90 | 0.95 | 73 |
Accuracy | 0.95 | 142 | ||
Macro Avg | 0.95 | 0.95 | 0.95 | 142 |
Weighted Avg | 0.96 | 0.95 | 0.95 | 142 |
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Not Fraud | 0.93 | 0.94 | 0.94 | 69 |
Fraud | 0.94 | 0.93 | 0.94 | 73 |
Accuracy | 0.94 | 142 | ||
Macro Avg | 0.94 | 0.94 | 0.94 | 142 |
Weighted Avg | 0.94 | 0.94 | 0.94 | 142 |
Precision | Recall | F1-Score | Support | |
---|---|---|---|---|
Not Fraud | 0.93 | 0.97 | 0.95 | 69 |
Fraud | 0.97 | 0.93 | 0.95 | 73 |
Accuracy | 0.95 | 142 | ||
Macro Avg | 0.95 | 0.95 | 0.95 | 142 |
Weighted Avg | 0.95 | 0.95 | 0.95 | 142 |