Skip to content

KaziTanvir/Stroke-Prediction-comparative-analysis-on-various-machine-learning-models

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Stroke Prediction comparative analysis on various machine learning models


## What is a Stroke?

A stroke, sometimes called a brain attack, occurs when something blocks blood supply to part of the brain or when a blood vessel in the brain bursts. In either case, parts of the brain become damaged or die. A stroke can cause lasting brain damage, long-term disability, or even death.

Why we need stroke prediction analysis?

This delay is particularly concerning when we consider that Ischemic stroke patients suffer the loss of two million brain cells every minute until blood flow is restored. In this context, an earlier diagnosis can mean faster treatment, with one Harvard Medical study breaking down the benefits as follows:

“For every 15-minute acceleration of time to…treatment, found:

  • 4% lower risk of in-hospital death
  • 4% better odds of walking independently after leaving the hospital
  • 3% better odds of being sent home instead of to an institution
  • 4% lower odds of a brain hemorrhage.”

Models Used:

  • Gaussian Naive Bayes :

    Naive Bayes methods are a set of supervised learning algorithms based on applying Bayes’ theorem with the “naive” assumption of conditional independence between every pair of features given the value of the class variable. Bayes’ theorem states the following relationship, given class variable y and dependent feature vector x1 through xn

  • Decision Tree :

    Decision Trees (DTs) are a non-parametric supervised learning method used for classification and regression. The goal is to create a model that predicts the value of a target variable by learning simple decision rules inferred from the data features. A tree can be seen as a piecewise constant approximation.

  • Random Forest Classifier :

    A random forest is a meta estimator that fits a number of classifying decision trees on various sub-samples of the dataset and uses averaging to improve the predictive accuracy and control over-fitting.

  • Multi-Layer Perceptron :

    MLPClassifier stands for Multi-layer Perceptron classifier which in the name itself connects to a Neural Network. Unlike other classification algorithms such as Support Vectors or Naive Bayes Classifier, MLPClassifier relies on an underlying Neural Network to perform the task of classification.

Results

  • Gaussian Naive Bayes : 97.57 %
  • Decision Trees : 96.00 %
  • Random Forest Classifier : 98.11 %
  • Multi-Layer Preceptron : 96.68 %

Conclusion

Random Forest Classifier works best on the said data compared to the other models.