Skip to content

yangshiteng/DS06-Naive-Bayes

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 

Repository files navigation

1. Video Tutorial

  • The intial guess can be any probability that we want, but a common guess is estimated from the training data
  • For example, since 8 of the 12 messages are normal messages, our initial guess will be 0.67
  • The initial guess that we observe a Normal messages is called a Prior Probability image
  • The thing that makes Naive Bayes so naive is that it treats all word orders the same image
  • Every language has grammar rules and common phrases, but naive bayes ignores all of that stuff. Instead, Naive Bayes treats language like it just a bag full of words and each message is random handful of them
  • Naive Bayes ignores all the rules because keeping track of every single reasonable phrase in a language would be impossbile
  • Gausian Naive Bayes is named after the Gaussian distributions that represent the data in the Training dataset
  • When we get really small numbers, it's good idea to take the log() of everything to prevent something called Underflow
  • The Underflow is every computer has a limit to how close a number can get to 0 before it can no longer accurately keep track of that number. When a number gets smaller than that limit, we run into Underflow problems and errors occur. So, we use log() function to avoid Underflow
  • The prior probability is found from training dataset
  • For each continuous variable among the same label, we calculate the mean and standard deviation for the parameter values of Gaussian normal distribution
  • Then, we calculate the product of prior probability and likelihoods of same label from Gaussian normal distribution, and compare the score

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published