Lecture 3: Probability and Naive Bayes Classification

Today’s Topic

Using Bayes’ theorem for machine learning. You should do some revision on the use of Bayes’ theorem in general. In this lecture you will look at how to use Bayes’ theorem to build a spam detector. One important idea to take away from this lecture is that there are a variety of ways implementing spam detection: in particular there are different feature models that you can use that give you different ways of calculating the relevant probabilities. It is important that you understand the difference between the different ways of implementing spam detection.

Reading Guide

What should I know by the end of this lecture?

  • How do I use Bayes’ theorem?
  • How do I use Bayes’ theorem to build a simple spam detector that only uses one word?
  • What is independence assumption in the naive Bayes’ algorithm?
  • What are the various models estimating the probabilities for spam detection?
  • What is Laplacian smoothing and how does it work in the context of spam detection?
  • When do you need to use logarithms to calculate the relevant probabilities, and how do you use them?
Previous
Next