What is the Naive Bayes classification? Name the method and goal
Goal: Find the class C that maximizes the conditional probability
P(C|A) -> Probability of class C given Attribute A
When is the application of the Bayes Theorem useful?
Bayes Theorem: P(C|A) = (P(C|A)*P(C)) / P(A)
Useful situations:
Whats the difference between prior and posterior probability (Bayes Theorem)
How do you apply Bayes Theorem to the classification task?
How to estimate the prior probability P(C)
Explain the independence assumption and its implications for estimating P(A|C) for the Naive Bayes
How to estimate the probabilities P(Ai|Cj)?
What are the names of the parts of the Bayes Theorem?
How to normalize the likelihood of the two classes (PC|A) ?
How should you handle numerical attributes when applying the Naive Bayes?
Option 1) Discretize the numerical attributes (apply categories to the numerical values)
Option2) Assume that numerical attributes have a normal distribution given the class
Which distribution parameters can be estimated from the training data?
- Standard deviation
How to handle missing values in the training data?
How to handle missing values in the test data?
Attribute will be omitted from calculation
Explain the zero-frequency problem
Solution: Laplace Estimator Add 1 to the count of every attribute class combination
LaPlace: P(Ai|C) = (Nic + 1) / (Nc + |Vi|)
Vi = number of values for the attribute in the training set
What are the characteristics of Naive bayes?
What is the technical advantage of Naive Bayes?
For which problems can you use Support Vector Machines ?
- Examples described by continuous attributes
When doe SVMs achieve good results?
How does SVMs work?
How does a SVM find the best Hyperplane?
How to deal with noise points in SVMs?
Goal: Have a large margin without ignoring too many data points
How to handle decision boundaries that are not linear with SVMs?
What are the characteristics of SVMs ?
What are the application areas for SVMs?