CS464 Machine Learning: Homework 1 solution

$24.99

Original Work ?
Category: You will Instantly receive a download link for .ZIP solution file upon Payment

Description

5/5 - (6 votes)

1 Slot Machine [10 pts]
You and your friend are playing the slot machine in a casino. Having played on two separate machines for
a while, you decide to switch machines to measure for differences in luck. The wins/losses of you and your
friend for each machine are tabulated below.
Machine 1 Wins Losses Machine 2 Wins Losses
You 40 60 You 212 828
Friend 25 75 Friend 18 72
Assuming that the outcome of playing the slot machine is independent of its history and that of the other
machine, answer the following questions:
Question 1.1 [2 points] If the outcome of a game is a success, what’s the probability that it was played
using machine 1?
Question 1.2 [5 points] What is the winning probability of you and your friend for each of the machines?
Compare your winning probability with your friend’s on different machines, who is more likely to win on
each machine?
Question 1.3 [3 points] Suppose you did not keep track of the wins/losses for each machine, but only of
the total number of wins/loses for the two machines. In this case, estimate the overall winning probability
of you and your friend in the casino (assume that there are only two slot machines in the casino). Who is
more likely to win?
1
2 Conditional Independence [10 pts]
Let’s say we have two dice: one is blue and the other one is red. You roll the dices and try to predict the
outcome. Let’s say the outcome of the blue die is b, and the outcome of the red is r. You can understand
that the outcome of the dices are independent of each other.
Let’s say we have an oracle who helps you to predict the probability of any outcome.
Question 2.1 [4 pts] Oracle gives you the following information about the outcomes:
C = ‘b is not equal to 6, and r is not equal to 1 ’
What is the probability of b = 1 and r = 3, which is P(b=1,r=3 | C)?
Question 2.2 [3 pts] Independent from the previous information, oracle gives you another information
in this time:
D = ‘sum of the outcomes (b+r) is an even number ’
What is the probability of b = 3 and r = 5, which is P(b=3,c=5 | D) ?
Question 2.3 [3 pts] What is the difference between the information C and D above? Explain your
reasoning in terms of conditional probability.
3 Packets [10 pts]
The Poisson distribution is a useful discrete distribution which can be used to model the number of occurrences of something per unit time. For example, in networking, the number of packets to arrive in a given time
window is often assumed to follow a Poisson distribution. If X is Poisson distributed, i.e. X ∼ P oisson(λ),
its probability mass function takes the following form:
P (X = x | λ) = λ
x
e
−λ
x!
Assume now we have n identically and independently drawn data points from P oisson(λ) : D = {x1, . . . , xn}.
Assume that prior distribution for λ is N(0, β2
), derive an expression for maximum a posterior (MAP)
estimate of λ.
4 Building a Spam Classifier with Naive Bayes [50 pts]
Your job is to build a spam classifier that can accurately predict whether an email is spam or not. The
questions summarizes the model, therefore, please read all questions before starting coding.
Dataset
Your dataset is a preprocessed and modified subset of the Ling-Spam Dataset [1]. It is based on 960 real
email messages from a linguistics mailing list. Emails have been preprocessed in the following ways:
• Stop word removal: Words like “and”,“the”, and “of”, are very common in all English sentences and
are therefore not very predictive in deciding spam/nonspam status. These words have been removed
from the emails.
2
• Lemmatization: Words that have the same meaning but different endings have been adjusted so
that they all have the same form. For example, “include”, “includes,” and “included,” would all be
represented as “include.” All words in the email body have also been converted to lower case.
• Removal of non-words: Numbers and punctuation have both been removed. All white spaces (tabs,
newlines, spaces) have all been trimmed to a single space character
The data has been already split into two subsets: a 700-email subset for training and a 260-email subset for
testing (consider this as your validation set and imagine there is another test set which is not given to you).
The features have been generated for you. You will use the following files:
• question4-train-features.txt
• question4-train-labels.txt
• question4-test-features.txt
• question4-test-labels.txt
The files that ends with features.txt contains the features and the files ending with labels.txt
contains the ground truth labels.
In the feature files each row contains the feature vector for an email. The j-th term in a row i is the
number of occurrences of the j-th vocabulary word in the i-th email. The size of the vocabulary is 2500.
The label files include the ground truth label for the corresponding email, the order of the emails (rows) are
the same as the features file. That is the i-th row in the files corresponds to the same email document. The
labels are indicated by 1 or 0, 1 stands for a spam email and 0 stands for the nonspam email.
Bag-of-Words Representation and Multinomial Naive Bayes Model
Recall the bag-of-words document representation makes the assumption that the probability that a word
appears in email is conditionally independent of the word position given the class of the email. If we have a
particular email document Di with ni words in it, we can compute the probability that Di comes from the
class yk as:
P (Di
| Y = yk) = P (X1 = x1, X2 = x2, .., Xni = xni
| Y = yk) = Yni
j=1
P (Xj = xj | Y = yk) (4.1)
In Eq. (4.1), Xj represents the j
th position in email Di and xj represents the actual word that appears
in the j
th position in the email, whereas ni represents the number of positions in the email. As a concrete
example, we might have the first email document (D1) which contains 200 words (n1 = 200). The document
might be of spam email (yk = 1) and the 15th position in the email might have the word “office” (xj =
“office”).
In the above formulation, the feature vector X~ has a length that depends on the number of words in the
email ni
. That means that the feature vector for each email will be of different sizes. Also, the above formal
definition of a feature vector ~x for a email says that xj = k if the j-th word in this email is the k-th word in
the dictionary. This does not exactly match our feature files, where the j-th term in a row i is the number
of occurrences of the j-th dictionary word in that email i. As shown in the lecture slides, we can slightly
change the representation, which makes it easier to implement:
P (Di
| Y = yk) = Y
V
j=1
P (Xj | Y = yk)
twj
,i (4.2)
,where V is the size of the vocabulary, Xj represents the appearing of the j-th vocabulary word and twj ,i
denotes how many times word wj appears in email Di
. As a concrete example, we might have a vocabulary
of size of 1309, V = 1309. The first email (D1) might be spam (yk = 1) and the 80-th word in the vocabulary,
w80, is “click” and tw80,1 = 2, which says the word “click” appears 2 times in email D1. Contemplate on
why these two models (Eq. (4.1) and Eq. (4.2)) are equivalent.
3
In the classification problem, we are interested in the probability distribution over the email classes (in this
case Spam and Nonspam) given a particular email Di
. We can use Bayes Rule to write:
P (Y = yk|Di) =
P (Y = yk)
QV
j=1 P (Xj | Y = y)
twj
,i
P
k P (Y = yk)
QV
j=1 P (Xj | Y = yk)
twj
,i
(4.3)
Note that, for the purposes of classification, we can actually ignore the denominator here and write:
P (Y = yk|Di) ∝ P (Y = yk)
Y
V
j=1
P (Xj | Y = y)
twj
,i (4.4)
yˆi = arg max
yk
P (Y = yk | Di) = arg max
yk
P (Y = yk)
Y
V
j=1
P (Xj | Y = yk)
twj
,i (4.5)
Question 4.1 [2 points] Why it is that we can ignore the denominator?
Probabilities are floating point numbers between 0 and 1, so when you are programming it is usually not a
good idea to use actual probability values as this might cause numerical underflow issues. As the logarithm
is a strictly monotonic function on [0,1] and all of the inputs are probabilities that must lie in [0,1], it does
not have an affect on which of the classes achieves a maximum. Taking the logarithm gives us:
yˆi = arg max
y

log P (Y = yk) +X
V
j=1
twj ,i ∗ log P (Xj | Y = yk)

 (4.6)
, where ˆyi
is the predicted label for the i-th example.
Question 4.2 [3 points] If the the ratio of the classes in a dataset is close to each other, it is a called “balanced” class distribution if not it is skewed. What is the percentage of spam emails in the train.labels.txt.
Is the training set balanced or skewed towards a one of the classes?
The parameters to learn and their MLE estimators are as follows:
θj | y=0 ≡ P
Tj,y=0
V
j=1 Tj,y=0
θj | y=1 ≡ P
Tj,y=1
V
j=1 Tj,y=1
πy=1 ≡ P (Y = 1) = N1
N
• Tj,0 is the number of occurrences of the word j in nonspam emails in the training set including the
multiple occurrences of the word in a single email.
• Tj,1 is the number of occurrences of the word j in spam emails in the training set including the multiple
occurrences of the word in a single email.
• N1 is the number of spam emails in the training set.
• N is the total number of emails in the training set.
• πy=1 estimates the probability that any particular email will be a spam email.
• θj | y=0 estimates the probability that a particular word in a nonspam email will be the j-th word of
the vocabulary, P (Xj | Y = 0)
• θj | y=1 estimates the probability that a particular word in a spam email will be the j-th word of the
vocabulary, P (Xj | Y = 1)
Question 4.3 (Coding) [20 points] Train a Naive Bayes classifier using all of the data in the training set (
train-features.txt and train-labels.txt). Test your classifier on the test data (test-features.txt
and test-labels.txt, and report the testing accuracy as well as how many wrong predictions were made.
In estimating the model parameters use the above MLE estimator. If it arises in your code, define 0∗log 0 = 0
4
(note that a ∗ log 0 is as it is, that is -inf ). In case of ties, you should predict “non-spam”. In the written
part of your report what your test set accuracy is? What did your classifier end up predicting? Why is using
the MLE estimate is a bad idea in this situation?
Question 4.4 (Coding) [5 points] Extend your classifier so that it can compute an MAP estimate of θ
parameters using a fair Dirichlet prior. This corresponds to additive smoothing. The prior is fair in the
sense that it “hallucinates” that each word appears additionally α times in the train set.
θj | y=0 ≡ P
Tj,y=0+α
V
j=1 Tj,y=0+α∗V
θj | y=1 ≡ P
Tj,y=1+α
V
j=1 Tj,y=1+α∗V
πy=1 ≡ P (Y = 1) = N1
N
For this question set α = 1. Train your classifier using all of the training set and have it classify all of
the test set and report test-set classification accuracy.
Question 4.5 (Coding) [10 points] Calculate Mutual information and rank features. Write indices and
mutual information scores of top 10 best features from highest mutual information to the lowest. (Hint:
You may want to refer mutual information estimation model provided in Stanford book chapter provided in
reference 3 below or provided in lecture slides.)
Question 4.6 (Coding) [10 points] Remove features one-by-one from the least informative one. Keep
removing until all features are removed. Plot test-set accuracy as a function of removed number of features.
5 Gausian Naive Bayes Classifier [20 pts]
In this question, you will implement the a Gausian Naive Bayes classifier when the features. The data is
consists of continious expression values of 5 different genes measured for each cancer patient. Class labels
are three different classes, which are kidney, breast and colon cancers. To ease the notation, we denoted the
class labels with discrete numbers as follows:
1 = Kidney Cancer, 2 = Breast Cancer, 3 = Colon Cancer
• question5-train.txt
• question5-test.txt
The first two columns of the data consists of the features, and the third column corresponds to class
label.
Fitting Gaussian Distribution to Data
In this question, you will fit Gaussian distributions to each class conditional probabilities. The class conditional Gausian distributions for each feature i are as follows:
p(Xi = x|Y = yk) = 1
σik


e
−(x−µik)
2
/2σ
2
ik
5
Figure 1: Maximum Likelihood Estimates of Parameter, Source:[4]
Maximum Likelihood Estimation of Parameters
To use the Gaussian distribution for class conditional probabilities, you need to find the parameters of the
Gaussian. To ease your work in this question, MLE estimation of µik and σik are provided below:
Question 5.1 [10 pts] Train your Gaussian Naive Bayes classifier, and estimate the parameters using
MLE and report them.
Question 5.2 [10 pts] In both train and test data, report the confusion matrices for the three different
cancer types.
References
1. Liang Spam dataset. http://csmining.org/index.php/ling-spam-datasets.html
2. ”On Discriminative vs. Generative Classifiers: A comparison of logistic regression and Naive Bayes” by
Andrew Ng and Michael I. Jordan.
3. Manning, C. D., Raghavan, P., and Schutze, H. (2008). Introduction to information retrieval. New York:
Cambridge University Press.
http://nlp.stanford.edu/IR-book/html/htmledition/mutual-information-1.html
4. CMU Lecture Notes.
http://www.cs.cmu.edu/˜epxing/Class/10701-10s/Lecture/lecture5.pdf
6