Homework 1 CS 436/580L solution

$29.99

Original Work ?
Category: You will Instantly receive a download link for .ZIP solution file upon Payment

Description

5/5 - (3 votes)

1 Instructions
1. You can use either C/C++, Java or Python to implement your algorithms.
2. Your implementations should compile on remote.cs.binghamton.edu.
3. Make sure remote.cs.binghamton.edu has the packages that you require
before starting to implement.
4. This homework requires you to implement decision tree. Using existing
decision package is not allowed.
5. Your homework should contain the following components:
(a) README.txt file with detailed instructions on how to compile
and run the code.
(b) Code source files
(c) Type-written document containing the results on the datasets.
6. Submit the homework as a single zip file: f irstname lastname hw1.zip.
Inducing Decision Trees
In this homework you will implement and test the decision tree learning
algorithm (See Mitchell, Chapter 3).
1
• Download the two datasets available on myCourses. Each data set is
divided into three sets: the training set, the validation set and the test
set. Data sets are in CSV format. The first line in the file gives the
attribute names. Each line after that is a training (or test) example
that contains a list of attribute values separated by a comma. The last
attribute is the class-variable. Assume that all attributes take values
from the domain {0,1}.
• Implement the decision tree learning algorithm. As discussed in class,
the main step in decision tree learning is choosing the next attribute to
split on. Implement the following two heuristics for selecting the next
attribute.
1. (30 points) Information gain heuristic (See Class slides, Mitchell
Chapter 3).
2. (50 points) Variance impurity heuristic described below.
Let K denote the number of examples in the training set. Let K0
denote the number of training examples that have class = 0 and
K1 denote the number of training examples that have class = 1.
The variance impurity of the training set S is defined as:
V I(S) = K0
K
K1
K
Notice that the impurity is 0 when the data is pure. The gain for
this impurity is defined as usual.
Gain(S, X) = V I(S) −
X
x∈V alues(X)
Pr(x)V I(Sx)
where X is an attribute, Sx denotes the set of training examples
that have X = x and Pr(x) is the fraction of the training examples
that have X = x (i.e., the number of training examples that have
X = x divided by the number of training examples in S).
• (5 points) Implement a function to print the decision tree to standard
output. We will use the following format.
wesley = 0 :
| honor = 0 :
2
| | barclay = 0 : 1
| | barclay = 1 : 0
| honor = 1 :
| | tea = 0 : 0
| | tea = 1 : 1
wesley = 1 : 0
According to this tree, if wesley = 0 and honor = 0 and barclay = 0,
then the class value of the corresponding instance should be 1. In other
words, the value appearing before a colon is an attribute value, and the
value appearing after a colon is a class value.
• (5 points) Once we compile your code, we should be able to run it from
the command line. Your program should take as input the following
six arguments:
.\program
to-print:{yes,no}
It should output the accuracies on the test set for decision trees constructed using the two heuristics. If to-print equals yes, it should print
the decision tree in the format described above to the standard output.
• (5 points) A README.txt file with detailed instructions on compiling
the code.
• (5 points) A document containing the accuracy on the test set in
both datasets for decision trees constructed using the two heuristics
mentioned above.
What to Turn in
• Your code and a Readme file for compiling the code.
• Document containing the accuracy on the two test sets for decision
trees constructed using the two heuristics.
3