Sale!

Assignment 6: Unsupervised Clustering

$35.00

Category:
Rate this product

Assignment 6: Unsupervised Clustering
UVA CS 4501-03 :
Machine Learning

1 Unsupervised Learning with Clustering
In this programming assignment, you are required to implement clustering algorithm: K-means Clustering.
A ZIP file has been provided (“data sets clustering.zip” ) that includes two different datasets. Please follow
all instructions for submitting source code.
You are required to submit a source-code file “clustering.py” containing the necessary functions for
training and evaluations. The maximum number of iterations to be performed for both algorithms is 1000.
DO NOT use scikit-learn package in this problem and please implement from scratch.
1.1 Data description
We have provided two different datasets for clustering tasks.
• Dataset 1 : The first dataset consists of height and weight data for average people and baseball
players. First column contains human height (inches) and second column has human weight (lbs),
while third column has true labels of samples that will be used only for evaluations.
• Dataset 2 : The second dataset is for a speech versus music classification task. This dataset has been
preprocessed and first 13 columns contain 13 features extracted from audio files. Last column has true
labels of samples that will be used only for evaluations.
1.2 load data
• (Q1) You are required to code the following function for loading datasets:
X = loadData(fileDj)
1
1.3 K-means Clustering
• (Q2) Next, code the following function to implement k-means clustering:
labels = kmeans(X, k, maxIter)
Here X is the input data matrix, k is the number of clusters and maxIter is the maximum number of
the iterations selected by you (max value =1000).
• (Q3) Implement k-means clustering for Dataset 1(use first two columns in the file as input) and
use scatter() function in the matplotlib package to visualize the result. The two clusters must be in
different colors.
• (Q4) Implement k knee-finding method for Dataset 1 and k = {1,2,…,6} to select value of k (number
of clusters) and plot graph for k versus objective function value (e.g. Slide 99, Lecture 20).
• (Q5) Now, code the following function to calculate the purity metric for the evaluation of results:
purityMetric = purity(labels, trueLabels)
Use this function to evaluate the results of (Q3)
1.4 How will your code be checked?
We will run the following command: “python clustering.py DatasetDirectoryFullPath” and your code should
print the following results:
• the scatter plots from (Q3)
• k knee-finding plot in (Q4)
• ALL purityMetric values for results obtained in (Q3)
1.5 Submission
Please submit your source code as ”clustering.py” and PDF report containing your written answers via
collab. In the report, you should include the following contents:
• ALL scatter plots generated in (Q3)
• k knee-finding plot in (Q4)
• ALL purityMetric values for results obtained in (Q3)
2
2 Sample Exam Questions:
Each assignment covers a few sample exam questions to help you prepare for the midterm and the final.
(Please do not bother by the information of points in some the exam questions.)
Question: 1. K-means and Gaussian Mixture Models
(a) Run k-means manually for the following dataset, where k = 2. Circles are data points and squares are
the initial cluster centers. Draw the cluster centers and the decision boundaries that define each cluster.
Use as many pictures as you need until convergence.
Note: Execute the algorithm such that if a mean has no points assigned to it, it stays where it is for
that iteration.
Answer:
3
Question: 2. K-means Clustering
There is a set S consisting of 6 points in the plane shown below where a = (0, 0), b = (8, 0), c =
(16, 0), d = (0, 6), e = (8, 6), f = (16, 6). Now we run the k-means algorithm on these points with k = 3.
The algorithm uses the Euclidian distance metric (i.e. the straight line distance between two points) to
assign each point to its nearest centroid. Ties are broken in favor of the centroid to the left/down. We
define the following two definitions:
• A k-starting configuration is a subset of k staring points from S that form the initial centroids,
e.g. {a, b, c}.
• A k-partition is a partition of S into k non-empty subsets, e.g. {a, b, e}, {c, d}, {f} is a 3-partition.
Clearly any k-partition induces a set of k centroids in the natural manner. A k-partition is called stable
if a repetition of the k-means iteration with the induced centroid leaves it unchanged.
(a) How many 3-starting configurations are there? (Remember, a 3-starting configuration is just a size 3
subset of the 6 datapoints.
Answer: C
3
6 = 20
(b) Fill in the following table:
Answer:
3-partition Is it stable?
An example 3-starting configuration
that can arrive at the 3-partition after
0 or more iterations of k-means
(or write “none” if no such 3-starting
configuration)
The number of unique
starting configurations
that can arrive at the
3-partition.
{a,b,e},{c,d},{f} N none 0
{a,b},{d,e},{c,f} Y {b,c,e} 4
{a,d},{b,e},{c,f} Y {a,b,c} 8
{a},{d},{b,c,e,f} Y {a,b,d} 2
{a,b},{d},{c,e,f} Y none 0
{a,b,d},{c},{e,f} Y {a,c,f} 1
4
Question: 3. Decision Trees The following dataset will be used to learn a decision tree for
predicting whether a person is happy (H) or sad (S) based on the color of their shoes, whether they wear
a wig and the number of ears they have.
(a) [2 points] What is H(Emotion|W ig = Y ) (where H is entropy)?
Answer: Answer: 1
(b) [2 points] What is H(Emotion|Ears = 3)?
Answer: Answer: 0
(c) [3 points] Which attribute would the decision-tree building algorithm choose to use for the root
of the tree (assume no pruning)?
Answer: Answer: Color
(d) [3 points] Draw the full decision tree that would be learned from this data (assume no pruning).
Answer: Answer: Color is root node, predict sad if green, happy if red, and 50/50 split if blue.
The next two parts do not use the previous example, but are still about decision tree classifiers.
(e) [3 points] Assuming that the output attribute can take two values (i.e. has arity 2) what is the
maximum training set error (expressed as a percentage) that any dataset could possibly have?
Answer: Answer: 50%
5
(f) [3 points] Construct an example dataset that achieves this maximum percentage training set error (it must have two or fewer inputs and five or fewer records).
Answer: Answer: x : {0, 0, 1, 1}, y : {0, 1, 0, 1}
Question: 4. Decision Trees
The following dataset will be used to learn a decision tree for predicting whether a mushroom is edible
or not based on its shape, color, and odor.
(a) [4 points] What is the entropy H(Edible|Odor = 1 or Odor = 3)?
Answer: Answer: 1
(b) [4 points] Which attribute would the ID3 algorithm choose to use for the root of the tree (no pruning)?
Answer: Answer: Odor
(c) [4 points] Draw the full decision tree that would be learned for this data (no pruning).
Answer: see figure:
6

Reviews

There are no reviews yet.

Be the first to review “Assignment 6: Unsupervised Clustering”

Your email address will not be published. Required fields are marked *

Scroll to Top