This is my revision of the great tutorial at http://brandonrose.org/clustering - many thanks to the author.
Data: Top 100 movies (http://www.imdb.com/list/ls055592025/) with title, genre, and synopsis (IMDB and Wiki)
Goal: Put 100 movies into 5 clusters by text-mining their synopses and plot the result as follows
First, clone the repo, go to the repo folder, setup the virtual environment, and install the required packages:
git clone https://github.com/harrywang/document_clustering.git
cd document_clustering
virtualenv venv
source venv/bin/activate
pip install -r requirements.txt
Second, use nltk.download() to download all nltk packages, which are saved to /Users/harrywang/nltk_data
ipython2
import nltk
nltk.download()
Lastly, view doc_clustering.ipynb directly on Github at https://github.com/harrywang/document_cluster/blob/master/doc_clustering.ipynb or locally by running ipython2 notebook
to learn the tutorial step-by-step.
- Read data: read titles, genres, synopses, rankings into four arrays
- Tokenize and stem: break paragraphs into sentences, then to words, stem the words (without removing stopwords) - each synopsis essentially becomes a bag of stemmed words.
- Generate tf-idf matrix: each row is a term (unigram, bigram, trigram...generated from the bag of words in 2.), each column is a synopsis.
- Generate clusters: based on the tf-idf matrix, 5 (or any number) clusters are generated using k-means. The top key terms are selected for each cluster.
- Calculate similarity: generate the cosine similarity matrix using the tf-idf matrix (100x100), then generate the distance matrix (1 - similarity matrix), so each pair of synopsis has a distance number between 0 and 1.
- Plot clusters: use multidimensional scaling (MDS) to convert distance matrix to a 2-dimensional array, each synopsis has (x, y) that represents their relative location based on the distance matrix. Plot the 100 points with their (x, y) using matplotlib (I added an example on using plotly.js).