/Playground

store my personal project

Primary LanguageJupyter Notebook

Playground

This depository stores some of my experimented projects. Welcome to check it out and give any feedback, so that we can improve the cases even better.

01-Classification-Modeling-on-Hotel-Scoring
Comparing the performance of models among MLP, Xgboost, and Logistic, and choose Logistic Model as the final choice. The final model reaches 0.62 in accuracy and 0.61 in F1 score in three classes scenario. If we look deeper, the share of each class are 35%, 29%, 36%, which compared to the precision from the predicted result for each class 68%, 47%, 65%, we can say that Logistic Model doubles the precision generally.

02-Mass-Transit-Analysis
By the end of this analysis, I am able to come up with a list of suggested stations for ads placement. Adhering to the stations, I also obtain a table of expected commute time for various subway routes.

04-Model-Comparison-Word2vec-Doc2vec-TfIdfWeighted
In this post, I'd like to test out the effect of different kinds of wordembeddings on text classifier. The wordembeddings investigated here include word2vec, TF-IDF weighted word2vec, pre-train GloVe word2vec and doc2vec. It turns out that there is no significant difference among their performance, in the meantime, pre-train Glove and doc2vec alone seem to under-perform a bit compared with others.

05-Try-out-Spacy-Pretrain
This post focuses on trying out spaCy's new BERT-style language pre-training feature. First, I set up a classifier without pre-training and then compare it with one initiated with pre-training. It turns out the classifier without pre-training has higher score on both f1 and accuracy. Nonetheless, take it with a grain of salt. If we look closely, the one with pre-training performs better at identifying rare classes. It still has its own merits if identifying rare classes matter.