- Benjamin Kosko, bkosko@seas.upenn.edu, bkosko
- Victor Lin, lin1129@seas.upenn.edu, victorHLin
- Chun-Fu Yeh, cfyeh@seas.upenn.edu, YehCF
- Sarah Payne, paynesa@seas.upenn.edu, paynesa
chart.js
3.6.1cors
2.8.5express
4.17.1mysql
2.18.1node-fetch
3.0.0nodemon
2.0.12supertest
6.1.6jest
27.1.0
ant-design/charts
1.2.14fortawesome/fontawesome-svg-core
1.2.36fortawesome/free-solid-svg-icons
5.15.4fortawesome/react-fontawesome
0.1.16testing-library/jest-dom
5.14.1testing-library/react
11.2.7testing-library/user-event
12.8.3antd
4.16.13antd-button-color
1.0.4bootstrap
5.1.3canvasjs
1.8.3chart.js
3.6.1colormap
2.3.2d3
7.1.1d3-format
3.0.1datamaps
0.5.9font-awesome
4.7.0query-string
7.0.1react
17.0.2react-bootstrap
2.0.3react-chartjs-2
4.0.0react-d3-library
1.1.8react-dom
17.0.2react-loading
2.0.3react-promise-tracker
2.1.0react-promise-tracker
2.1.0react-router-dom
5.3.0react-scripts
4.0.3react-usa-map
1.5.0react-vis
1.11.7reactstrap
9.0.1shards-react
1.0.3web-vitals
1.1.2
- R:
tidyverse
anddpylr
Open two terminal windows. In one, type the following commands:
cd server
npm install
npm start
In the other type:
cd client
npm install
npm start
In a few moments, the server should be running and a browser window should pop up. If no window pops up, open your
browser and go to http://localhost:3000/
.
Elections Data Wrangling Place preprocess_voting.R
and 1976-2020-senate.csv
in the same directory (both are in the preprocessing/voting_preprocessing
directory by default). Then either execute the R script on the command line or open preprocess_voting.R
in RStudio, set the session's working
directory to the source file location, and execute the entire script.
Stock Data Wrangling Run stock_preprocess.ipynb
to preprocess the original table downloaded from (https://www.kaggle.com/shannanl/sp500-dataset?select=sp500+agg.csv). Then, the preprocessed stock table can be retrieved.
COVID/Vaccine Data Wrangling In DataGrip, we replaced all slashes with hyphens so all the dates (in both files) followed the MM-DD-YYYY format. Vaccination data also had several negative values that needed to be corrected; we did this by sorting by case numbers, and then taking the absolute value of the clearly wrong 6 negative values.
Yelp data Wrangling
Get yelp_academic_dataset_business.json
, yelp_academic_dataset_user.json
and yelp_academic_dataset_review.json
from https://www.yelp.com/dataset.
To wrangle yelp_academic_dataset_review.json
and yelp_academic_dataset_user.json
, you need to execute chunk.sh
first. It will chunk the original file to smaller size files to speed up wrangling time.
Usage of chunk.sh
:
./chunck.sh {your_file_name}
Follow the program instruction to input the number of rows you want to store in a file. After chunking data, put all chunk files to the directory, and modify the path variable with directory path in yelp_review.py
and yelp_user.py
. Then execute yelp_review.py
to create the csv file for Review table and execute yelp_user.py
to create the csv file for User table.
To wrangle yelp_academic_dataset_business.json
, you just need to modify file path in both yelp_business.py
and yelp_categories.py
. yelp_business.py
will create the csv file for Business table. yelp_categories.py
will create two csv files, one for Categories table and the other for Business_Categories table.