for deep learning study 2022
- Clone this repository:
git clone https://github.com/arumaekawa/text-dataset-distillation.git
- Install packages:
pip install -r requirements.txt -f https://download.pytorch.org/whl/cu113/torch_stable.html
- In
src/env
, set the value of DATA_ROOT_DIR as data directory and MODEL_ROOT_DIR as model directory
Note: All command execute in src
directory
test example of distilled data (created by my experiment)
./examples/run_distilled_data.sh --random_init --label_type hard --n_inner_steps 1 --pretrained_distilled_data ../distilled_data_examples/distilbert_ag_news_1_random_init_hard_inner_step_1
src
├── all_dataset_attrs.json
├── distill.py
├── env
├── evaluate.py
├── examples
│ ├── run_distilled_data.sh
│ ├── run_full_data.sh
│ └── run_random_data.sh
├── full_data.py
├── main.py
├── model.py
├── random_data.py
├── requirements.txt
├── run.sh
├── settings.py
├── transformers_models
│ ├── __init__.py
│ └── modeling_distilbert.py
└── utils.py
3 directories, 20 files