!git clone https://github.com/ukairia777/LLM-Finetuning-tutorial.git
%cd LLM-Finetuning-tutorial
!pip install -r requirements.txt
!torchrun finetune.py \
--base_model beomi/llama-2-ko-7b \
--data-path '학습할 데이터셋' \
--output_dir ./fine-tuned-llama-2-ko-7b \
--batch_size 64 \
--micro_batch_size 1 \
--num_epochs 15 \
--learning_rate 1e-5 \
--cutoff_len 2048 \
--val_set_size 0 \
--lora_r 16 \
--lora_alpha 16 \
--lora_dropout 0.05 \
--lora_target_modules '[gate_proj, down_proj, up_proj]' \
--train_on_inputs False \
--add_eos_token True \
--group_by_length False \
--prompt_template_name alpaca \
--lr_scheduler 'linear' \
--warmup_steps 0
!wget https://raw.githubusercontent.com/ukairia777/LLM-Finetuning-tutorial/main/DPO.py
!pip install trl==0.7.9 peft==0.7.1 accelerate==0.26.1 datasets==2.16.1 bitsandbytes==0.42.0 scipy==1.11.4 sentencepiece==0.1.99 fire==0.5.0
!pip install transformers==4.37.2
!python DPO.py \
--base_model Qwen/Qwen1.5-72B \
--data-path Intel/orca_dpo_pairs \
--output_dir ./lora \
--num_epochs 3 \
--batch_size 16 \
--micro_batch_size 2 \
--learning_rate 1e-5 \
--lora_r 16 \
--lora_alpha 16 \
--lora_dropout 0.05 \
# --lora_target_modules ["embed_tokens", "q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj", "lm_head"] \
--lora_target_modules ["q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj", "lm_head"] \
--lr_scheduler 'linear' \
--warmup_ratio 0.1 \
--cutoff_len 4096 \
!wget https://raw.githubusercontent.com/ukairia777/LLM-Finetuning-tutorial/main/merge.py
!python merge.py \
--base_model_name_or_path Qwen/Qwen1.5-72B \
--peft_model_path checkpoint-80 \
--output_dir merged
from huggingface_hub import HfApi
api = HfApi()
username = "허깅페이스 ID"
MODEL_NAME = 'MyModel-Qwen72B-v1'
api.create_repo(
token="허깅페이스 토큰",
repo_id=f"{username}/{MODEL_NAME}",
repo_type="model"
)
api.upload_folder(
token="허깅페이스 토큰",
repo_id=f"{username}/{MODEL_NAME}",
folder_path="merged",
)