/Chinese-Alpaca-LoRA-and-Alpaca-7b-replication

A alpaca-7b-hf replication with no moderation filter and output watermark

Primary LanguagePythonApache License 2.0Apache-2.0

Huacaya: A Chinese-main multilingual Alpaca LoRA, and a replication of Alpaca-7b

This project is supported by AI community: AcceleratorI

A alpaca-7b-hf replication

A alpaca-7b-hf replication with no moderation filter and output watermark

Multilingual Instruction LLaMA LoRA

This is a multilingual instruction LLaMA LoRA that mainly focused on Simplified Chinese but also support Japanese and Traditional Chinese(HK and TW)
Note that the LLaMA model is not pre-trained in Chinese, so the model still has far better performance on the English task, so please try to use English if possible.

Version

Version 0.4 (Newest)
The repository now contains the latest version of our training results.

Training

We have fine-tuned a Multilingual Alpaca model baed on

Dataset

We used the following datasets:

  • JosephusCheung/GuanacoDateset
  • BelleGroup/generated_train_0.5M_CN
  • Baike_qa2019 (Partially used)

The fine-tuning process took approximately 30 hours on eight 80GB A100 GPUs.

Contributing

Contributions are welcome! Please read our contributing guidelines for more information on how to get started.

Contact

If you have any questions, issues, or suggestions, please open an issue or contact us at email.

Citation

@misc{alpaca,
  author = {Rohan Taori and Ishaan Gulrajani and Tianyi Zhang and Yann Dubois and Xuechen Li and Carlos Guestrin and Percy Liang and Tatsunori B. Hashimoto },
  title = {Stanford Alpaca: An Instruction-following LLaMA model},
  year = {2023},
  publisher = {GitHub},
  journal = {GitHub repository},
  howpublished = {\url{https://github.com/tatsu-lab/stanford_alpaca}},
}