lyuchenyang/Macaw-LLM

which llama tokenizer to use?

chatsci opened this issue · 1 comments

In the preprocessing file, we have tokenizer = AutoTokenizer.from_pretrained('trained_models/llama_tokenizer'). This seems won't get the llama tokenizer from HF. which llama tokenizer we should use? As there are several versions on HF. Thanks.

Hi, sorry for the confusion of the LLaMA tokenizer. We use the LLaMA model and tokenizer from this link: https://huggingface.co/decapoda-research/llama-7b-hf.