which llama tokenizer to use?
chatsci opened this issue · 1 comments
chatsci commented
In the preprocessing file, we have tokenizer = AutoTokenizer.from_pretrained('trained_models/llama_tokenizer'). This seems won't get the llama tokenizer from HF. which llama tokenizer we should use? As there are several versions on HF. Thanks.
lyuchenyang commented
Hi, sorry for the confusion of the LLaMA tokenizer. We use the LLaMA model and tokenizer from this link: https://huggingface.co/decapoda-research/llama-7b-hf
.