meta-llama/llama

The instructions to install Llama3 is horrible

Eyesun23 opened this issue · 2 comments

I followed the steps of getting access to the models; I received a link. But I am getting this error after I ran:

torchrun --nproc_per_node=1 example_chat_completion.py \ --ckpt_dir Meta-Llama-3-8B-Instruct/ \ --tokenizer_path Meta-Llama-3-8B-Instruct/tokenizer.model \ --max_seq_len 512 --max_batch_size 6

/opt/anaconda3/envs/llama3/bin/python: can't open file 'example_chat_completion.py': [Errno 2] No such file or directory
[2024-06-21 16:30:35,995] torch.distributed.elastic.multiprocessing.api: [ERROR] failed (exitcode: 2) local_rank: 0 (pid: 97659) of binary: /opt/anaconda3/envs/llama3/bin/python
Traceback (most recent call last):
  File "/opt/anaconda3/envs/llama3/bin/torchrun", line 8, in <module>
    sys.exit(main())
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 347, in wrapper
    return f(*args, **kwargs)
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/run.py", line 812, in main
    run(args)
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/run.py", line 803, in run
    elastic_launch(
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/launcher/api.py", line 135, in __call__
    return launch_agent(self._config, self._entrypoint, list(args))
  File "/opt/anaconda3/envs/llama3/lib/python3.8/site-packages/torch/distributed/launcher/api.py", line 268, in launch_agent
    raise ChildFailedError(
torch.distributed.elastic.multiprocessing.errors.ChildFailedError: 
============================================================
example_chat_completion.py FAILED
------------------------------------------------------------
Failures:
  <NO_OTHER_FAILURES>
------------------------------------------------------------
Root Cause (first observed failure):
[0]:
  time      : 2024-06-21_16:30:35
  host      : aysuns-mbp.attlocal.net
  rank      : 0 (local_rank: 0)
  exitcode  : 2 (pid: 97659)
  error_file: <N/A>
  traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html```

I suspect the issue lies in the first line: ‘’/opt/anaconda3/envs/llama3/bin/python: can't open file 'example_chat_completion.py': [Errno 2] No such file or directory‘’,you can check if there is any problem with the file path when you run this command.

Hi, I also followed this README.md procedure using a python virtual environment and I also got stuck at the PYTHONPATH=$(git rev-parse --show-toplevel) torchrun llama_models/scripts/example_chat_completion.py $CHECKPOINT_DIR since there is no such llama_models/ folder created.

Additionally I tried to follow this other installation tutorial but I also got stuck because 403 Forbidden