unable download

#97
by luckyrain67777 - opened

always get“The requested URL returned error: 403” in my terminal
i don't know why

kind of same problem
OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like meta-llama/Meta-Llama-3-8B-Instruct is not the path to a directory containing a file named config.json.
and Internet connection is good (curl -I https://huggingface.co is fine)

Facing same issue .Is there any solution

I also encountered the same problem, after getting the link again or prompt 403: Forbidden, how to solve this problem?

+1 Same issue here...

it is more easier to download the model from the official website https://llama.meta.com/llama-downloads/ . It would be the original pt model. There is a way to convert the original model to hf models. Just use the script 'convert_llama_weights_to_hf.py' in the newest transformers. It must be the newest version transformers, because for older version, the 'convert_llama_weights_to_hf.py' only supports llama or llama2

does anyone got their error resolved? @JiayinWang , @luckyrain67777

A less than ideal solution is to use the path to the snapshot, i.e. $HF_HOME/models--meta-llama--Meta-Llama-3-8B-Instruct/snapshots/e5e23bbe8e749ef0efcf16cad411a7d23bd23298

osanseviero changed discussion status to closed

If anyone is having fine grained access token, make sure to add the repo in the "Repositories permissions" session.

Hi there! Please make sure to read https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/discussions/130

I have been granted the model access , but I still got the error: OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like meta-llama/Meta-Llama-3-8B-Instruct is not the path to a directory containing a file named config.json.
Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.

i have same issue

Also having the same issue

I am having the same issue

Any solution?

Using HF token with "WRITE" permission instead of "FINEGRAINED" one solved my issue. Got the solution from https://discuss.huggingface.co/t/loading-llama-3/90492. Hope this helps.

You will have to edit the token access in your settings to let it have "WRITE" permissions.

You will have to edit the token access in your settings to let it have "WRITE" permissions.

This solved my problem, thanks

If anyone is having fine grained access token, make sure to add the repo in the "Repositories permissions" session.

The perfect solution for this issue ❤️.

If anyone is having fine grained access token, make sure to add the repo in the "Repositories permissions" session.

Yes. This is the solution. Thanks

Using HF token with "WRITE" permission instead of "FINEGRAINED" one solved my issue. Got the solution from https://discuss.huggingface.co/t/loading-llama-3/90492. Hope this helps.

It works~~

This solved the issue:

  1. request access to llama repo from hugging face
  2. generate access token and toggle "Read access to contents of all public gated repos you can access" in the access token permission
  3. login using access token in the python script
from huggingface_hub import login
login(token = <your_access_token>)

Sign up or log in to comment