inference api no longer working

#18
by skratos115 - opened

was testing this out last week and now im getting errors:The model meta-llama/Meta-Llama-3.1-405B-FP8 is too large to be loaded automatically (162GB > 10GB). Please use Spaces (https://huggingface.co/spaces) or Inference Endpoints (https://huggingface.co/inference-endpoints)."}
are pro users no longer able to run inference on these? it was the only reason i upgraded to pro.

Sign up or log in to comment