runtime error

call_function prediction = await anyio.to_thread.run_sync( # type: ignore File "/usr/local/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync return await get_async_backend().run_sync_in_worker_thread( File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2177, in run_sync_in_worker_thread return await future File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 859, in run result = context.run(func, *args) File "/usr/local/lib/python3.10/site-packages/gradio/utils.py", line 818, in wrapper response = f(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/gradio/external.py", line 371, in query_huggingface_inference_endpoints data = fn(*data) # type: ignore File "/usr/local/lib/python3.10/site-packages/gradio/external_utils.py", line 115, in text_generation_inner return input + client.text_generation(input) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 2173, in text_generation raise_text_generation_error(e) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_common.py", line 422, in raise_text_generation_error raise http_error File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 2143, in text_generation bytes_output = self.post(json=payload, model=model, task="text-generation", stream=stream) # type: ignore File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 304, in post hf_raise_for_status(response) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py", line 371, in hf_raise_for_status raise HfHubHTTPError(str(e), response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: 504 Server Error: Gateway Timeout for url: https://api-inference.huggingface.co/models/mistralai/Mistral-Nemo-Instruct-2407 (Request ID: QFJI--PzTa-19B-tPoByG) Model mistralai/Mistral-Nemo-Instruct-2407 time out

Container logs:

Fetching error logs...