runtime error
er/.local/lib/python3.10/site-packages/fsspec/asyn.py", line 56, in _runner result[0] = await coro File "/home/user/.local/lib/python3.10/site-packages/gradio/helpers.py", line 273, in create await self.cache() File "/home/user/.local/lib/python3.10/site-packages/gradio/helpers.py", line 308, in cache prediction = await Context.root_block.process_api( File "/home/user/.local/lib/python3.10/site-packages/gradio/blocks.py", line 1015, in process_api result = await self.call_function( File "/home/user/.local/lib/python3.10/site-packages/gradio/blocks.py", line 833, in call_function prediction = await anyio.to_thread.run_sync( File "/home/user/.local/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync return await get_async_backend().run_sync_in_worker_thread( File "/home/user/.local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2134, in run_sync_in_worker_thread return await future File "/home/user/.local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 851, in run result = context.run(func, *args) File "/home/user/.local/lib/python3.10/site-packages/gradio/external.py", line 286, in query_huggingface_api raise ValueError( ValueError: Could not complete request to HuggingFace API, Status Code: 500, Error: unknown error, Warnings: ['CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 14.75 GiB total capacity; 1.89 GiB already allocated; 13.06 MiB free; 1.94 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF', 'There was an inference error: CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 14.75 GiB total capacity; 1.89 GiB already allocated; 13.06 MiB free; 1.94 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF']
Container logs:
Fetching error logs...