runtime error

The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable. ā†‘ Those bitsandbytes warnings are expected on ZeroGPU ā†‘ The repository for microsoft/Phi-3-mini-128k-instruct contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/microsoft/Phi-3-mini-128k-instruct. You can avoid this prompt in future by passing the argument `trust_remote_code=True`. Traceback (most recent call last): File "/usr/local/lib/python3.10/site-packages/transformers/dynamic_module_utils.py", line 600, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/user/app/app.py", line 20, in <module> model = AutoModelForCausalLM.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 523, in from_pretrained config, kwargs = AutoConfig.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 932, in from_pretrained trust_remote_code = resolve_trust_remote_code( File "/usr/local/lib/python3.10/site-packages/transformers/dynamic_module_utils.py", line 613, in resolve_trust_remote_code raise ValueError( ValueError: The repository for microsoft/Phi-3-mini-128k-instruct contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/microsoft/Phi-3-mini-128k-instruct. Please pass the argument `trust_remote_code=True` to allow custom code to be run. Do you wish to run the custom code? [y/N]

Container logs:

Fetching error logs...