runtime error

Exit code: 1. Reason: r: - kv 11: bitnet-25.attention.layer_norm_rms_epsilon f32 = 0.000010 llama_model_loader: - kv 12: bitnet-25.rope.freq_base f32 = 500000.000000 llama_model_loader: - kv 13: general.file_type u32 = 137 llama_model_loader: - kv 14: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 15: tokenizer.ggml.tokens arr[str,128256] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 16: tokenizer.ggml.scores arr[f32,128256] = [0.000000, 0.000000, 0.000000, 0.0000... llama_model_loader: - kv 17: tokenizer.ggml.token_type arr[i32,128256] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 18: tokenizer.ggml.merges arr[str,280147] = ["Ä  Ä ", "Ä  Ä Ä Ä ", "Ä Ä  Ä Ä ", "... llama_model_loader: - kv 19: tokenizer.ggml.bos_token_id u32 = 128000 llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 128001 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 128001 llama_model_loader: - kv 22: tokenizer.chat_template str = {% for message in messages %}{% if lo... llama_model_loader: - kv 23: general.quantization_version u32 = 2 llama_model_loader: - type f32: 121 tensors llama_model_loader: - type q6_K: 1 tensors llama_model_loader: - type iq4_nl: 1 tensors llama_model_loader: - type iq2_bn: 210 tensors Traceback (most recent call last): File "/app/chat_demo.py", line 95, in <module> chat = gr.ChatInterface( TypeError: ChatInterface.__init__() got an unexpected keyword argument 'type' load: missing pre-tokenizer type, using: 'default' load: load: ************************************ load: GENERATION QUALITY WILL BE DEGRADED! load: CONSIDER REGENERATING THE MODEL load: ************************************ load:

Container logs:

Fetching error logs...