Spaces:
Running
Running
| import os | |
| import torch | |
| # --- AGGRESSIVE FIX: Bypass Security Check --- | |
| # We must import these modules specifically to patch the function where it is used | |
| import transformers.modeling_utils | |
| import transformers.utils.import_utils | |
| # Disable the check in both locations | |
| transformers.modeling_utils.check_torch_load_is_safe = lambda: None | |
| transformers.utils.import_utils.check_torch_load_is_safe = lambda: None | |
| # --------------------------------------------- | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| # 1. Path to your local PyTorch model | |
| input_path = r"B:\7B\!models--Gryphe--Tiamat-7b" | |
| # 2. Path where you want the SafeTensors version | |
| output_path = r"B:\7B\!models--Gryphe--Tiamat-7b\safe" | |
| print(f"Loading model from {input_path}...") | |
| # Load the model | |
| model = AutoModelForCausalLM.from_pretrained( | |
| input_path, | |
| torch_dtype=torch.bfloat16, | |
| device_map="cpu", | |
| low_cpu_mem_usage=True | |
| ) | |
| # Load the tokenizer | |
| tokenizer = AutoTokenizer.from_pretrained(input_path) | |
| print(f"Saving to {output_path}...") | |
| if not os.path.exists(output_path): | |
| os.makedirs(output_path) | |
| # 3. Save with safe_serialization=True | |
| model.save_pretrained( | |
| output_path, | |
| safe_serialization=True, | |
| max_shard_size="5GB" | |
| ) | |
| tokenizer.save_pretrained(output_path) | |
| print("Conversion complete.") |