| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import argparse |
| from pathlib import Path |
| from huggingface_hub import snapshot_download |
| |
| from onnxruntime_genai.models.builder import create_model |
| def main(): |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--name", required=False,default=None) |
| parser.add_argument("--token",required=False) |
| args = parser.parse_args() |
|
|
| token = args.token if args.token else None |
| |
| pwd = Path.cwd() |
| model_dir = pwd / "model" |
| onnx_dir = pwd / "onnx" |
| cache_dir = pwd / "cache" |
| model_dir.mkdir(exist_ok=True) |
| onnx_dir.mkdir(exist_ok=True) |
| cache_dir.mkdir(exist_ok=True) |
| |
| |
| print(">> Downloading model via huggingface_hub (Xet enabled if installed)...") |
|
|
| |
| |
| |
| |
| |
| |
|
|
| |
|
|
| |
| print(">> Converting to ONNX (GenAI format)...") |
|
|
| create_model( |
| model_name=args.name, |
| input_path=str(model_dir), |
| output_dir=str(onnx_dir), |
| precision="fp16", |
| execution_provider="cpu", |
| cache_dir=str(cache_dir), |
| extra_options={} |
| ) |
|
|
| print("\n✅ Done") |
| print(f"ONNX model at: {onnx_dir}") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|