{ "model_name": "finetuned-distilgptGGUF", "quantization_type": "q8_0", "source_model": "/tmp/inputs/merged_model_dir/data", "gguf_file": "finetuned-distilgptGGUF.gguf", "file_size_mb": 1.89, "conversion_tool": "llama.cpp" }