File size: 237 Bytes
19e1f45
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
{
  "model_name": "finetuned-distilgptGGUF",
  "quantization_type": "q8_0",
  "source_model": "/tmp/inputs/merged_model_dir/data",
  "gguf_file": "finetuned-distilgptGGUF.gguf",
  "file_size_mb": 87.06,
  "conversion_tool": "llama.cpp"
}