File size: 236 Bytes
19e1f45
 
 
 
 
81b1ec1
19e1f45
 
1
2
3
4
5
6
7
8
{
  "model_name": "finetuned-distilgptGGUF",
  "quantization_type": "q8_0",
  "source_model": "/tmp/inputs/merged_model_dir/data",
  "gguf_file": "finetuned-distilgptGGUF.gguf",
  "file_size_mb": 1.89,
  "conversion_tool": "llama.cpp"
}