LocLLM / coco /generation_config.json
kennethwdk
upload model weights
65baf7e
raw
history blame contribute delete
525 Bytes
{
"bos_token_id": 1,
"crop_size": 224,
"dino_path": "./checkpoints/model_weights/dinov2_vitl14_pretrain.pth",
"eos_token_id": 2,
"llama_path": "./checkpoints/model_weights/vicuna-7b-v1.5",
"lora_llm_alpha": 16,
"lora_llm_dropout": 0.05,
"lora_llm_enable": true,
"lora_llm_r": 8,
"lora_vision_alpha": 16,
"lora_vision_dropout": 0.05,
"lora_vision_enable": true,
"lora_vision_r": 8,
"max_length": 4096,
"pad_token_id": 0,
"temperature": 0.9,
"top_p": 0.6,
"transformers_version": "4.31.0"
}