CrPO-SFT-Llama-3.1-8B-Instruct

This is a Llama-3.1-8B-Instruct model supervised-finetuned (SFT) on the MuCE-SFT dataset from the Creative Preference Optimization paper.

Citation

@misc{ismayilzada2025creativepreferenceoptimization,
      title={Creative Preference Optimization}, 
      author={Mete Ismayilzada and Antonio Laverghetta Jr. and Simone A. Luchini and Reet Patel and Antoine Bosselut and Lonneke van der Plas and Roger E. Beaty},
      year={2025},
      eprint={2505.14442},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2505.14442}, 
}
Downloads last month
289
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for CNCL-Penn-State/CrPO-sft-llama-3.1-8b-instruct

Finetuned
(2587)
this model
Finetunes
10 models

Dataset used to train CNCL-Penn-State/CrPO-sft-llama-3.1-8b-instruct

Collections including CNCL-Penn-State/CrPO-sft-llama-3.1-8b-instruct

Paper for CNCL-Penn-State/CrPO-sft-llama-3.1-8b-instruct