Transformers
Safetensors
English

CrPO-SFT-Llama-3.1-8B-Instruct-div-qua

This is a CrPO-sft-llama-3.1-8b-instruct model preference-finetuned on the MuCE-Pref dataset from the Creative Preference Optimization paper. This model is optimized for high output diversity and quality.

Citation

@misc{ismayilzada2025creativepreferenceoptimization,
      title={Creative Preference Optimization}, 
      author={Mete Ismayilzada and Antonio Laverghetta Jr. and Simone A. Luchini and Reet Patel and Antoine Bosselut and Lonneke van der Plas and Roger E. Beaty},
      year={2025},
      eprint={2505.14442},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2505.14442}, 
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for CNCL-Penn-State/CrPO-llama-3.1-8b-instruct-div-qua

Dataset used to train CNCL-Penn-State/CrPO-llama-3.1-8b-instruct-div-qua

Collections including CNCL-Penn-State/CrPO-llama-3.1-8b-instruct-div-qua