Open Character Training: Shaping the Persona of AI Assistants through Constitutional AI
Paper β’ 2511.01689 β’ Published β’ 5
Method: SFT only (no DPO) LoRA rank: 4 | LoRA alpha: 8 Base model: meta-llama/Llama-3.1-8B-Instruct Pipeline: OpenCharacterTraining
A LoRA adapter fine-tuned to exhibit low Big Five Agreeableness via supervised fine-tuning only (SFT). This is the SFT-only variant with a small rank-4 adapter. For the full DPO+SFT pipeline version, see the companion model oct-low-agreeableness-llama3.1-8b-dpo-sft-r4.
| Model | Method | Description |
|---|---|---|
oct-low-agreeableness-llama3.1-8b-sft-r4 |
SFT only | This model. Single-stage character training. |
oct-low-agreeableness-llama3.1-8b-dpo-sft-r4 |
DPO + SFT | Full two-stage pipeline per the OCT paper. |
OCT = OpenCharacterTraining | r4 = LoRA rank 4 | sft = supervised fine-tuning only
OpenCharacterTraining β open-source Constitutional AI character training pipeline.
@article{OCT2024,
title={Open Character Training: Shaping the Persona of AI Assistants through Constitutional AI},
author={Maius Haiduc},
year={2024},
journal={arXiv preprint arXiv:2511.01689}
}
Base model
meta-llama/Llama-3.1-8B