Safetensors
English

YIELD Fine-Tuning Adapters

This repository contains the persona adapter models presented in the paper YIELD: A Large-Scale Dataset and Evaluation Framework for Information Elicitation Agents.

Model Information

All adapters in this repository are LoRA adapters trained on top of the Llama-3.1-8B-Instruct, Llama-3.2-3B-Instruct, and DeepSeek-R1-Distill-Llama-8B models. All models are fine-tuned using both the Supervised Fine-Tuning (SFT) and Offline Reinforcement-Learning (ORL) pipelines detailed in the paper.

Resources

Citing YIELD

If you use this resource in your projects, please cite the following paper:

@misc{De_Lima_YIELD_A_Large-Scale_2026,
author = {De Lima, Victor and Yang, Grace Hui},
doi = {10.48550/arXiv.2604.10968},
title = {{YIELD: A Large-Scale Dataset and Evaluation Framework for Information Elicitation Agents}},
url = {https://arxiv.org/abs/2604.10968},
year = {2026}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for infosense/yield-adapters

Finetuned
(2463)
this model

Dataset used to train infosense/yield-adapters

Paper for infosense/yield-adapters