·
AI & ML interests
None yet
Organizations
None yet
view article Simplifying Alignment: From RLHF to Direct Preference Optimization (DPO)
upvoted a paper 5 months ago view article Introducing RTEB: A New Standard for Retrieval Evaluation


- +4
view article Illustrating Reinforcement Learning from Human Feedback (RLHF)


- +2