Collections
Discover the best community collections!
Collections including paper arxiv:2406.11194
-
PockEngine: Sparse and Efficient Fine-tuning in a Pocket
Paper • 2310.17752 • Published • 15 -
S-LoRA: Serving Thousands of Concurrent LoRA Adapters
Paper • 2311.03285 • Published • 30 -
Parameter-Efficient Orthogonal Finetuning via Butterfly Factorization
Paper • 2311.06243 • Published • 21 -
Fine-tuning Language Models for Factuality
Paper • 2311.08401 • Published • 30
-
mDPO: Conditional Preference Optimization for Multimodal Large Language Models
Paper • 2406.11839 • Published • 40 -
Pandora: Towards General World Model with Natural Language Actions and Video States
Paper • 2406.09455 • Published • 16 -
WPO: Enhancing RLHF with Weighted Preference Optimization
Paper • 2406.11827 • Published • 17 -
In-Context Editing: Learning Knowledge from Self-Induced Distributions
Paper • 2406.11194 • Published • 20
-
Iterative Reasoning Preference Optimization
Paper • 2404.19733 • Published • 50 -
Better & Faster Large Language Models via Multi-token Prediction
Paper • 2404.19737 • Published • 80 -
ORPO: Monolithic Preference Optimization without Reference Model
Paper • 2403.07691 • Published • 72 -
KAN: Kolmogorov-Arnold Networks
Paper • 2404.19756 • Published • 116
-
mDPO: Conditional Preference Optimization for Multimodal Large Language Models
Paper • 2406.11839 • Published • 40 -
Pandora: Towards General World Model with Natural Language Actions and Video States
Paper • 2406.09455 • Published • 16 -
WPO: Enhancing RLHF with Weighted Preference Optimization
Paper • 2406.11827 • Published • 17 -
In-Context Editing: Learning Knowledge from Self-Induced Distributions
Paper • 2406.11194 • Published • 20
-
Iterative Reasoning Preference Optimization
Paper • 2404.19733 • Published • 50 -
Better & Faster Large Language Models via Multi-token Prediction
Paper • 2404.19737 • Published • 80 -
ORPO: Monolithic Preference Optimization without Reference Model
Paper • 2403.07691 • Published • 72 -
KAN: Kolmogorov-Arnold Networks
Paper • 2404.19756 • Published • 116
-
PockEngine: Sparse and Efficient Fine-tuning in a Pocket
Paper • 2310.17752 • Published • 15 -
S-LoRA: Serving Thousands of Concurrent LoRA Adapters
Paper • 2311.03285 • Published • 30 -
Parameter-Efficient Orthogonal Finetuning via Butterfly Factorization
Paper • 2311.06243 • Published • 21 -
Fine-tuning Language Models for Factuality
Paper • 2311.08401 • Published • 30