Why Does Self-Distillation (Sometimes) Degrade the Reasoning Capability of LLMs? Paper • 2603.24472 • Published 20 days ago • 53
Understanding Reasoning in LLMs through Strategic Information Allocation under Uncertainty Paper • 2603.15500 • Published 29 days ago • 12
CausalArmor: Efficient Indirect Prompt Injection Guardrails via Causal Attribution Paper • 2602.07918 • Published Feb 8 • 4
Drift: Decoding-time Personalized Alignments with Implicit User Preferences Paper • 2502.14289 • Published Feb 20, 2025 • 1
FlowRL: Matching Reward Distributions for LLM Reasoning Paper • 2509.15207 • Published Sep 18, 2025 • 118
Critic-Guided Decoding for Controlled Text Generation Paper • 2212.10938 • Published Dec 21, 2022 • 2