Understanding Reasoning in LLMs through Strategic Information Allocation under Uncertainty Paper • 2603.15500 • Published 8 days ago • 11
CausalArmor: Efficient Indirect Prompt Injection Guardrails via Causal Attribution Paper • 2602.07918 • Published Feb 8 • 3
Drift: Decoding-time Personalized Alignments with Implicit User Preferences Paper • 2502.14289 • Published Feb 20, 2025 • 1
FlowRL: Matching Reward Distributions for LLM Reasoning Paper • 2509.15207 • Published Sep 18, 2025 • 118
Critic-Guided Decoding for Controlled Text Generation Paper • 2212.10938 • Published Dec 21, 2022 • 2