Chain-of-Verification Reduces Hallucination in Large Language Models
Paper
• 2309.11495
• Published • 40
EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form
Narrative Text Generation
Paper
• 2310.08185
• Published • 8
The Consensus Game: Language Model Generation via Equilibrium Search
Paper
• 2310.09139
• Published • 14
In-Context Pretraining: Language Modeling Beyond Document Boundaries
Paper
• 2310.10638
• Published • 30
Reward-Augmented Decoding: Efficient Controlled Text Generation With a
Unidirectional Reward Model
Paper
• 2310.09520
• Published • 11
Self-RAG: Learning to Retrieve, Generate, and Critique through
Self-Reflection
Paper
• 2310.11511
• Published • 78
VeRA: Vector-based Random Matrix Adaptation
Paper
• 2310.11454
• Published • 30
Safe RLHF: Safe Reinforcement Learning from Human Feedback
Paper
• 2310.12773
• Published • 28
In-Context Learning Creates Task Vectors
Paper
• 2310.15916
• Published • 43
Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Paper
• 2310.17157
• Published • 14
Controlled Decoding from Language Models
Paper
• 2310.17022
• Published • 14
Tell Your Model Where to Attend: Post-hoc Attention Steering for LLMs
Paper
• 2311.02262
• Published • 14
S-LoRA: Serving Thousands of Concurrent LoRA Adapters
Paper
• 2311.03285
• Published • 30
Prompt Cache: Modular Attention Reuse for Low-Latency Inference
Paper
• 2311.04934
• Published • 32
System 2 Attention (is something you might need too)
Paper
• 2311.11829
• Published • 43
Adapters: A Unified Library for Parameter-Efficient and Modular Transfer
Learning
Paper
• 2311.11077
• Published • 29
Tuning Language Models by Proxy
Paper
• 2401.08565
• Published • 22
Self-Rewarding Language Models
Paper
• 2401.10020
• Published • 152
Collaborative Development of NLP models
Paper
• 2305.12219
• Published
Suppressing Pink Elephants with Direct Principle Feedback
Paper
• 2402.07896
• Published • 11
A Tale of Tails: Model Collapse as a Change of Scaling Laws
Paper
• 2402.07043
• Published • 15
Direct Language Model Alignment from Online AI Feedback
Paper
• 2402.04792
• Published • 35