HALoGEN: Fantastic LLM Hallucinations and Where to Find Them Paper • 2501.08292 • Published 5 days ago • 16
Inference-Time Policy Adapters (IPA): Tailoring Extreme-Scale LMs without Fine-tuning Paper • 2305.15065 • Published May 24, 2023 • 1
CONDAQA: A Contrastive Reading Comprehension Dataset for Reasoning about Negation Paper • 2211.00295 • Published Nov 1, 2022
The Art of Saying No: Contextual Noncompliance in Language Models Paper • 2407.12043 • Published Jul 2, 2024 • 4
WildHallucinations: Evaluating Long-form Factuality in LLMs with Real-World Entity Queries Paper • 2407.17468 • Published Jul 24, 2024
Question Answering for Privacy Policies: Combining Computational and Legal Perspectives Paper • 1911.00841 • Published Nov 3, 2019
HALoGEN: Fantastic LLM Hallucinations and Where to Find Them Paper • 2501.08292 • Published 5 days ago • 16
WildBench: Benchmarking LLMs with Challenging Tasks from Real Users in the Wild Paper • 2406.04770 • Published Jun 7, 2024 • 28
Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research Paper • 2402.00159 • Published Jan 31, 2024 • 62
The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning Paper • 2312.01552 • Published Dec 4, 2023 • 30
Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs Paper • 2311.05657 • Published Nov 9, 2023 • 27
The Generative AI Paradox: "What It Can Create, It May Not Understand" Paper • 2311.00059 • Published Oct 31, 2023 • 18