Information-Guided Identification of Training Data Imprint in (Proprietary) Large Language Models Paper • 2503.12072 • Published Mar 15
HALoGEN: Fantastic LLM Hallucinations and Where to Find Them Paper • 2501.08292 • Published Jan 14 • 17
Inference-Time Policy Adapters (IPA): Tailoring Extreme-Scale LMs without Fine-tuning Paper • 2305.15065 • Published May 24, 2023 • 1
CONDAQA: A Contrastive Reading Comprehension Dataset for Reasoning about Negation Paper • 2211.00295 • Published Nov 1, 2022
The Art of Saying No: Contextual Noncompliance in Language Models Paper • 2407.12043 • Published Jul 2, 2024 • 4
WildHallucinations: Evaluating Long-form Factuality in LLMs with Real-World Entity Queries Paper • 2407.17468 • Published Jul 24, 2024
Question Answering for Privacy Policies: Combining Computational and Legal Perspectives Paper • 1911.00841 • Published Nov 3, 2019
HALoGEN: Fantastic LLM Hallucinations and Where to Find Them Paper • 2501.08292 • Published Jan 14 • 17
WildBench: Benchmarking LLMs with Challenging Tasks from Real Users in the Wild Paper • 2406.04770 • Published Jun 7, 2024 • 31
Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research Paper • 2402.00159 • Published Jan 31, 2024 • 64
The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning Paper • 2312.01552 • Published Dec 4, 2023 • 33