WildTeaming at Scale: From In-the-Wild Jailbreaks to (Adversarially) Safer Language Models Paper • 2406.18510 • Published Jun 26, 2024 • 8
Why Has Predicting Downstream Capabilities of Frontier AI Models with Scale Remained Elusive? Paper • 2406.04391 • Published Jun 6, 2024 • 7
Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling Paper • 2304.01373 • Published Apr 3, 2023 • 9
Quality at a Glance: An Audit of Web-Crawled Multilingual Datasets Paper • 2103.12028 • Published Mar 22, 2021 • 3
BLOOM+1: Adding Language Support to BLOOM for Zero-Shot Prompting Paper • 2212.09535 • Published Dec 19, 2022 • 1
Documenting Geographically and Contextually Diverse Data Sources: The BigScience Catalogue of Language Data and Resources Paper • 2201.10066 • Published Jan 25, 2022
What Language Model to Train if You Have One Million GPU Hours? Paper • 2210.15424 • Published Oct 27, 2022 • 2
Recasting Self-Attention with Holographic Reduced Representations Paper • 2305.19534 • Published May 31, 2023 • 2
BigBIO: A Framework for Data-Centric Biomedical Natural Language Processing Paper • 2206.15076 • Published Jun 30, 2022 • 3
GAIA Search: Hugging Face and Pyserini Interoperability for NLP Training Data Exploration Paper • 2306.01481 • Published Jun 2, 2023 • 1
The Goldilocks of Pragmatic Understanding: Fine-Tuning Strategy Matters for Implicature Resolution by LLMs Paper • 2210.14986 • Published Oct 26, 2022 • 5
VQGAN-CLIP: Open Domain Image Generation and Editing with Natural Language Guidance Paper • 2204.08583 • Published Apr 18, 2022
The Case for Co-Designing Model Architectures with Hardware Paper • 2401.14489 • Published Jan 25, 2024 • 3
Suppressing Pink Elephants with Direct Principle Feedback Paper • 2402.07896 • Published Feb 12, 2024 • 9
Transformer-Based Models Are Not Yet Perfect At Learning to Emulate Structural Recursion Paper • 2401.12947 • Published Jan 23, 2024 • 2
EleutherAI: Going Beyond "Open Science" to "Science in the Open" Paper • 2210.06413 • Published Oct 12, 2022