S$^{2}$FT: Efficient, Scalable and Generalizable LLM Fine-tuning by Structured Sparsity Paper • 2412.06289 • Published 29 days ago
Taming Overconfidence in LLMs: Reward Calibration in RLHF Paper • 2410.09724 • Published Oct 13, 2024 • 2
Tuning Language Models as Training Data Generators for Augmentation-Enhanced Few-Shot Learning Paper • 2211.03044 • Published Nov 6, 2022 • 1
Optimizing Language Model's Reasoning Abilities with Weak Supervision Paper • 2405.04086 • Published May 7, 2024 • 1
Taming Overconfidence in LLMs: Reward Calibration in RLHF Paper • 2410.09724 • Published Oct 13, 2024 • 2
Divide, Reweight, and Conquer: A Logit Arithmetic Approach for In-Context Learning Paper • 2410.10074 • Published Oct 14, 2024
LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition Paper • 2307.13269 • Published Jul 25, 2023 • 31