L-Hongbin
's Collections
Search, Verify and Feedback: Towards Next Generation Post-training
Paradigm of Foundation Models via Verifier Engineering
Paper
•
2411.11504
•
Published
•
20
Top-nσ: Not All Logits Are You Need
Paper
•
2411.07641
•
Published
•
20
Adaptive Decoding via Latent Preference Optimization
Paper
•
2411.09661
•
Published
•
10
When Precision Meets Position: BFloat16 Breaks Down RoPE in Long-Context
Training
Paper
•
2411.13476
•
Published
•
15
Viewer
•
Updated
•
2.2M
•
6.31k
•
284
Hymba: A Hybrid-head Architecture for Small Language Models
Paper
•
2411.13676
•
Published
•
40
TÜLU 3: Pushing Frontiers in Open Language Model Post-Training
Paper
•
2411.15124
•
Published
•
58
Star Attention: Efficient LLM Inference over Long Sequences
Paper
•
2411.17116
•
Published
•
49
O1 Replication Journey -- Part 2: Surpassing O1-preview through Simple
Distillation, Big Progress or Bitter Lesson?
Paper
•
2411.16489
•
Published
•
42
MH-MoE:Multi-Head Mixture-of-Experts
Paper
•
2411.16205
•
Published
•
24
nGPT: Normalized Transformer with Representation Learning on the
Hypersphere
Paper
•
2410.01131
•
Published
•
9
Viewer
•
Updated
•
77.7k
•
2.19k
•
331
Viewer
•
Updated
•
860k
•
3.98k
•
327
Viewer
•
Updated
•
327
•
429
•
131
allenai/tulu-3-sft-mixture
Viewer
•
Updated
•
939k
•
4.12k
•
96
CASIA-LM/ChineseWebText2.0
Viewer
•
Updated
•
2k
•
2.57k
•
19
Yi-Lightning Technical Report
Paper
•
2412.01253
•
Published
•
25
Training Large Language Models to Reason in a Continuous Latent Space
Paper
•
2412.06769
•
Published
•
74
Weighted-Reward Preference Optimization for Implicit Model Fusion
Paper
•
2412.03187
•
Published
•
9
Paper
•
2412.08905
•
Published
•
103
SPaR: Self-Play with Tree-Search Refinement to Improve
Instruction-Following in Large Language Models
Paper
•
2412.11605
•
Published
•
17
Mix-LN: Unleashing the Power of Deeper Layers by Combining Pre-LN and
Post-LN
Paper
•
2412.13795
•
Published
•
19
Paper
•
2412.15115
•
Published
•
340
A Post-Training Enhanced Optimization Approach for Small Language Models
Paper
•
2411.02939
•
Published
Viewer
•
Updated
•
133k
•
1.28k
•
135
How to Synthesize Text Data without Model Collapse?
Paper
•
2412.14689
•
Published
•
48
Viewer
•
Updated
•
28M
•
1.15k
•
51
RobustFT: Robust Supervised Fine-tuning for Large Language Models under
Noisy Response
Paper
•
2412.14922
•
Published
•
85
DRT-o1: Optimized Deep Reasoning Translation via Long Chain-of-Thought
Paper
•
2412.17498
•
Published
•
21
B-STaR: Monitoring and Balancing Exploration and Exploitation in
Self-Taught Reasoners
Paper
•
2412.17256
•
Published
•
45
OpenRFT: Adapting Reasoning Foundation Model for Domain-specific Tasks
with Reinforcement Fine-Tuning
Paper
•
2412.16849
•
Published
•
9
rStar-Math: Small LLMs Can Master Math Reasoning with Self-Evolved Deep
Thinking
Paper
•
2501.04519
•
Published
•
232
MiniMax-01: Scaling Foundation Models with Lightning Attention
Paper
•
2501.08313
•
Published
•
259
OpenCSG Chinese Corpus: A Series of High-quality Chinese Datasets for
LLM Training
Paper
•
2501.08197
•
Published
•
7