Iterative Reasoning Preference Optimization
Abstract
Iterative preference optimization methods have recently been shown to perform well for general instruction tuning tasks, but typically make little improvement on reasoning tasks (Yuan et al., 2024, Chen et al., 2024). In this work we develop an iterative approach that optimizes the preference between competing generated Chain-of-Thought (CoT) candidates by optimizing for winning vs. losing reasoning steps that lead to the correct answer. We train using a modified DPO loss (Rafailov et al., 2023) with an additional negative log-likelihood term, which we find to be crucial. We show reasoning improves across repeated iterations of this scheme. While only relying on examples in the training set, our approach results in increasing accuracy for Llama-2-70B-Chat from 55.6% to 81.6% on GSM8K (and 88.7% with majority voting out of 32 samples), from 12.5% to 20.8% on MATH, and from 77.8% to 86.7% on ARC-Challenge, which outperforms other Llama-2-based models not relying on additionally sourced datasets.
Community
Very similar to my Memphis approach: https://huggingface.co/euclaise/Memphis-CoT-3B
I also experimented with an extra loss over the gold response, but ended up dropping it.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Self-Explore to Avoid the Pit: Improving the Reasoning Capabilities of Language Models with Fine-grained Rewards (2024)
- Improving Language Model Reasoning with Self-motivated Learning (2024)
- Advancing LLM Reasoning Generalists with Preference Trees (2024)
- Teaching Large Language Models to Reason with Reinforcement Learning (2024)
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Boosting AI Reasoning: Unraveling Iterative RPO for Better Logic
Links ๐:
๐ Subscribe: https://www.youtube.com/@Arxflix
๐ Twitter: https://x.com/arxflix
๐ LMNT (Partner): https://lmnt.com/
For the first term according to the equation, it seems they include the prompt in the loss as well? Seems a bit counter intuitive.
Hi! Thanks for noticing this issue. We updated the paper a few days after arXiv v1: https://arxiv.org/pdf/2404.19733
When you refine the reference model for the iterative reasoning, do you re-initialize the optimizer and learning rate scheduler?
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper