user
stringlengths
3
28
created_at
timestamp[us]
body
stringlengths
1
173k
issue_number
int64
1
2.53k
yiyepiaoling0715
2024-12-30T08:03:56
![image](https://github.com/user-attachments/assets/382ffa50-f3c6-4cd9-aabd-27e882409ed3)
2,532
qgallouedec
2024-12-30T14:49:29
> * [x] Any code provided is minimal, complete, and reproducible ([more on MREs](https://docs.github.com/en/get-started/writing-on-github/working-with-advanced-formatting/creating-and-highlighting-code-blocks)) can you please minimise your code? It seems like the error occurs at generation; what the input of the model here?: ``` | | 2024-12-30 10:53:44.559 | [rank4]: File "/opt/conda/lib/python3.11/site-packages/transformers/generation/utils.py", line 3254, in _sample | | | 2024-12-30 10:53:44.559 | [rank4]: outputs = model_forward(**model_inputs, return_dict=True) | ``` Can you reproduce the error without all the training logic?
2,532
yiyepiaoling0715
2024-12-30T04:55:00
same question,how to resolve thie?
2,529
HuggingFaceDocBuilderDev
2024-12-28T13:27:00
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2527). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,527
August-murr
2024-12-28T06:35:20
I recommend using GitHub Actions since they run the tests more reliably. Just enable it on your fork, push your changes, and it’ll automatically trigger the tests.
2,524
AMindToThink
2024-12-28T19:48:24
Does this mean that my environment is not set up incorrectly?
2,524
AMindToThink
2024-12-29T03:05:06
Thank you, took a while to figure out, but the tests that were triggered when I made an empty .py file in trl/trl worked. Somewhat bothersome that it tries and fails to post the results to slack, but the tests themselves pass. `Error: Need to provide at least one botToken or webhookUrl` I would appreciate if the [contributing](https://github.com/huggingface/trl/blob/main/CONTRIBUTING.md) document explained that the tests may not run properly locally and are auto-run by Github when changes are pushed to main. My workflow will be: Make changes to a branch of my fork. When I want to test, I'll merge my branch into main. Github will run the tests They'll fail. If on inspection the failure is because of the slack upload attempt, then everything is fine. If on inspection there was an error before the slack upload attempt, then there's a problem with my code. If my code is fine and my feature is ready, I can make a pull request.
2,524
qgallouedec
2024-12-29T10:19:28
Which tests fail locally?
2,524
AMindToThink
2024-12-30T19:00:10
Oddly, it says 6 failed when I only see 5. I'm on this commit: `commit aed5da580e9fcba6517460daf65106bc42fb6167 (upstream/main, origin/sac, sac) Author: Quentin Gallouédec <[email protected]> Date: Sun Dec 22 12:44:07 2024 +0100` ` 📦 Packing documentation (#2503)` These are the failures: ``` [gw2] FAILED tests/test_dpo_trainer.py::DPOTrainerTester::test_dpo_lora_bf16_autocast_llama [gw11] FAILED tests/test_gkd_trainer.py::GKDTrainerTester::test_gkd_trainer [gw12] FAILED tests/test_callbacks.py::WinRateCallbackTester::test_basic [gw11] FAILED tests/test_peft_models.py::PeftModelTester::test_create_bnb_peft_model_from_config [gw15] FAILED tests/test_xpo_trainer.py::TestXPOTrainer::test_training_with_peft | 0/50 [00:00<?, ?it/s] ================== 6 failed, 345 passed, 25 skipped, 242 warnings, 45 rerun in 113.62s (0:01:53) =================== ```
2,524
umbilnm
2024-12-27T09:11:21
Fixes #2400
2,521
umbilnm
2024-12-29T13:33:30
@qgallouedec Hello, can you merge? or something else needed from me?
2,521
HuggingFaceDocBuilderDev
2024-12-26T19:07:53
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2520). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,520
oliveiraeliel
2024-12-28T02:19:22
Hi, I have the same question as you do. I think that there must be some easy way to simply write a reward function as a `nn.Module`, so we don't have to refactor anything, but I didn't tried it yet. But I also think that `PPOTrainer` should accept a `custom_get_reward_function` as an optional parameter. In this case, anyone could define its own reward function, and would be a clean solution.
2,518
nityadav
2024-12-29T19:24:39
@yananchen1989 Thanks for posting this as I was stuck with a similar issue (but for `OnlineDPOTrainer`). The easiest workaround for me was to subclass the trainer class (`OnlineDPOTrainer`) and override the `training_step` with my custom `get_reward` logic, and rest of the implementation being the same as in the original method.
2,518
August-murr
2024-12-30T18:28:29
@yananchen1989 @oliveiraeliel @nityadav @hwhyyds @schmidtj3 This has been a recurring question, so before implementing a solution, I would like to ask you all for examples of when you would need this feature so that we can think of a good solution.
2,518
yananchen1989
2024-12-30T18:51:22
correct me if i am wrong. would like to know the primary motivation to rewrite the dpo from older version to current trainer unified version. maybe for better efficiency ? i understand that recent TRL versions wants to unify the pipeline in a more neat and organized manner across these different RL methods, where Trainer is the pivotal module and kick off the trainer.train() and all set. so for some methods like ppo the reward module is needed, it is also directed passed into the trainer. while for say dpo, sft, there is no provision for reward module. however. this could cause excessive encapsulation since it if hard to modularize the the reward module. the core reason is that in practical cases, reward module can be of any form not just a single torch.nn module which just score the whole output. the reward module may be a mixture and may be of dependence on external parameters, the prompt and most importantly it can not score the ppo trainer' outputs in a batch mode. anyway the flexibility is significantly reduced. although as you know the current unified pipeline is very fine with other mothods such as dpo as they do not have the reward concerns and the reward module is implicitly expressed within the algorithm. in my view, there is no need to rigidly transfer these rl methods into a unified training framework. pls advise.
2,518
August-murr
2024-12-31T06:52:17
Ultimately, TRL is a Hugging Face library built on top of Transformers and is part of the Hugging Face ecosystem. If the Trainer does limit flexibility, then Transformers will need to adapt; otherwise, we will have to maintain a much larger and more complex codebase. We'll come up with a way to add these features and prepare a PR soon!
2,518
August-murr
2024-12-31T06:52:44
@qgallouedec, do you want to comment?
2,518
qgallouedec
2024-12-31T07:30:41
Maybe having a `reward_func` arg of type `Callable` is an option. Alternatively, releasing the type of `reward_model` to accept any `Callable` is also an option. But given that a custom reward func won't return the same type/shape as proper `reward_model` I'm a bit afraid that it would require overcomplicated logic. In any case, I believe that the best approach is to discuss around a PR if anyone is willing to propose their approach
2,518
yananchen1989
2024-12-31T12:57:59
i hear u. thanks
2,518
dawidm
2024-12-27T20:40:20
Update: this approach (PR #2516) introduce another problem, because incrementing `self.state.global_step` by more than 1 needs parameters like `logging_steps` be divisible by the value of the increment. Solutions for this are: 1. Require `logging_steps` etc. to be divisible by `args.num_mini_batches * args.num_ppo_epochs`. 2. Change convention for what `step` is in RLOO - don't multiply `self.state.max_steps` by `args.num_mini_batches * args.num_ppo_epochs` (making `step` an equivalent of `episode`). I prefer the second one because it's simpler, but I'd appreciate comments on this. I'll update the PR. edit: 2. is also consistent with documentation: > episode: episode: The current global step or episode count in the training process.
2,515
dawidm
2024-12-29T13:12:02
Of course there's also 3. solution: update `global_step` after actual optimizer step (inside minibatch PPO loop), but also logging should have been moved here in this case. This will keep the most "correct" (i think) convention of steps but it requires the most changes.
2,515
SwayamInSync
2024-12-21T19:58:52
This accounted with `SFTTrainer` if this is a general issue with `Trainer` from transformers, can be re-located there
2,514
HuggingFaceDocBuilderDev
2024-12-21T12:12:26
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2513). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,513
HuggingFaceDocBuilderDev
2024-12-21T00:10:35
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2512). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,512
HuggingFaceDocBuilderDev
2024-12-20T23:42:15
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2511). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,511
HuggingFaceDocBuilderDev
2024-12-20T21:43:27
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2510). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,510
HuggingFaceDocBuilderDev
2024-12-20T16:10:32
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2509). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,509
metric-space
2024-12-21T21:30:33
@aivolcano There is a notebook that is related to this. The updated notebook is here: https://github.com/huggingface/trl/blob/main/examples/notebooks/best_of_n.ipynb
2,508
aivolcano
2024-12-27T08:53:25
thank u so much
2,508
HuggingFaceDocBuilderDev
2024-12-20T11:30:43
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2507). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,507
Mecoli1219
2024-12-20T06:46:11
Wait for https://github.com/linkedin/Liger-Kernel/pull/492
2,506
metric-space
2024-12-21T21:33:11
@nguyenhoa-uit I can help out with this as this was code I wrote more than a year ago. Mind you, I'll be very very slow. Let me take a look
2,505
metric-space
2024-12-23T09:46:31
@nguyenhoa-uit could you try this bit : https://github.com/huggingface/trl/blob/main/trl/trainer/ddpo_config.py#L64 ?
2,505
nguyenhoa-uit
2024-12-25T02:18:37
> @nguyenhoa-uit could you try this bit : https://github.com/huggingface/trl/blob/main/trl/trainer/ddpo_config.py#L64 ? When I used checkpoint resume from in config file, I ran and had a bug at https://github.com/huggingface/trl/blob/main/trl/trainer/ddpo_trainer.py#L541C20-L541C42 When I passed with try catch, it didnot use the parameters from this checkpoint but base model.
2,505
ggbetz
2024-12-20T15:19:13
It seems @philschmid has in implementation here: https://github.com/philschmid/deep-learning-pytorch-huggingface/blob/391f19ba06c128a2a290b3bdcb717ad6ff794fd7/training/scripts/run_sft.py#L54-L77 and the question is maybe just what's the best cleanest way to integrate this natively in trl?
2,504
anakin87
2024-12-21T16:25:29
This would be great and would prevent users from making mistakes in the manual implementation of this method: for example, [the code for integration with other libraries reported in the official repo](https://github.com/cognitivecomputations/spectrum?tab=readme-ov-file) has some problems. In contrast, the simple implementation in [my tutorial](https://huggingface.co/blog/anakin87/spectrum) and Philipp's code should be correct. BTW, Spectrum is quite agnostic with respect to training method (SFT, DPO...): the [models by VAGO solutions](https://huggingface.co/VAGOsolutions) show that it works well for DPO too. LMK what's the better way to proceed and help with this integration.
2,504
HuggingFaceDocBuilderDev
2024-12-19T10:50:45
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2503). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,503
HuggingFaceDocBuilderDev
2024-12-19T10:13:19
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2502). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,502
qgallouedec
2024-12-23T12:38:06
Can you screenshot a result?
2,501
HuggingFaceDocBuilderDev
2024-12-23T12:41:22
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2501). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,501
yaricom
2024-12-23T12:43:48
Sure, here is s screenshot from my account at Comet. <img width="2106" alt="Screenshot 2024-12-23 at 14 42 20" src="https://github.com/user-attachments/assets/69629fdb-77de-4a2d-b1d2-087889d96a4c" />
2,501
yaricom
2024-12-23T12:45:02
And this is a DataFrame encoded as CSV. [game_log.csv](https://github.com/user-attachments/files/18229453/game_log.csv)
2,501
yaricom
2024-12-23T13:08:10
The script I was using to test DPO trainer integration. ```python import os from datasets import load_dataset from transformers import AutoModelForCausalLM, AutoTokenizer from trl import DPOConfig, DPOTrainer os.environ["TOKENIZERS_PARALLELISM"] = "false" def main(): output_dir = "models/minimal/dpo_my" model_id = "trl-internal-testing/tiny-Qwen2ForCausalLM-2.5" # model_id = "Qwen/Qwen2-0.5B-Instruct" model = AutoModelForCausalLM.from_pretrained(model_id) ref_model = AutoModelForCausalLM.from_pretrained(model_id) tokenizer = AutoTokenizer.from_pretrained(model_id) tokenizer.pad_token = tokenizer.eos_token training_args = DPOConfig( output_dir=output_dir, per_device_train_batch_size=2, max_steps=1, remove_unused_columns=False, gradient_accumulation_steps=8, precompute_ref_log_probs=False, learning_rate=5.0e-7, eval_strategy="steps", eval_steps=1, report_to="all", generate_during_eval=True, max_length=1024, ) # dummy_dataset = load_dataset("trl-internal-testing/zen", "standard_preference") dummy_dataset = load_dataset("trl-lib/ultrafeedback_binarized", "default") dummy_dataset["train"] = dummy_dataset["train"].select(range(20)) dummy_dataset["test"] = dummy_dataset["test"].select(range(40)) trainer = DPOTrainer( model=model, ref_model=ref_model, args=training_args, processing_class=tokenizer, train_dataset=dummy_dataset["train"], eval_dataset=dummy_dataset["test"], ) trainer.train() trainer.evaluate() if __name__ == "__main__": main() ``` Do not forget to set `COMET_APY_KEY` environment variable while executing it.
2,501
asparius
2024-12-18T13:50:40
trl uses accelerate which supports FSDP. However there is no recommeded config of FSDP in the repo unlike DeepSpeed, so you could refer to this [page](https://huggingface.co/docs/accelerate/en/usage_guides/fsdp) for FSDP. All in all, DPO and trl supports FSDP but not for online algo like PPO #1726.
2,500
yingtongxiong
2024-12-19T05:57:01
> trl uses accelerate which supports FSDP. However there is no recommeded config of FSDP in the repo unlike DeepSpeed, so you could refer to this [page](https://huggingface.co/docs/accelerate/en/usage_guides/fsdp) for FSDP. All in all, DPO and trl supports FSDP but not for online algo like PPO #1726. @asparius Thank you very much
2,500
HuggingFaceDocBuilderDev
2024-12-17T23:16:28
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2499). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,499
HuggingFaceDocBuilderDev
2024-12-17T19:12:30
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2498). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,498
qgallouedec
2024-12-17T22:30:39
Yeah! thanks @sergiopaniego 🤘
2,498
asparius
2024-12-18T14:14:35
This has been noted previously #2281. I believe this was introduced in PPOv2 which was replication of the openai tldr paper which also contains this INVALID_LOGPROB=1.0 which does not break training because it cancels out at kl reward. Perhaps @vwxyzjn can tell why this was used, instead of masked_mean version
2,496
Mecoli1219
2024-12-20T05:30:02
Hi, I want to check that SimPO is in CPO instead of DPO, right?
2,495
qgallouedec
2024-12-20T11:01:35
> Hi, I want to check that SimPO is in CPO instead of DPO, right? Correct! Message modified
2,495
HuggingFaceDocBuilderDev
2024-12-17T08:16:37
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2494). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,494
qgallouedec
2024-12-17T11:19:51
Probably simpler: ```python from huggingface_hub import ModelCard model_card = ModelCard(""" --- tags: [trl] --- # Some title """) if script_args.push_to_hub: model_card.push_to_hub(script_args.repo_id, repo_type="dataset") ```
2,491
August-murr
2024-12-17T12:15:50
Well, that's one way to overengineer it I also opened [issue on datasets](https://github.com/huggingface/datasets/issues/7336) to clarify. I assume the next step is to add this to all the dataset scripts.
2,491
qgallouedec
2024-12-17T13:14:11
Very good like this
2,491
HuggingFaceDocBuilderDev
2024-12-25T17:41:32
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2491). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,491
August-murr
2024-12-29T14:01:58
it doesn't add all the details requested in the issue #2470 but It's an improvement
2,491
qgallouedec
2024-12-16T12:14:28
Thanks for reporting, please provide a *minimal* code/steps to reproduce this.
2,490
sagie-dekel
2024-12-16T12:48:53
pipeline.zip (edit by maintainer: remove link) thanks @qgallouedec The attached files constitute a pipeline that using the DPOTrainer with DeepSpeed. I am sorry that its ain't minimal but i don't see easy way to reproduce. if you prefer I can write the main steps.
2,490
qgallouedec
2024-12-16T13:52:12
Sorry but we don't use zip files. The easy way to provide a MRE is to go line by line, if the error remains when you remove it, then you can discard the line. When there is no line left to remove, you have your MRE
2,490
sagie-dekel
2024-12-16T16:16:47
Sorry @qgallouedec , here a minimal code of my pipeline: ```python import argparse import json import math import pandas as pd import torch from accelerate import Accelerator from datasets import Dataset from torch import optim, nn from torch.optim.lr_scheduler import LambdaLR from tqdm import tqdm from transformers import AutoTokenizer, AutoModelForMaskedLM, AutoModel, AutoModelForCausalLM from trl import PPOConfig, PPOTrainer, AutoModelForCausalLMWithValueHead, DPOConfig, DPOTrainer from huggingface_hub import login import os import itertools model_RLRF = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.1-8B-Instruct", torch_dtype=torch.float32) tokenizer_RLRF = AutoTokenizer.from_pretrained(model_RLRF_name_or_path) tokenizer_RLRF.add_special_tokens({'pad_token': tokenizer_RLRF.eos_token}) tokenizer_RLRF.padding_side = 'left' DPO_config = DPOConfig( report_to='tensorboard', logging_first_step=True, "per_device_train_batch_size": 4, "gradient_accumulation_steps": 1, "sync_ref_model": true, "ref_model_mixup_alpha": 0.6, "ref_model_sync_steps": 256, "bf16": True, ) # Create reference model: parameter_names = [n for n, _ in model_RLRF.named_parameters()] ref_model = deepcopy(model) # if no layers are shared, return copy of model for param_name in parameter_names: param = ref_model.get_parameter(param_name) param.requires_grad = False ref_model.eval() # Set optimizer for RLRF optimizer_RLRF = optim.AdamW(filter(lambda param: param.requires_grad, model_RLRF.parameters()), lr=1.41e-5) train_dataset= pd.read_csv("perfernces_dataset_from_ranker_train_queries_and_baseline_doc.csv") train_dataset= Dataset.from_pandas(train_dataset_RLRF) dpo_trainer = DPOTrainer(model=model, args=DPO_config, processing_class=tokenizer_RLRF, ref_model=ref_model, optimizers=(optimizer, None), train_dataset=train_dataset) dpo_trainer.train() ``` the loaded data file (train_dataset) is: [perfernces_dataset_from_ranker_train_queries_and_baseline_doc.csv](https://github.com/user-attachments/files/18153383/perfernces_dataset_from_ranker_train_queries_and_baseline_doc.csv)
2,490
qgallouedec
2024-12-16T11:04:09
Good point, given that for other trainers (like DPO), it's a truncation. In fact, the best thing would be to have a common behavior for all trainers (truncation), but the urgent thing is to clarify the documentation.
2,488
HuggingFaceDocBuilderDev
2024-12-16T09:16:23
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2487). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,487
Ciao-CA
2024-12-20T07:32:59
I have the same problem
2,486
HuggingFaceDocBuilderDev
2024-12-15T19:39:31
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2485). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,485
HuggingFaceDocBuilderDev
2024-12-15T18:22:29
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2484). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,484
HuggingFaceDocBuilderDev
2024-12-15T16:35:22
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2483). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,483
HuggingFaceDocBuilderDev
2024-12-15T12:58:48
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2482). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,482
qgallouedec
2024-12-15T15:34:21
2 questions/remarks: - can you run benchmark so that we can (1) quantify the improvement and (2) check that results with and without liger are the same - we could have an additional tag for the hub when a model is trained with liger
2,482
qgallouedec
2024-12-15T15:48:46
I think we should bump liger version to v0.5 (it doesn't include the loss before), see https://github.com/linkedin/Liger-Kernel/releases/tag/v0.5.0
2,482
kashif
2024-12-18T10:46:56
waiting on https://github.com/linkedin/Liger-Kernel/pull/486
2,482
kashif
2024-12-19T10:09:55
waiting on https://github.com/huggingface/trl/pull/2502
2,482
qgallouedec
2024-12-19T10:33:44
@kashif can you share the curves once it's ready?
2,482
kashif
2024-12-29T14:45:28
tests fail as they need: https://github.com/linkedin/Liger-Kernel/pull/503
2,482
kashif
2024-12-15T09:31:55
@hteague-qti so I wanted to get it working with this collator and then come back and make it more general after that.. so would you have a suggestion on what the next generalization could be? make it work for the SFT default collator?
2,481
hteague-qti
2024-12-16T19:28:17
I was thinking it could be made completely independent of the collator. First thing might be to warn users that even though they are providing a collator in the args, you are switching to a different one (for now). Seems to me that trainer should not care about the data preprocessing or the collator, just the output logits, etc. Making it work with default collator in SFT would be fine. This one is quite common for language: DataCollatorForCompletionOnlyLM
2,481
hteague-qti
2024-12-19T21:39:17
btw, appreciate the response.
2,481
HuggingFaceDocBuilderDev
2024-12-14T21:45:10
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2480). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,480
August-murr
2024-12-14T18:57:56
Before adding it to all the trainers, what do you think of the overall structure? Is it okay to include the tools in each trainer configuration?
2,479
qgallouedec
2024-12-14T19:05:11
Thanks for this addition! Let's keep things as separate as possible, and keep this PR for DPO only. The code as is looks good to me. The only question is: can this type (`Optional[list[Union[dict, Callable]]]`) being parsed. I'll try.
2,479
qgallouedec
2024-12-14T19:27:17
That's why I thought: ```python from trl import DPOConfig, TrlParser parser = TrlParser((DPOConfig,)) parser.parse_args_and_config() ``` ``` $ python 2479.py --output_dir out --tools "{'type': 'function', 'function': {'name': 'multiply', 'description': 'A function that multiplies two numbers', 'parameters': {'type': 'object', 'properties': {'a': {'type': 'number', 'description': 'The first number to multiply'}, 'b': {'type': 'number', 'description': 'The second number to multiply'}}, 'required': ['a', 'b']}}}" [...] 2479.py: error: argument --tools: invalid Union value: "{'type': 'function', 'function': {'name': 'multiply', 'description': 'A function that multiplies two numbers', 'parameters': {'type': 'object', 'properties': {'a': {'type': 'number', 'description': 'The first number to multiply'}, 'b': {'type': 'number', 'description': 'The second number to multiply'}}, 'required': ['a', 'b']}}}" ``` I'm not sure what the best way to handle it right now, I'll sleep on it.
2,479
August-murr
2024-12-15T08:51:45
> Let's keep things as separate as possible, and keep this PR for DPO only. a different PR for each trainer then? > can this type `(Optional[list[Union[dict, Callable]]])` being parsed. Adding tools to the CLI would be quite complicated. It wouldn't be practical to add all the tools into the CLI. My best guess is to read the functions from another source, like another script, if there’s a request for it later.
2,479
August-murr
2024-12-16T08:22:54
does this need anything else? test or docs?
2,479
August-murr
2024-12-25T13:16:53
I also wanted to add it to `SFTTrainer` but it doesn't use `maybe_apply_chat_template`
2,479
HuggingFaceDocBuilderDev
2024-12-13T20:46:51
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2476). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,476
HuggingFaceDocBuilderDev
2024-12-13T19:02:02
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2475). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,475
HuggingFaceDocBuilderDev
2024-12-13T17:43:29
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2474). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,474
asparius
2024-12-14T00:28:52
It utilizes `self.model`, which is defined in [[this line](https://github.com/huggingface/trl/blob/6d4ed070f1f53a87fb3cff2eb82a56db093bccc6/trl/trainer/rloo_trainer.py#L162)](https://github.com/huggingface/trl/blob/6d4ed070f1f53a87fb3cff2eb82a56db093bccc6/trl/trainer/rloo_trainer.py#L162). This approach is also adopted in `PPOTrainer`. I believe this is a deliberate nomenclature choice, designed to remain consistent across various preference learning frameworks without introducing the complexity of aligning with the diverse terminologies used in academic papers.
2,472
qgallouedec
2024-12-13T16:33:05
Yes, that's a good point! All datasets in [hf.co/trl-lib](https://huggingface.co/trl-lib) are taken from an original dataset. We should at least indicate this dataset in the readme with something like: ``` This dataset is a processed version of [openbmb/UltraFeedback](https://huggingface.co/datasets/openbmb/UltraFeedback) with this [script](https://github.com/huggingface/trl/blob/main/examples/datasets/ultrafeedback.py). ``` To do this, we should add to all script in https://github.com/huggingface/trl/blob/main/examples/datasets a model card that we push, like in https://github.com/huggingface/trl/blob/179ba5367181d9bd4bdaec70d50789b09754d04a/scripts/generate_tiny_models.py#L69-L97 We could also add the type/format of dataset with a link to the relevant section in this page of the documentation: https://huggingface.co/docs/trl/en/dataset_formats
2,470
qgallouedec
2024-12-13T16:44:51
What you're describing sounds closer to _padding-free_ than packing. We have a (currently draft) PR for this: #2437. Can you confirm that's it is what you're describing? --- At this point I'm not even sure that packing for DPO makes sense. How to ensure that you've as many chosen than rejected? How to ensure they match? How to handle partial sequences?
2,469
zhc7
2024-12-13T17:16:15
Hi, thank you for your response. I looked into the link you provided. I think we are talking about the same thing. I used the word "packing" from https://huggingface.co/blog/packing-with-FA2. The "packing" here actually means concatenating a fixed batch size of samples into one sequence, and use `position_ids` to mark the boundaries, rather than packing to a fixed length. So there won't be the problems you mentioned. I've also briefly read https://huggingface.co/blog/mayank-mishra/padding-free-transformer this blog, I think the ideas are the same. But I'm not sure how the latter is implemented. Maybe they are the same thing just with different names:) I breifly went through the pr, I see it is trying to add `position_ids` in the whole process, so I guess we are talking about the same thing.
2,469
qgallouedec
2024-12-13T16:51:33
That's a good point! Feel free to open a PR to fix this. I don't think adding a unittest for this is relevant. If possible, add plots (eg, with wandb) before/after to ensure that we aren't introducing a regression
2,468
zhc7
2024-12-13T17:17:59
Ofcourse! ![image](https://github.com/user-attachments/assets/2da93fdf-a29d-41a1-974a-2b640e3a6ee6) here's a graph for the same training with and without the modification. You can see the pink line is a lot more smoother. Especially the accuracy graph. My `per_device_batch_size` is 2 so the accuracy per device can only be 1, 0.5 or 0.
2,468
qgallouedec
2024-12-13T17:34:35
Perfect!
2,468
HuggingFaceDocBuilderDev
2024-12-12T14:04:44
The docs for this PR live [here](/static-proxy?url=https%3A%2F%2Fmoon-ci-docs.huggingface.co%2Fdocs%2Ftrl%2Fpr_2467). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
2,467
qgallouedec
2024-12-13T17:52:47
That's very interesting! It would be a nice improvement. If you want to tackle this problem, you should be aware that packing will be implemented differently (in a simpler way) in the near future, see #2405. You should branch from there.
2,466
qgallouedec
2024-12-12T10:44:50
First, SAC is designed for continuous action spaces, whereas NLP tasks involve discrete token outputs. (A discrete variant of SAC exists though.) Second, SAC lacks a mechanism to constrain the policy from deviating too far from the initial model. PPO, on the other hand, explicitly limits policy updates, which is crucial in RLHF to maintain alignment and preserve the pretrained model’s capabilities. Without such constraints, SAC could result in a policy that either drifts excessively or remains overly similar to the original model. Finally, SAC's entropy maximization encourages broad exploration, which may be counterproductive in RLHF. Finetuning typically involves domain-specific data, and excessive exploration could lead to unaligned or undesirable behaviors. This mechanism might inadvertently encourage unlearning of pretrained knowledge. That said, these points are speculative and based on intuition. I'd be eager to see papers or results that either confirm or challenge these hypotheses.
2,465
AMindToThink
2024-12-13T21:11:39
Thank you for the response. The [RLOO trainer](https://arxiv.org/pdf/2402.14740) also lacks PPO's clipping mechanism that constrains the policy from deviating too far from the previous policy. [It turns out](https://huggingface.co/blog/putting_rl_back_in_rlhf_with_rloo) that for RLHF on pretrained language models, that clipping step is not necessary. If you are referring to the reference policy, I don't see why a KL divergence term with a reference policy cannot be included into the SAC loss function. Mode collapse and loss of variety is a common problem for aligned models, so if SAC makes a different tradeoff, encouraging exploration, then that could be useful.
2,465
qgallouedec
2024-12-13T22:13:21
> lacks PPO's clipping mechanism that constrains the policy from deviating too far from the previous policy There is a KL term though > I don't see why a KL divergence term with a reference policy cannot be included into the SAC loss function. I guess you can, it's just that in its classic formulation, the SAC objective doesn't contain such a term. > encouraging exploration, then that could be useful I'm not so sure about that. But if you manage to produce or find results that help to see more clearly on this matter, please share them.
2,465

Stars

import requests
from datetime import datetime
from datasets import Dataset
import pyarrow as pa
import os

def get_stargazers(owner, repo, token):
    # Initialize the count and the page number
    page = 1
    stargazers = []
    while True:
        # Construct the URL for the stargazers with pagination
        stargazers_url = f"https://api.github.com/repos/{owner}/{repo}/stargazers?page={page}&per_page=100"

        # Send the request to GitHub API with appropriate headers
        headers = {"Accept": "application/vnd.github.v3.star+json", "Authorization": "token " + token}
        response = requests.get(stargazers_url, headers=headers)

        if response.status_code != 200:
            raise Exception(f"Failed to fetch stargazers with status code {response.status_code}: {response.text}")

        stargazers_page = response.json()

        if not stargazers_page:  # Exit the loop if there are no more stargazers to process
            break

        stargazers.extend(stargazers_page)
        page += 1  # Move to the next page

    return stargazers

token = os.environ.get("GITHUB_PAT")
stargazers = get_stargazers("huggingface", "trl", token)
stargazers = {key: [stargazer[key] for stargazer in stargazers] for key in stargazers[0].keys()}
dataset = Dataset.from_dict(stargazers)

def clean(example):
    starred_at = datetime.strptime(example["starred_at"], "%Y-%m-%dT%H:%M:%SZ")
    starred_at = pa.scalar(starred_at, type=pa.timestamp("s", tz="UTC"))
    return {"starred_at": starred_at, "user": example["user"]["login"]}

dataset = dataset.map(clean, remove_columns=dataset.column_names)
dataset.push_to_hub("qgallouedec/trl-metrics", config_name="stargazers")

Pypi downloads

from datasets import Dataset
from google.cloud import bigquery
import os

os.environ["GOOGLE_APPLICATION_CREDENTIALS"] = "propane-tree-432413-4c3e2b5e6b3c.json"

# Initialize a BigQuery client
client = bigquery.Client()

# Define your query
query = """
#standardSQL
WITH daily_downloads AS (
  SELECT
    DATE(timestamp) AS day,
    COUNT(*) AS num_downloads
  FROM
    `bigquery-public-data.pypi.file_downloads`
  WHERE
    file.project = 'trl'
    -- Filter for the last 12 months
    AND DATE(timestamp) BETWEEN DATE_SUB(CURRENT_DATE(), INTERVAL 54 MONTH) AND CURRENT_DATE()
  GROUP BY
    day
)
SELECT
  day,
  num_downloads
FROM
  daily_downloads
ORDER BY
  day DESC
"""

# Execute the query
query_job = client.query(query)

# Fetch the results
results = query_job.result()

# Convert the results to a pandas DataFrame and then to a Dataset
df = results.to_dataframe()
dataset = Dataset.from_pandas(df)

dataset.push_to_hub("qgallouedec/trl-metrics", config_name="pypi_downloads")

Models tagged

from huggingface_hub import HfApi
from datasets import Dataset

api = HfApi()
models = api.list_models(tags="trl")
dataset_list = [{"id": model.id, "created_at": model.created_at, "likes": model.likes, "downloads": model.downloads, "tags": model.tags} for model in models]
dataset_dict = {key: [d[key] for d in dataset_list] for key in dataset_list[0].keys()}
dataset = Dataset.from_dict(dataset_dict)
dataset.push_to_hub("qgallouedec/trl-metrics", config_name="models")

Issues and comments

import requests
from datetime import datetime
import os
from datasets import Dataset
from tqdm import tqdm

token = os.environ.get("GITHUB_PAT")

def get_full_response(url, headers, params=None):
    page = 1
    output = []
    params = params or {}
    while True:
        params = {**params, "page": page, "per_page": 100}
        response = requests.get(url, headers=headers, params=params)

        if response.status_code != 200:
            raise Exception(f"Failed to fetch issues: {response.text}")

        batch = response.json()
        if len(batch) == 0:
            break
        output.extend(batch)
        page += 1
    return output

# GitHub API URL for issues (closed and open)
issues_url = f"https://api.github.com/repos/huggingface/trl/issues"

# Set up headers for authentication
headers = {"Authorization": f"token {token}", "Accept": "application/vnd.github.v3+json"}

# Make the request
issues = get_full_response(issues_url, headers, params={"state": "all"})

issues_dataset_dict = {
    "number": [],
    "title": [],
    "user": [],
    "state": [],
    "created_at": [],
    "closed_at": [],
    "comments_count": [],
}
comments_dataset_dict = {
    "user": [],
    "created_at": [],
    "body": [],
    "issue_number": [],
}
for issue in tqdm(issues):
    # Extract relevant information
    issue_number = issue["number"]
    title = issue["title"]
    created_at = datetime.strptime(issue["created_at"], "%Y-%m-%dT%H:%M:%SZ")
    comments_count = issue["comments"]
    comments_url = issue["comments_url"]

    comments = get_full_response(comments_url, headers=headers)
    for comment in comments:
        comments_dataset_dict["user"].append(comment["user"]["login"])
        comments_dataset_dict["created_at"].append(datetime.strptime(comment["created_at"], "%Y-%m-%dT%H:%M:%SZ"))
        comments_dataset_dict["body"].append(comment["body"])
        comments_dataset_dict["issue_number"].append(issue_number)

    issues_dataset_dict["number"].append(issue_number)
    issues_dataset_dict["title"].append(title)
    issues_dataset_dict["user"].append(issue["user"]["login"])
    issues_dataset_dict["state"].append(issue["state"])
    issues_dataset_dict["created_at"].append(datetime.strptime(issue["created_at"], "%Y-%m-%dT%H:%M:%SZ"))
    issues_dataset_dict["closed_at"].append(datetime.strptime(issue["closed_at"], "%Y-%m-%dT%H:%M:%SZ") if issue["closed_at"] else None)
    issues_dataset_dict["comments_count"].append(comments_count)

issues_dataset = Dataset.from_dict(issues_dataset_dict)
comments_dataset = Dataset.from_dict(comments_dataset_dict)

issues_dataset.push_to_hub("qgallouedec/trl-metrics", config_name="issues")
comments_dataset.push_to_hub("qgallouedec/trl-metrics", config_name="issue_comments")
Downloads last month
285