--- license: apache-2.0 library_name: peft tags: - trl - dpo - generated_from_trainer base_model: mistralai/Mistral-7B-v0.1 model-index: - name: zephyr-7b results: [] --- # zephyr-7b This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.3625 - Rewards/chosen: -150.6127 - Rewards/rejected: -146.3050 - Rewards/accuracies: 0.2421 - Rewards/margins: -4.3077 - Logps/rejected: -14705.8975 - Logps/chosen: -15130.1680 - Logits/rejected: 13.5362 - Logits/chosen: 13.4716 - Use Label: 11165.9844 - Pred Label: 7522.0161 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 4 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - total_eval_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen | Use Label | Pred Label | |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|:----------:|:----------:| | 0.6637 | 0.1 | 100 | 0.6642 | -0.0947 | -0.1635 | 0.3254 | 0.0687 | -91.7446 | -78.3734 | -2.0927 | -2.1253 | 1838.9207 | 17.0794 | | 0.3902 | 0.21 | 200 | 0.3930 | -14.4219 | -14.0352 | 0.2560 | -0.3866 | -1478.9202 | -1511.0870 | 2.8471 | 2.7727 | 3444.6985 | 515.3016 | | 0.3845 | 0.31 | 300 | 0.3786 | -23.0869 | -24.5685 | 0.2520 | 1.4817 | -2532.2498 | -2377.5872 | 5.4283 | 5.3070 | 4579.4922 | 1484.5079 | | 0.3477 | 0.42 | 400 | 0.3622 | -111.3259 | -109.5294 | 0.25 | -1.7965 | -11028.3408 | -11201.4893 | 11.6816 | 11.5716 | 5682.4922 | 2485.5081 | | 0.3468 | 0.52 | 500 | 0.3613 | -144.7782 | -140.7408 | 0.2421 | -4.0373 | -14149.4824 | -14546.7158 | 13.8885 | 13.8347 | 6784.2383 | 3487.7620 | | 0.33 | 0.63 | 600 | 0.3605 | -143.0167 | -138.8336 | 0.2401 | -4.1831 | -13958.7627 | -14370.5693 | 12.5943 | 12.5399 | 7857.4287 | 4518.5713 | | 0.3665 | 0.73 | 700 | 0.3614 | -150.1877 | -145.8865 | 0.2421 | -4.3011 | -14664.0518 | -15087.6680 | 13.4024 | 13.3367 | 8936.4287 | 5543.5713 | | 0.3731 | 0.84 | 800 | 0.3623 | -150.4385 | -146.1303 | 0.2401 | -4.3082 | -14688.4258 | -15112.7539 | 13.5339 | 13.4696 | 10050.3330 | 6533.6665 | | 0.3696 | 0.94 | 900 | 0.3625 | -150.6127 | -146.3050 | 0.2421 | -4.3077 | -14705.8975 | -15130.1680 | 13.5362 | 13.4716 | 11165.9844 | 7522.0161 | ### Framework versions - PEFT 0.7.1 - Transformers 4.38.2 - Pytorch 2.1.1+cu121 - Datasets 2.14.6 - Tokenizers 0.15.2