diff --git a/checkpoint-1014/README.md b/checkpoint-1014/README.md new file mode 100644 index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6 --- /dev/null +++ b/checkpoint-1014/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-2-13b-chat-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.2 \ No newline at end of file diff --git a/checkpoint-1014/adapter_config.json b/checkpoint-1014/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06 --- /dev/null +++ b/checkpoint-1014/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 32, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-1014/adapter_model.safetensors b/checkpoint-1014/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f3b85154358f0ead8591b0893163cf7d7aafc789 --- /dev/null +++ b/checkpoint-1014/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7eb82e9832f7b4a2302c2a811a490862c5d2a86d1e8cea5da78dba68a3c8e4e +size 26235704 diff --git a/checkpoint-1014/optimizer.pt b/checkpoint-1014/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..9de84081ded3001bcff96eef621de201fac14323 --- /dev/null +++ b/checkpoint-1014/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79df9f3009a37edf2c4cdd8919d98bebd2e64369b8024b62149f5039bead3dfe +size 52563258 diff --git a/checkpoint-1014/rng_state.pth b/checkpoint-1014/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..5d968263657a30eea713bfe7c3ac18de2db73bca --- /dev/null +++ b/checkpoint-1014/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b4c4f0166ef9ca74027402d439827bc8ed2a02c96ba6ba9e51d0a9c0f291412 +size 14244 diff --git a/checkpoint-1014/scheduler.pt b/checkpoint-1014/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..8c15d820812361823eaae11e84a9769fc4b30d88 --- /dev/null +++ b/checkpoint-1014/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78fd9b36c5271bbf8da930a0ab7ea570971f4031660dcb029549ddb0db3d5123 +size 1064 diff --git a/checkpoint-1014/trainer_state.json b/checkpoint-1014/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..78d019e0c76a37176d298d06318c46fb02548b43 --- /dev/null +++ b/checkpoint-1014/trainer_state.json @@ -0,0 +1,173 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.991137370753323, + "eval_steps": 500, + "global_step": 1014, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.29542097488921715, + "grad_norm": 3.6940183639526367, + "learning_rate": 2.5e-05, + "loss": 3.3429, + "step": 50 + }, + { + "epoch": 0.5908419497784343, + "grad_norm": 0.5006749033927917, + "learning_rate": 5e-05, + "loss": 0.7796, + "step": 100 + }, + { + "epoch": 0.8862629246676514, + "grad_norm": 0.5715117454528809, + "learning_rate": 4.726477024070022e-05, + "loss": 0.22, + "step": 150 + }, + { + "epoch": 1.1816838995568686, + "grad_norm": 0.7143971920013428, + "learning_rate": 4.452954048140044e-05, + "loss": 0.1198, + "step": 200 + }, + { + "epoch": 1.4771048744460857, + "grad_norm": 0.380874902009964, + "learning_rate": 4.179431072210066e-05, + "loss": 0.117, + "step": 250 + }, + { + "epoch": 1.7725258493353029, + "grad_norm": 0.3196457028388977, + "learning_rate": 3.9059080962800876e-05, + "loss": 0.1134, + "step": 300 + }, + { + "epoch": 2.06794682422452, + "grad_norm": 0.40986520051956177, + "learning_rate": 3.6323851203501094e-05, + "loss": 0.1151, + "step": 350 + }, + { + "epoch": 2.363367799113737, + "grad_norm": 0.3212401568889618, + "learning_rate": 3.358862144420131e-05, + "loss": 0.1138, + "step": 400 + }, + { + "epoch": 2.658788774002954, + "grad_norm": 0.16917897760868073, + "learning_rate": 3.085339168490153e-05, + "loss": 0.1115, + "step": 450 + }, + { + "epoch": 2.9542097488921715, + "grad_norm": 0.22288767993450165, + "learning_rate": 2.811816192560175e-05, + "loss": 0.1155, + "step": 500 + }, + { + "epoch": 3.2496307237813884, + "grad_norm": 0.19413484632968903, + "learning_rate": 2.538293216630197e-05, + "loss": 0.1142, + "step": 550 + }, + { + "epoch": 3.5450516986706058, + "grad_norm": 0.22448602318763733, + "learning_rate": 2.264770240700219e-05, + "loss": 0.1134, + "step": 600 + }, + { + "epoch": 3.8404726735598227, + "grad_norm": 0.2704063057899475, + "learning_rate": 1.9912472647702408e-05, + "loss": 0.1121, + "step": 650 + }, + { + "epoch": 4.13589364844904, + "grad_norm": 0.3302886486053467, + "learning_rate": 1.7177242888402626e-05, + "loss": 0.1158, + "step": 700 + }, + { + "epoch": 4.431314623338257, + "grad_norm": 0.1943623274564743, + "learning_rate": 1.4442013129102846e-05, + "loss": 0.1143, + "step": 750 + }, + { + "epoch": 4.726735598227474, + "grad_norm": 0.21939796209335327, + "learning_rate": 1.1706783369803063e-05, + "loss": 0.1106, + "step": 800 + }, + { + "epoch": 5.022156573116692, + "grad_norm": 0.3173823654651642, + "learning_rate": 8.971553610503283e-06, + "loss": 0.1131, + "step": 850 + }, + { + "epoch": 5.317577548005908, + "grad_norm": 0.2860495150089264, + "learning_rate": 6.2363238512035015e-06, + "loss": 0.1127, + "step": 900 + }, + { + "epoch": 5.612998522895126, + "grad_norm": 0.18334375321865082, + "learning_rate": 3.50109409190372e-06, + "loss": 0.1102, + "step": 950 + }, + { + "epoch": 5.908419497784343, + "grad_norm": 0.18338781595230103, + "learning_rate": 7.658643326039388e-07, + "loss": 0.1116, + "step": 1000 + } + ], + "logging_steps": 50, + "max_steps": 1014, + "num_input_tokens_seen": 0, + "num_train_epochs": 6, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.6016891541848064e+17, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-1014/training_args.bin b/checkpoint-1014/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740 --- /dev/null +++ b/checkpoint-1014/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6 +size 5304 diff --git a/checkpoint-169/README.md b/checkpoint-169/README.md new file mode 100644 index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6 --- /dev/null +++ b/checkpoint-169/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-2-13b-chat-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.2 \ No newline at end of file diff --git a/checkpoint-169/adapter_config.json b/checkpoint-169/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06 --- /dev/null +++ b/checkpoint-169/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 32, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-169/adapter_model.safetensors b/checkpoint-169/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c063bb99959871a6ae13048da703260cca57677e --- /dev/null +++ b/checkpoint-169/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5574402ac6a72a2f4e3bbd7ef02f0c293abf98ace5d096071f7c827abae45ad4 +size 26235704 diff --git a/checkpoint-169/optimizer.pt b/checkpoint-169/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5e6018326869e7f5b0b9199b1a6a9000ba8533d9 --- /dev/null +++ b/checkpoint-169/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40a3a2910527f0bf62f5c79df157fa3fb70d39b63becd955f5e702652f341f1c +size 52563258 diff --git a/checkpoint-169/rng_state.pth b/checkpoint-169/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..8c3d7eaea723ec61d2e0b66cbd197d16544f0d37 --- /dev/null +++ b/checkpoint-169/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:671cb42ff173690da931c0dc9e5393050001d4311b9fa17e7f4b9742fc264993 +size 14244 diff --git a/checkpoint-169/scheduler.pt b/checkpoint-169/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ef22707cae3a3afe61900df410ff4cfb7caab5cf --- /dev/null +++ b/checkpoint-169/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6b9a90250d286d551e591117be4e8a759da87ed937533dfac8a7df81db37c5a +size 1064 diff --git a/checkpoint-169/trainer_state.json b/checkpoint-169/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..65c611f6b800783b69f0f403057025a4683280f3 --- /dev/null +++ b/checkpoint-169/trainer_state.json @@ -0,0 +1,54 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9985228951255539, + "eval_steps": 500, + "global_step": 169, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.29542097488921715, + "grad_norm": 3.6940183639526367, + "learning_rate": 2.5e-05, + "loss": 3.3429, + "step": 50 + }, + { + "epoch": 0.5908419497784343, + "grad_norm": 0.5006749033927917, + "learning_rate": 5e-05, + "loss": 0.7796, + "step": 100 + }, + { + "epoch": 0.8862629246676514, + "grad_norm": 0.5715117454528809, + "learning_rate": 4.726477024070022e-05, + "loss": 0.22, + "step": 150 + } + ], + "logging_steps": 50, + "max_steps": 1014, + "num_input_tokens_seen": 0, + "num_train_epochs": 6, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.673267488980992e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-169/training_args.bin b/checkpoint-169/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740 --- /dev/null +++ b/checkpoint-169/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6 +size 5304 diff --git a/checkpoint-338/README.md b/checkpoint-338/README.md new file mode 100644 index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6 --- /dev/null +++ b/checkpoint-338/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-2-13b-chat-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.2 \ No newline at end of file diff --git a/checkpoint-338/adapter_config.json b/checkpoint-338/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06 --- /dev/null +++ b/checkpoint-338/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 32, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-338/adapter_model.safetensors b/checkpoint-338/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c6dc752cf1c41909d2b7d6094061d0220d5755ed --- /dev/null +++ b/checkpoint-338/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:504ae73fa6ebfcee252bbed5309bc74dc621577205791a72d884997780045939 +size 26235704 diff --git a/checkpoint-338/optimizer.pt b/checkpoint-338/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..54ac431ab2691a9795d02db1a3414d8b9c58c363 --- /dev/null +++ b/checkpoint-338/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3d89c3861855f6d8d57ed090741e738c2ceb2df913eb60909e0e90221ca6805 +size 52563258 diff --git a/checkpoint-338/rng_state.pth b/checkpoint-338/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..28a15ce2f3f6b96a48280686bd9a851a0d33ae3e --- /dev/null +++ b/checkpoint-338/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a407ae22d91865b0c7b2f37be1ef15d17cbdabfdef79af88be1868bd0b57a9e +size 14244 diff --git a/checkpoint-338/scheduler.pt b/checkpoint-338/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..af352f337b969ed7e51b47fd8a09e0b7bbd50dab --- /dev/null +++ b/checkpoint-338/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e843bba70c519c5360d713ee493af48cef233f40f1d4fe81285921c728ccf34 +size 1064 diff --git a/checkpoint-338/trainer_state.json b/checkpoint-338/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..5f66aae13949be310913ee16331fd0072b448f12 --- /dev/null +++ b/checkpoint-338/trainer_state.json @@ -0,0 +1,75 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.9970457902511078, + "eval_steps": 500, + "global_step": 338, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.29542097488921715, + "grad_norm": 3.6940183639526367, + "learning_rate": 2.5e-05, + "loss": 3.3429, + "step": 50 + }, + { + "epoch": 0.5908419497784343, + "grad_norm": 0.5006749033927917, + "learning_rate": 5e-05, + "loss": 0.7796, + "step": 100 + }, + { + "epoch": 0.8862629246676514, + "grad_norm": 0.5715117454528809, + "learning_rate": 4.726477024070022e-05, + "loss": 0.22, + "step": 150 + }, + { + "epoch": 1.1816838995568686, + "grad_norm": 0.7143971920013428, + "learning_rate": 4.452954048140044e-05, + "loss": 0.1198, + "step": 200 + }, + { + "epoch": 1.4771048744460857, + "grad_norm": 0.380874902009964, + "learning_rate": 4.179431072210066e-05, + "loss": 0.117, + "step": 250 + }, + { + "epoch": 1.7725258493353029, + "grad_norm": 0.3196457028388977, + "learning_rate": 3.9059080962800876e-05, + "loss": 0.1134, + "step": 300 + } + ], + "logging_steps": 50, + "max_steps": 1014, + "num_input_tokens_seen": 0, + "num_train_epochs": 6, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 5.346534977961984e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-338/training_args.bin b/checkpoint-338/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740 --- /dev/null +++ b/checkpoint-338/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6 +size 5304 diff --git a/checkpoint-507/README.md b/checkpoint-507/README.md new file mode 100644 index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6 --- /dev/null +++ b/checkpoint-507/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-2-13b-chat-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.2 \ No newline at end of file diff --git a/checkpoint-507/adapter_config.json b/checkpoint-507/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06 --- /dev/null +++ b/checkpoint-507/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 32, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-507/adapter_model.safetensors b/checkpoint-507/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bb4fe68c29bdc8dc0c0db23c7aa51a6eea8564dd --- /dev/null +++ b/checkpoint-507/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62fda8bb16881c86b820ab5e00085fd39d2dcfda33a1a82bfc82a880fab4a164 +size 26235704 diff --git a/checkpoint-507/optimizer.pt b/checkpoint-507/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..ed20df75abb3bc0a23e88e8f44ad2da3aea9593f --- /dev/null +++ b/checkpoint-507/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6061ff0adba31434cd7707eb44812e65a50e3432039823041b1d7e4154a7491 +size 52563258 diff --git a/checkpoint-507/rng_state.pth b/checkpoint-507/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..518eebd4ed0e965d0147b7f42843fa71967f44ac --- /dev/null +++ b/checkpoint-507/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e835266e3a87be80aca307cad0a349799b85090ac5a34a85e773d1c8dc2560a +size 14244 diff --git a/checkpoint-507/scheduler.pt b/checkpoint-507/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..429c00f61c940436c48d3c586a7762becdfc310c --- /dev/null +++ b/checkpoint-507/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3fd4253769ea0aeaa0685369d70bcfb875725f58354f5399b1fa66d41b3ed04 +size 1064 diff --git a/checkpoint-507/trainer_state.json b/checkpoint-507/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..8d18abbf0874b7a6b919a8a96ac90b9b6164f68e --- /dev/null +++ b/checkpoint-507/trainer_state.json @@ -0,0 +1,103 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.9955686853766617, + "eval_steps": 500, + "global_step": 507, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.29542097488921715, + "grad_norm": 3.6940183639526367, + "learning_rate": 2.5e-05, + "loss": 3.3429, + "step": 50 + }, + { + "epoch": 0.5908419497784343, + "grad_norm": 0.5006749033927917, + "learning_rate": 5e-05, + "loss": 0.7796, + "step": 100 + }, + { + "epoch": 0.8862629246676514, + "grad_norm": 0.5715117454528809, + "learning_rate": 4.726477024070022e-05, + "loss": 0.22, + "step": 150 + }, + { + "epoch": 1.1816838995568686, + "grad_norm": 0.7143971920013428, + "learning_rate": 4.452954048140044e-05, + "loss": 0.1198, + "step": 200 + }, + { + "epoch": 1.4771048744460857, + "grad_norm": 0.380874902009964, + "learning_rate": 4.179431072210066e-05, + "loss": 0.117, + "step": 250 + }, + { + "epoch": 1.7725258493353029, + "grad_norm": 0.3196457028388977, + "learning_rate": 3.9059080962800876e-05, + "loss": 0.1134, + "step": 300 + }, + { + "epoch": 2.06794682422452, + "grad_norm": 0.40986520051956177, + "learning_rate": 3.6323851203501094e-05, + "loss": 0.1151, + "step": 350 + }, + { + "epoch": 2.363367799113737, + "grad_norm": 0.3212401568889618, + "learning_rate": 3.358862144420131e-05, + "loss": 0.1138, + "step": 400 + }, + { + "epoch": 2.658788774002954, + "grad_norm": 0.16917897760868073, + "learning_rate": 3.085339168490153e-05, + "loss": 0.1115, + "step": 450 + }, + { + "epoch": 2.9542097488921715, + "grad_norm": 0.22288767993450165, + "learning_rate": 2.811816192560175e-05, + "loss": 0.1155, + "step": 500 + } + ], + "logging_steps": 50, + "max_steps": 1014, + "num_input_tokens_seen": 0, + "num_train_epochs": 6, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 8.019802466942976e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-507/training_args.bin b/checkpoint-507/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740 --- /dev/null +++ b/checkpoint-507/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6 +size 5304 diff --git a/checkpoint-677/README.md b/checkpoint-677/README.md new file mode 100644 index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6 --- /dev/null +++ b/checkpoint-677/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-2-13b-chat-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.2 \ No newline at end of file diff --git a/checkpoint-677/adapter_config.json b/checkpoint-677/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06 --- /dev/null +++ b/checkpoint-677/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 32, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-677/adapter_model.safetensors b/checkpoint-677/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5b9f6473a20dc538e3e84929fc2f8e1468316347 --- /dev/null +++ b/checkpoint-677/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd197ea16a26f11ea213f0dd8360bd60fb21a6b1fd92588155fdc2a31e8eb2dd +size 26235704 diff --git a/checkpoint-677/optimizer.pt b/checkpoint-677/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..0a2c289f952978b7cc5919bdfb499f0f0823cad6 --- /dev/null +++ b/checkpoint-677/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e19050ed6ae2c00fc93ca9a564ea60545b039790a275b7755933d704d2eaf473 +size 52563258 diff --git a/checkpoint-677/rng_state.pth b/checkpoint-677/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..346707572282eff6e90aeff25f77cd75d6fde795 --- /dev/null +++ b/checkpoint-677/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6356d528bf3866558bd9dd722baf34c195d10152de27bd868852994ed582b79 +size 14244 diff --git a/checkpoint-677/scheduler.pt b/checkpoint-677/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..708df19cb736accef75470dddd81e209a8dd2ee1 --- /dev/null +++ b/checkpoint-677/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ddaadb35f74d04ab3c980cafc198834d2c4f383d5a8137825aa635ea88bf0035 +size 1064 diff --git a/checkpoint-677/trainer_state.json b/checkpoint-677/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b70a0e37320b21ecd7ca59209dd06ec3dd3fedbc --- /dev/null +++ b/checkpoint-677/trainer_state.json @@ -0,0 +1,124 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.0, + "eval_steps": 500, + "global_step": 677, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.29542097488921715, + "grad_norm": 3.6940183639526367, + "learning_rate": 2.5e-05, + "loss": 3.3429, + "step": 50 + }, + { + "epoch": 0.5908419497784343, + "grad_norm": 0.5006749033927917, + "learning_rate": 5e-05, + "loss": 0.7796, + "step": 100 + }, + { + "epoch": 0.8862629246676514, + "grad_norm": 0.5715117454528809, + "learning_rate": 4.726477024070022e-05, + "loss": 0.22, + "step": 150 + }, + { + "epoch": 1.1816838995568686, + "grad_norm": 0.7143971920013428, + "learning_rate": 4.452954048140044e-05, + "loss": 0.1198, + "step": 200 + }, + { + "epoch": 1.4771048744460857, + "grad_norm": 0.380874902009964, + "learning_rate": 4.179431072210066e-05, + "loss": 0.117, + "step": 250 + }, + { + "epoch": 1.7725258493353029, + "grad_norm": 0.3196457028388977, + "learning_rate": 3.9059080962800876e-05, + "loss": 0.1134, + "step": 300 + }, + { + "epoch": 2.06794682422452, + "grad_norm": 0.40986520051956177, + "learning_rate": 3.6323851203501094e-05, + "loss": 0.1151, + "step": 350 + }, + { + "epoch": 2.363367799113737, + "grad_norm": 0.3212401568889618, + "learning_rate": 3.358862144420131e-05, + "loss": 0.1138, + "step": 400 + }, + { + "epoch": 2.658788774002954, + "grad_norm": 0.16917897760868073, + "learning_rate": 3.085339168490153e-05, + "loss": 0.1115, + "step": 450 + }, + { + "epoch": 2.9542097488921715, + "grad_norm": 0.22288767993450165, + "learning_rate": 2.811816192560175e-05, + "loss": 0.1155, + "step": 500 + }, + { + "epoch": 3.2496307237813884, + "grad_norm": 0.19413484632968903, + "learning_rate": 2.538293216630197e-05, + "loss": 0.1142, + "step": 550 + }, + { + "epoch": 3.5450516986706058, + "grad_norm": 0.22448602318763733, + "learning_rate": 2.264770240700219e-05, + "loss": 0.1134, + "step": 600 + }, + { + "epoch": 3.8404726735598227, + "grad_norm": 0.2704063057899475, + "learning_rate": 1.9912472647702408e-05, + "loss": 0.1121, + "step": 650 + } + ], + "logging_steps": 50, + "max_steps": 1014, + "num_input_tokens_seen": 0, + "num_train_epochs": 6, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.0693069955923968e+17, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-677/training_args.bin b/checkpoint-677/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740 --- /dev/null +++ b/checkpoint-677/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6 +size 5304 diff --git a/checkpoint-846/README.md b/checkpoint-846/README.md new file mode 100644 index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6 --- /dev/null +++ b/checkpoint-846/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-2-13b-chat-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.13.2 \ No newline at end of file diff --git a/checkpoint-846/adapter_config.json b/checkpoint-846/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06 --- /dev/null +++ b/checkpoint-846/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 32, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoint-846/adapter_model.safetensors b/checkpoint-846/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e71523f26e01fc791eb69aa65ad84d0c7f170a15 --- /dev/null +++ b/checkpoint-846/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d723e468de941145d4bf0e7609ef0a0133c64296dcdedf82156aa474a64791d2 +size 26235704 diff --git a/checkpoint-846/optimizer.pt b/checkpoint-846/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..95fb2893b146de6a8c2fbb09f9da437a90fec7de --- /dev/null +++ b/checkpoint-846/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33f4d216498d0923edd0b3f274c57056179085aeb5a1ef7b3a4b597174e0bdf5 +size 52563258 diff --git a/checkpoint-846/rng_state.pth b/checkpoint-846/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..421edff8f71c55b5598ad5b7f1a12958255a7b35 --- /dev/null +++ b/checkpoint-846/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20c2eb11d2b683c85ee0e7da692f4b5b279e4d46d161a2bd6b46f8b649eb6709 +size 14244 diff --git a/checkpoint-846/scheduler.pt b/checkpoint-846/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ea2ea53af1469458e15d348d539cda3f453ffea2 --- /dev/null +++ b/checkpoint-846/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab8f16729e3a13f7c0b70a540f3fba6c55f3abffc4be248e85d11207b1e0846a +size 1064 diff --git a/checkpoint-846/trainer_state.json b/checkpoint-846/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..a3714a8c8e9502c7d55cb70b0291526a1bf97d91 --- /dev/null +++ b/checkpoint-846/trainer_state.json @@ -0,0 +1,145 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.998522895125554, + "eval_steps": 500, + "global_step": 846, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.29542097488921715, + "grad_norm": 3.6940183639526367, + "learning_rate": 2.5e-05, + "loss": 3.3429, + "step": 50 + }, + { + "epoch": 0.5908419497784343, + "grad_norm": 0.5006749033927917, + "learning_rate": 5e-05, + "loss": 0.7796, + "step": 100 + }, + { + "epoch": 0.8862629246676514, + "grad_norm": 0.5715117454528809, + "learning_rate": 4.726477024070022e-05, + "loss": 0.22, + "step": 150 + }, + { + "epoch": 1.1816838995568686, + "grad_norm": 0.7143971920013428, + "learning_rate": 4.452954048140044e-05, + "loss": 0.1198, + "step": 200 + }, + { + "epoch": 1.4771048744460857, + "grad_norm": 0.380874902009964, + "learning_rate": 4.179431072210066e-05, + "loss": 0.117, + "step": 250 + }, + { + "epoch": 1.7725258493353029, + "grad_norm": 0.3196457028388977, + "learning_rate": 3.9059080962800876e-05, + "loss": 0.1134, + "step": 300 + }, + { + "epoch": 2.06794682422452, + "grad_norm": 0.40986520051956177, + "learning_rate": 3.6323851203501094e-05, + "loss": 0.1151, + "step": 350 + }, + { + "epoch": 2.363367799113737, + "grad_norm": 0.3212401568889618, + "learning_rate": 3.358862144420131e-05, + "loss": 0.1138, + "step": 400 + }, + { + "epoch": 2.658788774002954, + "grad_norm": 0.16917897760868073, + "learning_rate": 3.085339168490153e-05, + "loss": 0.1115, + "step": 450 + }, + { + "epoch": 2.9542097488921715, + "grad_norm": 0.22288767993450165, + "learning_rate": 2.811816192560175e-05, + "loss": 0.1155, + "step": 500 + }, + { + "epoch": 3.2496307237813884, + "grad_norm": 0.19413484632968903, + "learning_rate": 2.538293216630197e-05, + "loss": 0.1142, + "step": 550 + }, + { + "epoch": 3.5450516986706058, + "grad_norm": 0.22448602318763733, + "learning_rate": 2.264770240700219e-05, + "loss": 0.1134, + "step": 600 + }, + { + "epoch": 3.8404726735598227, + "grad_norm": 0.2704063057899475, + "learning_rate": 1.9912472647702408e-05, + "loss": 0.1121, + "step": 650 + }, + { + "epoch": 4.13589364844904, + "grad_norm": 0.3302886486053467, + "learning_rate": 1.7177242888402626e-05, + "loss": 0.1158, + "step": 700 + }, + { + "epoch": 4.431314623338257, + "grad_norm": 0.1943623274564743, + "learning_rate": 1.4442013129102846e-05, + "loss": 0.1143, + "step": 750 + }, + { + "epoch": 4.726735598227474, + "grad_norm": 0.21939796209335327, + "learning_rate": 1.1706783369803063e-05, + "loss": 0.1106, + "step": 800 + } + ], + "logging_steps": 50, + "max_steps": 1014, + "num_input_tokens_seen": 0, + "num_train_epochs": 6, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.336633744490496e+17, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-846/training_args.bin b/checkpoint-846/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740 --- /dev/null +++ b/checkpoint-846/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6 +size 5304 diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..72ecfeeb7e14d244c936169d2ed139eeae235ef1 --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,24 @@ +{ + "bos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "", + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/tokenizer.model b/tokenizer.model new file mode 100644 index 0000000000000000000000000000000000000000..6c00c742ce03c627d6cd5b795984876fa49fa899 --- /dev/null +++ b/tokenizer.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347 +size 499723 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..80c6aa4e7dd57e0e78805f1ddec9b62f2132fa5d --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,44 @@ +{ + "add_bos_token": true, + "add_eos_token": false, + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "", + "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 and system_message != false %}{% set content = '<>\\n' + system_message + '\\n<>\\n\\n' + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ bos_token + '[INST] ' + content.strip() + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + content.strip() + ' ' + eos_token }}{% endif %}{% endfor %}", + "clean_up_tokenization_spaces": false, + "eos_token": "", + "legacy": false, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "padding_side": "right", + "sp_model_kwargs": {}, + "spaces_between_special_tokens": false, + "tokenizer_class": "LlamaTokenizer", + "unk_token": "", + "use_default_system_prompt": false +}