diff --git a/checkpoint-1014/README.md b/checkpoint-1014/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6
--- /dev/null
+++ b/checkpoint-1014/README.md
@@ -0,0 +1,202 @@
+---
+base_model: meta-llama/Llama-2-13b-chat-hf
+library_name: peft
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.13.2
\ No newline at end of file
diff --git a/checkpoint-1014/adapter_config.json b/checkpoint-1014/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06
--- /dev/null
+++ b/checkpoint-1014/adapter_config.json
@@ -0,0 +1,29 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 32,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "q_proj",
+ "v_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-1014/adapter_model.safetensors b/checkpoint-1014/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f3b85154358f0ead8591b0893163cf7d7aafc789
--- /dev/null
+++ b/checkpoint-1014/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b7eb82e9832f7b4a2302c2a811a490862c5d2a86d1e8cea5da78dba68a3c8e4e
+size 26235704
diff --git a/checkpoint-1014/optimizer.pt b/checkpoint-1014/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..9de84081ded3001bcff96eef621de201fac14323
--- /dev/null
+++ b/checkpoint-1014/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:79df9f3009a37edf2c4cdd8919d98bebd2e64369b8024b62149f5039bead3dfe
+size 52563258
diff --git a/checkpoint-1014/rng_state.pth b/checkpoint-1014/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5d968263657a30eea713bfe7c3ac18de2db73bca
--- /dev/null
+++ b/checkpoint-1014/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6b4c4f0166ef9ca74027402d439827bc8ed2a02c96ba6ba9e51d0a9c0f291412
+size 14244
diff --git a/checkpoint-1014/scheduler.pt b/checkpoint-1014/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..8c15d820812361823eaae11e84a9769fc4b30d88
--- /dev/null
+++ b/checkpoint-1014/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:78fd9b36c5271bbf8da930a0ab7ea570971f4031660dcb029549ddb0db3d5123
+size 1064
diff --git a/checkpoint-1014/trainer_state.json b/checkpoint-1014/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..78d019e0c76a37176d298d06318c46fb02548b43
--- /dev/null
+++ b/checkpoint-1014/trainer_state.json
@@ -0,0 +1,173 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 5.991137370753323,
+ "eval_steps": 500,
+ "global_step": 1014,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.29542097488921715,
+ "grad_norm": 3.6940183639526367,
+ "learning_rate": 2.5e-05,
+ "loss": 3.3429,
+ "step": 50
+ },
+ {
+ "epoch": 0.5908419497784343,
+ "grad_norm": 0.5006749033927917,
+ "learning_rate": 5e-05,
+ "loss": 0.7796,
+ "step": 100
+ },
+ {
+ "epoch": 0.8862629246676514,
+ "grad_norm": 0.5715117454528809,
+ "learning_rate": 4.726477024070022e-05,
+ "loss": 0.22,
+ "step": 150
+ },
+ {
+ "epoch": 1.1816838995568686,
+ "grad_norm": 0.7143971920013428,
+ "learning_rate": 4.452954048140044e-05,
+ "loss": 0.1198,
+ "step": 200
+ },
+ {
+ "epoch": 1.4771048744460857,
+ "grad_norm": 0.380874902009964,
+ "learning_rate": 4.179431072210066e-05,
+ "loss": 0.117,
+ "step": 250
+ },
+ {
+ "epoch": 1.7725258493353029,
+ "grad_norm": 0.3196457028388977,
+ "learning_rate": 3.9059080962800876e-05,
+ "loss": 0.1134,
+ "step": 300
+ },
+ {
+ "epoch": 2.06794682422452,
+ "grad_norm": 0.40986520051956177,
+ "learning_rate": 3.6323851203501094e-05,
+ "loss": 0.1151,
+ "step": 350
+ },
+ {
+ "epoch": 2.363367799113737,
+ "grad_norm": 0.3212401568889618,
+ "learning_rate": 3.358862144420131e-05,
+ "loss": 0.1138,
+ "step": 400
+ },
+ {
+ "epoch": 2.658788774002954,
+ "grad_norm": 0.16917897760868073,
+ "learning_rate": 3.085339168490153e-05,
+ "loss": 0.1115,
+ "step": 450
+ },
+ {
+ "epoch": 2.9542097488921715,
+ "grad_norm": 0.22288767993450165,
+ "learning_rate": 2.811816192560175e-05,
+ "loss": 0.1155,
+ "step": 500
+ },
+ {
+ "epoch": 3.2496307237813884,
+ "grad_norm": 0.19413484632968903,
+ "learning_rate": 2.538293216630197e-05,
+ "loss": 0.1142,
+ "step": 550
+ },
+ {
+ "epoch": 3.5450516986706058,
+ "grad_norm": 0.22448602318763733,
+ "learning_rate": 2.264770240700219e-05,
+ "loss": 0.1134,
+ "step": 600
+ },
+ {
+ "epoch": 3.8404726735598227,
+ "grad_norm": 0.2704063057899475,
+ "learning_rate": 1.9912472647702408e-05,
+ "loss": 0.1121,
+ "step": 650
+ },
+ {
+ "epoch": 4.13589364844904,
+ "grad_norm": 0.3302886486053467,
+ "learning_rate": 1.7177242888402626e-05,
+ "loss": 0.1158,
+ "step": 700
+ },
+ {
+ "epoch": 4.431314623338257,
+ "grad_norm": 0.1943623274564743,
+ "learning_rate": 1.4442013129102846e-05,
+ "loss": 0.1143,
+ "step": 750
+ },
+ {
+ "epoch": 4.726735598227474,
+ "grad_norm": 0.21939796209335327,
+ "learning_rate": 1.1706783369803063e-05,
+ "loss": 0.1106,
+ "step": 800
+ },
+ {
+ "epoch": 5.022156573116692,
+ "grad_norm": 0.3173823654651642,
+ "learning_rate": 8.971553610503283e-06,
+ "loss": 0.1131,
+ "step": 850
+ },
+ {
+ "epoch": 5.317577548005908,
+ "grad_norm": 0.2860495150089264,
+ "learning_rate": 6.2363238512035015e-06,
+ "loss": 0.1127,
+ "step": 900
+ },
+ {
+ "epoch": 5.612998522895126,
+ "grad_norm": 0.18334375321865082,
+ "learning_rate": 3.50109409190372e-06,
+ "loss": 0.1102,
+ "step": 950
+ },
+ {
+ "epoch": 5.908419497784343,
+ "grad_norm": 0.18338781595230103,
+ "learning_rate": 7.658643326039388e-07,
+ "loss": 0.1116,
+ "step": 1000
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 1014,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 6,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": true
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.6016891541848064e+17,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-1014/training_args.bin b/checkpoint-1014/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740
--- /dev/null
+++ b/checkpoint-1014/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6
+size 5304
diff --git a/checkpoint-169/README.md b/checkpoint-169/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6
--- /dev/null
+++ b/checkpoint-169/README.md
@@ -0,0 +1,202 @@
+---
+base_model: meta-llama/Llama-2-13b-chat-hf
+library_name: peft
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.13.2
\ No newline at end of file
diff --git a/checkpoint-169/adapter_config.json b/checkpoint-169/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06
--- /dev/null
+++ b/checkpoint-169/adapter_config.json
@@ -0,0 +1,29 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 32,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "q_proj",
+ "v_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-169/adapter_model.safetensors b/checkpoint-169/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..c063bb99959871a6ae13048da703260cca57677e
--- /dev/null
+++ b/checkpoint-169/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5574402ac6a72a2f4e3bbd7ef02f0c293abf98ace5d096071f7c827abae45ad4
+size 26235704
diff --git a/checkpoint-169/optimizer.pt b/checkpoint-169/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5e6018326869e7f5b0b9199b1a6a9000ba8533d9
--- /dev/null
+++ b/checkpoint-169/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40a3a2910527f0bf62f5c79df157fa3fb70d39b63becd955f5e702652f341f1c
+size 52563258
diff --git a/checkpoint-169/rng_state.pth b/checkpoint-169/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8c3d7eaea723ec61d2e0b66cbd197d16544f0d37
--- /dev/null
+++ b/checkpoint-169/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:671cb42ff173690da931c0dc9e5393050001d4311b9fa17e7f4b9742fc264993
+size 14244
diff --git a/checkpoint-169/scheduler.pt b/checkpoint-169/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..ef22707cae3a3afe61900df410ff4cfb7caab5cf
--- /dev/null
+++ b/checkpoint-169/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a6b9a90250d286d551e591117be4e8a759da87ed937533dfac8a7df81db37c5a
+size 1064
diff --git a/checkpoint-169/trainer_state.json b/checkpoint-169/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..65c611f6b800783b69f0f403057025a4683280f3
--- /dev/null
+++ b/checkpoint-169/trainer_state.json
@@ -0,0 +1,54 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 0.9985228951255539,
+ "eval_steps": 500,
+ "global_step": 169,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.29542097488921715,
+ "grad_norm": 3.6940183639526367,
+ "learning_rate": 2.5e-05,
+ "loss": 3.3429,
+ "step": 50
+ },
+ {
+ "epoch": 0.5908419497784343,
+ "grad_norm": 0.5006749033927917,
+ "learning_rate": 5e-05,
+ "loss": 0.7796,
+ "step": 100
+ },
+ {
+ "epoch": 0.8862629246676514,
+ "grad_norm": 0.5715117454528809,
+ "learning_rate": 4.726477024070022e-05,
+ "loss": 0.22,
+ "step": 150
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 1014,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 6,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 2.673267488980992e+16,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-169/training_args.bin b/checkpoint-169/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740
--- /dev/null
+++ b/checkpoint-169/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6
+size 5304
diff --git a/checkpoint-338/README.md b/checkpoint-338/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6
--- /dev/null
+++ b/checkpoint-338/README.md
@@ -0,0 +1,202 @@
+---
+base_model: meta-llama/Llama-2-13b-chat-hf
+library_name: peft
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.13.2
\ No newline at end of file
diff --git a/checkpoint-338/adapter_config.json b/checkpoint-338/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06
--- /dev/null
+++ b/checkpoint-338/adapter_config.json
@@ -0,0 +1,29 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 32,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "q_proj",
+ "v_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-338/adapter_model.safetensors b/checkpoint-338/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..c6dc752cf1c41909d2b7d6094061d0220d5755ed
--- /dev/null
+++ b/checkpoint-338/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:504ae73fa6ebfcee252bbed5309bc74dc621577205791a72d884997780045939
+size 26235704
diff --git a/checkpoint-338/optimizer.pt b/checkpoint-338/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..54ac431ab2691a9795d02db1a3414d8b9c58c363
--- /dev/null
+++ b/checkpoint-338/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e3d89c3861855f6d8d57ed090741e738c2ceb2df913eb60909e0e90221ca6805
+size 52563258
diff --git a/checkpoint-338/rng_state.pth b/checkpoint-338/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..28a15ce2f3f6b96a48280686bd9a851a0d33ae3e
--- /dev/null
+++ b/checkpoint-338/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4a407ae22d91865b0c7b2f37be1ef15d17cbdabfdef79af88be1868bd0b57a9e
+size 14244
diff --git a/checkpoint-338/scheduler.pt b/checkpoint-338/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..af352f337b969ed7e51b47fd8a09e0b7bbd50dab
--- /dev/null
+++ b/checkpoint-338/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0e843bba70c519c5360d713ee493af48cef233f40f1d4fe81285921c728ccf34
+size 1064
diff --git a/checkpoint-338/trainer_state.json b/checkpoint-338/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..5f66aae13949be310913ee16331fd0072b448f12
--- /dev/null
+++ b/checkpoint-338/trainer_state.json
@@ -0,0 +1,75 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 1.9970457902511078,
+ "eval_steps": 500,
+ "global_step": 338,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.29542097488921715,
+ "grad_norm": 3.6940183639526367,
+ "learning_rate": 2.5e-05,
+ "loss": 3.3429,
+ "step": 50
+ },
+ {
+ "epoch": 0.5908419497784343,
+ "grad_norm": 0.5006749033927917,
+ "learning_rate": 5e-05,
+ "loss": 0.7796,
+ "step": 100
+ },
+ {
+ "epoch": 0.8862629246676514,
+ "grad_norm": 0.5715117454528809,
+ "learning_rate": 4.726477024070022e-05,
+ "loss": 0.22,
+ "step": 150
+ },
+ {
+ "epoch": 1.1816838995568686,
+ "grad_norm": 0.7143971920013428,
+ "learning_rate": 4.452954048140044e-05,
+ "loss": 0.1198,
+ "step": 200
+ },
+ {
+ "epoch": 1.4771048744460857,
+ "grad_norm": 0.380874902009964,
+ "learning_rate": 4.179431072210066e-05,
+ "loss": 0.117,
+ "step": 250
+ },
+ {
+ "epoch": 1.7725258493353029,
+ "grad_norm": 0.3196457028388977,
+ "learning_rate": 3.9059080962800876e-05,
+ "loss": 0.1134,
+ "step": 300
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 1014,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 6,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 5.346534977961984e+16,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-338/training_args.bin b/checkpoint-338/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740
--- /dev/null
+++ b/checkpoint-338/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6
+size 5304
diff --git a/checkpoint-507/README.md b/checkpoint-507/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6
--- /dev/null
+++ b/checkpoint-507/README.md
@@ -0,0 +1,202 @@
+---
+base_model: meta-llama/Llama-2-13b-chat-hf
+library_name: peft
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.13.2
\ No newline at end of file
diff --git a/checkpoint-507/adapter_config.json b/checkpoint-507/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06
--- /dev/null
+++ b/checkpoint-507/adapter_config.json
@@ -0,0 +1,29 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 32,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "q_proj",
+ "v_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-507/adapter_model.safetensors b/checkpoint-507/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..bb4fe68c29bdc8dc0c0db23c7aa51a6eea8564dd
--- /dev/null
+++ b/checkpoint-507/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:62fda8bb16881c86b820ab5e00085fd39d2dcfda33a1a82bfc82a880fab4a164
+size 26235704
diff --git a/checkpoint-507/optimizer.pt b/checkpoint-507/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..ed20df75abb3bc0a23e88e8f44ad2da3aea9593f
--- /dev/null
+++ b/checkpoint-507/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a6061ff0adba31434cd7707eb44812e65a50e3432039823041b1d7e4154a7491
+size 52563258
diff --git a/checkpoint-507/rng_state.pth b/checkpoint-507/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..518eebd4ed0e965d0147b7f42843fa71967f44ac
--- /dev/null
+++ b/checkpoint-507/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5e835266e3a87be80aca307cad0a349799b85090ac5a34a85e773d1c8dc2560a
+size 14244
diff --git a/checkpoint-507/scheduler.pt b/checkpoint-507/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..429c00f61c940436c48d3c586a7762becdfc310c
--- /dev/null
+++ b/checkpoint-507/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e3fd4253769ea0aeaa0685369d70bcfb875725f58354f5399b1fa66d41b3ed04
+size 1064
diff --git a/checkpoint-507/trainer_state.json b/checkpoint-507/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..8d18abbf0874b7a6b919a8a96ac90b9b6164f68e
--- /dev/null
+++ b/checkpoint-507/trainer_state.json
@@ -0,0 +1,103 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 2.9955686853766617,
+ "eval_steps": 500,
+ "global_step": 507,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.29542097488921715,
+ "grad_norm": 3.6940183639526367,
+ "learning_rate": 2.5e-05,
+ "loss": 3.3429,
+ "step": 50
+ },
+ {
+ "epoch": 0.5908419497784343,
+ "grad_norm": 0.5006749033927917,
+ "learning_rate": 5e-05,
+ "loss": 0.7796,
+ "step": 100
+ },
+ {
+ "epoch": 0.8862629246676514,
+ "grad_norm": 0.5715117454528809,
+ "learning_rate": 4.726477024070022e-05,
+ "loss": 0.22,
+ "step": 150
+ },
+ {
+ "epoch": 1.1816838995568686,
+ "grad_norm": 0.7143971920013428,
+ "learning_rate": 4.452954048140044e-05,
+ "loss": 0.1198,
+ "step": 200
+ },
+ {
+ "epoch": 1.4771048744460857,
+ "grad_norm": 0.380874902009964,
+ "learning_rate": 4.179431072210066e-05,
+ "loss": 0.117,
+ "step": 250
+ },
+ {
+ "epoch": 1.7725258493353029,
+ "grad_norm": 0.3196457028388977,
+ "learning_rate": 3.9059080962800876e-05,
+ "loss": 0.1134,
+ "step": 300
+ },
+ {
+ "epoch": 2.06794682422452,
+ "grad_norm": 0.40986520051956177,
+ "learning_rate": 3.6323851203501094e-05,
+ "loss": 0.1151,
+ "step": 350
+ },
+ {
+ "epoch": 2.363367799113737,
+ "grad_norm": 0.3212401568889618,
+ "learning_rate": 3.358862144420131e-05,
+ "loss": 0.1138,
+ "step": 400
+ },
+ {
+ "epoch": 2.658788774002954,
+ "grad_norm": 0.16917897760868073,
+ "learning_rate": 3.085339168490153e-05,
+ "loss": 0.1115,
+ "step": 450
+ },
+ {
+ "epoch": 2.9542097488921715,
+ "grad_norm": 0.22288767993450165,
+ "learning_rate": 2.811816192560175e-05,
+ "loss": 0.1155,
+ "step": 500
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 1014,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 6,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 8.019802466942976e+16,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-507/training_args.bin b/checkpoint-507/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740
--- /dev/null
+++ b/checkpoint-507/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6
+size 5304
diff --git a/checkpoint-677/README.md b/checkpoint-677/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6
--- /dev/null
+++ b/checkpoint-677/README.md
@@ -0,0 +1,202 @@
+---
+base_model: meta-llama/Llama-2-13b-chat-hf
+library_name: peft
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.13.2
\ No newline at end of file
diff --git a/checkpoint-677/adapter_config.json b/checkpoint-677/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06
--- /dev/null
+++ b/checkpoint-677/adapter_config.json
@@ -0,0 +1,29 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 32,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "q_proj",
+ "v_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-677/adapter_model.safetensors b/checkpoint-677/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..5b9f6473a20dc538e3e84929fc2f8e1468316347
--- /dev/null
+++ b/checkpoint-677/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:dd197ea16a26f11ea213f0dd8360bd60fb21a6b1fd92588155fdc2a31e8eb2dd
+size 26235704
diff --git a/checkpoint-677/optimizer.pt b/checkpoint-677/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..0a2c289f952978b7cc5919bdfb499f0f0823cad6
--- /dev/null
+++ b/checkpoint-677/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e19050ed6ae2c00fc93ca9a564ea60545b039790a275b7755933d704d2eaf473
+size 52563258
diff --git a/checkpoint-677/rng_state.pth b/checkpoint-677/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..346707572282eff6e90aeff25f77cd75d6fde795
--- /dev/null
+++ b/checkpoint-677/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e6356d528bf3866558bd9dd722baf34c195d10152de27bd868852994ed582b79
+size 14244
diff --git a/checkpoint-677/scheduler.pt b/checkpoint-677/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..708df19cb736accef75470dddd81e209a8dd2ee1
--- /dev/null
+++ b/checkpoint-677/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ddaadb35f74d04ab3c980cafc198834d2c4f383d5a8137825aa635ea88bf0035
+size 1064
diff --git a/checkpoint-677/trainer_state.json b/checkpoint-677/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..b70a0e37320b21ecd7ca59209dd06ec3dd3fedbc
--- /dev/null
+++ b/checkpoint-677/trainer_state.json
@@ -0,0 +1,124 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 4.0,
+ "eval_steps": 500,
+ "global_step": 677,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.29542097488921715,
+ "grad_norm": 3.6940183639526367,
+ "learning_rate": 2.5e-05,
+ "loss": 3.3429,
+ "step": 50
+ },
+ {
+ "epoch": 0.5908419497784343,
+ "grad_norm": 0.5006749033927917,
+ "learning_rate": 5e-05,
+ "loss": 0.7796,
+ "step": 100
+ },
+ {
+ "epoch": 0.8862629246676514,
+ "grad_norm": 0.5715117454528809,
+ "learning_rate": 4.726477024070022e-05,
+ "loss": 0.22,
+ "step": 150
+ },
+ {
+ "epoch": 1.1816838995568686,
+ "grad_norm": 0.7143971920013428,
+ "learning_rate": 4.452954048140044e-05,
+ "loss": 0.1198,
+ "step": 200
+ },
+ {
+ "epoch": 1.4771048744460857,
+ "grad_norm": 0.380874902009964,
+ "learning_rate": 4.179431072210066e-05,
+ "loss": 0.117,
+ "step": 250
+ },
+ {
+ "epoch": 1.7725258493353029,
+ "grad_norm": 0.3196457028388977,
+ "learning_rate": 3.9059080962800876e-05,
+ "loss": 0.1134,
+ "step": 300
+ },
+ {
+ "epoch": 2.06794682422452,
+ "grad_norm": 0.40986520051956177,
+ "learning_rate": 3.6323851203501094e-05,
+ "loss": 0.1151,
+ "step": 350
+ },
+ {
+ "epoch": 2.363367799113737,
+ "grad_norm": 0.3212401568889618,
+ "learning_rate": 3.358862144420131e-05,
+ "loss": 0.1138,
+ "step": 400
+ },
+ {
+ "epoch": 2.658788774002954,
+ "grad_norm": 0.16917897760868073,
+ "learning_rate": 3.085339168490153e-05,
+ "loss": 0.1115,
+ "step": 450
+ },
+ {
+ "epoch": 2.9542097488921715,
+ "grad_norm": 0.22288767993450165,
+ "learning_rate": 2.811816192560175e-05,
+ "loss": 0.1155,
+ "step": 500
+ },
+ {
+ "epoch": 3.2496307237813884,
+ "grad_norm": 0.19413484632968903,
+ "learning_rate": 2.538293216630197e-05,
+ "loss": 0.1142,
+ "step": 550
+ },
+ {
+ "epoch": 3.5450516986706058,
+ "grad_norm": 0.22448602318763733,
+ "learning_rate": 2.264770240700219e-05,
+ "loss": 0.1134,
+ "step": 600
+ },
+ {
+ "epoch": 3.8404726735598227,
+ "grad_norm": 0.2704063057899475,
+ "learning_rate": 1.9912472647702408e-05,
+ "loss": 0.1121,
+ "step": 650
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 1014,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 6,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.0693069955923968e+17,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-677/training_args.bin b/checkpoint-677/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740
--- /dev/null
+++ b/checkpoint-677/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6
+size 5304
diff --git a/checkpoint-846/README.md b/checkpoint-846/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..e47b03ac2b3ec63bb9b693d5ea09a59bed58eec6
--- /dev/null
+++ b/checkpoint-846/README.md
@@ -0,0 +1,202 @@
+---
+base_model: meta-llama/Llama-2-13b-chat-hf
+library_name: peft
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.13.2
\ No newline at end of file
diff --git a/checkpoint-846/adapter_config.json b/checkpoint-846/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..adaaeb24374ba5c7059503a0ebd5378b39206f06
--- /dev/null
+++ b/checkpoint-846/adapter_config.json
@@ -0,0 +1,29 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 32,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "q_proj",
+ "v_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-846/adapter_model.safetensors b/checkpoint-846/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..e71523f26e01fc791eb69aa65ad84d0c7f170a15
--- /dev/null
+++ b/checkpoint-846/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d723e468de941145d4bf0e7609ef0a0133c64296dcdedf82156aa474a64791d2
+size 26235704
diff --git a/checkpoint-846/optimizer.pt b/checkpoint-846/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..95fb2893b146de6a8c2fbb09f9da437a90fec7de
--- /dev/null
+++ b/checkpoint-846/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:33f4d216498d0923edd0b3f274c57056179085aeb5a1ef7b3a4b597174e0bdf5
+size 52563258
diff --git a/checkpoint-846/rng_state.pth b/checkpoint-846/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..421edff8f71c55b5598ad5b7f1a12958255a7b35
--- /dev/null
+++ b/checkpoint-846/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:20c2eb11d2b683c85ee0e7da692f4b5b279e4d46d161a2bd6b46f8b649eb6709
+size 14244
diff --git a/checkpoint-846/scheduler.pt b/checkpoint-846/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..ea2ea53af1469458e15d348d539cda3f453ffea2
--- /dev/null
+++ b/checkpoint-846/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ab8f16729e3a13f7c0b70a540f3fba6c55f3abffc4be248e85d11207b1e0846a
+size 1064
diff --git a/checkpoint-846/trainer_state.json b/checkpoint-846/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..a3714a8c8e9502c7d55cb70b0291526a1bf97d91
--- /dev/null
+++ b/checkpoint-846/trainer_state.json
@@ -0,0 +1,145 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 4.998522895125554,
+ "eval_steps": 500,
+ "global_step": 846,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.29542097488921715,
+ "grad_norm": 3.6940183639526367,
+ "learning_rate": 2.5e-05,
+ "loss": 3.3429,
+ "step": 50
+ },
+ {
+ "epoch": 0.5908419497784343,
+ "grad_norm": 0.5006749033927917,
+ "learning_rate": 5e-05,
+ "loss": 0.7796,
+ "step": 100
+ },
+ {
+ "epoch": 0.8862629246676514,
+ "grad_norm": 0.5715117454528809,
+ "learning_rate": 4.726477024070022e-05,
+ "loss": 0.22,
+ "step": 150
+ },
+ {
+ "epoch": 1.1816838995568686,
+ "grad_norm": 0.7143971920013428,
+ "learning_rate": 4.452954048140044e-05,
+ "loss": 0.1198,
+ "step": 200
+ },
+ {
+ "epoch": 1.4771048744460857,
+ "grad_norm": 0.380874902009964,
+ "learning_rate": 4.179431072210066e-05,
+ "loss": 0.117,
+ "step": 250
+ },
+ {
+ "epoch": 1.7725258493353029,
+ "grad_norm": 0.3196457028388977,
+ "learning_rate": 3.9059080962800876e-05,
+ "loss": 0.1134,
+ "step": 300
+ },
+ {
+ "epoch": 2.06794682422452,
+ "grad_norm": 0.40986520051956177,
+ "learning_rate": 3.6323851203501094e-05,
+ "loss": 0.1151,
+ "step": 350
+ },
+ {
+ "epoch": 2.363367799113737,
+ "grad_norm": 0.3212401568889618,
+ "learning_rate": 3.358862144420131e-05,
+ "loss": 0.1138,
+ "step": 400
+ },
+ {
+ "epoch": 2.658788774002954,
+ "grad_norm": 0.16917897760868073,
+ "learning_rate": 3.085339168490153e-05,
+ "loss": 0.1115,
+ "step": 450
+ },
+ {
+ "epoch": 2.9542097488921715,
+ "grad_norm": 0.22288767993450165,
+ "learning_rate": 2.811816192560175e-05,
+ "loss": 0.1155,
+ "step": 500
+ },
+ {
+ "epoch": 3.2496307237813884,
+ "grad_norm": 0.19413484632968903,
+ "learning_rate": 2.538293216630197e-05,
+ "loss": 0.1142,
+ "step": 550
+ },
+ {
+ "epoch": 3.5450516986706058,
+ "grad_norm": 0.22448602318763733,
+ "learning_rate": 2.264770240700219e-05,
+ "loss": 0.1134,
+ "step": 600
+ },
+ {
+ "epoch": 3.8404726735598227,
+ "grad_norm": 0.2704063057899475,
+ "learning_rate": 1.9912472647702408e-05,
+ "loss": 0.1121,
+ "step": 650
+ },
+ {
+ "epoch": 4.13589364844904,
+ "grad_norm": 0.3302886486053467,
+ "learning_rate": 1.7177242888402626e-05,
+ "loss": 0.1158,
+ "step": 700
+ },
+ {
+ "epoch": 4.431314623338257,
+ "grad_norm": 0.1943623274564743,
+ "learning_rate": 1.4442013129102846e-05,
+ "loss": 0.1143,
+ "step": 750
+ },
+ {
+ "epoch": 4.726735598227474,
+ "grad_norm": 0.21939796209335327,
+ "learning_rate": 1.1706783369803063e-05,
+ "loss": 0.1106,
+ "step": 800
+ }
+ ],
+ "logging_steps": 50,
+ "max_steps": 1014,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 6,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": false,
+ "should_log": false,
+ "should_save": true,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.336633744490496e+17,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-846/training_args.bin b/checkpoint-846/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..cf9115e7b2b876874eb859d3c04d7eb96acdd740
--- /dev/null
+++ b/checkpoint-846/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:57d4014abc929a67eddc9035ef050123702cd59f738d5182b7fbaa477bc7a1a6
+size 5304
diff --git a/special_tokens_map.json b/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..72ecfeeb7e14d244c936169d2ed139eeae235ef1
--- /dev/null
+++ b/special_tokens_map.json
@@ -0,0 +1,24 @@
+{
+ "bos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "eos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "unk_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ }
+}
diff --git a/tokenizer.model b/tokenizer.model
new file mode 100644
index 0000000000000000000000000000000000000000..6c00c742ce03c627d6cd5b795984876fa49fa899
--- /dev/null
+++ b/tokenizer.model
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
+size 499723
diff --git a/tokenizer_config.json b/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..80c6aa4e7dd57e0e78805f1ddec9b62f2132fa5d
--- /dev/null
+++ b/tokenizer_config.json
@@ -0,0 +1,44 @@
+{
+ "add_bos_token": true,
+ "add_eos_token": false,
+ "add_prefix_space": true,
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 and system_message != false %}{% set content = '<>\\n' + system_message + '\\n<>\\n\\n' + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ bos_token + '[INST] ' + content.strip() + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + content.strip() + ' ' + eos_token }}{% endif %}{% endfor %}",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "legacy": false,
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "padding_side": "right",
+ "sp_model_kwargs": {},
+ "spaces_between_special_tokens": false,
+ "tokenizer_class": "LlamaTokenizer",
+ "unk_token": "",
+ "use_default_system_prompt": false
+}