sr5434 commited on
Commit
15e64a7
·
1 Parent(s): c59c371

CodegebraGPT-10B

Browse files
Files changed (22) hide show
  1. README.md +50 -0
  2. adapter_config.json +27 -0
  3. adapter_model.safetensors +3 -0
  4. runs/Dec29_15-53-54_104-171-202-28/events.out.tfevents.1703865235.104-171-202-28.2005.0 +3 -0
  5. runs/Dec29_15-54-07_104-171-202-28/events.out.tfevents.1703865248.104-171-202-28.2005.1 +3 -0
  6. runs/Dec29_16-03-29_104-171-202-28/events.out.tfevents.1703865810.104-171-202-28.3574.0 +3 -0
  7. runs/Dec29_16-15-15_104-171-202-28/events.out.tfevents.1703866516.104-171-202-28.6272.0 +3 -0
  8. runs/Dec29_16-18-16_104-171-202-28/events.out.tfevents.1703866697.104-171-202-28.6272.1 +3 -0
  9. runs/Dec29_16-18-35_104-171-202-28/events.out.tfevents.1703866716.104-171-202-28.6272.2 +3 -0
  10. runs/Dec29_16-19-38_104-171-202-28/events.out.tfevents.1703866779.104-171-202-28.6272.3 +3 -0
  11. runs/Dec29_16-23-37_104-171-202-28/events.out.tfevents.1703867018.104-171-202-28.1173.0 +3 -0
  12. runs/Dec29_16-24-18_104-171-202-28/events.out.tfevents.1703867058.104-171-202-28.1173.1 +3 -0
  13. runs/Dec29_16-28-10_104-171-202-28/events.out.tfevents.1703867290.104-171-202-28.1194.0 +3 -0
  14. runs/Dec29_16-32-24_104-171-202-28/events.out.tfevents.1703867545.104-171-202-28.1160.0 +3 -0
  15. runs/Dec29_17-12-50_104-171-202-28/events.out.tfevents.1703869970.104-171-202-28.1181.0 +3 -0
  16. runs/Dec29_17-28-16_104-171-202-28/events.out.tfevents.1703870897.104-171-202-28.1181.1 +3 -0
  17. runs/Dec29_17-35-30_104-171-202-28/events.out.tfevents.1703871331.104-171-202-28.1185.0 +3 -0
  18. runs/Dec29_17-47-10_104-171-202-28/events.out.tfevents.1703872031.104-171-202-28.1179.0 +3 -0
  19. runs/Dec29_18-47-13_104-171-202-28/events.out.tfevents.1703875634.104-171-202-28.1975.0 +3 -0
  20. runs/Dec29_18-52-06_104-171-202-28/events.out.tfevents.1703875927.104-171-202-28.1975.1 +3 -0
  21. runs/Dec29_18-52-52_104-171-202-28/events.out.tfevents.1703875973.104-171-202-28.1975.2 +3 -0
  22. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - generated_from_trainer
6
+ base_model: upstage/SOLAR-10.7B-v1.0
7
+ model-index:
8
+ - name: outputs
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # outputs
16
+
17
+ This model is a fine-tuned version of [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0) on an unknown dataset.
18
+
19
+ ## Model description
20
+
21
+ More information needed
22
+
23
+ ## Intended uses & limitations
24
+
25
+ More information needed
26
+
27
+ ## Training and evaluation data
28
+
29
+ More information needed
30
+
31
+ ## Training procedure
32
+
33
+ ### Training hyperparameters
34
+
35
+ The following hyperparameters were used during training:
36
+ - learning_rate: 0.0002
37
+ - train_batch_size: 1
38
+ - eval_batch_size: 8
39
+ - seed: 42
40
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
+ - lr_scheduler_type: linear
42
+ - num_epochs: 1
43
+
44
+ ### Framework versions
45
+
46
+ - PEFT 0.7.2.dev0
47
+ - Transformers 4.36.2
48
+ - Pytorch 2.0.1
49
+ - Datasets 2.16.0
50
+ - Tokenizers 0.15.0
adapter_config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "upstage/SOLAR-10.7B-v1.0",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layers_pattern": null,
10
+ "layers_to_transform": null,
11
+ "loftq_config": {},
12
+ "lora_alpha": 32,
13
+ "lora_dropout": 0.05,
14
+ "megatron_config": null,
15
+ "megatron_core": "megatron.core",
16
+ "modules_to_save": null,
17
+ "peft_type": "LORA",
18
+ "r": 8,
19
+ "rank_pattern": {},
20
+ "revision": null,
21
+ "target_modules": [
22
+ "q_proj",
23
+ "k_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM",
26
+ "use_rslora": false
27
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:463cbca58ceb375823affd6e5ac823abd254bfb671782041fe583229ee8f156d
3
+ size 20472752
runs/Dec29_15-53-54_104-171-202-28/events.out.tfevents.1703865235.104-171-202-28.2005.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5d13d37a8b0600cdfd5b244cd792a09904fa4a41968d8293e44f75599f2edbe
3
+ size 4758
runs/Dec29_15-54-07_104-171-202-28/events.out.tfevents.1703865248.104-171-202-28.2005.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be1693adb285da7b4eb7b39720784f99b25eec3b089535e703af8afe3fe56e80
3
+ size 4758
runs/Dec29_16-03-29_104-171-202-28/events.out.tfevents.1703865810.104-171-202-28.3574.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b91e5f8a85ca4ddf6e79b27dbf20da763d94e5151cad8e7d2cb510425f6565d0
3
+ size 4758
runs/Dec29_16-15-15_104-171-202-28/events.out.tfevents.1703866516.104-171-202-28.6272.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b42a5138ecdd78812a14a683010bc318f6dfd62df174a7cb082e5b5614b30143
3
+ size 4753
runs/Dec29_16-18-16_104-171-202-28/events.out.tfevents.1703866697.104-171-202-28.6272.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:729eed78d21618edcf9b652a5da5bd7b5f7232cca7cfd07a3f18bc017988d99d
3
+ size 4753
runs/Dec29_16-18-35_104-171-202-28/events.out.tfevents.1703866716.104-171-202-28.6272.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13680fb4ae04e456633e6b9001654b6067871fcfa8b61c976ec50d71e1173df1
3
+ size 4184
runs/Dec29_16-19-38_104-171-202-28/events.out.tfevents.1703866779.104-171-202-28.6272.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b72e2b04d4e9ad51d0909571dae78fa06bd3bcdb156414f057fd3acef87dbbf
3
+ size 4184
runs/Dec29_16-23-37_104-171-202-28/events.out.tfevents.1703867018.104-171-202-28.1173.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d9cb80ed29592d92066eab0a92b31f1acd291b6302439f49ab8fd5efb36a34e
3
+ size 4184
runs/Dec29_16-24-18_104-171-202-28/events.out.tfevents.1703867058.104-171-202-28.1173.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e81e195a77cd402817ea70fb3c8159cff032766c0c0cb7f09725385bc180dc58
3
+ size 4184
runs/Dec29_16-28-10_104-171-202-28/events.out.tfevents.1703867290.104-171-202-28.1194.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8afcb0c65120b7a080abb9a6919e9bada421e7a44027c83f07cd6a91095b7408
3
+ size 4184
runs/Dec29_16-32-24_104-171-202-28/events.out.tfevents.1703867545.104-171-202-28.1160.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:075979805879316cc57370f5a3f6367005744529caed952a4b721bfc2317a0b4
3
+ size 9065
runs/Dec29_17-12-50_104-171-202-28/events.out.tfevents.1703869970.104-171-202-28.1181.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19d238fc03a177d9bb5d93bb9311d2b08cc371742f9a174b1ad8822baf938acf
3
+ size 20922
runs/Dec29_17-28-16_104-171-202-28/events.out.tfevents.1703870897.104-171-202-28.1181.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c918c4bb5e93ea9de999b20625d3073004d3c450d3c579874899050aa04608d
3
+ size 41737
runs/Dec29_17-35-30_104-171-202-28/events.out.tfevents.1703871331.104-171-202-28.1185.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4175564fae19d924992be61ef45ff1e853c27373f6a86b16d4c0ae8555f9c3f0
3
+ size 82714
runs/Dec29_17-47-10_104-171-202-28/events.out.tfevents.1703872031.104-171-202-28.1179.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a8ea6ac1a886f9f5f9771c53031da964760df32f09cf5856ef0a2dac69e6bf3
3
+ size 498763
runs/Dec29_18-47-13_104-171-202-28/events.out.tfevents.1703875634.104-171-202-28.1975.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d713c9b019b30acfb15c7d443972a7d2e34219246a719b7c87873f6e4f5f67d
3
+ size 42521
runs/Dec29_18-52-06_104-171-202-28/events.out.tfevents.1703875927.104-171-202-28.1975.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aceae23be8964341a072f6ed0f9ed1143c73d17183b8208d9764e803154a64e4
3
+ size 9528
runs/Dec29_18-52-52_104-171-202-28/events.out.tfevents.1703875973.104-171-202-28.1975.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fdd4756cf4fc0e948429866c918d719cfb850c8d1cf1be90a20c3be499e08ce
3
+ size 5884662
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a37a6b7d599068cd4279d5d8cfc92f9bbbf89904800dd769ace41f7c13bdf4ef
3
+ size 4219