Sambaro commited on
Commit
6f0b364
·
verified ·
1 Parent(s): b299912

Upload folder using huggingface_hub

Browse files
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "o_proj",
24
- "gate_up_proj",
25
  "down_proj",
26
- "qkv_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "qkv_proj",
24
  "o_proj",
 
25
  "down_proj",
26
+ "gate_up_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf81191affb1d778af6a5658d18b5e7d6651fb71472854a45207af56c7a18862
3
  size 50366024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d10371d0a0226779ce3ec9c79d2f360f6df0ecebda48e645d61334e7041e834
3
  size 50366024
tokenizer.json CHANGED
@@ -150,12 +150,6 @@
150
  "post_processor": {
151
  "type": "TemplateProcessing",
152
  "single": [
153
- {
154
- "SpecialToken": {
155
- "id": "<s>",
156
- "type_id": 0
157
- }
158
- },
159
  {
160
  "Sequence": {
161
  "id": "A",
@@ -164,24 +158,12 @@
164
  }
165
  ],
166
  "pair": [
167
- {
168
- "SpecialToken": {
169
- "id": "<s>",
170
- "type_id": 0
171
- }
172
- },
173
  {
174
  "Sequence": {
175
  "id": "A",
176
  "type_id": 0
177
  }
178
  },
179
- {
180
- "SpecialToken": {
181
- "id": "<s>",
182
- "type_id": 1
183
- }
184
- },
185
  {
186
  "Sequence": {
187
  "id": "B",
@@ -189,17 +171,7 @@
189
  }
190
  }
191
  ],
192
- "special_tokens": {
193
- "<s>": {
194
- "id": "<s>",
195
- "ids": [
196
- 1
197
- ],
198
- "tokens": [
199
- "<s>"
200
- ]
201
- }
202
- }
203
  },
204
  "decoder": {
205
  "type": "Sequence",
 
150
  "post_processor": {
151
  "type": "TemplateProcessing",
152
  "single": [
 
 
 
 
 
 
153
  {
154
  "Sequence": {
155
  "id": "A",
 
158
  }
159
  ],
160
  "pair": [
 
 
 
 
 
 
161
  {
162
  "Sequence": {
163
  "id": "A",
164
  "type_id": 0
165
  }
166
  },
 
 
 
 
 
 
167
  {
168
  "Sequence": {
169
  "id": "B",
 
171
  }
172
  }
173
  ],
174
+ "special_tokens": {}
 
 
 
 
 
 
 
 
 
 
175
  },
176
  "decoder": {
177
  "type": "Sequence",
tokenizer_config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "add_bos_token": true,
3
  "add_eos_token": false,
4
  "added_tokens_decoder": {
5
  "0": {
@@ -116,7 +116,7 @@
116
  }
117
  },
118
  "bos_token": "<s>",
119
- "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif (message['role'] == 'assistant') %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}",
120
  "clean_up_tokenization_spaces": false,
121
  "eos_token": "<|endoftext|>",
122
  "legacy": false,
 
1
  {
2
+ "add_bos_token": false,
3
  "add_eos_token": false,
4
  "added_tokens_decoder": {
5
  "0": {
 
116
  }
117
  },
118
  "bos_token": "<s>",
119
+ "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}",
120
  "clean_up_tokenization_spaces": false,
121
  "eos_token": "<|endoftext|>",
122
  "legacy": false,