Upload folder using huggingface_hub
Browse files- README.md +4 -0
- README.mde +6 -2
- config.json +30 -23
- tokenizer.json +2 -29
- tokenizer.model +2 -2
README.md
CHANGED
@@ -31,6 +31,10 @@ This is Transformers/HF format fp16 weights for CodeLlama 7B-Python. It is the
|
|
31 |
|
32 |
Quantisations will be coming shortly.
|
33 |
|
|
|
|
|
|
|
|
|
34 |
## Prompt template: TBC
|
35 |
|
36 |
|
|
|
31 |
|
32 |
Quantisations will be coming shortly.
|
33 |
|
34 |
+
Please note that due to a change in the RoPE Theta value, for correct results you must load these FP16 models with `trust_remote_code=True`
|
35 |
+
|
36 |
+
Credit to @emozilla for creating the necessary modelling code to achieve this!
|
37 |
+
|
38 |
## Prompt template: TBC
|
39 |
|
40 |
|
README.mde
CHANGED
@@ -22,15 +22,19 @@ tags:
|
|
22 |
<hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
|
23 |
<!-- header end -->
|
24 |
|
25 |
-
# CodeLlama
|
26 |
- Model creator: [Meta](https://ai.meta.com/llama/)
|
27 |
|
28 |
## Description
|
29 |
|
30 |
-
This is Transformers/HF format fp16 weights for CodeLlama
|
31 |
|
32 |
Quantisations will be coming shortly.
|
33 |
|
|
|
|
|
|
|
|
|
34 |
## Prompt template: TBC
|
35 |
|
36 |
|
|
|
22 |
<hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
|
23 |
<!-- header end -->
|
24 |
|
25 |
+
# CodeLlama %%MODEL%% fp16
|
26 |
- Model creator: [Meta](https://ai.meta.com/llama/)
|
27 |
|
28 |
## Description
|
29 |
|
30 |
+
This is Transformers/HF format fp16 weights for CodeLlama %%MODEL%%. It is the result of downloading CodeLlama %%MODEL%% from [Meta](https://ai.meta.com/blog/code-llama-large-language-model-coding/) and converting to HF using `convert_llama_weights_to_hf.py`.
|
31 |
|
32 |
Quantisations will be coming shortly.
|
33 |
|
34 |
+
Please note that due to a change in the RoPE Theta value, for correct results you must load these FP16 models with `trust_remote_code=True`
|
35 |
+
|
36 |
+
Credit to @emozilla for creating the necessary modelling code to achieve this!
|
37 |
+
|
38 |
## Prompt template: TBC
|
39 |
|
40 |
|
config.json
CHANGED
@@ -1,24 +1,31 @@
|
|
1 |
{
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
{
|
2 |
+
"architectures": [
|
3 |
+
"LlamaForCausalLM"
|
4 |
+
],
|
5 |
+
"bos_token_id": 1,
|
6 |
+
"eos_token_id": 2,
|
7 |
+
"hidden_act": "silu",
|
8 |
+
"hidden_size": 4096,
|
9 |
+
"initializer_range": 0.02,
|
10 |
+
"intermediate_size": 11008,
|
11 |
+
"max_position_embeddings": 16384,
|
12 |
+
"model_type": "llama",
|
13 |
+
"num_attention_heads": 32,
|
14 |
+
"num_hidden_layers": 32,
|
15 |
+
"num_key_value_heads": 32,
|
16 |
+
"pretraining_tp": 1,
|
17 |
+
"rms_norm_eps": 1e-05,
|
18 |
+
"rope_scaling": null,
|
19 |
+
"tie_word_embeddings": false,
|
20 |
+
"torch_dtype": "float16",
|
21 |
+
"transformers_version": "4.32.0",
|
22 |
+
"use_cache": true,
|
23 |
+
"vocab_size": 32016,
|
24 |
+
"auto_map": {
|
25 |
+
"AutoConfig": "configuration_llama.LlamaConfig",
|
26 |
+
"AutoModel": "modeling_llama.LlamaModel",
|
27 |
+
"AutoModelForCausalLM": "modeling_llama.LlamaForCausalLM",
|
28 |
+
"AutoModelForSequenceClassification": "modeling_llama.LlamaForSequenceClassification"
|
29 |
+
},
|
30 |
+
"rope_theta": 1000000
|
31 |
+
}
|
tokenizer.json
CHANGED
@@ -32134,23 +32134,7 @@
|
|
32134 |
"μ": 31996,
|
32135 |
"ζΆ": 31997,
|
32136 |
"εΌ": 31998,
|
32137 |
-
"η»": 31999
|
32138 |
-
"β<SU": 32000,
|
32139 |
-
"β<SUF": 32001,
|
32140 |
-
"β<PRE": 32002,
|
32141 |
-
"β<M": 32003,
|
32142 |
-
"β<MID": 32004,
|
32143 |
-
"β<E": 32005,
|
32144 |
-
"β<EOT": 32006,
|
32145 |
-
"β<PRE>": 32007,
|
32146 |
-
"β<SUF>": 32008,
|
32147 |
-
"β<MID>": 32009,
|
32148 |
-
"β<EOT>": 32010,
|
32149 |
-
"β<EOT><EOT>": 32011,
|
32150 |
-
"β<EOT><EOT><EOT>": 32012,
|
32151 |
-
"β<EOT><EOT><EOT><EOT>": 32013,
|
32152 |
-
"β<EOT><EOT><EOT><EOT><EOT>": 32014,
|
32153 |
-
"β<EOT><EOT><EOT><EOT><EOT><EOT>": 32015
|
32154 |
},
|
32155 |
"merges": [
|
32156 |
"β t",
|
@@ -93401,18 +93385,7 @@
|
|
93401 |
"βββββββββ ββββββ",
|
93402 |
"βββββββ ββββββββ",
|
93403 |
"βββββββββββ ββββ",
|
93404 |
-
"β ββββββββββββββ"
|
93405 |
-
"β< SU",
|
93406 |
-
"β<SU F",
|
93407 |
-
"β< PRE",
|
93408 |
-
"β< M",
|
93409 |
-
"β<M ID",
|
93410 |
-
"β< E",
|
93411 |
-
"β<E OT",
|
93412 |
-
"β<PRE >",
|
93413 |
-
"β<SUF >",
|
93414 |
-
"β<MID >",
|
93415 |
-
"β<EOT >"
|
93416 |
]
|
93417 |
}
|
93418 |
}
|
|
|
32134 |
"μ": 31996,
|
32135 |
"ζΆ": 31997,
|
32136 |
"εΌ": 31998,
|
32137 |
+
"η»": 31999
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32138 |
},
|
32139 |
"merges": [
|
32140 |
"β t",
|
|
|
93385 |
"βββββββββ ββββββ",
|
93386 |
"βββββββ ββββββββ",
|
93387 |
"βββββββββββ ββββ",
|
93388 |
+
"β ββββββββββββββ"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93389 |
]
|
93390 |
}
|
93391 |
}
|
tokenizer.model
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
|
3 |
+
size 499723
|