fix eos token
Browse files- config.json +2 -2
config.json
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
{
|
2 |
"attention_dropout": 0.0,
|
3 |
-
"bos_token_id":
|
4 |
-
"eos_token_id":
|
5 |
"hidden_act": "gelu_pytorch_tanh",
|
6 |
"hidden_size": 3072,
|
7 |
"initializer_range": 0.018042,
|
|
|
1 |
{
|
2 |
"attention_dropout": 0.0,
|
3 |
+
"bos_token_id": 0,
|
4 |
+
"eos_token_id": 0,
|
5 |
"hidden_act": "gelu_pytorch_tanh",
|
6 |
"hidden_size": 3072,
|
7 |
"initializer_range": 0.018042,
|