RQ_RAG-Llama3.1-8B-Instruct / special_tokens_map.json
Bui1dMySea's picture
Upload tokenizer
6295fd7 verified
raw
history blame contribute delete
771 Bytes
{
"additional_special_tokens": [
"[A_Response]",
"[S_Rewritten_Query]",
"[S_Decomposed_Query]",
"[S_Disambiguated_Query]",
"[R_Evidences]",
"[/R_Evidences]",
"[S_Response]",
"[EOS]"
],
"bos_token": {
"content": "<s>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
},
"eos_token": {
"content": "</s>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
},
"pad_token": {
"content": "<pad>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
},
"unk_token": {
"content": "<unk>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
}
}