loiccabannes commited on
Commit
bcbc779
·
verified ·
1 Parent(s): 0a01110

Delete tmp-checkpoint-20

Browse files
tmp-checkpoint-20/emoji.json DELETED
The diff for this file is too large to render. See raw diff
 
tmp-checkpoint-20/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:58514597ee56ab00252fa4f3216f1b91e18b12706aa57efa0ae9c5427e0e8cd3
3
- size 587058052
 
 
 
 
tmp-checkpoint-20/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4800ccf99239238d25e2bbcc02745169cd458cc7934db3681fdd8f865a9c3f10
3
- size 460477330
 
 
 
 
tmp-checkpoint-20/special_tokens_map.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "bos_token": "<|startoftext|>",
3
- "eos_token": "<|endoftext|>",
4
- "pad_token": "<|endoftext|>",
5
- "unk_token": "<|endoftext|>"
6
- }
 
 
 
 
 
 
 
tmp-checkpoint-20/tokenizer_config.json DELETED
@@ -1,28 +0,0 @@
1
- {
2
- "added_tokens_decoder": {
3
- "31996": {
4
- "content": "<|startoftext|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "31999": {
12
- "content": "<|endoftext|>",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- }
19
- },
20
- "bos_token": "<|startoftext|>",
21
- "clean_up_tokenization_spaces": true,
22
- "do_clean_text": false,
23
- "eos_token": "<|endoftext|>",
24
- "model_max_length": 2048,
25
- "pad_token": "<|endoftext|>",
26
- "tokenizer_class": "GPTNeoXJapaneseTokenizer",
27
- "unk_token": "<|endoftext|>"
28
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tmp-checkpoint-20/vocab.txt DELETED
The diff for this file is too large to render. See raw diff