system HF staff commited on
Commit
9fc1aed
·
1 Parent(s): cadcfe7

Commit From AutoTrain

Browse files
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.bin.* filter=lfs diff=lfs merge=lfs -text
37
+ *.tar.gz filter=lfs diff=lfs merge=lfs -text
38
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - autotrain
4
+ - text-classification
5
+ language:
6
+ - unk
7
+ widget:
8
+ - text: "I love AutoTrain"
9
+ datasets:
10
+ - SH-W/autotrain-data-5000_koi
11
+ co2_eq_emissions:
12
+ emissions: 3.920765439350259
13
+ ---
14
+
15
+ # Model Trained Using AutoTrain
16
+
17
+ - Problem type: Multi-class Classification
18
+ - Model ID: 77927140735
19
+ - CO2 Emissions (in grams): 3.9208
20
+
21
+ ## Validation Metrics
22
+
23
+ - Loss: 2.432
24
+ - Accuracy: 0.415
25
+ - Macro F1: 0.410
26
+ - Micro F1: 0.415
27
+ - Weighted F1: 0.410
28
+ - Macro Precision: 0.459
29
+ - Micro Precision: 0.415
30
+ - Weighted Precision: 0.456
31
+ - Macro Recall: 0.413
32
+ - Micro Recall: 0.415
33
+ - Weighted Recall: 0.415
34
+
35
+
36
+ ## Usage
37
+
38
+ You can use cURL to access this model:
39
+
40
+ ```
41
+ $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoTrain"}' https://api-inference.huggingface.co/models/SH-W/autotrain-5000_koi-77927140735
42
+ ```
43
+
44
+ Or Python API:
45
+
46
+ ```
47
+ from transformers import AutoModelForSequenceClassification, AutoTokenizer
48
+
49
+ model = AutoModelForSequenceClassification.from_pretrained("SH-W/autotrain-5000_koi-77927140735", use_auth_token=True)
50
+
51
+ tokenizer = AutoTokenizer.from_pretrained("SH-W/autotrain-5000_koi-77927140735", use_auth_token=True)
52
+
53
+ inputs = tokenizer("I love AutoTrain", return_tensors="pt")
54
+
55
+ outputs = model(**inputs)
56
+ ```
config.json ADDED
@@ -0,0 +1,153 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "AutoTrain",
3
+ "_num_labels": 58,
4
+ "architectures": [
5
+ "RobertaForSequenceClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "bos_token_id": 0,
9
+ "classifier_dropout": null,
10
+ "eos_token_id": 2,
11
+ "gradient_checkpointing": false,
12
+ "hidden_act": "gelu",
13
+ "hidden_dropout_prob": 0.1,
14
+ "hidden_size": 1024,
15
+ "id2label": {
16
+ "0": "\uac00\ub09c\ud55c, \ubd88\uc6b0\ud55c",
17
+ "1": "\uac10\uc0ac\ud558\ub294",
18
+ "2": "\uac71\uc815\uc2a4\ub7ec\uc6b4",
19
+ "3": "\uace0\ub9bd\ub41c",
20
+ "4": "\uad34\ub85c\uc6cc\ud558\ub294",
21
+ "5": "\uad6c\uc5ed\uc9c8 \ub098\ub294",
22
+ "6": "\uae30\uc068",
23
+ "7": "\ub099\ub2f4\ud55c",
24
+ "8": "\ub0a8\uc758 \uc2dc\uc120\uc744 \uc758\uc2dd\ud558\ub294",
25
+ "9": "\ub178\uc5ec\uc6cc\ud558\ub294",
26
+ "10": "\ub208\ubb3c\uc774 \ub098\ub294",
27
+ "11": "\ub290\uae0b",
28
+ "12": "\ub2f9\ud639\uc2a4\ub7ec\uc6b4",
29
+ "13": "\ub2f9\ud669",
30
+ "14": "\ub450\ub824\uc6b4",
31
+ "15": "\ub9c8\ube44\ub41c",
32
+ "16": "\ub9cc\uc871\uc2a4\ub7ec\uc6b4",
33
+ "17": "\ubc29\uc5b4\uc801\uc778",
34
+ "18": "\ubc30\uc2e0\ub2f9\ud55c",
35
+ "19": "\ubc84\ub824\uc9c4",
36
+ "20": "\ubd80\ub044\ub7ec\uc6b4",
37
+ "21": "\ubd84\ub178",
38
+ "22": "\ubd88\uc548",
39
+ "23": "\ube44\ud1b5\ud55c",
40
+ "24": "\uc0c1\ucc98",
41
+ "25": "\uc131\uac00\uc2e0",
42
+ "26": "\uc2a4\ud2b8\ub808\uc2a4 \ubc1b\ub294",
43
+ "27": "\uc2ac\ud514",
44
+ "28": "\uc2e0\ub8b0\ud558\ub294",
45
+ "29": "\uc2e0\uc774 \ub09c",
46
+ "30": "\uc2e4\ub9dd\ud55c",
47
+ "31": "\uc545\uc758\uc801\uc778",
48
+ "32": "\uc548\ub2ec\ud558\ub294",
49
+ "33": "\uc548\ub3c4",
50
+ "34": "\uc5b5\uc6b8\ud55c",
51
+ "35": "\uc5f4\ub4f1\uac10",
52
+ "36": "\uc5fc\uc138\uc801\uc778",
53
+ "37": "\uc678\ub85c\uc6b4",
54
+ "38": "\uc6b0\uc6b8\ud55c",
55
+ "39": "\uc790\uc2e0\ud558\ub294",
56
+ "40": "\uc870\uc2ec\uc2a4\ub7ec\uc6b4",
57
+ "41": "\uc88c\uc808\ud55c",
58
+ "42": "\uc8c4\ucc45\uac10\uc758",
59
+ "43": "\uc9c8\ud22c\ud558\ub294",
60
+ "44": "\uc9dc\uc99d\ub0b4\ub294",
61
+ "45": "\ucd08\uc870\ud55c",
62
+ "46": "\ucda9\uaca9 \ubc1b\uc740",
63
+ "47": "\ucde8\uc57d\ud55c",
64
+ "48": "\ud234\ud234\ub300\ub294",
65
+ "49": "\ud3b8\uc548\ud55c",
66
+ "50": "\ud55c\uc2ec\ud55c",
67
+ "51": "\ud610\uc624\uc2a4\ub7ec\uc6b4",
68
+ "52": "\ud63c\ub780\uc2a4\ub7ec\uc6b4",
69
+ "53": "\ud658\uba78\uc744 \ub290\ub07c\ub294",
70
+ "54": "\ud68c\uc758\uc801\uc778",
71
+ "55": "\ud6c4\ud68c\ub418\ub294",
72
+ "56": "\ud765\ubd84",
73
+ "57": "\ud76c\uc0dd\ub41c"
74
+ },
75
+ "initializer_range": 0.02,
76
+ "intermediate_size": 4096,
77
+ "label2id": {
78
+ "\uac00\ub09c\ud55c, \ubd88\uc6b0\ud55c": 0,
79
+ "\uac10\uc0ac\ud558\ub294": 1,
80
+ "\uac71\uc815\uc2a4\ub7ec\uc6b4": 2,
81
+ "\uace0\ub9bd\ub41c": 3,
82
+ "\uad34\ub85c\uc6cc\ud558\ub294": 4,
83
+ "\uad6c\uc5ed\uc9c8 \ub098\ub294": 5,
84
+ "\uae30\uc068": 6,
85
+ "\ub099\ub2f4\ud55c": 7,
86
+ "\ub0a8\uc758 \uc2dc\uc120\uc744 \uc758\uc2dd\ud558\ub294": 8,
87
+ "\ub178\uc5ec\uc6cc\ud558\ub294": 9,
88
+ "\ub208\ubb3c\uc774 \ub098\ub294": 10,
89
+ "\ub290\uae0b": 11,
90
+ "\ub2f9\ud639\uc2a4\ub7ec\uc6b4": 12,
91
+ "\ub2f9\ud669": 13,
92
+ "\ub450\ub824\uc6b4": 14,
93
+ "\ub9c8\ube44\ub41c": 15,
94
+ "\ub9cc\uc871\uc2a4\ub7ec\uc6b4": 16,
95
+ "\ubc29\uc5b4\uc801\uc778": 17,
96
+ "\ubc30\uc2e0\ub2f9\ud55c": 18,
97
+ "\ubc84\ub824\uc9c4": 19,
98
+ "\ubd80\ub044\ub7ec\uc6b4": 20,
99
+ "\ubd84\ub178": 21,
100
+ "\ubd88\uc548": 22,
101
+ "\ube44\ud1b5\ud55c": 23,
102
+ "\uc0c1\ucc98": 24,
103
+ "\uc131\uac00\uc2e0": 25,
104
+ "\uc2a4\ud2b8\ub808\uc2a4 \ubc1b\ub294": 26,
105
+ "\uc2ac\ud514": 27,
106
+ "\uc2e0\ub8b0\ud558\ub294": 28,
107
+ "\uc2e0\uc774 \ub09c": 29,
108
+ "\uc2e4\ub9dd\ud55c": 30,
109
+ "\uc545\uc758\uc801\uc778": 31,
110
+ "\uc548\ub2ec\ud558\ub294": 32,
111
+ "\uc548\ub3c4": 33,
112
+ "\uc5b5\uc6b8\ud55c": 34,
113
+ "\uc5f4\ub4f1\uac10": 35,
114
+ "\uc5fc\uc138\uc801\uc778": 36,
115
+ "\uc678\ub85c\uc6b4": 37,
116
+ "\uc6b0\uc6b8\ud55c": 38,
117
+ "\uc790\uc2e0\ud558\ub294": 39,
118
+ "\uc870\uc2ec\uc2a4\ub7ec\uc6b4": 40,
119
+ "\uc88c\uc808\ud55c": 41,
120
+ "\uc8c4\ucc45\uac10\uc758": 42,
121
+ "\uc9c8\ud22c\ud558\ub294": 43,
122
+ "\uc9dc\uc99d\ub0b4\ub294": 44,
123
+ "\ucd08\uc870\ud55c": 45,
124
+ "\ucda9\uaca9 \ubc1b\uc740": 46,
125
+ "\ucde8\uc57d\ud55c": 47,
126
+ "\ud234\ud234\ub300\ub294": 48,
127
+ "\ud3b8\uc548\ud55c": 49,
128
+ "\ud55c\uc2ec\ud55c": 50,
129
+ "\ud610\uc624\uc2a4\ub7ec\uc6b4": 51,
130
+ "\ud63c\ub780\uc2a4\ub7ec\uc6b4": 52,
131
+ "\ud658\uba78\uc744 \ub290\ub07c\ub294": 53,
132
+ "\ud68c\uc758\uc801\uc778": 54,
133
+ "\ud6c4\ud68c\ub418\ub294": 55,
134
+ "\ud765\ubd84": 56,
135
+ "\ud76c\uc0dd\ub41c": 57
136
+ },
137
+ "layer_norm_eps": 1e-05,
138
+ "max_length": 192,
139
+ "max_position_embeddings": 514,
140
+ "model_type": "roberta",
141
+ "num_attention_heads": 16,
142
+ "num_hidden_layers": 24,
143
+ "pad_token_id": 1,
144
+ "padding": "max_length",
145
+ "position_embedding_type": "absolute",
146
+ "problem_type": "single_label_classification",
147
+ "tokenizer_class": "BertTokenizer",
148
+ "torch_dtype": "float32",
149
+ "transformers_version": "4.29.2",
150
+ "type_vocab_size": 1,
151
+ "use_cache": true,
152
+ "vocab_size": 39255
153
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e91d1fcae808e9c44105e8994ebe9c1c32221bb5adcd604112ea65ef2677acd
3
+ size 1376632224
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac84d9dfa8f0a4abea053ef8ad7580797316c7404fd37c63dc0512da36fd923e
3
+ size 1376719797
special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:371a3b718af27ecbf30617e517ef10320ea2fa90e130be7f1f0a4e7a07bb2320
3
+ size 893531
tokenizer_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "clean_up_tokenization_spaces": true,
4
+ "cls_token": "[CLS]",
5
+ "do_basic_tokenize": true,
6
+ "do_lower_case": false,
7
+ "eos_token": "[SEP]",
8
+ "mask_token": "[MASK]",
9
+ "model_max_length": 512,
10
+ "never_split": null,
11
+ "pad_token": "[PAD]",
12
+ "sep_token": "[SEP]",
13
+ "strip_accents": null,
14
+ "tokenize_chinese_chars": true,
15
+ "tokenizer_class": "BertTokenizer",
16
+ "unk_token": "[UNK]"
17
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff