vikas117 commited on
Commit
3c8c81a
·
verified ·
1 Parent(s): 3901454

Model save

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
- license: cc-by-4.0
4
- base_model: umm-maybe/AI-image-detector
5
  tags:
6
  - generated_from_trainer
7
  datasets:
@@ -23,7 +23,7 @@ model-index:
23
  metrics:
24
  - name: Accuracy
25
  type: accuracy
26
- value: 1.0
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -31,10 +31,10 @@ should probably proofread and complete it, then remove this comment. -->
31
 
32
  # finetuned-ai-real-swin
33
 
34
- This model is a fine-tuned version of [umm-maybe/AI-image-detector](https://huggingface.co/umm-maybe/AI-image-detector) on the imagefolder dataset.
35
  It achieves the following results on the evaluation set:
36
- - Loss: 0.0001
37
- - Accuracy: 1.0
38
 
39
  ## Model description
40
 
@@ -66,7 +66,7 @@ The following hyperparameters were used during training:
66
 
67
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
68
  |:-------------:|:------:|:----:|:---------------:|:--------:|
69
- | 0.0009 | 2.2727 | 50 | 0.0001 | 1.0 |
70
 
71
 
72
  ### Framework versions
 
1
  ---
2
  library_name: transformers
3
+ license: cc-by-nc-3.0
4
+ base_model: Organika/sdxl-detector
5
  tags:
6
  - generated_from_trainer
7
  datasets:
 
23
  metrics:
24
  - name: Accuracy
25
  type: accuracy
26
+ value: 0.9752066115702479
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
31
 
32
  # finetuned-ai-real-swin
33
 
34
+ This model is a fine-tuned version of [Organika/sdxl-detector](https://huggingface.co/Organika/sdxl-detector) on the imagefolder dataset.
35
  It achieves the following results on the evaluation set:
36
+ - Loss: 0.2485
37
+ - Accuracy: 0.9752
38
 
39
  ## Model description
40
 
 
66
 
67
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
68
  |:-------------:|:------:|:----:|:---------------:|:--------:|
69
+ | 0.0141 | 2.2727 | 50 | 0.2485 | 0.9752 |
70
 
71
 
72
  ### Framework versions
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
- "epoch": 5.0,
3
- "eval_accuracy": 0.9586776859504132,
4
- "eval_loss": 0.22044861316680908,
5
- "eval_runtime": 1.7849,
6
- "eval_samples_per_second": 67.79,
7
- "eval_steps_per_second": 8.964,
8
- "total_flos": 2.663748891353088e+17,
9
- "train_loss": 0.05316784412346103,
10
- "train_runtime": 106.9013,
11
- "train_samples_per_second": 31.805,
12
- "train_steps_per_second": 1.029
13
  }
 
1
  {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 1.0,
4
+ "eval_loss": 0.000129295774968341,
5
+ "eval_runtime": 3.1734,
6
+ "eval_samples_per_second": 38.129,
7
+ "eval_steps_per_second": 5.042,
8
+ "total_flos": 1.5982493348118528e+17,
9
+ "train_loss": 0.07911565561186183,
10
+ "train_runtime": 65.9385,
11
+ "train_samples_per_second": 30.938,
12
+ "train_steps_per_second": 1.001
13
  }
config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
- "_name_or_path": "umm-maybe/AI-image-detector",
 
3
  "architectures": [
4
  "SwinForImageClassification"
5
  ],
@@ -23,8 +24,8 @@
23
  "image_size": 224,
24
  "initializer_range": 0.02,
25
  "label2id": {
26
- "artificial": "0",
27
- "human": "1"
28
  },
29
  "layer_norm_eps": 1e-05,
30
  "max_length": 128,
 
1
  {
2
+ "_name_or_path": "Organika/sdxl-detector",
3
+ "_num_labels": 2,
4
  "architectures": [
5
  "SwinForImageClassification"
6
  ],
 
24
  "image_size": 224,
25
  "initializer_range": 0.02,
26
  "label2id": {
27
+ "artificial": 0,
28
+ "human": 1
29
  },
30
  "layer_norm_eps": 1e-05,
31
  "max_length": 128,
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 5.0,
3
- "eval_accuracy": 0.9586776859504132,
4
- "eval_loss": 0.22044861316680908,
5
- "eval_runtime": 1.7849,
6
- "eval_samples_per_second": 67.79,
7
- "eval_steps_per_second": 8.964
8
  }
 
1
  {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 1.0,
4
+ "eval_loss": 0.000129295774968341,
5
+ "eval_runtime": 3.1734,
6
+ "eval_samples_per_second": 38.129,
7
+ "eval_steps_per_second": 5.042
8
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56d3f70bc13428ce7fcf6fc08ebcd51f6a472e32692d2b0534a7f0b3d14c2990
3
  size 347498816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:229caa9409e25ff1f3b2271f08fbe14e69b015b83d5549a2028075de87b318d4
3
  size 347498816
runs/Dec31_08-31-31_3dcd8bc8b582/events.out.tfevents.1735633989.3dcd8bc8b582.14185.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c8e30dfc3efad574aa5f782848e37bef497706c0dbe12275532eff1ba8ec80f
3
+ size 405
runs/Dec31_08-40-59_3dcd8bc8b582/events.out.tfevents.1735634464.3dcd8bc8b582.14185.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b628681e4df5bd3465ba5232b6a5070e78f173b35fbfb666b4a9b0e51ea91f77
3
+ size 7438
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 5.0,
3
- "total_flos": 2.663748891353088e+17,
4
- "train_loss": 0.05316784412346103,
5
- "train_runtime": 106.9013,
6
- "train_samples_per_second": 31.805,
7
- "train_steps_per_second": 1.029
8
  }
 
1
  {
2
+ "epoch": 3.0,
3
+ "total_flos": 1.5982493348118528e+17,
4
+ "train_loss": 0.07911565561186183,
5
+ "train_runtime": 65.9385,
6
+ "train_samples_per_second": 30.938,
7
+ "train_steps_per_second": 1.001
8
  }
trainer_state.json CHANGED
@@ -1,122 +1,78 @@
1
  {
2
- "best_metric": 0.1993160843849182,
3
  "best_model_checkpoint": "finetuned-ai-real-swin/checkpoint-50",
4
- "epoch": 5.0,
5
  "eval_steps": 50,
6
- "global_step": 110,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.45454545454545453,
13
- "grad_norm": 0.0013060887577012181,
14
- "learning_rate": 0.00018545454545454545,
15
- "loss": 0.2915,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.9090909090909091,
20
- "grad_norm": 0.018985752016305923,
21
- "learning_rate": 0.00016727272727272728,
22
- "loss": 0.0872,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 1.3636363636363638,
27
- "grad_norm": 3.094077328569256e-05,
28
- "learning_rate": 0.0001490909090909091,
29
- "loss": 0.0103,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 1.8181818181818183,
34
- "grad_norm": 1.7741825580596924,
35
- "learning_rate": 0.00013090909090909093,
36
- "loss": 0.0615,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 2.2727272727272725,
41
- "grad_norm": 0.0014310627011582255,
42
- "learning_rate": 0.00011272727272727272,
43
- "loss": 0.0133,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 2.2727272727272725,
48
- "eval_accuracy": 0.9834710743801653,
49
- "eval_loss": 0.1993160843849182,
50
- "eval_runtime": 1.6646,
51
- "eval_samples_per_second": 72.69,
52
- "eval_steps_per_second": 9.612,
53
  "step": 50
54
  },
55
  {
56
  "epoch": 2.7272727272727275,
57
- "grad_norm": 0.028763145208358765,
58
- "learning_rate": 9.454545454545455e-05,
59
- "loss": 0.0267,
60
  "step": 60
61
  },
62
  {
63
- "epoch": 3.1818181818181817,
64
- "grad_norm": 0.3835113048553467,
65
- "learning_rate": 7.636363636363637e-05,
66
- "loss": 0.0458,
67
- "step": 70
68
- },
69
- {
70
- "epoch": 3.6363636363636362,
71
- "grad_norm": 0.016482913866639137,
72
- "learning_rate": 5.818181818181818e-05,
73
- "loss": 0.0047,
74
- "step": 80
75
- },
76
- {
77
- "epoch": 4.090909090909091,
78
- "grad_norm": 0.0009892478119581938,
79
- "learning_rate": 4e-05,
80
- "loss": 0.0001,
81
- "step": 90
82
- },
83
- {
84
- "epoch": 4.545454545454545,
85
- "grad_norm": 0.0006707807769998908,
86
- "learning_rate": 2.1818181818181818e-05,
87
- "loss": 0.0007,
88
- "step": 100
89
- },
90
- {
91
- "epoch": 4.545454545454545,
92
- "eval_accuracy": 0.9586776859504132,
93
- "eval_loss": 0.25598978996276855,
94
- "eval_runtime": 2.0569,
95
- "eval_samples_per_second": 58.827,
96
- "eval_steps_per_second": 7.779,
97
- "step": 100
98
- },
99
- {
100
- "epoch": 5.0,
101
- "grad_norm": 13.271048545837402,
102
- "learning_rate": 3.636363636363636e-06,
103
- "loss": 0.0432,
104
- "step": 110
105
- },
106
- {
107
- "epoch": 5.0,
108
- "step": 110,
109
- "total_flos": 2.663748891353088e+17,
110
- "train_loss": 0.05316784412346103,
111
- "train_runtime": 106.9013,
112
- "train_samples_per_second": 31.805,
113
- "train_steps_per_second": 1.029
114
  }
115
  ],
116
  "logging_steps": 10,
117
- "max_steps": 110,
118
  "num_input_tokens_seen": 0,
119
- "num_train_epochs": 5,
120
  "save_steps": 100,
121
  "stateful_callbacks": {
122
  "TrainerControl": {
@@ -130,7 +86,7 @@
130
  "attributes": {}
131
  }
132
  },
133
- "total_flos": 2.663748891353088e+17,
134
  "train_batch_size": 32,
135
  "trial_name": null,
136
  "trial_params": null
 
1
  {
2
+ "best_metric": 8.1218960986007e-05,
3
  "best_model_checkpoint": "finetuned-ai-real-swin/checkpoint-50",
4
+ "epoch": 3.0,
5
  "eval_steps": 50,
6
+ "global_step": 66,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.45454545454545453,
13
+ "grad_norm": 15.097875595092773,
14
+ "learning_rate": 0.00017272727272727275,
15
+ "loss": 0.2428,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.9090909090909091,
20
+ "grad_norm": 2.583950996398926,
21
+ "learning_rate": 0.00014242424242424243,
22
+ "loss": 0.032,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 1.3636363636363638,
27
+ "grad_norm": 0.0005581710720434785,
28
+ "learning_rate": 0.0001181818181818182,
29
+ "loss": 0.0741,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 1.8181818181818183,
34
+ "grad_norm": 0.25307825207710266,
35
+ "learning_rate": 8.787878787878789e-05,
36
+ "loss": 0.1721,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 2.2727272727272725,
41
+ "grad_norm": 0.11153484135866165,
42
+ "learning_rate": 5.757575757575758e-05,
43
+ "loss": 0.0009,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 2.2727272727272725,
48
+ "eval_accuracy": 1.0,
49
+ "eval_loss": 8.1218960986007e-05,
50
+ "eval_runtime": 2.2287,
51
+ "eval_samples_per_second": 54.292,
52
+ "eval_steps_per_second": 7.179,
53
  "step": 50
54
  },
55
  {
56
  "epoch": 2.7272727272727275,
57
+ "grad_norm": 0.003770518582314253,
58
+ "learning_rate": 2.7272727272727273e-05,
59
+ "loss": 0.0002,
60
  "step": 60
61
  },
62
  {
63
+ "epoch": 3.0,
64
+ "step": 66,
65
+ "total_flos": 1.5982493348118528e+17,
66
+ "train_loss": 0.07911565561186183,
67
+ "train_runtime": 65.9385,
68
+ "train_samples_per_second": 30.938,
69
+ "train_steps_per_second": 1.001
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
70
  }
71
  ],
72
  "logging_steps": 10,
73
+ "max_steps": 66,
74
  "num_input_tokens_seen": 0,
75
+ "num_train_epochs": 3,
76
  "save_steps": 100,
77
  "stateful_callbacks": {
78
  "TrainerControl": {
 
86
  "attributes": {}
87
  }
88
  },
89
+ "total_flos": 1.5982493348118528e+17,
90
  "train_batch_size": 32,
91
  "trial_name": null,
92
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2b65e74682131468fa2d5609238fbb268468f3dbf38eb1c751b2c6e50ee26fc
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aac1e3d03f13e22039e314aa587e786c0df1cabd5fbeb7624e362f11b696d660
3
  size 5304