qassim227 commited on
Commit
fe6f643
·
verified ·
1 Parent(s): 5b237e3

End of training

Browse files
Files changed (5) hide show
  1. README.md +2 -3
  2. config.json +8 -2
  3. generation_config.json +6 -1
  4. model.safetensors +2 -2
  5. training_args.bin +1 -1
README.md CHANGED
@@ -5,7 +5,6 @@ tags:
5
  model-index:
6
  - name: Auto-pharmacy-V3
7
  results: []
8
- pipeline_tag: image-to-text
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -40,7 +39,7 @@ The following hyperparameters were used during training:
40
  - total_train_batch_size: 64
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
- - num_epochs: 1
44
  - mixed_precision_training: Native AMP
45
 
46
  ### Framework versions
@@ -48,4 +47,4 @@ The following hyperparameters were used during training:
48
  - Transformers 4.39.3
49
  - Pytorch 2.1.2
50
  - Datasets 2.18.0
51
- - Tokenizers 0.15.2
 
5
  model-index:
6
  - name: Auto-pharmacy-V3
7
  results: []
 
8
  ---
9
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
39
  - total_train_batch_size: 64
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: linear
42
+ - num_epochs: 10
43
  - mixed_precision_training: Native AMP
44
 
45
  ### Framework versions
 
47
  - Transformers 4.39.3
48
  - Pytorch 2.1.2
49
  - Datasets 2.18.0
50
+ - Tokenizers 0.15.2
config.json CHANGED
@@ -81,9 +81,10 @@
81
  "use_bfloat16": false,
82
  "use_cache": false,
83
  "use_learned_position_embeddings": true,
84
- "vocab_size": 64044
85
  },
86
  "decoder_start_token_id": 0,
 
87
  "encoder": {
88
  "_name_or_path": "",
89
  "add_cross_attention": false,
@@ -163,9 +164,14 @@
163
  },
164
  "eos_token_id": 2,
165
  "is_encoder_decoder": true,
 
 
166
  "model_type": "vision-encoder-decoder",
 
 
167
  "pad_token_id": 1,
168
  "tie_word_embeddings": false,
169
  "torch_dtype": "float32",
170
- "transformers_version": "4.39.3"
 
171
  }
 
81
  "use_bfloat16": false,
82
  "use_cache": false,
83
  "use_learned_position_embeddings": true,
84
+ "vocab_size": 50265
85
  },
86
  "decoder_start_token_id": 0,
87
+ "early_stopping": true,
88
  "encoder": {
89
  "_name_or_path": "",
90
  "add_cross_attention": false,
 
164
  },
165
  "eos_token_id": 2,
166
  "is_encoder_decoder": true,
167
+ "length_penalty": 2.0,
168
+ "max_length": 11,
169
  "model_type": "vision-encoder-decoder",
170
+ "no_repeat_ngram_size": 3,
171
+ "num_beams": 8,
172
  "pad_token_id": 1,
173
  "tie_word_embeddings": false,
174
  "torch_dtype": "float32",
175
+ "transformers_version": "4.39.3",
176
+ "vocab_size": 50265
177
  }
generation_config.json CHANGED
@@ -1,8 +1,13 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
- "decoder_start_token_id": 2,
 
5
  "eos_token_id": 2,
 
 
 
 
6
  "pad_token_id": 1,
7
  "transformers_version": "4.39.3",
8
  "use_cache": false
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
+ "decoder_start_token_id": 0,
5
+ "early_stopping": true,
6
  "eos_token_id": 2,
7
+ "length_penalty": 2.0,
8
+ "max_length": 11,
9
+ "no_repeat_ngram_size": 3,
10
+ "num_beams": 8,
11
  "pad_token_id": 1,
12
  "transformers_version": "4.39.3",
13
  "use_cache": false
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc2e35eac96d01cc46277959afa1c09067399ef495b47a745d723fcf2ddc684e
3
- size 246430696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92309973080c1e1a42d856ba651dec7767cdba4f5507ebddae04e2ec4cebe821
3
+ size 218412460
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8be483e22de4bd4aa70b490f0c953632049122e2ca8e063f9b6fe018d8845301
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1240abebcd17ba83fc49d694d73c06e15c35f9775eab27fe87e40a22fbef7861
3
  size 4920