taufiqsyed commited on
Commit
d054974
·
verified ·
1 Parent(s): ef877ab

Model save

Browse files
README.md CHANGED
@@ -3,8 +3,6 @@ library_name: peft
3
  license: cc-by-nc-4.0
4
  base_model: facebook/musicgen-small
5
  tags:
6
- - text-to-audio
7
- - taufiqsyed/salami_cleaned_sampled
8
  - generated_from_trainer
9
  model-index:
10
  - name: salami_truncsplit_model_trial2
@@ -16,10 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # salami_truncsplit_model_trial2
18
 
19
- This model is a fine-tuned version of [facebook/musicgen-small](https://huggingface.co/facebook/musicgen-small) on the TAUFIQSYED/SALAMI_CLEANED_SAMPLED - DEFAULT dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 5.6239
22
- - Clap: 0.0428
23
 
24
  ## Model description
25
 
@@ -46,7 +41,7 @@ The following hyperparameters were used during training:
46
  - total_train_batch_size: 16
47
  - optimizer: Use adamw_torch with betas=(0.9,0.99) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: constant
49
- - num_epochs: 1.0
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
 
3
  license: cc-by-nc-4.0
4
  base_model: facebook/musicgen-small
5
  tags:
 
 
6
  - generated_from_trainer
7
  model-index:
8
  - name: salami_truncsplit_model_trial2
 
14
 
15
  # salami_truncsplit_model_trial2
16
 
17
+ This model is a fine-tuned version of [facebook/musicgen-small](https://huggingface.co/facebook/musicgen-small) on an unknown dataset.
 
 
 
18
 
19
  ## Model description
20
 
 
41
  - total_train_batch_size: 16
42
  - optimizer: Use adamw_torch with betas=(0.9,0.99) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: constant
44
+ - num_epochs: 0.0
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
adapter_config.json CHANGED
@@ -23,22 +23,22 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "fc2",
27
- "embed_tokens.2",
28
  "lm_heads.0",
29
- "k_proj",
30
- "enc_to_dec_proj",
31
- "lm_heads.2",
32
- "embed_tokens.1",
33
- "q_proj",
34
  "audio_enc_to_dec_proj",
35
- "lm_heads.3",
 
36
  "embed_tokens.0",
 
37
  "embed_tokens.3",
38
- "v_proj",
39
- "lm_heads.1",
 
 
 
40
  "out_proj",
41
- "fc1"
 
 
42
  ],
43
  "task_type": null,
44
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
26
  "lm_heads.0",
 
 
 
 
 
27
  "audio_enc_to_dec_proj",
28
+ "lm_heads.1",
29
+ "q_proj",
30
  "embed_tokens.0",
31
+ "embed_tokens.1",
32
  "embed_tokens.3",
33
+ "embed_tokens.2",
34
+ "k_proj",
35
+ "fc1",
36
+ "lm_heads.3",
37
+ "lm_heads.2",
38
  "out_proj",
39
+ "enc_to_dec_proj",
40
+ "v_proj",
41
+ "fc2"
42
  ],
43
  "task_type": null,
44
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a1140c66270323fd2bd7b8565dd87bdb71e767cb253cf70c0d77bffebc70b07
3
  size 85238600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:399130fbff589b1b3383eac463649c5aa103043497bcbc8b0c60dc547eac433a
3
  size 85238600
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ff2f849e7eefc5afba4b07dad74bbe23e7f1c5570e7a1382aa0b517cc30dd16
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9962517366aac47f52e1e796f6e10d85eb150e1e9a2735ecdada083f29e823ae
3
  size 5496