File size: 10,769 Bytes
ef5fb31
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
2022-01-30 17:54:17,518 INFO    MainThread:22087 [wandb_setup.py:_flush():71] setting env: {}
2022-01-30 17:54:17,518 INFO    MainThread:22087 [wandb_setup.py:_flush():71] setting login settings: {}
2022-01-30 17:54:17,518 INFO    MainThread:22087 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220130_175417-3n44neww/logs/debug.log
2022-01-30 17:54:17,518 INFO    MainThread:22087 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220130_175417-3n44neww/logs/debug-internal.log
2022-01-30 17:54:17,518 INFO    MainThread:22087 [wandb_init.py:init():404] calling init triggers
2022-01-30 17:54:17,518 INFO    MainThread:22087 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
config: {}
2022-01-30 17:54:17,518 INFO    MainThread:22087 [wandb_init.py:init():460] starting backend
2022-01-30 17:54:17,518 INFO    MainThread:22087 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-01-30 17:54:17,583 INFO    MainThread:22087 [backend.py:ensure_launched():216] starting backend process...
2022-01-30 17:54:17,675 INFO    MainThread:22087 [backend.py:ensure_launched():221] started backend process with pid: 22568
2022-01-30 17:54:17,678 INFO    MainThread:22087 [wandb_init.py:init():469] backend started and connected
2022-01-30 17:54:17,687 INFO    MainThread:22087 [wandb_init.py:init():533] updated telemetry
2022-01-30 17:54:17,831 INFO    MainThread:22087 [wandb_init.py:init():563] communicating current version
2022-01-30 17:54:18,586 INFO    MainThread:22087 [wandb_init.py:init():568] got version response 
2022-01-30 17:54:18,586 INFO    MainThread:22087 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
2022-01-30 17:54:18,779 INFO    MainThread:22087 [wandb_init.py:init():606] starting run threads in backend
2022-01-30 17:54:19,412 INFO    MainThread:22087 [wandb_run.py:_console_start():1810] atexit reg
2022-01-30 17:54:19,413 INFO    MainThread:22087 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
2022-01-30 17:54:19,414 INFO    MainThread:22087 [wandb_run.py:_redirect():1689] Redirecting console.
2022-01-30 17:54:19,420 INFO    MainThread:22087 [wandb_run.py:_redirect():1745] Redirects installed.
2022-01-30 17:54:19,420 INFO    MainThread:22087 [wandb_init.py:init():633] run started, returning control to user process
2022-01-30 17:54:19,424 INFO    MainThread:22087 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 218, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 219, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.6, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 4.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jan30_17-53-36_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 5, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 20, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 5, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
2022-01-30 17:54:19,428 INFO    MainThread:22087 [wandb_watch.py:watch():43] Watching
2022-01-30 17:55:51,472 INFO    MainThread:22087 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 255
2022-01-30 17:55:51,473 INFO    MainThread:22087 [wandb_run.py:_restore():1752] restore
2022-01-30 17:55:54,076 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2211
  total_bytes: 2211
}

2022-01-30 17:55:54,346 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2211
  total_bytes: 2211
}

2022-01-30 17:55:54,871 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 4
}
pusher_stats {
  uploaded_bytes: 2211
  total_bytes: 68995
}

2022-01-30 17:55:54,975 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 2211
  total_bytes: 72128
}

2022-01-30 17:55:55,079 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 12363
  total_bytes: 72128
}

2022-01-30 17:55:55,184 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 72128
  total_bytes: 72128
}

2022-01-30 17:55:55,288 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 72128
  total_bytes: 72128
}

2022-01-30 17:55:55,391 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 72128
  total_bytes: 72128
}

2022-01-30 17:55:55,495 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 72128
  total_bytes: 72128
}

2022-01-30 17:55:55,598 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 72128
  total_bytes: 72128
}

2022-01-30 17:55:56,189 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 72128
  total_bytes: 72128
}

2022-01-30 17:55:56,598 INFO    MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
exit_result {
}
file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 72128
  total_bytes: 72128
}
local_info {
}

2022-01-30 17:55:57,751 INFO    MainThread:22087 [wandb_run.py:_append_history():2130] rendering history
2022-01-30 17:55:57,752 INFO    MainThread:22087 [wandb_run.py:_append_summary():2085] rendering summary
2022-01-30 17:55:57,752 INFO    MainThread:22087 [wandb_run.py:_append_files():2180] logging synced files