File size: 10,769 Bytes
ef5fb31 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 |
2022-01-30 17:54:17,518 INFO MainThread:22087 [wandb_setup.py:_flush():71] setting env: {}
2022-01-30 17:54:17,518 INFO MainThread:22087 [wandb_setup.py:_flush():71] setting login settings: {}
2022-01-30 17:54:17,518 INFO MainThread:22087 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220130_175417-3n44neww/logs/debug.log
2022-01-30 17:54:17,518 INFO MainThread:22087 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220130_175417-3n44neww/logs/debug-internal.log
2022-01-30 17:54:17,518 INFO MainThread:22087 [wandb_init.py:init():404] calling init triggers
2022-01-30 17:54:17,518 INFO MainThread:22087 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
config: {}
2022-01-30 17:54:17,518 INFO MainThread:22087 [wandb_init.py:init():460] starting backend
2022-01-30 17:54:17,518 INFO MainThread:22087 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-01-30 17:54:17,583 INFO MainThread:22087 [backend.py:ensure_launched():216] starting backend process...
2022-01-30 17:54:17,675 INFO MainThread:22087 [backend.py:ensure_launched():221] started backend process with pid: 22568
2022-01-30 17:54:17,678 INFO MainThread:22087 [wandb_init.py:init():469] backend started and connected
2022-01-30 17:54:17,687 INFO MainThread:22087 [wandb_init.py:init():533] updated telemetry
2022-01-30 17:54:17,831 INFO MainThread:22087 [wandb_init.py:init():563] communicating current version
2022-01-30 17:54:18,586 INFO MainThread:22087 [wandb_init.py:init():568] got version response
2022-01-30 17:54:18,586 INFO MainThread:22087 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
2022-01-30 17:54:18,779 INFO MainThread:22087 [wandb_init.py:init():606] starting run threads in backend
2022-01-30 17:54:19,412 INFO MainThread:22087 [wandb_run.py:_console_start():1810] atexit reg
2022-01-30 17:54:19,413 INFO MainThread:22087 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
2022-01-30 17:54:19,414 INFO MainThread:22087 [wandb_run.py:_redirect():1689] Redirecting console.
2022-01-30 17:54:19,420 INFO MainThread:22087 [wandb_run.py:_redirect():1745] Redirects installed.
2022-01-30 17:54:19,420 INFO MainThread:22087 [wandb_init.py:init():633] run started, returning control to user process
2022-01-30 17:54:19,424 INFO MainThread:22087 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 218, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 219, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.6, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 4.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jan30_17-53-36_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 5, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 20, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 5, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
2022-01-30 17:54:19,428 INFO MainThread:22087 [wandb_watch.py:watch():43] Watching
2022-01-30 17:55:51,472 INFO MainThread:22087 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 255
2022-01-30 17:55:51,473 INFO MainThread:22087 [wandb_run.py:_restore():1752] restore
2022-01-30 17:55:54,076 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 1
}
pusher_stats {
uploaded_bytes: 2211
total_bytes: 2211
}
2022-01-30 17:55:54,346 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 1
}
pusher_stats {
uploaded_bytes: 2211
total_bytes: 2211
}
2022-01-30 17:55:54,871 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 4
}
pusher_stats {
uploaded_bytes: 2211
total_bytes: 68995
}
2022-01-30 17:55:54,975 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2211
total_bytes: 72128
}
2022-01-30 17:55:55,079 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 12363
total_bytes: 72128
}
2022-01-30 17:55:55,184 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 72128
total_bytes: 72128
}
2022-01-30 17:55:55,288 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 72128
total_bytes: 72128
}
2022-01-30 17:55:55,391 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 72128
total_bytes: 72128
}
2022-01-30 17:55:55,495 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 72128
total_bytes: 72128
}
2022-01-30 17:55:55,598 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 72128
total_bytes: 72128
}
2022-01-30 17:55:56,189 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 72128
total_bytes: 72128
}
2022-01-30 17:55:56,598 INFO MainThread:22087 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
exit_result {
}
file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 72128
total_bytes: 72128
}
local_info {
}
2022-01-30 17:55:57,751 INFO MainThread:22087 [wandb_run.py:_append_history():2130] rendering history
2022-01-30 17:55:57,752 INFO MainThread:22087 [wandb_run.py:_append_summary():2085] rendering summary
2022-01-30 17:55:57,752 INFO MainThread:22087 [wandb_run.py:_append_files():2180] logging synced files
|