Mardiyyah commited on
Commit
b83c3fe
·
verified ·
1 Parent(s): 032073f

Mlr-shared-task-ewc_stabilised-no-date

Browse files
Files changed (35) hide show
  1. .gitattributes +1 -0
  2. README.md +63 -0
  3. config.json +53 -0
  4. model.safetensors +3 -0
  5. sentencepiece.bpe.model +3 -0
  6. special_tokens_map.json +51 -0
  7. tokenizer.json +3 -0
  8. tokenizer_config.json +54 -0
  9. training_args.bin +3 -0
  10. wandb/debug-internal.log +0 -0
  11. wandb/debug.log +49 -0
  12. wandb/run-20240917_120410-nm6gjq7r/files/config.yaml +693 -0
  13. wandb/run-20240917_120410-nm6gjq7r/files/output.log +4 -0
  14. wandb/run-20240917_120410-nm6gjq7r/files/requirements.txt +313 -0
  15. wandb/run-20240917_120410-nm6gjq7r/files/wandb-metadata.json +281 -0
  16. wandb/run-20240917_120410-nm6gjq7r/files/wandb-summary.json +1 -0
  17. wandb/run-20240917_120410-nm6gjq7r/logs/debug-internal.log +654 -0
  18. wandb/run-20240917_120410-nm6gjq7r/logs/debug.log +72 -0
  19. wandb/run-20240917_120410-nm6gjq7r/run-nm6gjq7r.wandb +0 -0
  20. wandb/run-20240917_124325-7dvmn2ta/files/config.yaml +741 -0
  21. wandb/run-20240917_124325-7dvmn2ta/files/output.log +13 -0
  22. wandb/run-20240917_124325-7dvmn2ta/files/requirements.txt +313 -0
  23. wandb/run-20240917_124325-7dvmn2ta/files/wandb-metadata.json +281 -0
  24. wandb/run-20240917_124325-7dvmn2ta/files/wandb-summary.json +1 -0
  25. wandb/run-20240917_124325-7dvmn2ta/logs/debug-internal.log +0 -0
  26. wandb/run-20240917_124325-7dvmn2ta/logs/debug.log +60 -0
  27. wandb/run-20240917_124325-7dvmn2ta/run-7dvmn2ta.wandb +0 -0
  28. wandb/run-20240917_140929-fx32bo52/files/config.yaml +741 -0
  29. wandb/run-20240917_140929-fx32bo52/files/output.log +75 -0
  30. wandb/run-20240917_140929-fx32bo52/files/requirements.txt +313 -0
  31. wandb/run-20240917_140929-fx32bo52/files/wandb-metadata.json +281 -0
  32. wandb/run-20240917_140929-fx32bo52/files/wandb-summary.json +1 -0
  33. wandb/run-20240917_140929-fx32bo52/logs/debug-internal.log +0 -0
  34. wandb/run-20240917_140929-fx32bo52/logs/debug.log +49 -0
  35. wandb/run-20240917_140929-fx32bo52/run-fx32bo52.wandb +0 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: afl-3.0
3
+ base_model: masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: ewc_stabilised_no_date
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # ewc_stabilised_no_date
15
+
16
+ This model is a fine-tuned version of [masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0](https://huggingface.co/masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0) on the None dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - eval_loss: 0.1801
19
+ - eval_f1: 0.8373
20
+ - eval_precision: 0.8346
21
+ - eval_recall: 0.8399
22
+ - eval_accuracy: 0.9648
23
+ - eval_runtime: 74.7153
24
+ - eval_samples_per_second: 150.197
25
+ - eval_steps_per_second: 18.778
26
+ - epoch: 8.9993
27
+ - step: 6313
28
+
29
+ ## Model description
30
+
31
+ More information needed
32
+
33
+ ## Intended uses & limitations
34
+
35
+ More information needed
36
+
37
+ ## Training and evaluation data
38
+
39
+ More information needed
40
+
41
+ ## Training procedure
42
+
43
+ ### Training hyperparameters
44
+
45
+ The following hyperparameters were used during training:
46
+ - learning_rate: 2e-05
47
+ - train_batch_size: 16
48
+ - eval_batch_size: 8
49
+ - seed: 3407
50
+ - gradient_accumulation_steps: 4
51
+ - total_train_batch_size: 64
52
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
+ - lr_scheduler_type: linear
54
+ - lr_scheduler_warmup_steps: 1000
55
+ - num_epochs: 10
56
+ - mixed_precision_training: Native AMP
57
+
58
+ ### Framework versions
59
+
60
+ - Transformers 4.43.4
61
+ - Pytorch 2.4.1+cu121
62
+ - Datasets 2.20.0
63
+ - Tokenizers 0.19.1
config.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0",
3
+ "adapters": {
4
+ "adapters": {},
5
+ "config_map": {},
6
+ "fusion_config_map": {},
7
+ "fusions": {}
8
+ },
9
+ "architectures": [
10
+ "XLMRobertaForTokenClassification"
11
+ ],
12
+ "attention_probs_dropout_prob": 0.1,
13
+ "bos_token_id": 0,
14
+ "classifier_dropout": null,
15
+ "eos_token_id": 2,
16
+ "gradient_checkpointing": false,
17
+ "hidden_act": "gelu",
18
+ "hidden_dropout_prob": 0.1,
19
+ "hidden_size": 1024,
20
+ "id2label": {
21
+ "0": "O",
22
+ "1": "B-PER",
23
+ "2": "I-PER",
24
+ "3": "B-ORG",
25
+ "4": "I-ORG",
26
+ "5": "B-LOC",
27
+ "6": "I-LOC"
28
+ },
29
+ "initializer_range": 0.02,
30
+ "intermediate_size": 4096,
31
+ "label2id": {
32
+ "B-LOC": 5,
33
+ "B-ORG": 3,
34
+ "B-PER": 1,
35
+ "I-LOC": 6,
36
+ "I-ORG": 4,
37
+ "I-PER": 2,
38
+ "O": 0
39
+ },
40
+ "layer_norm_eps": 1e-05,
41
+ "max_position_embeddings": 514,
42
+ "model_type": "xlm-roberta",
43
+ "num_attention_heads": 16,
44
+ "num_hidden_layers": 24,
45
+ "output_past": true,
46
+ "pad_token_id": 1,
47
+ "position_embedding_type": "absolute",
48
+ "torch_dtype": "float32",
49
+ "transformers_version": "4.43.4",
50
+ "type_vocab_size": 1,
51
+ "use_cache": true,
52
+ "vocab_size": 250002
53
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e941f0cc2579d9cbfb110910229f6cac163dbdb14c72bc2d548b24eb2d2ad7c
3
+ size 2235440556
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:883b037111086fd4dfebbbc9b7cee11e1517b5e0c0514879478661440f137085
3
+ size 17082987
tokenizer_config.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "250001": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "eos_token": "</s>",
48
+ "mask_token": "<mask>",
49
+ "model_max_length": 512,
50
+ "pad_token": "<pad>",
51
+ "sep_token": "</s>",
52
+ "tokenizer_class": "XLMRobertaTokenizer",
53
+ "unk_token": "<unk>"
54
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92a4df8ddec76bfb710014b760ac91af4d01d2f058abc581a01e3eae2e539d63
3
+ size 5304
wandb/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-09-17 14:09:29,476 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Current SDK version is 0.17.9
2
+ 2024-09-17 14:09:29,476 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Configure stats pid to 1271154
3
+ 2024-09-17 14:09:29,476 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Loading settings from /homes/amrufai/.config/wandb/settings
4
+ 2024-09-17 14:09:29,476 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Loading settings from /nfs/production/literature/amina-mardiyyah/wandb/settings
5
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Loading settings from environment variables: {}
6
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Inferring run settings from compute environment: {'program': '<python with no main file>'}
7
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Applying login settings: {}
8
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Applying login settings: {}
9
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:_log_setup():524] Logging user logs to /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_140929-fx32bo52/logs/debug.log
10
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:_log_setup():525] Logging internal logs to /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_140929-fx32bo52/logs/debug-internal.log
11
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:_jupyter_setup():470] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7fbb5e686b90>
12
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:init():608] calling init triggers
13
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:init():615] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:init():658] starting backend
16
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:init():662] setting up manager
17
+ 2024-09-17 14:09:29,479 INFO MainThread:1271154 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-09-17 14:09:29,485 INFO MainThread:1271154 [wandb_init.py:init():670] backend started and connected
19
+ 2024-09-17 14:09:29,490 INFO MainThread:1271154 [wandb_run.py:_label_probe_notebook():1344] probe notebook
20
+ 2024-09-17 14:09:29,504 INFO MainThread:1271154 [wandb_run.py:_label_probe_notebook():1354] Unable to probe notebook: 'NoneType' object has no attribute 'get'
21
+ 2024-09-17 14:09:29,504 INFO MainThread:1271154 [wandb_init.py:init():768] updated telemetry
22
+ 2024-09-17 14:09:29,514 INFO MainThread:1271154 [wandb_init.py:init():801] communicating run to backend with 90.0 second timeout
23
+ 2024-09-17 14:09:29,907 INFO MainThread:1271154 [wandb_init.py:init():852] starting run threads in backend
24
+ 2024-09-17 14:09:30,803 INFO MainThread:1271154 [wandb_run.py:_console_start():2465] atexit reg
25
+ 2024-09-17 14:09:30,803 INFO MainThread:1271154 [wandb_run.py:_redirect():2311] redirect: wrap_raw
26
+ 2024-09-17 14:09:30,803 INFO MainThread:1271154 [wandb_run.py:_redirect():2376] Wrapping output streams.
27
+ 2024-09-17 14:09:30,803 INFO MainThread:1271154 [wandb_run.py:_redirect():2401] Redirects installed.
28
+ 2024-09-17 14:09:30,811 INFO MainThread:1271154 [wandb_init.py:init():895] run started, returning control to user process
29
+ 2024-09-17 14:09:30,821 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
30
+ 2024-09-17 14:09:30,821 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
31
+ 2024-09-17 14:11:21,293 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
32
+ 2024-09-17 14:11:21,353 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
33
+ 2024-09-17 14:11:21,353 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
34
+ 2024-09-17 14:11:23,083 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
35
+ 2024-09-17 14:11:23,089 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
36
+ 2024-09-17 14:11:23,090 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
37
+ 2024-09-17 14:11:28,594 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
38
+ 2024-09-17 14:11:36,241 INFO MainThread:1271154 [wandb_run.py:_config_callback():1392] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['XLMRobertaForTokenClassification'], 'finetuning_task': None, 'id2label': {0: 'O', 1: 'B-PER', 2: 'I-PER', 3: 'B-ORG', 4: 'I-ORG', 5: 'B-LOC', 6: 'I-LOC'}, 'label2id': {'O': 0, 'B-PER': 1, 'I-PER': 2, 'B-ORG': 3, 'I-ORG': 4, 'B-LOC': 5, 'I-LOC': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0', 'transformers_version': '4.43.4', 'adapters': {'adapters': {}, 'config_map': {}, 'fusion_config_map': {}, 'fusions': {}}, 'gradient_checkpointing': False, 'model_type': 'xlm-roberta', 'output_past': True, 'vocab_size': 250002, 'hidden_size': 1024, 'num_hidden_layers': 24, 'num_attention_heads': 16, 'hidden_act': 'gelu', 'intermediate_size': 4096, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 514, 'type_vocab_size': 1, 'initializer_range': 0.02, 'layer_norm_eps': 1e-05, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 2e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/logs', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 3407, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'eval_f1', 'greater_is_better': True, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'eval_use_gather_object': False}
39
+ 2024-09-17 14:11:36,243 INFO MainThread:1271154 [wandb_config.py:__setitem__():154] config set model/num_parameters = 558848007 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7fbad9edf090>>
40
+ 2024-09-17 14:11:36,243 INFO MainThread:1271154 [wandb_run.py:_config_callback():1392] config_cb model/num_parameters 558848007 None
41
+ 2024-09-17 15:31:39,471 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
42
+ 2024-09-17 15:31:39,472 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
43
+ 2024-09-17 15:31:41,931 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
44
+ 2024-09-17 15:31:41,938 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
45
+ 2024-09-17 15:31:41,938 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
46
+ 2024-09-17 15:31:49,020 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
47
+ 2024-09-17 15:31:49,568 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
48
+ 2024-09-17 15:31:49,568 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
49
+ 2024-09-17 15:32:22,919 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
wandb/run-20240917_120410-nm6gjq7r/files/config.yaml ADDED
@@ -0,0 +1,693 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.11.4
7
+ cli_version: 0.17.9
8
+ framework: huggingface
9
+ huggingface_version: 4.43.4
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: false
12
+ start_time: 1726571050
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 5
17
+ - 11
18
+ - 41
19
+ - 49
20
+ - 51
21
+ - 53
22
+ - 55
23
+ - 71
24
+ - 98
25
+ - 100
26
+ - 105
27
+ 2:
28
+ - 1
29
+ - 5
30
+ - 11
31
+ - 41
32
+ - 49
33
+ - 51
34
+ - 53
35
+ - 55
36
+ - 71
37
+ - 98
38
+ - 100
39
+ - 105
40
+ 3:
41
+ - 7
42
+ - 19
43
+ - 23
44
+ - 66
45
+ 4: 3.11.4
46
+ 5: 0.17.9
47
+ 6: 4.43.4
48
+ 8:
49
+ - 1
50
+ - 5
51
+ 9:
52
+ 1: transformers_trainer
53
+ 13: linux-x86_64
54
+ m:
55
+ - 1: train/global_step
56
+ 6:
57
+ - 3
58
+ return_dict:
59
+ desc: null
60
+ value: true
61
+ output_hidden_states:
62
+ desc: null
63
+ value: false
64
+ output_attentions:
65
+ desc: null
66
+ value: false
67
+ torchscript:
68
+ desc: null
69
+ value: false
70
+ torch_dtype:
71
+ desc: null
72
+ value: float32
73
+ use_bfloat16:
74
+ desc: null
75
+ value: false
76
+ tf_legacy_loss:
77
+ desc: null
78
+ value: false
79
+ pruned_heads:
80
+ desc: null
81
+ value: {}
82
+ tie_word_embeddings:
83
+ desc: null
84
+ value: true
85
+ chunk_size_feed_forward:
86
+ desc: null
87
+ value: 0
88
+ is_encoder_decoder:
89
+ desc: null
90
+ value: false
91
+ is_decoder:
92
+ desc: null
93
+ value: false
94
+ cross_attention_hidden_size:
95
+ desc: null
96
+ value: null
97
+ add_cross_attention:
98
+ desc: null
99
+ value: false
100
+ tie_encoder_decoder:
101
+ desc: null
102
+ value: false
103
+ max_length:
104
+ desc: null
105
+ value: 20
106
+ min_length:
107
+ desc: null
108
+ value: 0
109
+ do_sample:
110
+ desc: null
111
+ value: false
112
+ early_stopping:
113
+ desc: null
114
+ value: false
115
+ num_beams:
116
+ desc: null
117
+ value: 1
118
+ num_beam_groups:
119
+ desc: null
120
+ value: 1
121
+ diversity_penalty:
122
+ desc: null
123
+ value: 0.0
124
+ temperature:
125
+ desc: null
126
+ value: 1.0
127
+ top_k:
128
+ desc: null
129
+ value: 50
130
+ top_p:
131
+ desc: null
132
+ value: 1.0
133
+ typical_p:
134
+ desc: null
135
+ value: 1.0
136
+ repetition_penalty:
137
+ desc: null
138
+ value: 1.0
139
+ length_penalty:
140
+ desc: null
141
+ value: 1.0
142
+ no_repeat_ngram_size:
143
+ desc: null
144
+ value: 0
145
+ encoder_no_repeat_ngram_size:
146
+ desc: null
147
+ value: 0
148
+ bad_words_ids:
149
+ desc: null
150
+ value: null
151
+ num_return_sequences:
152
+ desc: null
153
+ value: 1
154
+ output_scores:
155
+ desc: null
156
+ value: false
157
+ return_dict_in_generate:
158
+ desc: null
159
+ value: false
160
+ forced_bos_token_id:
161
+ desc: null
162
+ value: null
163
+ forced_eos_token_id:
164
+ desc: null
165
+ value: null
166
+ remove_invalid_values:
167
+ desc: null
168
+ value: false
169
+ exponential_decay_length_penalty:
170
+ desc: null
171
+ value: null
172
+ suppress_tokens:
173
+ desc: null
174
+ value: null
175
+ begin_suppress_tokens:
176
+ desc: null
177
+ value: null
178
+ architectures:
179
+ desc: null
180
+ value:
181
+ - XLMRobertaForTokenClassification
182
+ finetuning_task:
183
+ desc: null
184
+ value: null
185
+ id2label:
186
+ desc: null
187
+ value:
188
+ '0': O
189
+ '1': B-PER
190
+ '2': I-PER
191
+ '3': B-ORG
192
+ '4': I-ORG
193
+ '5': B-LOC
194
+ '6': I-LOC
195
+ label2id:
196
+ desc: null
197
+ value:
198
+ O: 0
199
+ B-PER: 1
200
+ I-PER: 2
201
+ B-ORG: 3
202
+ I-ORG: 4
203
+ B-LOC: 5
204
+ I-LOC: 6
205
+ tokenizer_class:
206
+ desc: null
207
+ value: null
208
+ prefix:
209
+ desc: null
210
+ value: null
211
+ bos_token_id:
212
+ desc: null
213
+ value: 0
214
+ pad_token_id:
215
+ desc: null
216
+ value: 1
217
+ eos_token_id:
218
+ desc: null
219
+ value: 2
220
+ sep_token_id:
221
+ desc: null
222
+ value: null
223
+ decoder_start_token_id:
224
+ desc: null
225
+ value: null
226
+ task_specific_params:
227
+ desc: null
228
+ value: null
229
+ problem_type:
230
+ desc: null
231
+ value: null
232
+ _name_or_path:
233
+ desc: null
234
+ value: masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0
235
+ transformers_version:
236
+ desc: null
237
+ value: 4.43.4
238
+ adapters:
239
+ desc: null
240
+ value:
241
+ adapters: {}
242
+ config_map: {}
243
+ fusion_config_map: {}
244
+ fusions: {}
245
+ gradient_checkpointing:
246
+ desc: null
247
+ value: false
248
+ model_type:
249
+ desc: null
250
+ value: xlm-roberta
251
+ output_past:
252
+ desc: null
253
+ value: true
254
+ vocab_size:
255
+ desc: null
256
+ value: 250002
257
+ hidden_size:
258
+ desc: null
259
+ value: 1024
260
+ num_hidden_layers:
261
+ desc: null
262
+ value: 24
263
+ num_attention_heads:
264
+ desc: null
265
+ value: 16
266
+ hidden_act:
267
+ desc: null
268
+ value: gelu
269
+ intermediate_size:
270
+ desc: null
271
+ value: 4096
272
+ hidden_dropout_prob:
273
+ desc: null
274
+ value: 0.1
275
+ attention_probs_dropout_prob:
276
+ desc: null
277
+ value: 0.1
278
+ max_position_embeddings:
279
+ desc: null
280
+ value: 514
281
+ type_vocab_size:
282
+ desc: null
283
+ value: 1
284
+ initializer_range:
285
+ desc: null
286
+ value: 0.02
287
+ layer_norm_eps:
288
+ desc: null
289
+ value: 1.0e-05
290
+ position_embedding_type:
291
+ desc: null
292
+ value: absolute
293
+ use_cache:
294
+ desc: null
295
+ value: true
296
+ classifier_dropout:
297
+ desc: null
298
+ value: null
299
+ output_dir:
300
+ desc: null
301
+ value: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date
302
+ overwrite_output_dir:
303
+ desc: null
304
+ value: false
305
+ do_train:
306
+ desc: null
307
+ value: false
308
+ do_eval:
309
+ desc: null
310
+ value: true
311
+ do_predict:
312
+ desc: null
313
+ value: false
314
+ eval_strategy:
315
+ desc: null
316
+ value: epoch
317
+ prediction_loss_only:
318
+ desc: null
319
+ value: false
320
+ per_device_train_batch_size:
321
+ desc: null
322
+ value: 16
323
+ per_device_eval_batch_size:
324
+ desc: null
325
+ value: 8
326
+ per_gpu_train_batch_size:
327
+ desc: null
328
+ value: null
329
+ per_gpu_eval_batch_size:
330
+ desc: null
331
+ value: null
332
+ gradient_accumulation_steps:
333
+ desc: null
334
+ value: 4
335
+ eval_accumulation_steps:
336
+ desc: null
337
+ value: null
338
+ eval_delay:
339
+ desc: null
340
+ value: 0
341
+ torch_empty_cache_steps:
342
+ desc: null
343
+ value: null
344
+ learning_rate:
345
+ desc: null
346
+ value: 2.0e-05
347
+ weight_decay:
348
+ desc: null
349
+ value: 0.001
350
+ adam_beta1:
351
+ desc: null
352
+ value: 0.9
353
+ adam_beta2:
354
+ desc: null
355
+ value: 0.999
356
+ adam_epsilon:
357
+ desc: null
358
+ value: 1.0e-08
359
+ max_grad_norm:
360
+ desc: null
361
+ value: 1.0
362
+ num_train_epochs:
363
+ desc: null
364
+ value: 10
365
+ max_steps:
366
+ desc: null
367
+ value: -1
368
+ lr_scheduler_type:
369
+ desc: null
370
+ value: linear
371
+ lr_scheduler_kwargs:
372
+ desc: null
373
+ value: {}
374
+ warmup_ratio:
375
+ desc: null
376
+ value: 0.0
377
+ warmup_steps:
378
+ desc: null
379
+ value: 1000
380
+ log_level:
381
+ desc: null
382
+ value: passive
383
+ log_level_replica:
384
+ desc: null
385
+ value: warning
386
+ log_on_each_node:
387
+ desc: null
388
+ value: true
389
+ logging_dir:
390
+ desc: null
391
+ value: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/logs
392
+ logging_strategy:
393
+ desc: null
394
+ value: epoch
395
+ logging_first_step:
396
+ desc: null
397
+ value: false
398
+ logging_steps:
399
+ desc: null
400
+ value: 500
401
+ logging_nan_inf_filter:
402
+ desc: null
403
+ value: true
404
+ save_strategy:
405
+ desc: null
406
+ value: epoch
407
+ save_steps:
408
+ desc: null
409
+ value: 500
410
+ save_total_limit:
411
+ desc: null
412
+ value: 2
413
+ save_safetensors:
414
+ desc: null
415
+ value: true
416
+ save_on_each_node:
417
+ desc: null
418
+ value: false
419
+ save_only_model:
420
+ desc: null
421
+ value: false
422
+ restore_callback_states_from_checkpoint:
423
+ desc: null
424
+ value: false
425
+ no_cuda:
426
+ desc: null
427
+ value: false
428
+ use_cpu:
429
+ desc: null
430
+ value: false
431
+ use_mps_device:
432
+ desc: null
433
+ value: false
434
+ seed:
435
+ desc: null
436
+ value: 3407
437
+ data_seed:
438
+ desc: null
439
+ value: null
440
+ jit_mode_eval:
441
+ desc: null
442
+ value: false
443
+ use_ipex:
444
+ desc: null
445
+ value: false
446
+ bf16:
447
+ desc: null
448
+ value: false
449
+ fp16:
450
+ desc: null
451
+ value: true
452
+ fp16_opt_level:
453
+ desc: null
454
+ value: O1
455
+ half_precision_backend:
456
+ desc: null
457
+ value: auto
458
+ bf16_full_eval:
459
+ desc: null
460
+ value: false
461
+ fp16_full_eval:
462
+ desc: null
463
+ value: false
464
+ tf32:
465
+ desc: null
466
+ value: null
467
+ local_rank:
468
+ desc: null
469
+ value: 0
470
+ ddp_backend:
471
+ desc: null
472
+ value: null
473
+ tpu_num_cores:
474
+ desc: null
475
+ value: null
476
+ tpu_metrics_debug:
477
+ desc: null
478
+ value: false
479
+ debug:
480
+ desc: null
481
+ value: []
482
+ dataloader_drop_last:
483
+ desc: null
484
+ value: false
485
+ eval_steps:
486
+ desc: null
487
+ value: null
488
+ dataloader_num_workers:
489
+ desc: null
490
+ value: 0
491
+ dataloader_prefetch_factor:
492
+ desc: null
493
+ value: null
494
+ past_index:
495
+ desc: null
496
+ value: -1
497
+ run_name:
498
+ desc: null
499
+ value: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date
500
+ disable_tqdm:
501
+ desc: null
502
+ value: false
503
+ remove_unused_columns:
504
+ desc: null
505
+ value: true
506
+ label_names:
507
+ desc: null
508
+ value: null
509
+ load_best_model_at_end:
510
+ desc: null
511
+ value: true
512
+ metric_for_best_model:
513
+ desc: null
514
+ value: eval_f1
515
+ greater_is_better:
516
+ desc: null
517
+ value: true
518
+ ignore_data_skip:
519
+ desc: null
520
+ value: false
521
+ fsdp:
522
+ desc: null
523
+ value: []
524
+ fsdp_min_num_params:
525
+ desc: null
526
+ value: 0
527
+ fsdp_config:
528
+ desc: null
529
+ value:
530
+ min_num_params: 0
531
+ xla: false
532
+ xla_fsdp_v2: false
533
+ xla_fsdp_grad_ckpt: false
534
+ fsdp_transformer_layer_cls_to_wrap:
535
+ desc: null
536
+ value: null
537
+ accelerator_config:
538
+ desc: null
539
+ value:
540
+ split_batches: false
541
+ dispatch_batches: null
542
+ even_batches: true
543
+ use_seedable_sampler: true
544
+ non_blocking: false
545
+ gradient_accumulation_kwargs: null
546
+ deepspeed:
547
+ desc: null
548
+ value: null
549
+ label_smoothing_factor:
550
+ desc: null
551
+ value: 0.0
552
+ optim:
553
+ desc: null
554
+ value: adamw_torch
555
+ optim_args:
556
+ desc: null
557
+ value: null
558
+ adafactor:
559
+ desc: null
560
+ value: false
561
+ group_by_length:
562
+ desc: null
563
+ value: true
564
+ length_column_name:
565
+ desc: null
566
+ value: length
567
+ report_to:
568
+ desc: null
569
+ value:
570
+ - wandb
571
+ ddp_find_unused_parameters:
572
+ desc: null
573
+ value: null
574
+ ddp_bucket_cap_mb:
575
+ desc: null
576
+ value: null
577
+ ddp_broadcast_buffers:
578
+ desc: null
579
+ value: null
580
+ dataloader_pin_memory:
581
+ desc: null
582
+ value: true
583
+ dataloader_persistent_workers:
584
+ desc: null
585
+ value: false
586
+ skip_memory_metrics:
587
+ desc: null
588
+ value: true
589
+ use_legacy_prediction_loop:
590
+ desc: null
591
+ value: false
592
+ push_to_hub:
593
+ desc: null
594
+ value: false
595
+ resume_from_checkpoint:
596
+ desc: null
597
+ value: null
598
+ hub_model_id:
599
+ desc: null
600
+ value: null
601
+ hub_strategy:
602
+ desc: null
603
+ value: every_save
604
+ hub_token:
605
+ desc: null
606
+ value: <HUB_TOKEN>
607
+ hub_private_repo:
608
+ desc: null
609
+ value: false
610
+ hub_always_push:
611
+ desc: null
612
+ value: false
613
+ gradient_checkpointing_kwargs:
614
+ desc: null
615
+ value: null
616
+ include_inputs_for_metrics:
617
+ desc: null
618
+ value: false
619
+ eval_do_concat_batches:
620
+ desc: null
621
+ value: true
622
+ fp16_backend:
623
+ desc: null
624
+ value: auto
625
+ evaluation_strategy:
626
+ desc: null
627
+ value: null
628
+ push_to_hub_model_id:
629
+ desc: null
630
+ value: null
631
+ push_to_hub_organization:
632
+ desc: null
633
+ value: null
634
+ push_to_hub_token:
635
+ desc: null
636
+ value: <PUSH_TO_HUB_TOKEN>
637
+ mp_parameters:
638
+ desc: null
639
+ value: ''
640
+ auto_find_batch_size:
641
+ desc: null
642
+ value: false
643
+ full_determinism:
644
+ desc: null
645
+ value: false
646
+ torchdynamo:
647
+ desc: null
648
+ value: null
649
+ ray_scope:
650
+ desc: null
651
+ value: last
652
+ ddp_timeout:
653
+ desc: null
654
+ value: 1800
655
+ torch_compile:
656
+ desc: null
657
+ value: false
658
+ torch_compile_backend:
659
+ desc: null
660
+ value: null
661
+ torch_compile_mode:
662
+ desc: null
663
+ value: null
664
+ dispatch_batches:
665
+ desc: null
666
+ value: null
667
+ split_batches:
668
+ desc: null
669
+ value: null
670
+ include_tokens_per_second:
671
+ desc: null
672
+ value: false
673
+ include_num_input_tokens_seen:
674
+ desc: null
675
+ value: false
676
+ neftune_noise_alpha:
677
+ desc: null
678
+ value: null
679
+ optim_target_modules:
680
+ desc: null
681
+ value: null
682
+ batch_eval_metrics:
683
+ desc: null
684
+ value: false
685
+ eval_on_start:
686
+ desc: null
687
+ value: false
688
+ eval_use_gather_object:
689
+ desc: null
690
+ value: false
691
+ model/num_parameters:
692
+ desc: null
693
+ value: 558848007
wandb/run-20240917_120410-nm6gjq7r/files/output.log ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ /nfs/production/literature/amina-mardiyyah/envs/llm-prompt/lib/python3.11/site-packages/accelerate/accelerator.py:488: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead.
2
+ self.scaler = torch.cuda.amp.GradScaler(**kwargs)
3
+ Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
4
+ wandb: WARNING The `run_name` is currently set to the same value as `TrainingArguments.output_dir`. If this was not intended, please specify a different run name by setting the `TrainingArguments.run_name` parameter.
wandb/run-20240917_120410-nm6gjq7r/files/requirements.txt ADDED
@@ -0,0 +1,313 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ AutoROM.accept-rom-license==0.6.1
2
+ AutoROM==0.6.1
3
+ Farama-Notifications==0.0.4
4
+ GitPython==3.1.43
5
+ Jinja2==3.1.4
6
+ Mako==1.3.5
7
+ Markdown==3.7
8
+ MarkupSafe==2.1.5
9
+ PyYAML==6.0.2
10
+ Pygments==2.18.0
11
+ QtPy==2.4.1
12
+ SQLAlchemy==2.0.32
13
+ Send2Trash==1.8.3
14
+ Shimmy==1.3.0
15
+ Werkzeug==3.0.4
16
+ absl-py==2.1.0
17
+ accelerate==0.33.0
18
+ aiohappyeyeballs==2.3.5
19
+ aiohttp-retry==2.8.3
20
+ aiohttp==3.10.3
21
+ aiosignal==1.3.1
22
+ ale-py==0.8.1
23
+ alembic==1.13.2
24
+ amqp==5.2.0
25
+ annotated-types==0.7.0
26
+ antlr4-python3-runtime==4.9.3
27
+ anyio==4.4.0
28
+ appdirs==1.4.4
29
+ argon2-cffi-bindings==21.2.0
30
+ argon2-cffi==23.1.0
31
+ arrow==1.3.0
32
+ asttokens==2.4.1
33
+ async-generator==1.10
34
+ async-lru==2.0.4
35
+ asyncssh==2.17.0
36
+ atpublic==5.0
37
+ attrs==24.2.0
38
+ audioread==3.0.1
39
+ babel==2.16.0
40
+ beautifulsoup4==4.12.3
41
+ billiard==4.2.0
42
+ bitsandbytes==0.43.3
43
+ bleach==6.1.0
44
+ blis==0.7.11
45
+ catalogue==2.0.10
46
+ celery==5.4.0
47
+ certifi==2024.7.4
48
+ certipy==0.1.3
49
+ cffi==1.17.0
50
+ charset-normalizer==3.3.2
51
+ click-didyoumean==0.3.1
52
+ click-plugins==1.1.1
53
+ click-repl==0.3.0
54
+ click==8.1.7
55
+ cloudpathlib==0.18.1
56
+ cloudpickle==3.0.0
57
+ colorama==0.4.6
58
+ coloredlogs==15.0.1
59
+ comm==0.2.2
60
+ conda-store==2024.6.1
61
+ confection==0.1.5
62
+ configobj==5.0.8
63
+ contourpy==1.2.1
64
+ cryptography==43.0.0
65
+ cuda-python==11.8.3
66
+ cycler==0.12.1
67
+ cymem==2.0.8
68
+ datasets==2.20.0
69
+ debugpy==1.8.5
70
+ decorator==5.1.1
71
+ defusedxml==0.7.1
72
+ dictdiffer==0.9.0
73
+ dill==0.3.8
74
+ diskcache==5.6.3
75
+ distro==1.9.0
76
+ docker-pycreds==0.4.0
77
+ docstring_parser==0.16
78
+ dpath==2.2.0
79
+ dulwich==0.22.1
80
+ dvc-data==3.16.5
81
+ dvc-http==2.32.0
82
+ dvc-objects==5.1.0
83
+ dvc-render==1.0.2
84
+ dvc-studio-client==0.21.0
85
+ dvc-task==0.4.0
86
+ dvc==3.55.2
87
+ en-core-web-lg==3.7.1
88
+ en-core-web-sm==3.7.1
89
+ entrypoints==0.4
90
+ evaluate==0.4.2
91
+ executing==2.0.1
92
+ fastjsonschema==2.20.0
93
+ filelock==3.15.4
94
+ flatbuffers==24.3.25
95
+ flatten-dict==0.4.2
96
+ flufl.lock==8.1.0
97
+ fonttools==4.53.1
98
+ fqdn==1.5.1
99
+ frozenlist==1.4.1
100
+ fsspec==2024.5.0
101
+ funcy==2.0
102
+ gitdb==4.0.11
103
+ grandalf==0.8
104
+ greenlet==3.0.3
105
+ grpcio==1.66.0
106
+ gto==1.7.1
107
+ gymnasium==0.29.1
108
+ h11==0.14.0
109
+ httpcore==1.0.5
110
+ httpx==0.27.0
111
+ huggingface-hub==0.24.5
112
+ humanfriendly==10.0
113
+ hydra-core==1.3.2
114
+ idna==3.7
115
+ importlib_resources==6.4.5
116
+ ipykernel==6.29.5
117
+ ipython==8.26.0
118
+ ipywidgets==8.1.3
119
+ isoduration==20.11.0
120
+ iterative-telemetry==0.0.8
121
+ jedi==0.19.1
122
+ joblib==1.4.2
123
+ json5==0.9.25
124
+ jsonpointer==3.0.0
125
+ jsonschema-specifications==2023.12.1
126
+ jsonschema==4.23.0
127
+ jupyter-console==6.6.3
128
+ jupyter-events==0.10.0
129
+ jupyter-launcher-shortcuts==4.0.3
130
+ jupyter-lsp==2.2.5
131
+ jupyter-telemetry==0.1.0
132
+ jupyter==1.0.0
133
+ jupyter_client==8.6.2
134
+ jupyter_core==5.7.2
135
+ jupyter_server==2.14.2
136
+ jupyter_server_terminals==0.5.3
137
+ jupyterhub==1.5.1
138
+ jupyterlab-conda-store==2024.6.1
139
+ jupyterlab==4.2.4
140
+ jupyterlab_pygments==0.3.0
141
+ jupyterlab_server==2.27.3
142
+ jupyterlab_widgets==3.0.11
143
+ kiwisolver==1.4.5
144
+ kombu==5.4.0
145
+ lab==8.2
146
+ langcodes==3.4.0
147
+ language_data==1.2.0
148
+ lazy_loader==0.4
149
+ librosa==0.10.2.post1
150
+ llvmlite==0.43.0
151
+ lxml==5.3.0
152
+ marisa-trie==1.2.0
153
+ markdown-it-py==3.0.0
154
+ matplotlib-inline==0.1.7
155
+ matplotlib==3.9.1.post1
156
+ mdurl==0.1.2
157
+ mistune==3.0.2
158
+ mpmath==1.3.0
159
+ msgpack==1.1.0
160
+ multidict==6.0.5
161
+ multiprocess==0.70.16
162
+ murmurhash==1.0.10
163
+ nb_conda_store_kernels==0.1.5
164
+ nbclient==0.10.0
165
+ nbconvert==7.16.4
166
+ nbformat==5.10.4
167
+ nest-asyncio==1.6.0
168
+ networkx==3.3
169
+ nodejs==0.1.1
170
+ notebook==7.2.1
171
+ notebook_shim==0.2.4
172
+ numba==0.60.0
173
+ numpy==1.26.4
174
+ nvidia-cublas-cu12==12.1.3.1
175
+ nvidia-cuda-cupti-cu12==12.1.105
176
+ nvidia-cuda-nvrtc-cu12==12.1.105
177
+ nvidia-cuda-runtime-cu12==12.1.105
178
+ nvidia-cudnn-cu12==8.9.2.26
179
+ nvidia-cudnn-cu12==9.1.0.70
180
+ nvidia-cufft-cu12==11.0.2.54
181
+ nvidia-curand-cu12==10.3.2.106
182
+ nvidia-cusolver-cu12==11.4.5.107
183
+ nvidia-cusparse-cu12==12.1.0.106
184
+ nvidia-nccl-cu12==2.19.3
185
+ nvidia-nccl-cu12==2.20.5
186
+ nvidia-nvjitlink-cu12==12.6.20
187
+ nvidia-nvtx-cu12==12.1.105
188
+ oauthlib==3.2.2
189
+ omegaconf==2.3.0
190
+ onnx==1.16.2
191
+ onnxruntime==1.19.0
192
+ opencv-python==4.10.0.84
193
+ optimum==1.21.4
194
+ optional-django==0.1.0
195
+ orjson==3.10.7
196
+ overrides==7.7.0
197
+ packaging==24.1
198
+ pamela==1.2.0
199
+ pandas==2.2.2
200
+ pandocfilters==1.5.1
201
+ parso==0.8.4
202
+ pathspec==0.12.1
203
+ peft==0.12.0
204
+ pexpect==4.9.0
205
+ pillow==10.4.0
206
+ pip==23.1.2
207
+ platformdirs==3.11.0
208
+ pooch==1.8.2
209
+ preshed==3.0.9
210
+ prometheus_client==0.20.0
211
+ prompt_toolkit==3.0.47
212
+ protobuf==5.27.3
213
+ psutil==6.0.0
214
+ ptyprocess==0.7.0
215
+ pure_eval==0.2.3
216
+ pyOpenSSL==24.2.1
217
+ pyarrow-hotfix==0.6
218
+ pyarrow==17.0.0
219
+ pycparser==2.22
220
+ pydantic==2.8.2
221
+ pydantic_core==2.20.1
222
+ pydot==3.0.1
223
+ pygame==2.6.0
224
+ pygit2==1.15.1
225
+ pygtrie==2.5.0
226
+ pyparsing==3.1.2
227
+ python-dateutil==2.9.0.post0
228
+ python-dotenv==1.0.1
229
+ python-json-logger==2.0.7
230
+ pytz==2024.1
231
+ pyzmq==26.1.0
232
+ qtconsole==5.5.2
233
+ referencing==0.35.1
234
+ regex==2024.7.24
235
+ requests==2.32.3
236
+ rfc3339-validator==0.1.4
237
+ rfc3986-validator==0.1.1
238
+ rich==13.7.1
239
+ rpds-py==0.20.0
240
+ ruamel.yaml.clib==0.2.8
241
+ ruamel.yaml==0.18.6
242
+ safetensors==0.4.4
243
+ scikit-learn==1.5.1
244
+ scipy==1.14.0
245
+ scmrepo==3.3.7
246
+ seaborn==0.13.2
247
+ semver==3.0.2
248
+ sentencepiece==0.2.0
249
+ sentry-sdk==2.14.0
250
+ seqeval==1.2.2
251
+ setproctitle==1.3.3
252
+ setuptools==65.5.0
253
+ shellingham==1.5.4
254
+ shortuuid==1.0.13
255
+ shtab==1.7.1
256
+ simplejson==3.19.2
257
+ six==1.16.0
258
+ smart-open==7.0.4
259
+ smmap==5.0.1
260
+ sniffio==1.3.1
261
+ soundfile==0.12.1
262
+ soupsieve==2.5
263
+ sox==1.5.0
264
+ soxr==0.5.0.post1
265
+ spacy-legacy==3.0.12
266
+ spacy-loggers==1.0.5
267
+ spacy==3.7.5
268
+ sqltrie==0.11.1
269
+ srsly==2.4.8
270
+ stable_baselines3==2.3.2
271
+ stack-data==0.6.3
272
+ sympy==1.13.2
273
+ tabulate==0.9.0
274
+ tensorboard-data-server==0.7.2
275
+ tensorboard==2.17.1
276
+ terminado==0.18.1
277
+ thinc==8.2.5
278
+ threadpoolctl==3.5.0
279
+ tinycss2==1.3.0
280
+ tokenizers==0.19.1
281
+ tomlkit==0.13.2
282
+ torch==2.4.1
283
+ torchaudio==2.4.1
284
+ torchvision==0.19.1
285
+ tornado==6.4.1
286
+ tqdm==4.66.5
287
+ traitlets==5.14.3
288
+ transformers==4.43.4
289
+ triton==3.0.0
290
+ trl==0.9.6
291
+ txt2tags==3.9
292
+ typer==0.12.3
293
+ types-python-dateutil==2.9.0.20240316
294
+ typing_extensions==4.12.2
295
+ tyro==0.8.6
296
+ tzdata==2024.1
297
+ uri-template==1.3.0
298
+ urllib3==2.2.2
299
+ vine==5.1.0
300
+ voluptuous==0.15.2
301
+ wandb==0.17.9
302
+ wasabi==1.1.3
303
+ wcwidth==0.2.13
304
+ weasel==0.4.1
305
+ webcolors==24.8.0
306
+ webencodings==0.5.1
307
+ websocket-client==1.8.0
308
+ widgetsnbextension==4.0.11
309
+ wrapt==1.16.0
310
+ xlrd==2.0.1
311
+ xxhash==3.4.1
312
+ yarl==1.9.4
313
+ zc.lockfile==3.0.post1
wandb/run-20240917_120410-nm6gjq7r/files/wandb-metadata.json ADDED
@@ -0,0 +1,281 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.18.0-513.24.1.el8_9.x86_64-x86_64-with-glibc2.28",
3
+ "python": "3.11.4",
4
+ "heartbeatAt": "2024-09-17T11:04:11.273502",
5
+ "startedAt": "2024-09-17T11:04:10.689415",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "<python with no main file>",
11
+ "codePathLocal": null,
12
+ "host": "codon-gpu-014.ebi.ac.uk",
13
+ "username": "amrufai",
14
+ "executable": "/nfs/production/literature/amina-mardiyyah/envs/llm-prompt/bin/python",
15
+ "cpu_count": 48,
16
+ "cpu_count_logical": 48,
17
+ "cpu_freq": {
18
+ "current": 2728.292645833333,
19
+ "min": 0.0,
20
+ "max": 0.0
21
+ },
22
+ "cpu_freq_per_core": [
23
+ {
24
+ "current": 2800.0,
25
+ "min": 0.0,
26
+ "max": 0.0
27
+ },
28
+ {
29
+ "current": 2800.0,
30
+ "min": 0.0,
31
+ "max": 0.0
32
+ },
33
+ {
34
+ "current": 2800.0,
35
+ "min": 0.0,
36
+ "max": 0.0
37
+ },
38
+ {
39
+ "current": 2800.0,
40
+ "min": 0.0,
41
+ "max": 0.0
42
+ },
43
+ {
44
+ "current": 2800.0,
45
+ "min": 0.0,
46
+ "max": 0.0
47
+ },
48
+ {
49
+ "current": 2800.0,
50
+ "min": 0.0,
51
+ "max": 0.0
52
+ },
53
+ {
54
+ "current": 1182.367,
55
+ "min": 0.0,
56
+ "max": 0.0
57
+ },
58
+ {
59
+ "current": 2800.0,
60
+ "min": 0.0,
61
+ "max": 0.0
62
+ },
63
+ {
64
+ "current": 2800.0,
65
+ "min": 0.0,
66
+ "max": 0.0
67
+ },
68
+ {
69
+ "current": 2800.0,
70
+ "min": 0.0,
71
+ "max": 0.0
72
+ },
73
+ {
74
+ "current": 2800.0,
75
+ "min": 0.0,
76
+ "max": 0.0
77
+ },
78
+ {
79
+ "current": 2800.0,
80
+ "min": 0.0,
81
+ "max": 0.0
82
+ },
83
+ {
84
+ "current": 2800.0,
85
+ "min": 0.0,
86
+ "max": 0.0
87
+ },
88
+ {
89
+ "current": 2800.0,
90
+ "min": 0.0,
91
+ "max": 0.0
92
+ },
93
+ {
94
+ "current": 2800.0,
95
+ "min": 0.0,
96
+ "max": 0.0
97
+ },
98
+ {
99
+ "current": 2800.0,
100
+ "min": 0.0,
101
+ "max": 0.0
102
+ },
103
+ {
104
+ "current": 2800.0,
105
+ "min": 0.0,
106
+ "max": 0.0
107
+ },
108
+ {
109
+ "current": 2800.0,
110
+ "min": 0.0,
111
+ "max": 0.0
112
+ },
113
+ {
114
+ "current": 2800.0,
115
+ "min": 0.0,
116
+ "max": 0.0
117
+ },
118
+ {
119
+ "current": 2800.0,
120
+ "min": 0.0,
121
+ "max": 0.0
122
+ },
123
+ {
124
+ "current": 975.68,
125
+ "min": 0.0,
126
+ "max": 0.0
127
+ },
128
+ {
129
+ "current": 2800.0,
130
+ "min": 0.0,
131
+ "max": 0.0
132
+ },
133
+ {
134
+ "current": 2800.0,
135
+ "min": 0.0,
136
+ "max": 0.0
137
+ },
138
+ {
139
+ "current": 2800.0,
140
+ "min": 0.0,
141
+ "max": 0.0
142
+ },
143
+ {
144
+ "current": 2800.0,
145
+ "min": 0.0,
146
+ "max": 0.0
147
+ },
148
+ {
149
+ "current": 2800.0,
150
+ "min": 0.0,
151
+ "max": 0.0
152
+ },
153
+ {
154
+ "current": 2800.0,
155
+ "min": 0.0,
156
+ "max": 0.0
157
+ },
158
+ {
159
+ "current": 2800.0,
160
+ "min": 0.0,
161
+ "max": 0.0
162
+ },
163
+ {
164
+ "current": 2800.0,
165
+ "min": 0.0,
166
+ "max": 0.0
167
+ },
168
+ {
169
+ "current": 2800.0,
170
+ "min": 0.0,
171
+ "max": 0.0
172
+ },
173
+ {
174
+ "current": 2800.0,
175
+ "min": 0.0,
176
+ "max": 0.0
177
+ },
178
+ {
179
+ "current": 2800.0,
180
+ "min": 0.0,
181
+ "max": 0.0
182
+ },
183
+ {
184
+ "current": 2800.0,
185
+ "min": 0.0,
186
+ "max": 0.0
187
+ },
188
+ {
189
+ "current": 2800.0,
190
+ "min": 0.0,
191
+ "max": 0.0
192
+ },
193
+ {
194
+ "current": 2800.0,
195
+ "min": 0.0,
196
+ "max": 0.0
197
+ },
198
+ {
199
+ "current": 2800.0,
200
+ "min": 0.0,
201
+ "max": 0.0
202
+ },
203
+ {
204
+ "current": 2800.0,
205
+ "min": 0.0,
206
+ "max": 0.0
207
+ },
208
+ {
209
+ "current": 2800.0,
210
+ "min": 0.0,
211
+ "max": 0.0
212
+ },
213
+ {
214
+ "current": 2800.0,
215
+ "min": 0.0,
216
+ "max": 0.0
217
+ },
218
+ {
219
+ "current": 2800.0,
220
+ "min": 0.0,
221
+ "max": 0.0
222
+ },
223
+ {
224
+ "current": 2800.0,
225
+ "min": 0.0,
226
+ "max": 0.0
227
+ },
228
+ {
229
+ "current": 2800.0,
230
+ "min": 0.0,
231
+ "max": 0.0
232
+ },
233
+ {
234
+ "current": 2800.0,
235
+ "min": 0.0,
236
+ "max": 0.0
237
+ },
238
+ {
239
+ "current": 2800.0,
240
+ "min": 0.0,
241
+ "max": 0.0
242
+ },
243
+ {
244
+ "current": 2800.0,
245
+ "min": 0.0,
246
+ "max": 0.0
247
+ },
248
+ {
249
+ "current": 2800.0,
250
+ "min": 0.0,
251
+ "max": 0.0
252
+ },
253
+ {
254
+ "current": 2800.0,
255
+ "min": 0.0,
256
+ "max": 0.0
257
+ },
258
+ {
259
+ "current": 2800.0,
260
+ "min": 0.0,
261
+ "max": 0.0
262
+ }
263
+ ],
264
+ "disk": {
265
+ "/": {
266
+ "total": 47.760292053222656,
267
+ "used": 15.848091125488281
268
+ }
269
+ },
270
+ "gpu": "NVIDIA A100 80GB PCIe",
271
+ "gpu_count": 1,
272
+ "gpu_devices": [
273
+ {
274
+ "name": "NVIDIA A100 80GB PCIe",
275
+ "memory_total": 85899345920
276
+ }
277
+ ],
278
+ "memory": {
279
+ "total": 502.83758544921875
280
+ }
281
+ }
wandb/run-20240917_120410-nm6gjq7r/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 19}}
wandb/run-20240917_120410-nm6gjq7r/logs/debug-internal.log ADDED
@@ -0,0 +1,654 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-09-17 12:04:10,722 INFO StreamThr :1253420 [internal.py:wandb_internal():85] W&B internal server running at pid: 1253420, started at: 2024-09-17 12:04:10.719948
2
+ 2024-09-17 12:04:10,726 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status
3
+ 2024-09-17 12:04:10,740 INFO WriterThread:1253420 [datastore.py:open_for_write():87] open: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/run-nm6gjq7r.wandb
4
+ 2024-09-17 12:04:10,747 DEBUG SenderThread:1253420 [sender.py:send():391] send: header
5
+ 2024-09-17 12:04:10,748 DEBUG SenderThread:1253420 [sender.py:send():391] send: run
6
+ 2024-09-17 12:04:11,130 INFO SenderThread:1253420 [dir_watcher.py:__init__():211] watching files in: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files
7
+ 2024-09-17 12:04:11,130 INFO SenderThread:1253420 [sender.py:_start_run_threads():1200] run started: nm6gjq7r with start time 1726571050.72261
8
+ 2024-09-17 12:04:11,138 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: run_start
9
+ 2024-09-17 12:04:11,244 DEBUG HandlerThread:1253420 [system_info.py:__init__():26] System info init
10
+ 2024-09-17 12:04:11,244 DEBUG HandlerThread:1253420 [system_info.py:__init__():41] System info init done
11
+ 2024-09-17 12:04:11,244 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
12
+ 2024-09-17 12:04:11,245 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
13
+ 2024-09-17 12:04:11,245 INFO HandlerThread:1253420 [system_monitor.py:probe():214] Collecting system info
14
+ 2024-09-17 12:04:11,246 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
15
+ 2024-09-17 12:04:11,250 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
16
+ 2024-09-17 12:04:11,252 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
17
+ 2024-09-17 12:04:11,258 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
18
+ 2024-09-17 12:04:11,261 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
19
+ 2024-09-17 12:04:11,273 DEBUG HandlerThread:1253420 [system_info.py:probe():152] Probing system
20
+ 2024-09-17 12:04:11,277 DEBUG HandlerThread:1253420 [gitlib.py:_init_repo():56] git repository is invalid
21
+ 2024-09-17 12:04:11,277 DEBUG HandlerThread:1253420 [system_info.py:probe():200] Probing system done
22
+ 2024-09-17 12:04:11,277 DEBUG HandlerThread:1253420 [system_monitor.py:probe():223] {'os': 'Linux-4.18.0-513.24.1.el8_9.x86_64-x86_64-with-glibc2.28', 'python': '3.11.4', 'heartbeatAt': '2024-09-17T11:04:11.273502', 'startedAt': '2024-09-17T11:04:10.689415', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': '<python with no main file>', 'codePathLocal': None, 'host': 'codon-gpu-014.ebi.ac.uk', 'username': 'amrufai', 'executable': '/nfs/production/literature/amina-mardiyyah/envs/llm-prompt/bin/python', 'cpu_count': 48, 'cpu_count_logical': 48, 'cpu_freq': {'current': 2728.292645833333, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 1182.367, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 975.68, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}, {'current': 2800.0, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 47.760292053222656, 'used': 15.848091125488281}}, 'gpu': 'NVIDIA A100 80GB PCIe', 'gpu_count': 1, 'gpu_devices': [{'name': 'NVIDIA A100 80GB PCIe', 'memory_total': 85899345920}], 'memory': {'total': 502.83758544921875}}
23
+ 2024-09-17 12:04:11,277 INFO HandlerThread:1253420 [system_monitor.py:probe():224] Finished collecting system info
24
+ 2024-09-17 12:04:11,277 INFO HandlerThread:1253420 [system_monitor.py:probe():227] Publishing system info
25
+ 2024-09-17 12:04:11,281 INFO HandlerThread:1253420 [system_monitor.py:probe():229] Finished publishing system info
26
+ 2024-09-17 12:04:11,322 DEBUG SenderThread:1253420 [sender.py:send():391] send: files
27
+ 2024-09-17 12:04:11,323 INFO SenderThread:1253420 [sender.py:_save_file():1466] saving file wandb-metadata.json with policy now
28
+ 2024-09-17 12:04:11,920 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: python_packages
29
+ 2024-09-17 12:04:11,921 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: python_packages
30
+ 2024-09-17 12:04:11,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
31
+ 2024-09-17 12:04:11,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
32
+ 2024-09-17 12:04:11,926 DEBUG SenderThread:1253420 [sender.py:send():391] send: telemetry
33
+ 2024-09-17 12:04:11,927 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
34
+ 2024-09-17 12:04:11,938 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
35
+ 2024-09-17 12:04:11,938 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
36
+ 2024-09-17 12:04:11,938 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
37
+ 2024-09-17 12:04:11,939 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
38
+ 2024-09-17 12:04:11,940 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
39
+ 2024-09-17 12:04:11,942 INFO wandb-upload_0:1253420 [upload_job.py:push():130] Uploaded file /tmp/tmp39_l9k7owandb/2kmhtdvh-wandb-metadata.json
40
+ 2024-09-17 12:04:11,943 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
41
+ 2024-09-17 12:04:11,944 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
42
+ 2024-09-17 12:04:11,944 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
43
+ 2024-09-17 12:04:12,004 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
44
+ 2024-09-17 12:04:12,005 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
45
+ 2024-09-17 12:04:12,005 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
46
+ 2024-09-17 12:04:12,134 INFO Thread-12 :1253420 [dir_watcher.py:_on_file_created():271] file/dir created: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/wandb-metadata.json
47
+ 2024-09-17 12:04:12,135 INFO Thread-12 :1253420 [dir_watcher.py:_on_file_created():271] file/dir created: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/requirements.txt
48
+ 2024-09-17 12:04:12,156 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
49
+ 2024-09-17 12:04:16,158 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
50
+ 2024-09-17 12:04:21,923 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
51
+ 2024-09-17 12:04:21,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
52
+ 2024-09-17 12:04:26,921 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
53
+ 2024-09-17 12:04:26,922 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
54
+ 2024-09-17 12:04:27,156 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
55
+ 2024-09-17 12:04:31,923 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
56
+ 2024-09-17 12:04:32,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
57
+ 2024-09-17 12:04:38,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
58
+ 2024-09-17 12:04:41,921 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
59
+ 2024-09-17 12:04:41,922 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
60
+ 2024-09-17 12:04:41,962 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
61
+ 2024-09-17 12:04:44,140 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
62
+ 2024-09-17 12:04:45,171 INFO Thread-12 :1253420 [dir_watcher.py:_on_file_modified():288] file/dir modified: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/config.yaml
63
+ 2024-09-17 12:04:47,842 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
64
+ 2024-09-17 12:04:47,842 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
65
+ 2024-09-17 12:04:47,842 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
66
+ 2024-09-17 12:04:47,842 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
67
+ 2024-09-17 12:04:47,843 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
68
+ 2024-09-17 12:04:47,844 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
69
+ 2024-09-17 12:04:47,846 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
70
+ 2024-09-17 12:04:47,849 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
71
+ 2024-09-17 12:04:47,853 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
72
+ 2024-09-17 12:04:47,893 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
73
+ 2024-09-17 12:04:47,893 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
74
+ 2024-09-17 12:04:47,894 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
75
+ 2024-09-17 12:04:47,894 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
76
+ 2024-09-17 12:04:47,896 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
77
+ 2024-09-17 12:04:47,896 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
78
+ 2024-09-17 12:04:47,896 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
79
+ 2024-09-17 12:04:47,897 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
80
+ 2024-09-17 12:04:47,957 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
81
+ 2024-09-17 12:04:47,957 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
82
+ 2024-09-17 12:04:47,958 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
83
+ 2024-09-17 12:04:47,959 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
84
+ 2024-09-17 12:04:48,687 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
85
+ 2024-09-17 12:04:48,687 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
86
+ 2024-09-17 12:04:48,687 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
87
+ 2024-09-17 12:04:48,688 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
88
+ 2024-09-17 12:04:48,689 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
89
+ 2024-09-17 12:04:48,691 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
90
+ 2024-09-17 12:04:48,695 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
91
+ 2024-09-17 12:04:48,696 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
92
+ 2024-09-17 12:04:48,696 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
93
+ 2024-09-17 12:04:48,697 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
94
+ 2024-09-17 12:04:48,697 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
95
+ 2024-09-17 12:04:48,697 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
96
+ 2024-09-17 12:04:48,697 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
97
+ 2024-09-17 12:04:48,699 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
98
+ 2024-09-17 12:04:48,701 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
99
+ 2024-09-17 12:04:48,765 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
100
+ 2024-09-17 12:04:48,767 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
101
+ 2024-09-17 12:04:49,768 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
102
+ 2024-09-17 12:04:51,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
103
+ 2024-09-17 12:04:54,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
104
+ 2024-09-17 12:04:56,921 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
105
+ 2024-09-17 12:04:56,922 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
106
+ 2024-09-17 12:05:00,082 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
107
+ 2024-09-17 12:05:01,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
108
+ 2024-09-17 12:05:05,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
109
+ 2024-09-17 12:05:07,802 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
110
+ 2024-09-17 12:05:07,802 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
111
+ 2024-09-17 12:05:07,802 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
112
+ 2024-09-17 12:05:07,803 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
113
+ 2024-09-17 12:05:07,804 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
114
+ 2024-09-17 12:05:07,807 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
115
+ 2024-09-17 12:05:07,809 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
116
+ 2024-09-17 12:05:07,810 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
117
+ 2024-09-17 12:05:07,811 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
118
+ 2024-09-17 12:05:07,843 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
119
+ 2024-09-17 12:05:07,843 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
120
+ 2024-09-17 12:05:07,843 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
121
+ 2024-09-17 12:05:07,844 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
122
+ 2024-09-17 12:05:07,844 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
123
+ 2024-09-17 12:05:07,845 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
124
+ 2024-09-17 12:05:07,845 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
125
+ 2024-09-17 12:05:07,848 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
126
+ 2024-09-17 12:05:07,906 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
127
+ 2024-09-17 12:05:07,906 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
128
+ 2024-09-17 12:05:07,907 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
129
+ 2024-09-17 12:05:07,908 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
130
+ 2024-09-17 12:05:11,910 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
131
+ 2024-09-17 12:05:11,922 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
132
+ 2024-09-17 12:05:11,922 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
133
+ 2024-09-17 12:05:11,928 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
134
+ 2024-09-17 12:05:16,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
135
+ 2024-09-17 12:05:21,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
136
+ 2024-09-17 12:05:21,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
137
+ 2024-09-17 12:05:26,922 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
138
+ 2024-09-17 12:05:26,923 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
139
+ 2024-09-17 12:05:27,086 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
140
+ 2024-09-17 12:05:31,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
141
+ 2024-09-17 12:05:32,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
142
+ 2024-09-17 12:05:38,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
143
+ 2024-09-17 12:05:41,922 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
144
+ 2024-09-17 12:05:41,923 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
145
+ 2024-09-17 12:05:41,963 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
146
+ 2024-09-17 12:05:44,139 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
147
+ 2024-09-17 12:05:49,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
148
+ 2024-09-17 12:05:50,657 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
149
+ 2024-09-17 12:05:50,657 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
150
+ 2024-09-17 12:05:50,657 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
151
+ 2024-09-17 12:05:50,658 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
152
+ 2024-09-17 12:05:50,659 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
153
+ 2024-09-17 12:05:50,661 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
154
+ 2024-09-17 12:05:50,665 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
155
+ 2024-09-17 12:05:50,666 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
156
+ 2024-09-17 12:05:50,667 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
157
+ 2024-09-17 12:05:50,698 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
158
+ 2024-09-17 12:05:50,698 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
159
+ 2024-09-17 12:05:50,698 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
160
+ 2024-09-17 12:05:50,700 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
161
+ 2024-09-17 12:05:50,700 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
162
+ 2024-09-17 12:05:50,701 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
163
+ 2024-09-17 12:05:50,701 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
164
+ 2024-09-17 12:05:50,701 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
165
+ 2024-09-17 12:05:50,760 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
166
+ 2024-09-17 12:05:50,761 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
167
+ 2024-09-17 12:05:50,761 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
168
+ 2024-09-17 12:05:50,762 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
169
+ 2024-09-17 12:05:51,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
170
+ 2024-09-17 12:05:53,407 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
171
+ 2024-09-17 12:05:53,407 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
172
+ 2024-09-17 12:05:53,407 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
173
+ 2024-09-17 12:05:53,408 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
174
+ 2024-09-17 12:05:53,409 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
175
+ 2024-09-17 12:05:53,411 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
176
+ 2024-09-17 12:05:53,413 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
177
+ 2024-09-17 12:05:53,415 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
178
+ 2024-09-17 12:05:53,417 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
179
+ 2024-09-17 12:05:53,448 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
180
+ 2024-09-17 12:05:53,448 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
181
+ 2024-09-17 12:05:53,448 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
182
+ 2024-09-17 12:05:53,449 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
183
+ 2024-09-17 12:05:53,450 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
184
+ 2024-09-17 12:05:53,451 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
185
+ 2024-09-17 12:05:53,451 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
186
+ 2024-09-17 12:05:53,451 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
187
+ 2024-09-17 12:05:53,510 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
188
+ 2024-09-17 12:05:53,510 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
189
+ 2024-09-17 12:05:53,511 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
190
+ 2024-09-17 12:05:53,512 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
191
+ 2024-09-17 12:05:54,529 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
192
+ 2024-09-17 12:05:54,529 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
193
+ 2024-09-17 12:05:54,529 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
194
+ 2024-09-17 12:05:54,530 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
195
+ 2024-09-17 12:05:54,531 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
196
+ 2024-09-17 12:05:54,534 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
197
+ 2024-09-17 12:05:54,535 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
198
+ 2024-09-17 12:05:54,537 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
199
+ 2024-09-17 12:05:54,538 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
200
+ 2024-09-17 12:05:54,702 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
201
+ 2024-09-17 12:05:54,702 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
202
+ 2024-09-17 12:05:54,702 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
203
+ 2024-09-17 12:05:54,703 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
204
+ 2024-09-17 12:05:54,703 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
205
+ 2024-09-17 12:05:54,703 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
206
+ 2024-09-17 12:05:54,706 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
207
+ 2024-09-17 12:05:54,706 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
208
+ 2024-09-17 12:05:54,763 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
209
+ 2024-09-17 12:05:54,764 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
210
+ 2024-09-17 12:05:54,764 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
211
+ 2024-09-17 12:05:54,765 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
212
+ 2024-09-17 12:05:55,767 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
213
+ 2024-09-17 12:05:56,922 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
214
+ 2024-09-17 12:05:56,923 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
215
+ 2024-09-17 12:06:01,129 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
216
+ 2024-09-17 12:06:01,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
217
+ 2024-09-17 12:06:04,941 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
218
+ 2024-09-17 12:06:04,942 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
219
+ 2024-09-17 12:06:04,942 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
220
+ 2024-09-17 12:06:04,942 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
221
+ 2024-09-17 12:06:04,944 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
222
+ 2024-09-17 12:06:04,946 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
223
+ 2024-09-17 12:06:04,947 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
224
+ 2024-09-17 12:06:04,950 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
225
+ 2024-09-17 12:06:04,952 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
226
+ 2024-09-17 12:06:05,237 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
227
+ 2024-09-17 12:06:05,238 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
228
+ 2024-09-17 12:06:05,238 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
229
+ 2024-09-17 12:06:05,239 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
230
+ 2024-09-17 12:06:05,240 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
231
+ 2024-09-17 12:06:05,241 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
232
+ 2024-09-17 12:06:05,241 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
233
+ 2024-09-17 12:06:05,241 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
234
+ 2024-09-17 12:06:05,301 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
235
+ 2024-09-17 12:06:05,301 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
236
+ 2024-09-17 12:06:05,302 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
237
+ 2024-09-17 12:06:05,303 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
238
+ 2024-09-17 12:06:06,304 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
239
+ 2024-09-17 12:06:11,922 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
240
+ 2024-09-17 12:06:11,924 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
241
+ 2024-09-17 12:06:11,963 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
242
+ 2024-09-17 12:06:12,050 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
243
+ 2024-09-17 12:06:13,487 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
244
+ 2024-09-17 12:06:13,488 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
245
+ 2024-09-17 12:06:13,488 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
246
+ 2024-09-17 12:06:13,488 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
247
+ 2024-09-17 12:06:13,491 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
248
+ 2024-09-17 12:06:13,492 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
249
+ 2024-09-17 12:06:13,493 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
250
+ 2024-09-17 12:06:13,494 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
251
+ 2024-09-17 12:06:13,496 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
252
+ 2024-09-17 12:06:13,528 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
253
+ 2024-09-17 12:06:13,528 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
254
+ 2024-09-17 12:06:13,528 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
255
+ 2024-09-17 12:06:13,529 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
256
+ 2024-09-17 12:06:13,530 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
257
+ 2024-09-17 12:06:13,530 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
258
+ 2024-09-17 12:06:13,530 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
259
+ 2024-09-17 12:06:13,532 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
260
+ 2024-09-17 12:06:13,589 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
261
+ 2024-09-17 12:06:13,590 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
262
+ 2024-09-17 12:06:13,590 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
263
+ 2024-09-17 12:06:13,591 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
264
+ 2024-09-17 12:06:14,617 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
265
+ 2024-09-17 12:06:14,617 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
266
+ 2024-09-17 12:06:14,617 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
267
+ 2024-09-17 12:06:14,617 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
268
+ 2024-09-17 12:06:14,619 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
269
+ 2024-09-17 12:06:14,620 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
270
+ 2024-09-17 12:06:14,621 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
271
+ 2024-09-17 12:06:14,622 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
272
+ 2024-09-17 12:06:14,624 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
273
+ 2024-09-17 12:06:14,624 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
274
+ 2024-09-17 12:06:14,624 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
275
+ 2024-09-17 12:06:14,628 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
276
+ 2024-09-17 12:06:14,628 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
277
+ 2024-09-17 12:06:14,628 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
278
+ 2024-09-17 12:06:14,628 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
279
+ 2024-09-17 12:06:14,630 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
280
+ 2024-09-17 12:06:14,632 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
281
+ 2024-09-17 12:06:14,687 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
282
+ 2024-09-17 12:06:14,687 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
283
+ 2024-09-17 12:06:14,688 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
284
+ 2024-09-17 12:06:14,688 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
285
+ 2024-09-17 12:06:14,930 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
286
+ 2024-09-17 12:06:14,931 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
287
+ 2024-09-17 12:06:14,931 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
288
+ 2024-09-17 12:06:14,931 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
289
+ 2024-09-17 12:06:14,932 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
290
+ 2024-09-17 12:06:14,935 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
291
+ 2024-09-17 12:06:14,938 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
292
+ 2024-09-17 12:06:14,941 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
293
+ 2024-09-17 12:06:14,942 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
294
+ 2024-09-17 12:06:17,928 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
295
+ 2024-09-17 12:06:21,979 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
296
+ 2024-09-17 12:06:22,979 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
297
+ 2024-09-17 12:06:24,010 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
298
+ 2024-09-17 12:06:24,011 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
299
+ 2024-09-17 12:06:24,011 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
300
+ 2024-09-17 12:06:24,013 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
301
+ 2024-09-17 12:06:24,013 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
302
+ 2024-09-17 12:06:24,014 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
303
+ 2024-09-17 12:06:24,014 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
304
+ 2024-09-17 12:06:24,014 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
305
+ 2024-09-17 12:06:24,076 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
306
+ 2024-09-17 12:06:24,077 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
307
+ 2024-09-17 12:06:24,077 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
308
+ 2024-09-17 12:06:24,078 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
309
+ 2024-09-17 12:06:25,314 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
310
+ 2024-09-17 12:06:25,314 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
311
+ 2024-09-17 12:06:25,314 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
312
+ 2024-09-17 12:06:25,314 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
313
+ 2024-09-17 12:06:25,316 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
314
+ 2024-09-17 12:06:25,319 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
315
+ 2024-09-17 12:06:25,320 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
316
+ 2024-09-17 12:06:25,323 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
317
+ 2024-09-17 12:06:25,325 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
318
+ 2024-09-17 12:06:25,325 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
319
+ 2024-09-17 12:06:25,325 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
320
+ 2024-09-17 12:06:25,326 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
321
+ 2024-09-17 12:06:25,327 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
322
+ 2024-09-17 12:06:25,328 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
323
+ 2024-09-17 12:06:25,329 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
324
+ 2024-09-17 12:06:25,329 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
325
+ 2024-09-17 12:06:25,332 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
326
+ 2024-09-17 12:06:25,390 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
327
+ 2024-09-17 12:06:25,390 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
328
+ 2024-09-17 12:06:25,391 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
329
+ 2024-09-17 12:06:25,391 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
330
+ 2024-09-17 12:06:26,068 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
331
+ 2024-09-17 12:06:26,068 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
332
+ 2024-09-17 12:06:26,068 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
333
+ 2024-09-17 12:06:26,069 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
334
+ 2024-09-17 12:06:26,070 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
335
+ 2024-09-17 12:06:26,071 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
336
+ 2024-09-17 12:06:26,074 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
337
+ 2024-09-17 12:06:26,076 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
338
+ 2024-09-17 12:06:26,079 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
339
+ 2024-09-17 12:06:26,080 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
340
+ 2024-09-17 12:06:26,080 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
341
+ 2024-09-17 12:06:26,081 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
342
+ 2024-09-17 12:06:26,083 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
343
+ 2024-09-17 12:06:26,084 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
344
+ 2024-09-17 12:06:26,084 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
345
+ 2024-09-17 12:06:26,087 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
346
+ 2024-09-17 12:06:26,088 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
347
+ 2024-09-17 12:06:26,145 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
348
+ 2024-09-17 12:06:26,145 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
349
+ 2024-09-17 12:06:26,145 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
350
+ 2024-09-17 12:06:26,146 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
351
+ 2024-09-17 12:06:26,922 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
352
+ 2024-09-17 12:06:26,923 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
353
+ 2024-09-17 12:06:28,136 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
354
+ 2024-09-17 12:06:31,944 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
355
+ 2024-09-17 12:06:33,956 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
356
+ 2024-09-17 12:06:34,983 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: resume
357
+ 2024-09-17 12:06:34,984 INFO HandlerThread:1253420 [handler.py:handle_request_resume():719] starting system metrics thread
358
+ 2024-09-17 12:06:34,984 INFO HandlerThread:1253420 [system_monitor.py:start():194] Starting system monitor
359
+ 2024-09-17 12:06:34,984 INFO SystemMonitor:1253420 [system_monitor.py:_start():158] Starting system asset monitoring threads
360
+ 2024-09-17 12:06:34,985 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started cpu monitoring
361
+ 2024-09-17 12:06:34,989 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started disk monitoring
362
+ 2024-09-17 12:06:34,991 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started gpu monitoring
363
+ 2024-09-17 12:06:34,993 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started memory monitoring
364
+ 2024-09-17 12:06:34,995 INFO SystemMonitor:1253420 [interfaces.py:start():188] Started network monitoring
365
+ 2024-09-17 12:06:35,298 INFO Thread-12 :1253420 [dir_watcher.py:_on_file_created():271] file/dir created: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/output.log
366
+ 2024-09-17 12:06:37,301 INFO Thread-12 :1253420 [dir_watcher.py:_on_file_modified():288] file/dir modified: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/output.log
367
+ 2024-09-17 12:06:39,145 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
368
+ 2024-09-17 12:06:41,651 DEBUG SenderThread:1253420 [sender.py:send():391] send: config
369
+ 2024-09-17 12:06:41,652 DEBUG SenderThread:1253420 [sender.py:send():391] send: telemetry
370
+ 2024-09-17 12:06:41,652 DEBUG SenderThread:1253420 [sender.py:send():391] send: metric
371
+ 2024-09-17 12:06:41,652 DEBUG SenderThread:1253420 [sender.py:send():391] send: telemetry
372
+ 2024-09-17 12:06:41,652 DEBUG SenderThread:1253420 [sender.py:send():391] send: metric
373
+ 2024-09-17 12:06:41,652 WARNING SenderThread:1253420 [sender.py:send_metric():1417] Seen metric with glob (shouldn't happen)
374
+ 2024-09-17 12:06:41,654 DEBUG SenderThread:1253420 [sender.py:send():391] send: telemetry
375
+ 2024-09-17 12:06:41,654 DEBUG SenderThread:1253420 [sender.py:send():391] send: telemetry
376
+ 2024-09-17 12:06:41,654 DEBUG SenderThread:1253420 [sender.py:send():391] send: config
377
+ 2024-09-17 12:06:41,923 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
378
+ 2024-09-17 12:06:41,924 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
379
+ 2024-09-17 12:06:41,944 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
380
+ 2024-09-17 12:06:42,726 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: pause
381
+ 2024-09-17 12:06:42,726 INFO HandlerThread:1253420 [handler.py:handle_request_pause():728] stopping system metrics thread
382
+ 2024-09-17 12:06:42,726 INFO HandlerThread:1253420 [system_monitor.py:finish():203] Stopping system monitor
383
+ 2024-09-17 12:06:42,727 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined cpu monitor
384
+ 2024-09-17 12:06:42,728 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined disk monitor
385
+ 2024-09-17 12:06:42,729 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():172] Starting system metrics aggregation loop
386
+ 2024-09-17 12:06:42,729 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():179] Finished system metrics aggregation loop
387
+ 2024-09-17 12:06:42,729 DEBUG SystemMonitor:1253420 [system_monitor.py:_start():183] Publishing last batch of metrics
388
+ 2024-09-17 12:06:42,791 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined gpu monitor
389
+ 2024-09-17 12:06:42,791 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined memory monitor
390
+ 2024-09-17 12:06:42,792 INFO HandlerThread:1253420 [interfaces.py:finish():200] Joined network monitor
391
+ 2024-09-17 12:06:42,793 DEBUG SenderThread:1253420 [sender.py:send():391] send: stats
392
+ 2024-09-17 12:06:43,307 INFO Thread-12 :1253420 [dir_watcher.py:_on_file_modified():288] file/dir modified: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/output.log
393
+ 2024-09-17 12:06:44,794 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
394
+ 2024-09-17 12:06:49,996 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
395
+ 2024-09-17 12:06:51,944 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
396
+ 2024-09-17 12:06:56,006 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
397
+ 2024-09-17 12:06:56,323 INFO Thread-12 :1253420 [dir_watcher.py:_on_file_modified():288] file/dir modified: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/config.yaml
398
+ 2024-09-17 12:06:56,923 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
399
+ 2024-09-17 12:06:56,924 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
400
+ 2024-09-17 12:07:01,944 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
401
+ 2024-09-17 12:07:01,996 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
402
+ 2024-09-17 12:07:06,996 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
403
+ 2024-09-17 12:07:11,923 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
404
+ 2024-09-17 12:07:11,924 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
405
+ 2024-09-17 12:07:11,964 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
406
+ 2024-09-17 12:07:12,094 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
407
+ 2024-09-17 12:07:17,997 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
408
+ 2024-09-17 12:07:21,945 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
409
+ 2024-09-17 12:07:22,997 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
410
+ 2024-09-17 12:07:26,923 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
411
+ 2024-09-17 12:07:26,924 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
412
+ 2024-09-17 12:07:28,087 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
413
+ 2024-09-17 12:07:31,945 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
414
+ 2024-09-17 12:07:33,997 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
415
+ 2024-09-17 12:07:38,997 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
416
+ 2024-09-17 12:07:41,923 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
417
+ 2024-09-17 12:07:41,924 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
418
+ 2024-09-17 12:07:41,964 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
419
+ 2024-09-17 12:07:44,112 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
420
+ 2024-09-17 12:07:49,998 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
421
+ 2024-09-17 12:07:51,945 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
422
+ 2024-09-17 12:07:54,998 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
423
+ 2024-09-17 12:07:56,923 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
424
+ 2024-09-17 12:07:56,924 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
425
+ 2024-09-17 12:08:00,093 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
426
+ 2024-09-17 12:08:01,945 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
427
+ 2024-09-17 12:08:05,999 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
428
+ 2024-09-17 12:08:10,999 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
429
+ 2024-09-17 12:08:11,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
430
+ 2024-09-17 12:08:11,924 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
431
+ 2024-09-17 12:08:11,964 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
432
+ 2024-09-17 12:08:16,090 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
433
+ 2024-09-17 12:08:21,946 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
434
+ 2024-09-17 12:08:21,999 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
435
+ 2024-09-17 12:08:26,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
436
+ 2024-09-17 12:08:26,925 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
437
+ 2024-09-17 12:08:27,086 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
438
+ 2024-09-17 12:08:31,946 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
439
+ 2024-09-17 12:08:32,999 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
440
+ 2024-09-17 12:08:38,000 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
441
+ 2024-09-17 12:08:41,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
442
+ 2024-09-17 12:08:41,925 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
443
+ 2024-09-17 12:08:41,965 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
444
+ 2024-09-17 12:08:43,076 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
445
+ 2024-09-17 12:08:49,000 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
446
+ 2024-09-17 12:08:51,946 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
447
+ 2024-09-17 12:08:54,000 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
448
+ 2024-09-17 12:08:56,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
449
+ 2024-09-17 12:08:56,925 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
450
+ 2024-09-17 12:08:59,063 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
451
+ 2024-09-17 12:09:01,946 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
452
+ 2024-09-17 12:09:05,001 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
453
+ 2024-09-17 12:09:10,001 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
454
+ 2024-09-17 12:09:11,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
455
+ 2024-09-17 12:09:11,925 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
456
+ 2024-09-17 12:09:11,965 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
457
+ 2024-09-17 12:09:15,094 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
458
+ 2024-09-17 12:09:21,001 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
459
+ 2024-09-17 12:09:21,947 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
460
+ 2024-09-17 12:09:26,001 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
461
+ 2024-09-17 12:09:26,924 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
462
+ 2024-09-17 12:09:26,925 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
463
+ 2024-09-17 12:09:31,084 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
464
+ 2024-09-17 12:09:31,947 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
465
+ 2024-09-17 12:09:37,001 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
466
+ 2024-09-17 12:09:41,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
467
+ 2024-09-17 12:09:41,925 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
468
+ 2024-09-17 12:09:41,965 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
469
+ 2024-09-17 12:09:42,133 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
470
+ 2024-09-17 12:09:48,001 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
471
+ 2024-09-17 12:09:51,947 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
472
+ 2024-09-17 12:09:53,002 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
473
+ 2024-09-17 12:09:56,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
474
+ 2024-09-17 12:09:56,926 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
475
+ 2024-09-17 12:09:58,140 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
476
+ 2024-09-17 12:10:01,947 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
477
+ 2024-09-17 12:10:04,002 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
478
+ 2024-09-17 12:10:09,003 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
479
+ 2024-09-17 12:10:11,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
480
+ 2024-09-17 12:10:11,926 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
481
+ 2024-09-17 12:10:11,966 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
482
+ 2024-09-17 12:10:14,095 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
483
+ 2024-09-17 12:10:20,003 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
484
+ 2024-09-17 12:10:21,948 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
485
+ 2024-09-17 12:10:25,003 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
486
+ 2024-09-17 12:10:26,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
487
+ 2024-09-17 12:10:26,926 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
488
+ 2024-09-17 12:10:30,096 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
489
+ 2024-09-17 12:10:31,948 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
490
+ 2024-09-17 12:10:36,003 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
491
+ 2024-09-17 12:10:41,004 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
492
+ 2024-09-17 12:10:41,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
493
+ 2024-09-17 12:10:41,926 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
494
+ 2024-09-17 12:10:41,966 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
495
+ 2024-09-17 12:10:46,112 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
496
+ 2024-09-17 12:10:51,948 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
497
+ 2024-09-17 12:10:52,003 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
498
+ 2024-09-17 12:10:56,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
499
+ 2024-09-17 12:10:56,926 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
500
+ 2024-09-17 12:10:57,143 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
501
+ 2024-09-17 12:11:01,948 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
502
+ 2024-09-17 12:11:03,004 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
503
+ 2024-09-17 12:11:08,004 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
504
+ 2024-09-17 12:11:11,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
505
+ 2024-09-17 12:11:11,926 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
506
+ 2024-09-17 12:11:11,966 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
507
+ 2024-09-17 12:11:13,079 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
508
+ 2024-09-17 12:11:19,005 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
509
+ 2024-09-17 12:11:21,948 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
510
+ 2024-09-17 12:11:24,005 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
511
+ 2024-09-17 12:11:26,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
512
+ 2024-09-17 12:11:26,927 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
513
+ 2024-09-17 12:11:29,082 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
514
+ 2024-09-17 12:11:31,949 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
515
+ 2024-09-17 12:11:35,005 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
516
+ 2024-09-17 12:11:40,005 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
517
+ 2024-09-17 12:11:41,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
518
+ 2024-09-17 12:11:41,927 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
519
+ 2024-09-17 12:11:41,967 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
520
+ 2024-09-17 12:11:45,065 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
521
+ 2024-09-17 12:11:51,006 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
522
+ 2024-09-17 12:11:51,949 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
523
+ 2024-09-17 12:11:56,006 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
524
+ 2024-09-17 12:11:56,926 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: stop_status
525
+ 2024-09-17 12:11:56,927 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: stop_status
526
+ 2024-09-17 12:11:59,448 DEBUG SenderThread:1253420 [sender.py:send():391] send: exit
527
+ 2024-09-17 12:11:59,449 INFO SenderThread:1253420 [sender.py:send_exit():598] handling exit code: 0
528
+ 2024-09-17 12:11:59,449 INFO SenderThread:1253420 [sender.py:send_exit():600] handling runtime: 19
529
+ 2024-09-17 12:11:59,454 INFO SenderThread:1253420 [sender.py:_save_file():1466] saving file wandb-summary.json with policy end
530
+ 2024-09-17 12:11:59,455 INFO SenderThread:1253420 [sender.py:send_exit():606] send defer
531
+ 2024-09-17 12:11:59,455 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
532
+ 2024-09-17 12:11:59,455 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 0
533
+ 2024-09-17 12:11:59,456 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
534
+ 2024-09-17 12:11:59,456 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 0
535
+ 2024-09-17 12:11:59,456 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 1
536
+ 2024-09-17 12:11:59,457 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
537
+ 2024-09-17 12:11:59,457 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 1
538
+ 2024-09-17 12:11:59,457 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
539
+ 2024-09-17 12:11:59,457 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 1
540
+ 2024-09-17 12:11:59,457 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 2
541
+ 2024-09-17 12:11:59,458 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
542
+ 2024-09-17 12:11:59,458 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 2
543
+ 2024-09-17 12:11:59,459 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
544
+ 2024-09-17 12:11:59,459 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 2
545
+ 2024-09-17 12:11:59,459 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 3
546
+ 2024-09-17 12:11:59,459 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
547
+ 2024-09-17 12:11:59,459 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 3
548
+ 2024-09-17 12:11:59,460 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
549
+ 2024-09-17 12:11:59,460 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 3
550
+ 2024-09-17 12:11:59,460 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 4
551
+ 2024-09-17 12:11:59,460 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
552
+ 2024-09-17 12:11:59,461 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 4
553
+ 2024-09-17 12:11:59,461 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
554
+ 2024-09-17 12:11:59,461 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 4
555
+ 2024-09-17 12:11:59,461 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 5
556
+ 2024-09-17 12:11:59,462 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
557
+ 2024-09-17 12:11:59,462 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 5
558
+ 2024-09-17 12:11:59,463 DEBUG SenderThread:1253420 [sender.py:send():391] send: summary
559
+ 2024-09-17 12:11:59,466 INFO SenderThread:1253420 [sender.py:_save_file():1466] saving file wandb-summary.json with policy end
560
+ 2024-09-17 12:11:59,467 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
561
+ 2024-09-17 12:11:59,467 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 5
562
+ 2024-09-17 12:11:59,467 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 6
563
+ 2024-09-17 12:11:59,467 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
564
+ 2024-09-17 12:11:59,468 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 6
565
+ 2024-09-17 12:11:59,468 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
566
+ 2024-09-17 12:11:59,468 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 6
567
+ 2024-09-17 12:11:59,468 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 7
568
+ 2024-09-17 12:11:59,469 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: status_report
569
+ 2024-09-17 12:11:59,469 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
570
+ 2024-09-17 12:11:59,469 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 7
571
+ 2024-09-17 12:11:59,470 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
572
+ 2024-09-17 12:11:59,470 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 7
573
+ 2024-09-17 12:11:59,690 INFO Thread-12 :1253420 [dir_watcher.py:_on_file_created():271] file/dir created: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/wandb-summary.json
574
+ 2024-09-17 12:12:00,448 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: poll_exit
575
+ 2024-09-17 12:12:01,273 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 8
576
+ 2024-09-17 12:12:01,274 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: poll_exit
577
+ 2024-09-17 12:12:01,274 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
578
+ 2024-09-17 12:12:01,275 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 8
579
+ 2024-09-17 12:12:01,275 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
580
+ 2024-09-17 12:12:01,275 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 8
581
+ 2024-09-17 12:12:01,276 INFO SenderThread:1253420 [job_builder.py:build():443] Attempting to build job artifact
582
+ 2024-09-17 12:12:01,280 INFO SenderThread:1253420 [job_builder.py:_get_source_type():583] no source found
583
+ 2024-09-17 12:12:01,280 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 9
584
+ 2024-09-17 12:12:01,280 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
585
+ 2024-09-17 12:12:01,281 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 9
586
+ 2024-09-17 12:12:01,281 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
587
+ 2024-09-17 12:12:01,281 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 9
588
+ 2024-09-17 12:12:01,282 INFO SenderThread:1253420 [dir_watcher.py:finish():358] shutting down directory watcher
589
+ 2024-09-17 12:12:01,451 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: poll_exit
590
+ 2024-09-17 12:12:01,692 INFO SenderThread:1253420 [dir_watcher.py:_on_file_modified():288] file/dir modified: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/output.log
591
+ 2024-09-17 12:12:01,693 INFO SenderThread:1253420 [dir_watcher.py:finish():388] scan: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files
592
+ 2024-09-17 12:12:01,694 INFO SenderThread:1253420 [dir_watcher.py:finish():402] scan save: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/config.yaml config.yaml
593
+ 2024-09-17 12:12:01,694 INFO SenderThread:1253420 [dir_watcher.py:finish():402] scan save: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/wandb-metadata.json wandb-metadata.json
594
+ 2024-09-17 12:12:01,695 INFO SenderThread:1253420 [dir_watcher.py:finish():402] scan save: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/requirements.txt requirements.txt
595
+ 2024-09-17 12:12:01,695 INFO SenderThread:1253420 [dir_watcher.py:finish():402] scan save: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/output.log output.log
596
+ 2024-09-17 12:12:01,695 INFO SenderThread:1253420 [dir_watcher.py:finish():402] scan save: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/wandb-summary.json wandb-summary.json
597
+ 2024-09-17 12:12:01,695 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 10
598
+ 2024-09-17 12:12:01,696 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: poll_exit
599
+ 2024-09-17 12:12:01,697 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
600
+ 2024-09-17 12:12:01,707 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 10
601
+ 2024-09-17 12:12:01,718 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
602
+ 2024-09-17 12:12:01,718 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 10
603
+ 2024-09-17 12:12:01,719 INFO SenderThread:1253420 [file_pusher.py:finish():169] shutting down file pusher
604
+ 2024-09-17 12:12:02,063 INFO wandb-upload_0:1253420 [upload_job.py:push():130] Uploaded file /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/config.yaml
605
+ 2024-09-17 12:12:02,089 INFO wandb-upload_3:1253420 [upload_job.py:push():130] Uploaded file /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/wandb-summary.json
606
+ 2024-09-17 12:12:02,094 INFO wandb-upload_1:1253420 [upload_job.py:push():130] Uploaded file /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/requirements.txt
607
+ 2024-09-17 12:12:02,107 INFO wandb-upload_2:1253420 [upload_job.py:push():130] Uploaded file /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/files/output.log
608
+ 2024-09-17 12:12:02,308 INFO Thread-11 (_thread_body):1253420 [sender.py:transition_state():626] send defer: 11
609
+ 2024-09-17 12:12:02,309 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
610
+ 2024-09-17 12:12:02,309 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 11
611
+ 2024-09-17 12:12:02,310 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
612
+ 2024-09-17 12:12:02,311 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 11
613
+ 2024-09-17 12:12:02,311 INFO SenderThread:1253420 [file_pusher.py:join():175] waiting for file pusher
614
+ 2024-09-17 12:12:02,311 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 12
615
+ 2024-09-17 12:12:02,311 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
616
+ 2024-09-17 12:12:02,312 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 12
617
+ 2024-09-17 12:12:02,312 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
618
+ 2024-09-17 12:12:02,313 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 12
619
+ 2024-09-17 12:12:02,313 INFO SenderThread:1253420 [file_stream.py:finish():601] file stream finish called
620
+ 2024-09-17 12:12:02,452 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: poll_exit
621
+ 2024-09-17 12:12:02,801 INFO SenderThread:1253420 [file_stream.py:finish():605] file stream finish is done
622
+ 2024-09-17 12:12:02,801 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 13
623
+ 2024-09-17 12:12:02,801 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: poll_exit
624
+ 2024-09-17 12:12:02,802 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
625
+ 2024-09-17 12:12:02,802 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 13
626
+ 2024-09-17 12:12:02,803 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
627
+ 2024-09-17 12:12:02,803 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 13
628
+ 2024-09-17 12:12:02,803 INFO SenderThread:1253420 [sender.py:transition_state():626] send defer: 14
629
+ 2024-09-17 12:12:02,804 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: defer
630
+ 2024-09-17 12:12:02,804 INFO HandlerThread:1253420 [handler.py:handle_request_defer():184] handle defer: 14
631
+ 2024-09-17 12:12:02,805 DEBUG SenderThread:1253420 [sender.py:send():391] send: final
632
+ 2024-09-17 12:12:02,805 DEBUG SenderThread:1253420 [sender.py:send():391] send: footer
633
+ 2024-09-17 12:12:02,805 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: defer
634
+ 2024-09-17 12:12:02,805 INFO SenderThread:1253420 [sender.py:send_request_defer():622] handle sender defer: 14
635
+ 2024-09-17 12:12:02,807 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: poll_exit
636
+ 2024-09-17 12:12:02,807 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: server_info
637
+ 2024-09-17 12:12:02,808 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: poll_exit
638
+ 2024-09-17 12:12:02,809 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: server_info
639
+ 2024-09-17 12:12:02,814 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: poll_exit
640
+ 2024-09-17 12:12:02,815 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: get_summary
641
+ 2024-09-17 12:12:02,816 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: sampled_history
642
+ 2024-09-17 12:12:02,817 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: internal_messages
643
+ 2024-09-17 12:12:02,921 DEBUG SenderThread:1253420 [sender.py:send_request():418] send_request: poll_exit
644
+ 2024-09-17 12:12:02,922 INFO MainThread:1253420 [wandb_run.py:_footer_history_summary_info():4088] rendering history
645
+ 2024-09-17 12:12:02,922 INFO MainThread:1253420 [wandb_run.py:_footer_history_summary_info():4120] rendering summary
646
+ 2024-09-17 12:12:02,922 INFO MainThread:1253420 [wandb_run.py:_footer_sync_info():4047] logging synced files
647
+ 2024-09-17 12:12:02,925 DEBUG HandlerThread:1253420 [handler.py:handle_request():158] handle_request: shutdown
648
+ 2024-09-17 12:12:02,925 INFO HandlerThread:1253420 [handler.py:finish():884] shutting down handler
649
+ 2024-09-17 12:12:03,797 WARNING StreamThr :1253420 [internal.py:is_dead():413] Internal process exiting, parent pid 1253167 disappeared
650
+ 2024-09-17 12:12:03,798 ERROR StreamThr :1253420 [internal.py:wandb_internal():151] Internal process shutdown.
651
+ 2024-09-17 12:12:03,815 INFO WriterThread:1253420 [datastore.py:close():296] close: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/run-nm6gjq7r.wandb
652
+ 2024-09-17 12:12:03,921 INFO SenderThread:1253420 [sender.py:finish():1631] shutting down sender
653
+ 2024-09-17 12:12:03,922 INFO SenderThread:1253420 [file_pusher.py:finish():169] shutting down file pusher
654
+ 2024-09-17 12:12:03,922 INFO SenderThread:1253420 [file_pusher.py:join():175] waiting for file pusher
wandb/run-20240917_120410-nm6gjq7r/logs/debug.log ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_setup.py:_flush():77] Current SDK version is 0.17.9
2
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_setup.py:_flush():77] Configure stats pid to 1253167
3
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_setup.py:_flush():77] Loading settings from /homes/amrufai/.config/wandb/settings
4
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_setup.py:_flush():77] Loading settings from /nfs/production/literature/amina-mardiyyah/wandb/settings
5
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_setup.py:_flush():77] Loading settings from environment variables: {}
6
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_setup.py:_flush():77] Inferring run settings from compute environment: {'program': '<python with no main file>'}
7
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_setup.py:_flush():77] Applying login settings: {}
8
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_setup.py:_flush():77] Applying login settings: {}
9
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_init.py:_log_setup():524] Logging user logs to /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/logs/debug.log
10
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_init.py:_log_setup():525] Logging internal logs to /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_120410-nm6gjq7r/logs/debug-internal.log
11
+ 2024-09-17 12:04:10,714 INFO MainThread:1253167 [wandb_init.py:_jupyter_setup():470] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7fa9acf55510>
12
+ 2024-09-17 12:04:10,715 INFO MainThread:1253167 [wandb_init.py:init():608] calling init triggers
13
+ 2024-09-17 12:04:10,715 INFO MainThread:1253167 [wandb_init.py:init():615] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-09-17 12:04:10,715 INFO MainThread:1253167 [wandb_init.py:init():658] starting backend
16
+ 2024-09-17 12:04:10,715 INFO MainThread:1253167 [wandb_init.py:init():662] setting up manager
17
+ 2024-09-17 12:04:10,717 INFO MainThread:1253167 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-09-17 12:04:10,722 INFO MainThread:1253167 [wandb_init.py:init():670] backend started and connected
19
+ 2024-09-17 12:04:10,729 INFO MainThread:1253167 [wandb_run.py:_label_probe_notebook():1344] probe notebook
20
+ 2024-09-17 12:04:10,739 INFO MainThread:1253167 [wandb_run.py:_label_probe_notebook():1354] Unable to probe notebook: 'NoneType' object has no attribute 'get'
21
+ 2024-09-17 12:04:10,739 INFO MainThread:1253167 [wandb_init.py:init():768] updated telemetry
22
+ 2024-09-17 12:04:10,746 INFO MainThread:1253167 [wandb_init.py:init():801] communicating run to backend with 90.0 second timeout
23
+ 2024-09-17 12:04:11,130 INFO MainThread:1253167 [wandb_init.py:init():852] starting run threads in backend
24
+ 2024-09-17 12:04:11,922 INFO MainThread:1253167 [wandb_run.py:_console_start():2465] atexit reg
25
+ 2024-09-17 12:04:11,922 INFO MainThread:1253167 [wandb_run.py:_redirect():2311] redirect: wrap_raw
26
+ 2024-09-17 12:04:11,922 INFO MainThread:1253167 [wandb_run.py:_redirect():2376] Wrapping output streams.
27
+ 2024-09-17 12:04:11,922 INFO MainThread:1253167 [wandb_run.py:_redirect():2401] Redirects installed.
28
+ 2024-09-17 12:04:11,930 INFO MainThread:1253167 [wandb_init.py:init():895] run started, returning control to user process
29
+ 2024-09-17 12:04:11,937 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
30
+ 2024-09-17 12:04:11,937 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
31
+ 2024-09-17 12:04:47,840 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
32
+ 2024-09-17 12:04:47,892 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
33
+ 2024-09-17 12:04:47,893 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
34
+ 2024-09-17 12:04:48,686 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
35
+ 2024-09-17 12:04:48,692 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
36
+ 2024-09-17 12:04:48,692 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
37
+ 2024-09-17 12:05:07,800 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
38
+ 2024-09-17 12:05:07,832 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
39
+ 2024-09-17 12:05:07,832 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
40
+ 2024-09-17 12:05:50,656 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
41
+ 2024-09-17 12:05:50,659 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
42
+ 2024-09-17 12:05:50,659 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
43
+ 2024-09-17 12:05:53,406 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
44
+ 2024-09-17 12:05:53,414 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
45
+ 2024-09-17 12:05:53,414 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
46
+ 2024-09-17 12:05:54,528 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
47
+ 2024-09-17 12:05:54,701 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
48
+ 2024-09-17 12:05:54,701 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
49
+ 2024-09-17 12:06:04,941 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
50
+ 2024-09-17 12:06:05,237 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
51
+ 2024-09-17 12:06:05,237 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
52
+ 2024-09-17 12:06:13,487 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
53
+ 2024-09-17 12:06:13,491 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
54
+ 2024-09-17 12:06:13,491 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
55
+ 2024-09-17 12:06:14,616 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
56
+ 2024-09-17 12:06:14,622 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
57
+ 2024-09-17 12:06:14,622 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
58
+ 2024-09-17 12:06:14,930 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
59
+ 2024-09-17 12:06:24,010 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
60
+ 2024-09-17 12:06:24,010 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
61
+ 2024-09-17 12:06:25,313 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
62
+ 2024-09-17 12:06:25,321 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
63
+ 2024-09-17 12:06:25,321 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
64
+ 2024-09-17 12:06:26,067 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
65
+ 2024-09-17 12:06:26,074 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
66
+ 2024-09-17 12:06:26,075 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
67
+ 2024-09-17 12:06:34,983 INFO MainThread:1253167 [wandb_init.py:_resume_backend():440] resuming backend
68
+ 2024-09-17 12:06:41,613 INFO MainThread:1253167 [wandb_run.py:_config_callback():1392] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['XLMRobertaForTokenClassification'], 'finetuning_task': None, 'id2label': {0: 'O', 1: 'B-PER', 2: 'I-PER', 3: 'B-ORG', 4: 'I-ORG', 5: 'B-LOC', 6: 'I-LOC'}, 'label2id': {'O': 0, 'B-PER': 1, 'I-PER': 2, 'B-ORG': 3, 'I-ORG': 4, 'B-LOC': 5, 'I-LOC': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0', 'transformers_version': '4.43.4', 'adapters': {'adapters': {}, 'config_map': {}, 'fusion_config_map': {}, 'fusions': {}}, 'gradient_checkpointing': False, 'model_type': 'xlm-roberta', 'output_past': True, 'vocab_size': 250002, 'hidden_size': 1024, 'num_hidden_layers': 24, 'num_attention_heads': 16, 'hidden_act': 'gelu', 'intermediate_size': 4096, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 514, 'type_vocab_size': 1, 'initializer_range': 0.02, 'layer_norm_eps': 1e-05, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 2e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/logs', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 3407, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'eval_f1', 'greater_is_better': True, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'eval_use_gather_object': False}
69
+ 2024-09-17 12:06:41,616 INFO MainThread:1253167 [wandb_config.py:__setitem__():154] config set model/num_parameters = 558848007 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7fa9b1668e50>>
70
+ 2024-09-17 12:06:41,616 INFO MainThread:1253167 [wandb_run.py:_config_callback():1392] config_cb model/num_parameters 558848007 None
71
+ 2024-09-17 12:06:42,725 INFO MainThread:1253167 [jupyter.py:save_ipynb():372] not saving jupyter notebook
72
+ 2024-09-17 12:06:42,725 INFO MainThread:1253167 [wandb_init.py:_pause_backend():435] pausing backend
wandb/run-20240917_120410-nm6gjq7r/run-nm6gjq7r.wandb ADDED
Binary file (35.5 kB). View file
 
wandb/run-20240917_124325-7dvmn2ta/files/config.yaml ADDED
@@ -0,0 +1,741 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.11.4
7
+ cli_version: 0.17.9
8
+ framework: huggingface
9
+ huggingface_version: 4.43.4
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: false
12
+ start_time: 1726573405
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 5
17
+ - 11
18
+ - 41
19
+ - 49
20
+ - 51
21
+ - 53
22
+ - 55
23
+ - 71
24
+ - 98
25
+ - 100
26
+ - 105
27
+ 2:
28
+ - 1
29
+ - 5
30
+ - 11
31
+ - 41
32
+ - 49
33
+ - 51
34
+ - 53
35
+ - 55
36
+ - 71
37
+ - 98
38
+ - 100
39
+ - 105
40
+ 3:
41
+ - 7
42
+ - 19
43
+ - 23
44
+ - 66
45
+ 4: 3.11.4
46
+ 5: 0.17.9
47
+ 6: 4.43.4
48
+ 8:
49
+ - 1
50
+ - 5
51
+ 9:
52
+ 1: transformers_trainer
53
+ 13: linux-x86_64
54
+ m:
55
+ - 1: train/global_step
56
+ 6:
57
+ - 3
58
+ - 1: train/loss
59
+ 5: 1
60
+ 6:
61
+ - 1
62
+ - 1: train/grad_norm
63
+ 5: 1
64
+ 6:
65
+ - 1
66
+ - 1: train/learning_rate
67
+ 5: 1
68
+ 6:
69
+ - 1
70
+ - 1: train/epoch
71
+ 5: 1
72
+ 6:
73
+ - 1
74
+ - 1: eval/loss
75
+ 5: 1
76
+ 6:
77
+ - 1
78
+ - 1: eval/f1
79
+ 5: 1
80
+ 6:
81
+ - 1
82
+ - 1: eval/precision
83
+ 5: 1
84
+ 6:
85
+ - 1
86
+ - 1: eval/recall
87
+ 5: 1
88
+ 6:
89
+ - 1
90
+ - 1: eval/accuracy
91
+ 5: 1
92
+ 6:
93
+ - 1
94
+ - 1: eval/runtime
95
+ 5: 1
96
+ 6:
97
+ - 1
98
+ - 1: eval/samples_per_second
99
+ 5: 1
100
+ 6:
101
+ - 1
102
+ - 1: eval/steps_per_second
103
+ 5: 1
104
+ 6:
105
+ - 1
106
+ return_dict:
107
+ desc: null
108
+ value: true
109
+ output_hidden_states:
110
+ desc: null
111
+ value: false
112
+ output_attentions:
113
+ desc: null
114
+ value: false
115
+ torchscript:
116
+ desc: null
117
+ value: false
118
+ torch_dtype:
119
+ desc: null
120
+ value: float32
121
+ use_bfloat16:
122
+ desc: null
123
+ value: false
124
+ tf_legacy_loss:
125
+ desc: null
126
+ value: false
127
+ pruned_heads:
128
+ desc: null
129
+ value: {}
130
+ tie_word_embeddings:
131
+ desc: null
132
+ value: true
133
+ chunk_size_feed_forward:
134
+ desc: null
135
+ value: 0
136
+ is_encoder_decoder:
137
+ desc: null
138
+ value: false
139
+ is_decoder:
140
+ desc: null
141
+ value: false
142
+ cross_attention_hidden_size:
143
+ desc: null
144
+ value: null
145
+ add_cross_attention:
146
+ desc: null
147
+ value: false
148
+ tie_encoder_decoder:
149
+ desc: null
150
+ value: false
151
+ max_length:
152
+ desc: null
153
+ value: 20
154
+ min_length:
155
+ desc: null
156
+ value: 0
157
+ do_sample:
158
+ desc: null
159
+ value: false
160
+ early_stopping:
161
+ desc: null
162
+ value: false
163
+ num_beams:
164
+ desc: null
165
+ value: 1
166
+ num_beam_groups:
167
+ desc: null
168
+ value: 1
169
+ diversity_penalty:
170
+ desc: null
171
+ value: 0.0
172
+ temperature:
173
+ desc: null
174
+ value: 1.0
175
+ top_k:
176
+ desc: null
177
+ value: 50
178
+ top_p:
179
+ desc: null
180
+ value: 1.0
181
+ typical_p:
182
+ desc: null
183
+ value: 1.0
184
+ repetition_penalty:
185
+ desc: null
186
+ value: 1.0
187
+ length_penalty:
188
+ desc: null
189
+ value: 1.0
190
+ no_repeat_ngram_size:
191
+ desc: null
192
+ value: 0
193
+ encoder_no_repeat_ngram_size:
194
+ desc: null
195
+ value: 0
196
+ bad_words_ids:
197
+ desc: null
198
+ value: null
199
+ num_return_sequences:
200
+ desc: null
201
+ value: 1
202
+ output_scores:
203
+ desc: null
204
+ value: false
205
+ return_dict_in_generate:
206
+ desc: null
207
+ value: false
208
+ forced_bos_token_id:
209
+ desc: null
210
+ value: null
211
+ forced_eos_token_id:
212
+ desc: null
213
+ value: null
214
+ remove_invalid_values:
215
+ desc: null
216
+ value: false
217
+ exponential_decay_length_penalty:
218
+ desc: null
219
+ value: null
220
+ suppress_tokens:
221
+ desc: null
222
+ value: null
223
+ begin_suppress_tokens:
224
+ desc: null
225
+ value: null
226
+ architectures:
227
+ desc: null
228
+ value:
229
+ - XLMRobertaForTokenClassification
230
+ finetuning_task:
231
+ desc: null
232
+ value: null
233
+ id2label:
234
+ desc: null
235
+ value:
236
+ '0': O
237
+ '1': B-PER
238
+ '2': I-PER
239
+ '3': B-ORG
240
+ '4': I-ORG
241
+ '5': B-LOC
242
+ '6': I-LOC
243
+ label2id:
244
+ desc: null
245
+ value:
246
+ O: 0
247
+ B-PER: 1
248
+ I-PER: 2
249
+ B-ORG: 3
250
+ I-ORG: 4
251
+ B-LOC: 5
252
+ I-LOC: 6
253
+ tokenizer_class:
254
+ desc: null
255
+ value: null
256
+ prefix:
257
+ desc: null
258
+ value: null
259
+ bos_token_id:
260
+ desc: null
261
+ value: 0
262
+ pad_token_id:
263
+ desc: null
264
+ value: 1
265
+ eos_token_id:
266
+ desc: null
267
+ value: 2
268
+ sep_token_id:
269
+ desc: null
270
+ value: null
271
+ decoder_start_token_id:
272
+ desc: null
273
+ value: null
274
+ task_specific_params:
275
+ desc: null
276
+ value: null
277
+ problem_type:
278
+ desc: null
279
+ value: null
280
+ _name_or_path:
281
+ desc: null
282
+ value: masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0
283
+ transformers_version:
284
+ desc: null
285
+ value: 4.43.4
286
+ adapters:
287
+ desc: null
288
+ value:
289
+ adapters: {}
290
+ config_map: {}
291
+ fusion_config_map: {}
292
+ fusions: {}
293
+ gradient_checkpointing:
294
+ desc: null
295
+ value: false
296
+ model_type:
297
+ desc: null
298
+ value: xlm-roberta
299
+ output_past:
300
+ desc: null
301
+ value: true
302
+ vocab_size:
303
+ desc: null
304
+ value: 250002
305
+ hidden_size:
306
+ desc: null
307
+ value: 1024
308
+ num_hidden_layers:
309
+ desc: null
310
+ value: 24
311
+ num_attention_heads:
312
+ desc: null
313
+ value: 16
314
+ hidden_act:
315
+ desc: null
316
+ value: gelu
317
+ intermediate_size:
318
+ desc: null
319
+ value: 4096
320
+ hidden_dropout_prob:
321
+ desc: null
322
+ value: 0.1
323
+ attention_probs_dropout_prob:
324
+ desc: null
325
+ value: 0.1
326
+ max_position_embeddings:
327
+ desc: null
328
+ value: 514
329
+ type_vocab_size:
330
+ desc: null
331
+ value: 1
332
+ initializer_range:
333
+ desc: null
334
+ value: 0.02
335
+ layer_norm_eps:
336
+ desc: null
337
+ value: 1.0e-05
338
+ position_embedding_type:
339
+ desc: null
340
+ value: absolute
341
+ use_cache:
342
+ desc: null
343
+ value: true
344
+ classifier_dropout:
345
+ desc: null
346
+ value: null
347
+ output_dir:
348
+ desc: null
349
+ value: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date
350
+ overwrite_output_dir:
351
+ desc: null
352
+ value: false
353
+ do_train:
354
+ desc: null
355
+ value: false
356
+ do_eval:
357
+ desc: null
358
+ value: true
359
+ do_predict:
360
+ desc: null
361
+ value: false
362
+ eval_strategy:
363
+ desc: null
364
+ value: epoch
365
+ prediction_loss_only:
366
+ desc: null
367
+ value: false
368
+ per_device_train_batch_size:
369
+ desc: null
370
+ value: 16
371
+ per_device_eval_batch_size:
372
+ desc: null
373
+ value: 8
374
+ per_gpu_train_batch_size:
375
+ desc: null
376
+ value: null
377
+ per_gpu_eval_batch_size:
378
+ desc: null
379
+ value: null
380
+ gradient_accumulation_steps:
381
+ desc: null
382
+ value: 4
383
+ eval_accumulation_steps:
384
+ desc: null
385
+ value: null
386
+ eval_delay:
387
+ desc: null
388
+ value: 0
389
+ torch_empty_cache_steps:
390
+ desc: null
391
+ value: null
392
+ learning_rate:
393
+ desc: null
394
+ value: 2.0e-05
395
+ weight_decay:
396
+ desc: null
397
+ value: 0.001
398
+ adam_beta1:
399
+ desc: null
400
+ value: 0.9
401
+ adam_beta2:
402
+ desc: null
403
+ value: 0.999
404
+ adam_epsilon:
405
+ desc: null
406
+ value: 1.0e-08
407
+ max_grad_norm:
408
+ desc: null
409
+ value: 1.0
410
+ num_train_epochs:
411
+ desc: null
412
+ value: 10
413
+ max_steps:
414
+ desc: null
415
+ value: -1
416
+ lr_scheduler_type:
417
+ desc: null
418
+ value: linear
419
+ lr_scheduler_kwargs:
420
+ desc: null
421
+ value: {}
422
+ warmup_ratio:
423
+ desc: null
424
+ value: 0.0
425
+ warmup_steps:
426
+ desc: null
427
+ value: 1000
428
+ log_level:
429
+ desc: null
430
+ value: passive
431
+ log_level_replica:
432
+ desc: null
433
+ value: warning
434
+ log_on_each_node:
435
+ desc: null
436
+ value: true
437
+ logging_dir:
438
+ desc: null
439
+ value: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/logs
440
+ logging_strategy:
441
+ desc: null
442
+ value: epoch
443
+ logging_first_step:
444
+ desc: null
445
+ value: false
446
+ logging_steps:
447
+ desc: null
448
+ value: 500
449
+ logging_nan_inf_filter:
450
+ desc: null
451
+ value: true
452
+ save_strategy:
453
+ desc: null
454
+ value: epoch
455
+ save_steps:
456
+ desc: null
457
+ value: 500
458
+ save_total_limit:
459
+ desc: null
460
+ value: 2
461
+ save_safetensors:
462
+ desc: null
463
+ value: true
464
+ save_on_each_node:
465
+ desc: null
466
+ value: false
467
+ save_only_model:
468
+ desc: null
469
+ value: false
470
+ restore_callback_states_from_checkpoint:
471
+ desc: null
472
+ value: false
473
+ no_cuda:
474
+ desc: null
475
+ value: false
476
+ use_cpu:
477
+ desc: null
478
+ value: false
479
+ use_mps_device:
480
+ desc: null
481
+ value: false
482
+ seed:
483
+ desc: null
484
+ value: 3407
485
+ data_seed:
486
+ desc: null
487
+ value: null
488
+ jit_mode_eval:
489
+ desc: null
490
+ value: false
491
+ use_ipex:
492
+ desc: null
493
+ value: false
494
+ bf16:
495
+ desc: null
496
+ value: false
497
+ fp16:
498
+ desc: null
499
+ value: true
500
+ fp16_opt_level:
501
+ desc: null
502
+ value: O1
503
+ half_precision_backend:
504
+ desc: null
505
+ value: auto
506
+ bf16_full_eval:
507
+ desc: null
508
+ value: false
509
+ fp16_full_eval:
510
+ desc: null
511
+ value: false
512
+ tf32:
513
+ desc: null
514
+ value: null
515
+ local_rank:
516
+ desc: null
517
+ value: 0
518
+ ddp_backend:
519
+ desc: null
520
+ value: null
521
+ tpu_num_cores:
522
+ desc: null
523
+ value: null
524
+ tpu_metrics_debug:
525
+ desc: null
526
+ value: false
527
+ debug:
528
+ desc: null
529
+ value: []
530
+ dataloader_drop_last:
531
+ desc: null
532
+ value: false
533
+ eval_steps:
534
+ desc: null
535
+ value: null
536
+ dataloader_num_workers:
537
+ desc: null
538
+ value: 0
539
+ dataloader_prefetch_factor:
540
+ desc: null
541
+ value: null
542
+ past_index:
543
+ desc: null
544
+ value: -1
545
+ run_name:
546
+ desc: null
547
+ value: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date
548
+ disable_tqdm:
549
+ desc: null
550
+ value: false
551
+ remove_unused_columns:
552
+ desc: null
553
+ value: true
554
+ label_names:
555
+ desc: null
556
+ value: null
557
+ load_best_model_at_end:
558
+ desc: null
559
+ value: true
560
+ metric_for_best_model:
561
+ desc: null
562
+ value: eval_f1
563
+ greater_is_better:
564
+ desc: null
565
+ value: true
566
+ ignore_data_skip:
567
+ desc: null
568
+ value: false
569
+ fsdp:
570
+ desc: null
571
+ value: []
572
+ fsdp_min_num_params:
573
+ desc: null
574
+ value: 0
575
+ fsdp_config:
576
+ desc: null
577
+ value:
578
+ min_num_params: 0
579
+ xla: false
580
+ xla_fsdp_v2: false
581
+ xla_fsdp_grad_ckpt: false
582
+ fsdp_transformer_layer_cls_to_wrap:
583
+ desc: null
584
+ value: null
585
+ accelerator_config:
586
+ desc: null
587
+ value:
588
+ split_batches: false
589
+ dispatch_batches: null
590
+ even_batches: true
591
+ use_seedable_sampler: true
592
+ non_blocking: false
593
+ gradient_accumulation_kwargs: null
594
+ deepspeed:
595
+ desc: null
596
+ value: null
597
+ label_smoothing_factor:
598
+ desc: null
599
+ value: 0.0
600
+ optim:
601
+ desc: null
602
+ value: adamw_torch
603
+ optim_args:
604
+ desc: null
605
+ value: null
606
+ adafactor:
607
+ desc: null
608
+ value: false
609
+ group_by_length:
610
+ desc: null
611
+ value: true
612
+ length_column_name:
613
+ desc: null
614
+ value: length
615
+ report_to:
616
+ desc: null
617
+ value:
618
+ - wandb
619
+ ddp_find_unused_parameters:
620
+ desc: null
621
+ value: null
622
+ ddp_bucket_cap_mb:
623
+ desc: null
624
+ value: null
625
+ ddp_broadcast_buffers:
626
+ desc: null
627
+ value: null
628
+ dataloader_pin_memory:
629
+ desc: null
630
+ value: true
631
+ dataloader_persistent_workers:
632
+ desc: null
633
+ value: false
634
+ skip_memory_metrics:
635
+ desc: null
636
+ value: true
637
+ use_legacy_prediction_loop:
638
+ desc: null
639
+ value: false
640
+ push_to_hub:
641
+ desc: null
642
+ value: false
643
+ resume_from_checkpoint:
644
+ desc: null
645
+ value: null
646
+ hub_model_id:
647
+ desc: null
648
+ value: null
649
+ hub_strategy:
650
+ desc: null
651
+ value: every_save
652
+ hub_token:
653
+ desc: null
654
+ value: <HUB_TOKEN>
655
+ hub_private_repo:
656
+ desc: null
657
+ value: false
658
+ hub_always_push:
659
+ desc: null
660
+ value: false
661
+ gradient_checkpointing_kwargs:
662
+ desc: null
663
+ value: null
664
+ include_inputs_for_metrics:
665
+ desc: null
666
+ value: false
667
+ eval_do_concat_batches:
668
+ desc: null
669
+ value: true
670
+ fp16_backend:
671
+ desc: null
672
+ value: auto
673
+ evaluation_strategy:
674
+ desc: null
675
+ value: null
676
+ push_to_hub_model_id:
677
+ desc: null
678
+ value: null
679
+ push_to_hub_organization:
680
+ desc: null
681
+ value: null
682
+ push_to_hub_token:
683
+ desc: null
684
+ value: <PUSH_TO_HUB_TOKEN>
685
+ mp_parameters:
686
+ desc: null
687
+ value: ''
688
+ auto_find_batch_size:
689
+ desc: null
690
+ value: false
691
+ full_determinism:
692
+ desc: null
693
+ value: false
694
+ torchdynamo:
695
+ desc: null
696
+ value: null
697
+ ray_scope:
698
+ desc: null
699
+ value: last
700
+ ddp_timeout:
701
+ desc: null
702
+ value: 1800
703
+ torch_compile:
704
+ desc: null
705
+ value: false
706
+ torch_compile_backend:
707
+ desc: null
708
+ value: null
709
+ torch_compile_mode:
710
+ desc: null
711
+ value: null
712
+ dispatch_batches:
713
+ desc: null
714
+ value: null
715
+ split_batches:
716
+ desc: null
717
+ value: null
718
+ include_tokens_per_second:
719
+ desc: null
720
+ value: false
721
+ include_num_input_tokens_seen:
722
+ desc: null
723
+ value: false
724
+ neftune_noise_alpha:
725
+ desc: null
726
+ value: null
727
+ optim_target_modules:
728
+ desc: null
729
+ value: null
730
+ batch_eval_metrics:
731
+ desc: null
732
+ value: false
733
+ eval_on_start:
734
+ desc: null
735
+ value: false
736
+ eval_use_gather_object:
737
+ desc: null
738
+ value: false
739
+ model/num_parameters:
740
+ desc: null
741
+ value: 558848007
wandb/run-20240917_124325-7dvmn2ta/files/output.log ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /nfs/production/literature/amina-mardiyyah/envs/llm-prompt/lib/python3.11/site-packages/accelerate/accelerator.py:488: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead.
2
+ self.scaler = torch.cuda.amp.GradScaler(**kwargs)
3
+ Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
4
+ wandb: WARNING The `run_name` is currently set to the same value as `TrainingArguments.output_dir`. If this was not intended, please specify a different run name by setting the `TrainingArguments.run_name` parameter.
5
+ Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
6
+ Classification Report:
7
+ precision recall f1-score support
8
+ LOC 0.73 0.79 0.76 9327
9
+ ORG 0.74 0.68 0.71 7547
10
+ PER 0.88 0.88 0.88 11204
11
+ micro avg 0.79 0.80 0.80 28078
12
+ macro avg 0.79 0.78 0.78 28078
13
+ weighted avg 0.80 0.80 0.80 28078
wandb/run-20240917_124325-7dvmn2ta/files/requirements.txt ADDED
@@ -0,0 +1,313 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ AutoROM.accept-rom-license==0.6.1
2
+ AutoROM==0.6.1
3
+ Farama-Notifications==0.0.4
4
+ GitPython==3.1.43
5
+ Jinja2==3.1.4
6
+ Mako==1.3.5
7
+ Markdown==3.7
8
+ MarkupSafe==2.1.5
9
+ PyYAML==6.0.2
10
+ Pygments==2.18.0
11
+ QtPy==2.4.1
12
+ SQLAlchemy==2.0.32
13
+ Send2Trash==1.8.3
14
+ Shimmy==1.3.0
15
+ Werkzeug==3.0.4
16
+ absl-py==2.1.0
17
+ accelerate==0.33.0
18
+ aiohappyeyeballs==2.3.5
19
+ aiohttp-retry==2.8.3
20
+ aiohttp==3.10.3
21
+ aiosignal==1.3.1
22
+ ale-py==0.8.1
23
+ alembic==1.13.2
24
+ amqp==5.2.0
25
+ annotated-types==0.7.0
26
+ antlr4-python3-runtime==4.9.3
27
+ anyio==4.4.0
28
+ appdirs==1.4.4
29
+ argon2-cffi-bindings==21.2.0
30
+ argon2-cffi==23.1.0
31
+ arrow==1.3.0
32
+ asttokens==2.4.1
33
+ async-generator==1.10
34
+ async-lru==2.0.4
35
+ asyncssh==2.17.0
36
+ atpublic==5.0
37
+ attrs==24.2.0
38
+ audioread==3.0.1
39
+ babel==2.16.0
40
+ beautifulsoup4==4.12.3
41
+ billiard==4.2.0
42
+ bitsandbytes==0.43.3
43
+ bleach==6.1.0
44
+ blis==0.7.11
45
+ catalogue==2.0.10
46
+ celery==5.4.0
47
+ certifi==2024.7.4
48
+ certipy==0.1.3
49
+ cffi==1.17.0
50
+ charset-normalizer==3.3.2
51
+ click-didyoumean==0.3.1
52
+ click-plugins==1.1.1
53
+ click-repl==0.3.0
54
+ click==8.1.7
55
+ cloudpathlib==0.18.1
56
+ cloudpickle==3.0.0
57
+ colorama==0.4.6
58
+ coloredlogs==15.0.1
59
+ comm==0.2.2
60
+ conda-store==2024.6.1
61
+ confection==0.1.5
62
+ configobj==5.0.8
63
+ contourpy==1.2.1
64
+ cryptography==43.0.0
65
+ cuda-python==11.8.3
66
+ cycler==0.12.1
67
+ cymem==2.0.8
68
+ datasets==2.20.0
69
+ debugpy==1.8.5
70
+ decorator==5.1.1
71
+ defusedxml==0.7.1
72
+ dictdiffer==0.9.0
73
+ dill==0.3.8
74
+ diskcache==5.6.3
75
+ distro==1.9.0
76
+ docker-pycreds==0.4.0
77
+ docstring_parser==0.16
78
+ dpath==2.2.0
79
+ dulwich==0.22.1
80
+ dvc-data==3.16.5
81
+ dvc-http==2.32.0
82
+ dvc-objects==5.1.0
83
+ dvc-render==1.0.2
84
+ dvc-studio-client==0.21.0
85
+ dvc-task==0.4.0
86
+ dvc==3.55.2
87
+ en-core-web-lg==3.7.1
88
+ en-core-web-sm==3.7.1
89
+ entrypoints==0.4
90
+ evaluate==0.4.2
91
+ executing==2.0.1
92
+ fastjsonschema==2.20.0
93
+ filelock==3.15.4
94
+ flatbuffers==24.3.25
95
+ flatten-dict==0.4.2
96
+ flufl.lock==8.1.0
97
+ fonttools==4.53.1
98
+ fqdn==1.5.1
99
+ frozenlist==1.4.1
100
+ fsspec==2024.5.0
101
+ funcy==2.0
102
+ gitdb==4.0.11
103
+ grandalf==0.8
104
+ greenlet==3.0.3
105
+ grpcio==1.66.0
106
+ gto==1.7.1
107
+ gymnasium==0.29.1
108
+ h11==0.14.0
109
+ httpcore==1.0.5
110
+ httpx==0.27.0
111
+ huggingface-hub==0.24.5
112
+ humanfriendly==10.0
113
+ hydra-core==1.3.2
114
+ idna==3.7
115
+ importlib_resources==6.4.5
116
+ ipykernel==6.29.5
117
+ ipython==8.26.0
118
+ ipywidgets==8.1.3
119
+ isoduration==20.11.0
120
+ iterative-telemetry==0.0.8
121
+ jedi==0.19.1
122
+ joblib==1.4.2
123
+ json5==0.9.25
124
+ jsonpointer==3.0.0
125
+ jsonschema-specifications==2023.12.1
126
+ jsonschema==4.23.0
127
+ jupyter-console==6.6.3
128
+ jupyter-events==0.10.0
129
+ jupyter-launcher-shortcuts==4.0.3
130
+ jupyter-lsp==2.2.5
131
+ jupyter-telemetry==0.1.0
132
+ jupyter==1.0.0
133
+ jupyter_client==8.6.2
134
+ jupyter_core==5.7.2
135
+ jupyter_server==2.14.2
136
+ jupyter_server_terminals==0.5.3
137
+ jupyterhub==1.5.1
138
+ jupyterlab-conda-store==2024.6.1
139
+ jupyterlab==4.2.4
140
+ jupyterlab_pygments==0.3.0
141
+ jupyterlab_server==2.27.3
142
+ jupyterlab_widgets==3.0.11
143
+ kiwisolver==1.4.5
144
+ kombu==5.4.0
145
+ lab==8.2
146
+ langcodes==3.4.0
147
+ language_data==1.2.0
148
+ lazy_loader==0.4
149
+ librosa==0.10.2.post1
150
+ llvmlite==0.43.0
151
+ lxml==5.3.0
152
+ marisa-trie==1.2.0
153
+ markdown-it-py==3.0.0
154
+ matplotlib-inline==0.1.7
155
+ matplotlib==3.9.1.post1
156
+ mdurl==0.1.2
157
+ mistune==3.0.2
158
+ mpmath==1.3.0
159
+ msgpack==1.1.0
160
+ multidict==6.0.5
161
+ multiprocess==0.70.16
162
+ murmurhash==1.0.10
163
+ nb_conda_store_kernels==0.1.5
164
+ nbclient==0.10.0
165
+ nbconvert==7.16.4
166
+ nbformat==5.10.4
167
+ nest-asyncio==1.6.0
168
+ networkx==3.3
169
+ nodejs==0.1.1
170
+ notebook==7.2.1
171
+ notebook_shim==0.2.4
172
+ numba==0.60.0
173
+ numpy==1.26.4
174
+ nvidia-cublas-cu12==12.1.3.1
175
+ nvidia-cuda-cupti-cu12==12.1.105
176
+ nvidia-cuda-nvrtc-cu12==12.1.105
177
+ nvidia-cuda-runtime-cu12==12.1.105
178
+ nvidia-cudnn-cu12==8.9.2.26
179
+ nvidia-cudnn-cu12==9.1.0.70
180
+ nvidia-cufft-cu12==11.0.2.54
181
+ nvidia-curand-cu12==10.3.2.106
182
+ nvidia-cusolver-cu12==11.4.5.107
183
+ nvidia-cusparse-cu12==12.1.0.106
184
+ nvidia-nccl-cu12==2.19.3
185
+ nvidia-nccl-cu12==2.20.5
186
+ nvidia-nvjitlink-cu12==12.6.20
187
+ nvidia-nvtx-cu12==12.1.105
188
+ oauthlib==3.2.2
189
+ omegaconf==2.3.0
190
+ onnx==1.16.2
191
+ onnxruntime==1.19.0
192
+ opencv-python==4.10.0.84
193
+ optimum==1.21.4
194
+ optional-django==0.1.0
195
+ orjson==3.10.7
196
+ overrides==7.7.0
197
+ packaging==24.1
198
+ pamela==1.2.0
199
+ pandas==2.2.2
200
+ pandocfilters==1.5.1
201
+ parso==0.8.4
202
+ pathspec==0.12.1
203
+ peft==0.12.0
204
+ pexpect==4.9.0
205
+ pillow==10.4.0
206
+ pip==23.1.2
207
+ platformdirs==3.11.0
208
+ pooch==1.8.2
209
+ preshed==3.0.9
210
+ prometheus_client==0.20.0
211
+ prompt_toolkit==3.0.47
212
+ protobuf==5.27.3
213
+ psutil==6.0.0
214
+ ptyprocess==0.7.0
215
+ pure_eval==0.2.3
216
+ pyOpenSSL==24.2.1
217
+ pyarrow-hotfix==0.6
218
+ pyarrow==17.0.0
219
+ pycparser==2.22
220
+ pydantic==2.8.2
221
+ pydantic_core==2.20.1
222
+ pydot==3.0.1
223
+ pygame==2.6.0
224
+ pygit2==1.15.1
225
+ pygtrie==2.5.0
226
+ pyparsing==3.1.2
227
+ python-dateutil==2.9.0.post0
228
+ python-dotenv==1.0.1
229
+ python-json-logger==2.0.7
230
+ pytz==2024.1
231
+ pyzmq==26.1.0
232
+ qtconsole==5.5.2
233
+ referencing==0.35.1
234
+ regex==2024.7.24
235
+ requests==2.32.3
236
+ rfc3339-validator==0.1.4
237
+ rfc3986-validator==0.1.1
238
+ rich==13.7.1
239
+ rpds-py==0.20.0
240
+ ruamel.yaml.clib==0.2.8
241
+ ruamel.yaml==0.18.6
242
+ safetensors==0.4.4
243
+ scikit-learn==1.5.1
244
+ scipy==1.14.0
245
+ scmrepo==3.3.7
246
+ seaborn==0.13.2
247
+ semver==3.0.2
248
+ sentencepiece==0.2.0
249
+ sentry-sdk==2.14.0
250
+ seqeval==1.2.2
251
+ setproctitle==1.3.3
252
+ setuptools==65.5.0
253
+ shellingham==1.5.4
254
+ shortuuid==1.0.13
255
+ shtab==1.7.1
256
+ simplejson==3.19.2
257
+ six==1.16.0
258
+ smart-open==7.0.4
259
+ smmap==5.0.1
260
+ sniffio==1.3.1
261
+ soundfile==0.12.1
262
+ soupsieve==2.5
263
+ sox==1.5.0
264
+ soxr==0.5.0.post1
265
+ spacy-legacy==3.0.12
266
+ spacy-loggers==1.0.5
267
+ spacy==3.7.5
268
+ sqltrie==0.11.1
269
+ srsly==2.4.8
270
+ stable_baselines3==2.3.2
271
+ stack-data==0.6.3
272
+ sympy==1.13.2
273
+ tabulate==0.9.0
274
+ tensorboard-data-server==0.7.2
275
+ tensorboard==2.17.1
276
+ terminado==0.18.1
277
+ thinc==8.2.5
278
+ threadpoolctl==3.5.0
279
+ tinycss2==1.3.0
280
+ tokenizers==0.19.1
281
+ tomlkit==0.13.2
282
+ torch==2.4.1
283
+ torchaudio==2.4.1
284
+ torchvision==0.19.1
285
+ tornado==6.4.1
286
+ tqdm==4.66.5
287
+ traitlets==5.14.3
288
+ transformers==4.43.4
289
+ triton==3.0.0
290
+ trl==0.9.6
291
+ txt2tags==3.9
292
+ typer==0.12.3
293
+ types-python-dateutil==2.9.0.20240316
294
+ typing_extensions==4.12.2
295
+ tyro==0.8.6
296
+ tzdata==2024.1
297
+ uri-template==1.3.0
298
+ urllib3==2.2.2
299
+ vine==5.1.0
300
+ voluptuous==0.15.2
301
+ wandb==0.17.9
302
+ wasabi==1.1.3
303
+ wcwidth==0.2.13
304
+ weasel==0.4.1
305
+ webcolors==24.8.0
306
+ webencodings==0.5.1
307
+ websocket-client==1.8.0
308
+ widgetsnbextension==4.0.11
309
+ wrapt==1.16.0
310
+ xlrd==2.0.1
311
+ xxhash==3.4.1
312
+ yarl==1.9.4
313
+ zc.lockfile==3.0.post1
wandb/run-20240917_124325-7dvmn2ta/files/wandb-metadata.json ADDED
@@ -0,0 +1,281 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.18.0-513.24.1.el8_9.x86_64-x86_64-with-glibc2.28",
3
+ "python": "3.11.4",
4
+ "heartbeatAt": "2024-09-17T11:43:25.936017",
5
+ "startedAt": "2024-09-17T11:43:25.358772",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "<python with no main file>",
11
+ "codePathLocal": null,
12
+ "host": "codon-gpu-014.ebi.ac.uk",
13
+ "username": "amrufai",
14
+ "executable": "/nfs/production/literature/amina-mardiyyah/envs/llm-prompt/bin/python",
15
+ "cpu_count": 48,
16
+ "cpu_count_logical": 48,
17
+ "cpu_freq": {
18
+ "current": 2758.421791666666,
19
+ "min": 0.0,
20
+ "max": 0.0
21
+ },
22
+ "cpu_freq_per_core": [
23
+ {
24
+ "current": 2800.0,
25
+ "min": 0.0,
26
+ "max": 0.0
27
+ },
28
+ {
29
+ "current": 2800.0,
30
+ "min": 0.0,
31
+ "max": 0.0
32
+ },
33
+ {
34
+ "current": 2800.0,
35
+ "min": 0.0,
36
+ "max": 0.0
37
+ },
38
+ {
39
+ "current": 2800.0,
40
+ "min": 0.0,
41
+ "max": 0.0
42
+ },
43
+ {
44
+ "current": 2800.0,
45
+ "min": 0.0,
46
+ "max": 0.0
47
+ },
48
+ {
49
+ "current": 2800.0,
50
+ "min": 0.0,
51
+ "max": 0.0
52
+ },
53
+ {
54
+ "current": 2800.0,
55
+ "min": 0.0,
56
+ "max": 0.0
57
+ },
58
+ {
59
+ "current": 2800.0,
60
+ "min": 0.0,
61
+ "max": 0.0
62
+ },
63
+ {
64
+ "current": 2800.0,
65
+ "min": 0.0,
66
+ "max": 0.0
67
+ },
68
+ {
69
+ "current": 2800.0,
70
+ "min": 0.0,
71
+ "max": 0.0
72
+ },
73
+ {
74
+ "current": 2800.0,
75
+ "min": 0.0,
76
+ "max": 0.0
77
+ },
78
+ {
79
+ "current": 2800.0,
80
+ "min": 0.0,
81
+ "max": 0.0
82
+ },
83
+ {
84
+ "current": 804.246,
85
+ "min": 0.0,
86
+ "max": 0.0
87
+ },
88
+ {
89
+ "current": 2800.0,
90
+ "min": 0.0,
91
+ "max": 0.0
92
+ },
93
+ {
94
+ "current": 2800.0,
95
+ "min": 0.0,
96
+ "max": 0.0
97
+ },
98
+ {
99
+ "current": 2800.0,
100
+ "min": 0.0,
101
+ "max": 0.0
102
+ },
103
+ {
104
+ "current": 2800.0,
105
+ "min": 0.0,
106
+ "max": 0.0
107
+ },
108
+ {
109
+ "current": 2800.0,
110
+ "min": 0.0,
111
+ "max": 0.0
112
+ },
113
+ {
114
+ "current": 2800.0,
115
+ "min": 0.0,
116
+ "max": 0.0
117
+ },
118
+ {
119
+ "current": 2800.0,
120
+ "min": 0.0,
121
+ "max": 0.0
122
+ },
123
+ {
124
+ "current": 2800.0,
125
+ "min": 0.0,
126
+ "max": 0.0
127
+ },
128
+ {
129
+ "current": 2800.0,
130
+ "min": 0.0,
131
+ "max": 0.0
132
+ },
133
+ {
134
+ "current": 2800.0,
135
+ "min": 0.0,
136
+ "max": 0.0
137
+ },
138
+ {
139
+ "current": 2800.0,
140
+ "min": 0.0,
141
+ "max": 0.0
142
+ },
143
+ {
144
+ "current": 2800.0,
145
+ "min": 0.0,
146
+ "max": 0.0
147
+ },
148
+ {
149
+ "current": 2800.0,
150
+ "min": 0.0,
151
+ "max": 0.0
152
+ },
153
+ {
154
+ "current": 2800.0,
155
+ "min": 0.0,
156
+ "max": 0.0
157
+ },
158
+ {
159
+ "current": 2800.0,
160
+ "min": 0.0,
161
+ "max": 0.0
162
+ },
163
+ {
164
+ "current": 2800.0,
165
+ "min": 0.0,
166
+ "max": 0.0
167
+ },
168
+ {
169
+ "current": 2800.0,
170
+ "min": 0.0,
171
+ "max": 0.0
172
+ },
173
+ {
174
+ "current": 2800.0,
175
+ "min": 0.0,
176
+ "max": 0.0
177
+ },
178
+ {
179
+ "current": 2800.0,
180
+ "min": 0.0,
181
+ "max": 0.0
182
+ },
183
+ {
184
+ "current": 2800.0,
185
+ "min": 0.0,
186
+ "max": 0.0
187
+ },
188
+ {
189
+ "current": 2800.0,
190
+ "min": 0.0,
191
+ "max": 0.0
192
+ },
193
+ {
194
+ "current": 2800.0,
195
+ "min": 0.0,
196
+ "max": 0.0
197
+ },
198
+ {
199
+ "current": 2800.0,
200
+ "min": 0.0,
201
+ "max": 0.0
202
+ },
203
+ {
204
+ "current": 2800.0,
205
+ "min": 0.0,
206
+ "max": 0.0
207
+ },
208
+ {
209
+ "current": 2800.0,
210
+ "min": 0.0,
211
+ "max": 0.0
212
+ },
213
+ {
214
+ "current": 2800.0,
215
+ "min": 0.0,
216
+ "max": 0.0
217
+ },
218
+ {
219
+ "current": 2800.0,
220
+ "min": 0.0,
221
+ "max": 0.0
222
+ },
223
+ {
224
+ "current": 2800.0,
225
+ "min": 0.0,
226
+ "max": 0.0
227
+ },
228
+ {
229
+ "current": 2800.0,
230
+ "min": 0.0,
231
+ "max": 0.0
232
+ },
233
+ {
234
+ "current": 2800.0,
235
+ "min": 0.0,
236
+ "max": 0.0
237
+ },
238
+ {
239
+ "current": 2800.0,
240
+ "min": 0.0,
241
+ "max": 0.0
242
+ },
243
+ {
244
+ "current": 2800.0,
245
+ "min": 0.0,
246
+ "max": 0.0
247
+ },
248
+ {
249
+ "current": 2800.0,
250
+ "min": 0.0,
251
+ "max": 0.0
252
+ },
253
+ {
254
+ "current": 2800.0,
255
+ "min": 0.0,
256
+ "max": 0.0
257
+ },
258
+ {
259
+ "current": 2800.0,
260
+ "min": 0.0,
261
+ "max": 0.0
262
+ }
263
+ ],
264
+ "disk": {
265
+ "/": {
266
+ "total": 47.760292053222656,
267
+ "used": 15.848091125488281
268
+ }
269
+ },
270
+ "gpu": "NVIDIA A100 80GB PCIe",
271
+ "gpu_count": 1,
272
+ "gpu_devices": [
273
+ {
274
+ "name": "NVIDIA A100 80GB PCIe",
275
+ "memory_total": 85899345920
276
+ }
277
+ ],
278
+ "memory": {
279
+ "total": 502.83758544921875
280
+ }
281
+ }
wandb/run-20240917_124325-7dvmn2ta/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 0.3296, "train/grad_norm": 2.798435688018799, "train/learning_rate": 1.398e-05, "train/epoch": 0.9992872416250891, "train/global_step": 701, "_timestamp": 1726577567.027478, "_runtime": 4161.6403028965, "_step": 1, "eval/loss": 0.13608910143375397, "eval/f1": 0.7956595214800319, "eval/precision": 0.7948253189750151, "eval/recall": 0.7964954768858181, "eval/accuracy": 0.9563037319665285, "eval/runtime": 75.5067, "eval/samples_per_second": 148.623, "eval/steps_per_second": 18.581, "_wandb": {"runtime": 618}}
wandb/run-20240917_124325-7dvmn2ta/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240917_124325-7dvmn2ta/logs/debug.log ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-09-17 12:43:25,380 INFO MainThread:1260347 [wandb_setup.py:_flush():77] Current SDK version is 0.17.9
2
+ 2024-09-17 12:43:25,380 INFO MainThread:1260347 [wandb_setup.py:_flush():77] Configure stats pid to 1260347
3
+ 2024-09-17 12:43:25,380 INFO MainThread:1260347 [wandb_setup.py:_flush():77] Loading settings from /homes/amrufai/.config/wandb/settings
4
+ 2024-09-17 12:43:25,380 INFO MainThread:1260347 [wandb_setup.py:_flush():77] Loading settings from /nfs/production/literature/amina-mardiyyah/wandb/settings
5
+ 2024-09-17 12:43:25,380 INFO MainThread:1260347 [wandb_setup.py:_flush():77] Loading settings from environment variables: {}
6
+ 2024-09-17 12:43:25,380 INFO MainThread:1260347 [wandb_setup.py:_flush():77] Inferring run settings from compute environment: {'program': '<python with no main file>'}
7
+ 2024-09-17 12:43:25,380 INFO MainThread:1260347 [wandb_setup.py:_flush():77] Applying login settings: {}
8
+ 2024-09-17 12:43:25,380 INFO MainThread:1260347 [wandb_setup.py:_flush():77] Applying login settings: {}
9
+ 2024-09-17 12:43:25,381 INFO MainThread:1260347 [wandb_init.py:_log_setup():524] Logging user logs to /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_124325-7dvmn2ta/logs/debug.log
10
+ 2024-09-17 12:43:25,381 INFO MainThread:1260347 [wandb_init.py:_log_setup():525] Logging internal logs to /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_124325-7dvmn2ta/logs/debug-internal.log
11
+ 2024-09-17 12:43:25,381 INFO MainThread:1260347 [wandb_init.py:_jupyter_setup():470] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7f7c10823490>
12
+ 2024-09-17 12:43:25,381 INFO MainThread:1260347 [wandb_init.py:init():608] calling init triggers
13
+ 2024-09-17 12:43:25,381 INFO MainThread:1260347 [wandb_init.py:init():615] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-09-17 12:43:25,381 INFO MainThread:1260347 [wandb_init.py:init():658] starting backend
16
+ 2024-09-17 12:43:25,381 INFO MainThread:1260347 [wandb_init.py:init():662] setting up manager
17
+ 2024-09-17 12:43:25,383 INFO MainThread:1260347 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-09-17 12:43:25,386 INFO MainThread:1260347 [wandb_init.py:init():670] backend started and connected
19
+ 2024-09-17 12:43:25,392 INFO MainThread:1260347 [wandb_run.py:_label_probe_notebook():1344] probe notebook
20
+ 2024-09-17 12:43:25,407 INFO MainThread:1260347 [wandb_run.py:_label_probe_notebook():1354] Unable to probe notebook: 'NoneType' object has no attribute 'get'
21
+ 2024-09-17 12:43:25,408 INFO MainThread:1260347 [wandb_init.py:init():768] updated telemetry
22
+ 2024-09-17 12:43:25,418 INFO MainThread:1260347 [wandb_init.py:init():801] communicating run to backend with 90.0 second timeout
23
+ 2024-09-17 12:43:25,789 INFO MainThread:1260347 [wandb_init.py:init():852] starting run threads in backend
24
+ 2024-09-17 12:43:26,460 INFO MainThread:1260347 [wandb_run.py:_console_start():2465] atexit reg
25
+ 2024-09-17 12:43:26,460 INFO MainThread:1260347 [wandb_run.py:_redirect():2311] redirect: wrap_raw
26
+ 2024-09-17 12:43:26,460 INFO MainThread:1260347 [wandb_run.py:_redirect():2376] Wrapping output streams.
27
+ 2024-09-17 12:43:26,460 INFO MainThread:1260347 [wandb_run.py:_redirect():2401] Redirects installed.
28
+ 2024-09-17 12:43:26,465 INFO MainThread:1260347 [wandb_init.py:init():895] run started, returning control to user process
29
+ 2024-09-17 12:43:26,469 INFO MainThread:1260347 [jupyter.py:save_ipynb():372] not saving jupyter notebook
30
+ 2024-09-17 12:43:26,469 INFO MainThread:1260347 [wandb_init.py:_pause_backend():435] pausing backend
31
+ 2024-09-17 12:43:35,145 INFO MainThread:1260347 [wandb_init.py:_resume_backend():440] resuming backend
32
+ 2024-09-17 12:43:35,191 INFO MainThread:1260347 [jupyter.py:save_ipynb():372] not saving jupyter notebook
33
+ 2024-09-17 12:43:35,191 INFO MainThread:1260347 [wandb_init.py:_pause_backend():435] pausing backend
34
+ 2024-09-17 12:43:36,136 INFO MainThread:1260347 [wandb_init.py:_resume_backend():440] resuming backend
35
+ 2024-09-17 12:43:36,141 INFO MainThread:1260347 [jupyter.py:save_ipynb():372] not saving jupyter notebook
36
+ 2024-09-17 12:43:36,141 INFO MainThread:1260347 [wandb_init.py:_pause_backend():435] pausing backend
37
+ 2024-09-17 12:43:39,689 INFO MainThread:1260347 [wandb_init.py:_resume_backend():440] resuming backend
38
+ 2024-09-17 12:43:46,727 INFO MainThread:1260347 [wandb_run.py:_config_callback():1392] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['XLMRobertaForTokenClassification'], 'finetuning_task': None, 'id2label': {0: 'O', 1: 'B-PER', 2: 'I-PER', 3: 'B-ORG', 4: 'I-ORG', 5: 'B-LOC', 6: 'I-LOC'}, 'label2id': {'O': 0, 'B-PER': 1, 'I-PER': 2, 'B-ORG': 3, 'I-ORG': 4, 'B-LOC': 5, 'I-LOC': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0', 'transformers_version': '4.43.4', 'adapters': {'adapters': {}, 'config_map': {}, 'fusion_config_map': {}, 'fusions': {}}, 'gradient_checkpointing': False, 'model_type': 'xlm-roberta', 'output_past': True, 'vocab_size': 250002, 'hidden_size': 1024, 'num_hidden_layers': 24, 'num_attention_heads': 16, 'hidden_act': 'gelu', 'intermediate_size': 4096, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 514, 'type_vocab_size': 1, 'initializer_range': 0.02, 'layer_norm_eps': 1e-05, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 2e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/logs', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 3407, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'eval_f1', 'greater_is_better': True, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'eval_use_gather_object': False}
39
+ 2024-09-17 12:43:46,730 INFO MainThread:1260347 [wandb_config.py:__setitem__():154] config set model/num_parameters = 558848007 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7f7c10c10310>>
40
+ 2024-09-17 12:43:46,730 INFO MainThread:1260347 [wandb_run.py:_config_callback():1392] config_cb model/num_parameters 558848007 None
41
+ 2024-09-17 12:43:47,226 INFO MainThread:1260347 [jupyter.py:save_ipynb():372] not saving jupyter notebook
42
+ 2024-09-17 12:43:47,226 INFO MainThread:1260347 [wandb_init.py:_pause_backend():435] pausing backend
43
+ 2024-09-17 13:43:52,107 INFO MainThread:1260347 [wandb_init.py:_resume_backend():440] resuming backend
44
+ 2024-09-17 13:43:52,110 INFO MainThread:1260347 [jupyter.py:save_ipynb():372] not saving jupyter notebook
45
+ 2024-09-17 13:43:52,110 INFO MainThread:1260347 [wandb_init.py:_pause_backend():435] pausing backend
46
+ 2024-09-17 13:44:05,681 INFO MainThread:1260347 [wandb_init.py:_resume_backend():440] resuming backend
47
+ 2024-09-17 13:44:05,711 INFO MainThread:1260347 [jupyter.py:save_ipynb():372] not saving jupyter notebook
48
+ 2024-09-17 13:44:05,711 INFO MainThread:1260347 [wandb_init.py:_pause_backend():435] pausing backend
49
+ 2024-09-17 13:44:07,997 INFO MainThread:1260347 [wandb_init.py:_resume_backend():440] resuming backend
50
+ 2024-09-17 13:44:08,003 INFO MainThread:1260347 [jupyter.py:save_ipynb():372] not saving jupyter notebook
51
+ 2024-09-17 13:44:08,003 INFO MainThread:1260347 [wandb_init.py:_pause_backend():435] pausing backend
52
+ 2024-09-17 13:44:08,415 INFO MainThread:1260347 [wandb_init.py:_resume_backend():440] resuming backend
53
+ 2024-09-17 13:44:08,418 INFO MainThread:1260347 [jupyter.py:save_ipynb():372] not saving jupyter notebook
54
+ 2024-09-17 13:44:08,419 INFO MainThread:1260347 [wandb_init.py:_pause_backend():435] pausing backend
55
+ 2024-09-17 13:44:15,367 INFO MainThread:1260347 [wandb_init.py:_resume_backend():440] resuming backend
56
+ 2024-09-17 13:44:22,185 INFO MainThread:1260347 [wandb_run.py:_config_callback():1392] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['XLMRobertaForTokenClassification'], 'finetuning_task': None, 'id2label': {0: 'O', 1: 'B-PER', 2: 'I-PER', 3: 'B-ORG', 4: 'I-ORG', 5: 'B-LOC', 6: 'I-LOC'}, 'label2id': {'O': 0, 'B-PER': 1, 'I-PER': 2, 'B-ORG': 3, 'I-ORG': 4, 'B-LOC': 5, 'I-LOC': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0', 'transformers_version': '4.43.4', 'adapters': {'adapters': {}, 'config_map': {}, 'fusion_config_map': {}, 'fusions': {}}, 'gradient_checkpointing': False, 'model_type': 'xlm-roberta', 'output_past': True, 'vocab_size': 250002, 'hidden_size': 1024, 'num_hidden_layers': 24, 'num_attention_heads': 16, 'hidden_act': 'gelu', 'intermediate_size': 4096, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 514, 'type_vocab_size': 1, 'initializer_range': 0.02, 'layer_norm_eps': 1e-05, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 2e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/logs', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 3407, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'eval_f1', 'greater_is_better': True, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'eval_use_gather_object': False}
57
+ 2024-09-17 13:44:22,187 INFO MainThread:1260347 [wandb_config.py:__setitem__():154] config set model/num_parameters = 558848007 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7f7c10c10310>>
58
+ 2024-09-17 13:44:22,187 INFO MainThread:1260347 [wandb_run.py:_config_callback():1392] config_cb model/num_parameters 558848007 None
59
+ 2024-09-17 13:54:25,378 INFO MainThread:1260347 [jupyter.py:save_ipynb():372] not saving jupyter notebook
60
+ 2024-09-17 13:54:25,378 INFO MainThread:1260347 [wandb_init.py:_pause_backend():435] pausing backend
wandb/run-20240917_124325-7dvmn2ta/run-7dvmn2ta.wandb ADDED
Binary file (75.3 kB). View file
 
wandb/run-20240917_140929-fx32bo52/files/config.yaml ADDED
@@ -0,0 +1,741 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.11.4
7
+ cli_version: 0.17.9
8
+ framework: huggingface
9
+ huggingface_version: 4.43.4
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: false
12
+ start_time: 1726578569
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 5
17
+ - 11
18
+ - 41
19
+ - 49
20
+ - 51
21
+ - 53
22
+ - 55
23
+ - 71
24
+ - 98
25
+ - 100
26
+ - 105
27
+ 2:
28
+ - 1
29
+ - 5
30
+ - 11
31
+ - 41
32
+ - 49
33
+ - 51
34
+ - 53
35
+ - 55
36
+ - 71
37
+ - 98
38
+ - 100
39
+ - 105
40
+ 3:
41
+ - 7
42
+ - 19
43
+ - 23
44
+ - 66
45
+ 4: 3.11.4
46
+ 5: 0.17.9
47
+ 6: 4.43.4
48
+ 8:
49
+ - 1
50
+ - 5
51
+ 9:
52
+ 1: transformers_trainer
53
+ 13: linux-x86_64
54
+ m:
55
+ - 1: train/global_step
56
+ 6:
57
+ - 3
58
+ - 1: train/loss
59
+ 5: 1
60
+ 6:
61
+ - 1
62
+ - 1: train/grad_norm
63
+ 5: 1
64
+ 6:
65
+ - 1
66
+ - 1: train/learning_rate
67
+ 5: 1
68
+ 6:
69
+ - 1
70
+ - 1: train/epoch
71
+ 5: 1
72
+ 6:
73
+ - 1
74
+ - 1: eval/loss
75
+ 5: 1
76
+ 6:
77
+ - 1
78
+ - 1: eval/f1
79
+ 5: 1
80
+ 6:
81
+ - 1
82
+ - 1: eval/precision
83
+ 5: 1
84
+ 6:
85
+ - 1
86
+ - 1: eval/recall
87
+ 5: 1
88
+ 6:
89
+ - 1
90
+ - 1: eval/accuracy
91
+ 5: 1
92
+ 6:
93
+ - 1
94
+ - 1: eval/runtime
95
+ 5: 1
96
+ 6:
97
+ - 1
98
+ - 1: eval/samples_per_second
99
+ 5: 1
100
+ 6:
101
+ - 1
102
+ - 1: eval/steps_per_second
103
+ 5: 1
104
+ 6:
105
+ - 1
106
+ return_dict:
107
+ desc: null
108
+ value: true
109
+ output_hidden_states:
110
+ desc: null
111
+ value: false
112
+ output_attentions:
113
+ desc: null
114
+ value: false
115
+ torchscript:
116
+ desc: null
117
+ value: false
118
+ torch_dtype:
119
+ desc: null
120
+ value: float32
121
+ use_bfloat16:
122
+ desc: null
123
+ value: false
124
+ tf_legacy_loss:
125
+ desc: null
126
+ value: false
127
+ pruned_heads:
128
+ desc: null
129
+ value: {}
130
+ tie_word_embeddings:
131
+ desc: null
132
+ value: true
133
+ chunk_size_feed_forward:
134
+ desc: null
135
+ value: 0
136
+ is_encoder_decoder:
137
+ desc: null
138
+ value: false
139
+ is_decoder:
140
+ desc: null
141
+ value: false
142
+ cross_attention_hidden_size:
143
+ desc: null
144
+ value: null
145
+ add_cross_attention:
146
+ desc: null
147
+ value: false
148
+ tie_encoder_decoder:
149
+ desc: null
150
+ value: false
151
+ max_length:
152
+ desc: null
153
+ value: 20
154
+ min_length:
155
+ desc: null
156
+ value: 0
157
+ do_sample:
158
+ desc: null
159
+ value: false
160
+ early_stopping:
161
+ desc: null
162
+ value: false
163
+ num_beams:
164
+ desc: null
165
+ value: 1
166
+ num_beam_groups:
167
+ desc: null
168
+ value: 1
169
+ diversity_penalty:
170
+ desc: null
171
+ value: 0.0
172
+ temperature:
173
+ desc: null
174
+ value: 1.0
175
+ top_k:
176
+ desc: null
177
+ value: 50
178
+ top_p:
179
+ desc: null
180
+ value: 1.0
181
+ typical_p:
182
+ desc: null
183
+ value: 1.0
184
+ repetition_penalty:
185
+ desc: null
186
+ value: 1.0
187
+ length_penalty:
188
+ desc: null
189
+ value: 1.0
190
+ no_repeat_ngram_size:
191
+ desc: null
192
+ value: 0
193
+ encoder_no_repeat_ngram_size:
194
+ desc: null
195
+ value: 0
196
+ bad_words_ids:
197
+ desc: null
198
+ value: null
199
+ num_return_sequences:
200
+ desc: null
201
+ value: 1
202
+ output_scores:
203
+ desc: null
204
+ value: false
205
+ return_dict_in_generate:
206
+ desc: null
207
+ value: false
208
+ forced_bos_token_id:
209
+ desc: null
210
+ value: null
211
+ forced_eos_token_id:
212
+ desc: null
213
+ value: null
214
+ remove_invalid_values:
215
+ desc: null
216
+ value: false
217
+ exponential_decay_length_penalty:
218
+ desc: null
219
+ value: null
220
+ suppress_tokens:
221
+ desc: null
222
+ value: null
223
+ begin_suppress_tokens:
224
+ desc: null
225
+ value: null
226
+ architectures:
227
+ desc: null
228
+ value:
229
+ - XLMRobertaForTokenClassification
230
+ finetuning_task:
231
+ desc: null
232
+ value: null
233
+ id2label:
234
+ desc: null
235
+ value:
236
+ '0': O
237
+ '1': B-PER
238
+ '2': I-PER
239
+ '3': B-ORG
240
+ '4': I-ORG
241
+ '5': B-LOC
242
+ '6': I-LOC
243
+ label2id:
244
+ desc: null
245
+ value:
246
+ O: 0
247
+ B-PER: 1
248
+ I-PER: 2
249
+ B-ORG: 3
250
+ I-ORG: 4
251
+ B-LOC: 5
252
+ I-LOC: 6
253
+ tokenizer_class:
254
+ desc: null
255
+ value: null
256
+ prefix:
257
+ desc: null
258
+ value: null
259
+ bos_token_id:
260
+ desc: null
261
+ value: 0
262
+ pad_token_id:
263
+ desc: null
264
+ value: 1
265
+ eos_token_id:
266
+ desc: null
267
+ value: 2
268
+ sep_token_id:
269
+ desc: null
270
+ value: null
271
+ decoder_start_token_id:
272
+ desc: null
273
+ value: null
274
+ task_specific_params:
275
+ desc: null
276
+ value: null
277
+ problem_type:
278
+ desc: null
279
+ value: null
280
+ _name_or_path:
281
+ desc: null
282
+ value: masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0
283
+ transformers_version:
284
+ desc: null
285
+ value: 4.43.4
286
+ adapters:
287
+ desc: null
288
+ value:
289
+ adapters: {}
290
+ config_map: {}
291
+ fusion_config_map: {}
292
+ fusions: {}
293
+ gradient_checkpointing:
294
+ desc: null
295
+ value: false
296
+ model_type:
297
+ desc: null
298
+ value: xlm-roberta
299
+ output_past:
300
+ desc: null
301
+ value: true
302
+ vocab_size:
303
+ desc: null
304
+ value: 250002
305
+ hidden_size:
306
+ desc: null
307
+ value: 1024
308
+ num_hidden_layers:
309
+ desc: null
310
+ value: 24
311
+ num_attention_heads:
312
+ desc: null
313
+ value: 16
314
+ hidden_act:
315
+ desc: null
316
+ value: gelu
317
+ intermediate_size:
318
+ desc: null
319
+ value: 4096
320
+ hidden_dropout_prob:
321
+ desc: null
322
+ value: 0.1
323
+ attention_probs_dropout_prob:
324
+ desc: null
325
+ value: 0.1
326
+ max_position_embeddings:
327
+ desc: null
328
+ value: 514
329
+ type_vocab_size:
330
+ desc: null
331
+ value: 1
332
+ initializer_range:
333
+ desc: null
334
+ value: 0.02
335
+ layer_norm_eps:
336
+ desc: null
337
+ value: 1.0e-05
338
+ position_embedding_type:
339
+ desc: null
340
+ value: absolute
341
+ use_cache:
342
+ desc: null
343
+ value: true
344
+ classifier_dropout:
345
+ desc: null
346
+ value: null
347
+ output_dir:
348
+ desc: null
349
+ value: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date
350
+ overwrite_output_dir:
351
+ desc: null
352
+ value: false
353
+ do_train:
354
+ desc: null
355
+ value: false
356
+ do_eval:
357
+ desc: null
358
+ value: true
359
+ do_predict:
360
+ desc: null
361
+ value: false
362
+ eval_strategy:
363
+ desc: null
364
+ value: epoch
365
+ prediction_loss_only:
366
+ desc: null
367
+ value: false
368
+ per_device_train_batch_size:
369
+ desc: null
370
+ value: 16
371
+ per_device_eval_batch_size:
372
+ desc: null
373
+ value: 8
374
+ per_gpu_train_batch_size:
375
+ desc: null
376
+ value: null
377
+ per_gpu_eval_batch_size:
378
+ desc: null
379
+ value: null
380
+ gradient_accumulation_steps:
381
+ desc: null
382
+ value: 4
383
+ eval_accumulation_steps:
384
+ desc: null
385
+ value: null
386
+ eval_delay:
387
+ desc: null
388
+ value: 0
389
+ torch_empty_cache_steps:
390
+ desc: null
391
+ value: null
392
+ learning_rate:
393
+ desc: null
394
+ value: 2.0e-05
395
+ weight_decay:
396
+ desc: null
397
+ value: 0.01
398
+ adam_beta1:
399
+ desc: null
400
+ value: 0.9
401
+ adam_beta2:
402
+ desc: null
403
+ value: 0.999
404
+ adam_epsilon:
405
+ desc: null
406
+ value: 1.0e-08
407
+ max_grad_norm:
408
+ desc: null
409
+ value: 1.0
410
+ num_train_epochs:
411
+ desc: null
412
+ value: 10
413
+ max_steps:
414
+ desc: null
415
+ value: -1
416
+ lr_scheduler_type:
417
+ desc: null
418
+ value: linear
419
+ lr_scheduler_kwargs:
420
+ desc: null
421
+ value: {}
422
+ warmup_ratio:
423
+ desc: null
424
+ value: 0.0
425
+ warmup_steps:
426
+ desc: null
427
+ value: 1000
428
+ log_level:
429
+ desc: null
430
+ value: passive
431
+ log_level_replica:
432
+ desc: null
433
+ value: warning
434
+ log_on_each_node:
435
+ desc: null
436
+ value: true
437
+ logging_dir:
438
+ desc: null
439
+ value: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/logs
440
+ logging_strategy:
441
+ desc: null
442
+ value: epoch
443
+ logging_first_step:
444
+ desc: null
445
+ value: false
446
+ logging_steps:
447
+ desc: null
448
+ value: 500
449
+ logging_nan_inf_filter:
450
+ desc: null
451
+ value: true
452
+ save_strategy:
453
+ desc: null
454
+ value: epoch
455
+ save_steps:
456
+ desc: null
457
+ value: 500
458
+ save_total_limit:
459
+ desc: null
460
+ value: 2
461
+ save_safetensors:
462
+ desc: null
463
+ value: true
464
+ save_on_each_node:
465
+ desc: null
466
+ value: false
467
+ save_only_model:
468
+ desc: null
469
+ value: false
470
+ restore_callback_states_from_checkpoint:
471
+ desc: null
472
+ value: false
473
+ no_cuda:
474
+ desc: null
475
+ value: false
476
+ use_cpu:
477
+ desc: null
478
+ value: false
479
+ use_mps_device:
480
+ desc: null
481
+ value: false
482
+ seed:
483
+ desc: null
484
+ value: 3407
485
+ data_seed:
486
+ desc: null
487
+ value: null
488
+ jit_mode_eval:
489
+ desc: null
490
+ value: false
491
+ use_ipex:
492
+ desc: null
493
+ value: false
494
+ bf16:
495
+ desc: null
496
+ value: false
497
+ fp16:
498
+ desc: null
499
+ value: true
500
+ fp16_opt_level:
501
+ desc: null
502
+ value: O1
503
+ half_precision_backend:
504
+ desc: null
505
+ value: auto
506
+ bf16_full_eval:
507
+ desc: null
508
+ value: false
509
+ fp16_full_eval:
510
+ desc: null
511
+ value: false
512
+ tf32:
513
+ desc: null
514
+ value: null
515
+ local_rank:
516
+ desc: null
517
+ value: 0
518
+ ddp_backend:
519
+ desc: null
520
+ value: null
521
+ tpu_num_cores:
522
+ desc: null
523
+ value: null
524
+ tpu_metrics_debug:
525
+ desc: null
526
+ value: false
527
+ debug:
528
+ desc: null
529
+ value: []
530
+ dataloader_drop_last:
531
+ desc: null
532
+ value: false
533
+ eval_steps:
534
+ desc: null
535
+ value: null
536
+ dataloader_num_workers:
537
+ desc: null
538
+ value: 0
539
+ dataloader_prefetch_factor:
540
+ desc: null
541
+ value: null
542
+ past_index:
543
+ desc: null
544
+ value: -1
545
+ run_name:
546
+ desc: null
547
+ value: /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date
548
+ disable_tqdm:
549
+ desc: null
550
+ value: false
551
+ remove_unused_columns:
552
+ desc: null
553
+ value: true
554
+ label_names:
555
+ desc: null
556
+ value: null
557
+ load_best_model_at_end:
558
+ desc: null
559
+ value: true
560
+ metric_for_best_model:
561
+ desc: null
562
+ value: eval_f1
563
+ greater_is_better:
564
+ desc: null
565
+ value: true
566
+ ignore_data_skip:
567
+ desc: null
568
+ value: false
569
+ fsdp:
570
+ desc: null
571
+ value: []
572
+ fsdp_min_num_params:
573
+ desc: null
574
+ value: 0
575
+ fsdp_config:
576
+ desc: null
577
+ value:
578
+ min_num_params: 0
579
+ xla: false
580
+ xla_fsdp_v2: false
581
+ xla_fsdp_grad_ckpt: false
582
+ fsdp_transformer_layer_cls_to_wrap:
583
+ desc: null
584
+ value: null
585
+ accelerator_config:
586
+ desc: null
587
+ value:
588
+ split_batches: false
589
+ dispatch_batches: null
590
+ even_batches: true
591
+ use_seedable_sampler: true
592
+ non_blocking: false
593
+ gradient_accumulation_kwargs: null
594
+ deepspeed:
595
+ desc: null
596
+ value: null
597
+ label_smoothing_factor:
598
+ desc: null
599
+ value: 0.0
600
+ optim:
601
+ desc: null
602
+ value: adamw_torch
603
+ optim_args:
604
+ desc: null
605
+ value: null
606
+ adafactor:
607
+ desc: null
608
+ value: false
609
+ group_by_length:
610
+ desc: null
611
+ value: true
612
+ length_column_name:
613
+ desc: null
614
+ value: length
615
+ report_to:
616
+ desc: null
617
+ value:
618
+ - wandb
619
+ ddp_find_unused_parameters:
620
+ desc: null
621
+ value: null
622
+ ddp_bucket_cap_mb:
623
+ desc: null
624
+ value: null
625
+ ddp_broadcast_buffers:
626
+ desc: null
627
+ value: null
628
+ dataloader_pin_memory:
629
+ desc: null
630
+ value: true
631
+ dataloader_persistent_workers:
632
+ desc: null
633
+ value: false
634
+ skip_memory_metrics:
635
+ desc: null
636
+ value: true
637
+ use_legacy_prediction_loop:
638
+ desc: null
639
+ value: false
640
+ push_to_hub:
641
+ desc: null
642
+ value: false
643
+ resume_from_checkpoint:
644
+ desc: null
645
+ value: null
646
+ hub_model_id:
647
+ desc: null
648
+ value: null
649
+ hub_strategy:
650
+ desc: null
651
+ value: every_save
652
+ hub_token:
653
+ desc: null
654
+ value: <HUB_TOKEN>
655
+ hub_private_repo:
656
+ desc: null
657
+ value: false
658
+ hub_always_push:
659
+ desc: null
660
+ value: false
661
+ gradient_checkpointing_kwargs:
662
+ desc: null
663
+ value: null
664
+ include_inputs_for_metrics:
665
+ desc: null
666
+ value: false
667
+ eval_do_concat_batches:
668
+ desc: null
669
+ value: true
670
+ fp16_backend:
671
+ desc: null
672
+ value: auto
673
+ evaluation_strategy:
674
+ desc: null
675
+ value: null
676
+ push_to_hub_model_id:
677
+ desc: null
678
+ value: null
679
+ push_to_hub_organization:
680
+ desc: null
681
+ value: null
682
+ push_to_hub_token:
683
+ desc: null
684
+ value: <PUSH_TO_HUB_TOKEN>
685
+ mp_parameters:
686
+ desc: null
687
+ value: ''
688
+ auto_find_batch_size:
689
+ desc: null
690
+ value: false
691
+ full_determinism:
692
+ desc: null
693
+ value: false
694
+ torchdynamo:
695
+ desc: null
696
+ value: null
697
+ ray_scope:
698
+ desc: null
699
+ value: last
700
+ ddp_timeout:
701
+ desc: null
702
+ value: 1800
703
+ torch_compile:
704
+ desc: null
705
+ value: false
706
+ torch_compile_backend:
707
+ desc: null
708
+ value: null
709
+ torch_compile_mode:
710
+ desc: null
711
+ value: null
712
+ dispatch_batches:
713
+ desc: null
714
+ value: null
715
+ split_batches:
716
+ desc: null
717
+ value: null
718
+ include_tokens_per_second:
719
+ desc: null
720
+ value: false
721
+ include_num_input_tokens_seen:
722
+ desc: null
723
+ value: false
724
+ neftune_noise_alpha:
725
+ desc: null
726
+ value: null
727
+ optim_target_modules:
728
+ desc: null
729
+ value: null
730
+ batch_eval_metrics:
731
+ desc: null
732
+ value: false
733
+ eval_on_start:
734
+ desc: null
735
+ value: false
736
+ eval_use_gather_object:
737
+ desc: null
738
+ value: false
739
+ model/num_parameters:
740
+ desc: null
741
+ value: 558848007
wandb/run-20240917_140929-fx32bo52/files/output.log ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /nfs/production/literature/amina-mardiyyah/envs/llm-prompt/lib/python3.11/site-packages/accelerate/accelerator.py:488: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead.
2
+ self.scaler = torch.cuda.amp.GradScaler(**kwargs)
3
+ Detected kernel version 4.18.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
4
+ wandb: WARNING The `run_name` is currently set to the same value as `TrainingArguments.output_dir`. If this was not intended, please specify a different run name by setting the `TrainingArguments.run_name` parameter.
5
+ Classification Report:
6
+ precision recall f1-score support
7
+ LOC 0.73 0.79 0.76 9327
8
+ ORG 0.75 0.68 0.71 7547
9
+ PER 0.89 0.88 0.88 11204
10
+ micro avg 0.80 0.80 0.80 28078
11
+ macro avg 0.79 0.78 0.78 28078
12
+ weighted avg 0.80 0.80 0.80 28078
13
+ Classification Report:
14
+ precision recall f1-score support
15
+ LOC 0.81 0.79 0.80 9327
16
+ ORG 0.71 0.80 0.75 7547
17
+ PER 0.91 0.89 0.90 11204
18
+ micro avg 0.82 0.83 0.83 28078
19
+ macro avg 0.81 0.83 0.82 28078
20
+ weighted avg 0.82 0.83 0.83 28078
21
+ Classification Report:
22
+ precision recall f1-score support
23
+ LOC 0.81 0.80 0.81 9327
24
+ ORG 0.73 0.80 0.77 7547
25
+ PER 0.90 0.90 0.90 11204
26
+ micro avg 0.82 0.84 0.83 28078
27
+ macro avg 0.81 0.84 0.82 28078
28
+ weighted avg 0.82 0.84 0.83 28078
29
+ Classification Report:
30
+ precision recall f1-score support
31
+ LOC 0.80 0.82 0.81 9327
32
+ ORG 0.78 0.78 0.78 7547
33
+ PER 0.91 0.90 0.90 11204
34
+ micro avg 0.84 0.84 0.84 28078
35
+ macro avg 0.83 0.84 0.83 28078
36
+ weighted avg 0.84 0.84 0.84 28078
37
+ Classification Report:
38
+ precision recall f1-score support
39
+ LOC 0.80 0.81 0.80 9327
40
+ ORG 0.77 0.79 0.78 7547
41
+ PER 0.90 0.90 0.90 11204
42
+ micro avg 0.83 0.84 0.84 28078
43
+ macro avg 0.82 0.83 0.83 28078
44
+ weighted avg 0.83 0.84 0.84 28078
45
+ Classification Report:
46
+ precision recall f1-score support
47
+ LOC 0.79 0.81 0.80 9327
48
+ ORG 0.77 0.80 0.78 7547
49
+ PER 0.89 0.91 0.90 11204
50
+ micro avg 0.82 0.85 0.84 28078
51
+ macro avg 0.82 0.84 0.83 28078
52
+ weighted avg 0.83 0.85 0.84 28078
53
+ Classification Report:
54
+ precision recall f1-score support
55
+ LOC 0.81 0.81 0.81 9327
56
+ ORG 0.76 0.80 0.78 7547
57
+ PER 0.91 0.90 0.91 11204
58
+ micro avg 0.84 0.84 0.84 28078
59
+ macro avg 0.83 0.84 0.83 28078
60
+ weighted avg 0.84 0.84 0.84 28078
61
+ Classification Report:
62
+ precision recall f1-score support
63
+ LOC 0.81 0.79 0.80 9327
64
+ ORG 0.76 0.80 0.78 7547
65
+ PER 0.90 0.91 0.90 11204
66
+ micro avg 0.83 0.84 0.84 28078
67
+ macro avg 0.83 0.83 0.83 28078
68
+ weighted avg 0.83 0.84 0.84 28078
69
+ Classification Report:
70
+ precision recall f1-score support
71
+ LOC 0.80 0.80 0.80 9327
72
+ ORG 0.77 0.79 0.78 7547
73
+ PER 0.91 0.91 0.91 11204
74
+ micro avg 0.83 0.84 0.84 28078
75
+ macro avg 0.83 0.83 0.83 28078
wandb/run-20240917_140929-fx32bo52/files/requirements.txt ADDED
@@ -0,0 +1,313 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ AutoROM.accept-rom-license==0.6.1
2
+ AutoROM==0.6.1
3
+ Farama-Notifications==0.0.4
4
+ GitPython==3.1.43
5
+ Jinja2==3.1.4
6
+ Mako==1.3.5
7
+ Markdown==3.7
8
+ MarkupSafe==2.1.5
9
+ PyYAML==6.0.2
10
+ Pygments==2.18.0
11
+ QtPy==2.4.1
12
+ SQLAlchemy==2.0.32
13
+ Send2Trash==1.8.3
14
+ Shimmy==1.3.0
15
+ Werkzeug==3.0.4
16
+ absl-py==2.1.0
17
+ accelerate==0.33.0
18
+ aiohappyeyeballs==2.3.5
19
+ aiohttp-retry==2.8.3
20
+ aiohttp==3.10.3
21
+ aiosignal==1.3.1
22
+ ale-py==0.8.1
23
+ alembic==1.13.2
24
+ amqp==5.2.0
25
+ annotated-types==0.7.0
26
+ antlr4-python3-runtime==4.9.3
27
+ anyio==4.4.0
28
+ appdirs==1.4.4
29
+ argon2-cffi-bindings==21.2.0
30
+ argon2-cffi==23.1.0
31
+ arrow==1.3.0
32
+ asttokens==2.4.1
33
+ async-generator==1.10
34
+ async-lru==2.0.4
35
+ asyncssh==2.17.0
36
+ atpublic==5.0
37
+ attrs==24.2.0
38
+ audioread==3.0.1
39
+ babel==2.16.0
40
+ beautifulsoup4==4.12.3
41
+ billiard==4.2.0
42
+ bitsandbytes==0.43.3
43
+ bleach==6.1.0
44
+ blis==0.7.11
45
+ catalogue==2.0.10
46
+ celery==5.4.0
47
+ certifi==2024.7.4
48
+ certipy==0.1.3
49
+ cffi==1.17.0
50
+ charset-normalizer==3.3.2
51
+ click-didyoumean==0.3.1
52
+ click-plugins==1.1.1
53
+ click-repl==0.3.0
54
+ click==8.1.7
55
+ cloudpathlib==0.18.1
56
+ cloudpickle==3.0.0
57
+ colorama==0.4.6
58
+ coloredlogs==15.0.1
59
+ comm==0.2.2
60
+ conda-store==2024.6.1
61
+ confection==0.1.5
62
+ configobj==5.0.8
63
+ contourpy==1.2.1
64
+ cryptography==43.0.0
65
+ cuda-python==11.8.3
66
+ cycler==0.12.1
67
+ cymem==2.0.8
68
+ datasets==2.20.0
69
+ debugpy==1.8.5
70
+ decorator==5.1.1
71
+ defusedxml==0.7.1
72
+ dictdiffer==0.9.0
73
+ dill==0.3.8
74
+ diskcache==5.6.3
75
+ distro==1.9.0
76
+ docker-pycreds==0.4.0
77
+ docstring_parser==0.16
78
+ dpath==2.2.0
79
+ dulwich==0.22.1
80
+ dvc-data==3.16.5
81
+ dvc-http==2.32.0
82
+ dvc-objects==5.1.0
83
+ dvc-render==1.0.2
84
+ dvc-studio-client==0.21.0
85
+ dvc-task==0.4.0
86
+ dvc==3.55.2
87
+ en-core-web-lg==3.7.1
88
+ en-core-web-sm==3.7.1
89
+ entrypoints==0.4
90
+ evaluate==0.4.2
91
+ executing==2.0.1
92
+ fastjsonschema==2.20.0
93
+ filelock==3.15.4
94
+ flatbuffers==24.3.25
95
+ flatten-dict==0.4.2
96
+ flufl.lock==8.1.0
97
+ fonttools==4.53.1
98
+ fqdn==1.5.1
99
+ frozenlist==1.4.1
100
+ fsspec==2024.5.0
101
+ funcy==2.0
102
+ gitdb==4.0.11
103
+ grandalf==0.8
104
+ greenlet==3.0.3
105
+ grpcio==1.66.0
106
+ gto==1.7.1
107
+ gymnasium==0.29.1
108
+ h11==0.14.0
109
+ httpcore==1.0.5
110
+ httpx==0.27.0
111
+ huggingface-hub==0.24.5
112
+ humanfriendly==10.0
113
+ hydra-core==1.3.2
114
+ idna==3.7
115
+ importlib_resources==6.4.5
116
+ ipykernel==6.29.5
117
+ ipython==8.26.0
118
+ ipywidgets==8.1.3
119
+ isoduration==20.11.0
120
+ iterative-telemetry==0.0.8
121
+ jedi==0.19.1
122
+ joblib==1.4.2
123
+ json5==0.9.25
124
+ jsonpointer==3.0.0
125
+ jsonschema-specifications==2023.12.1
126
+ jsonschema==4.23.0
127
+ jupyter-console==6.6.3
128
+ jupyter-events==0.10.0
129
+ jupyter-launcher-shortcuts==4.0.3
130
+ jupyter-lsp==2.2.5
131
+ jupyter-telemetry==0.1.0
132
+ jupyter==1.0.0
133
+ jupyter_client==8.6.2
134
+ jupyter_core==5.7.2
135
+ jupyter_server==2.14.2
136
+ jupyter_server_terminals==0.5.3
137
+ jupyterhub==1.5.1
138
+ jupyterlab-conda-store==2024.6.1
139
+ jupyterlab==4.2.4
140
+ jupyterlab_pygments==0.3.0
141
+ jupyterlab_server==2.27.3
142
+ jupyterlab_widgets==3.0.11
143
+ kiwisolver==1.4.5
144
+ kombu==5.4.0
145
+ lab==8.2
146
+ langcodes==3.4.0
147
+ language_data==1.2.0
148
+ lazy_loader==0.4
149
+ librosa==0.10.2.post1
150
+ llvmlite==0.43.0
151
+ lxml==5.3.0
152
+ marisa-trie==1.2.0
153
+ markdown-it-py==3.0.0
154
+ matplotlib-inline==0.1.7
155
+ matplotlib==3.9.1.post1
156
+ mdurl==0.1.2
157
+ mistune==3.0.2
158
+ mpmath==1.3.0
159
+ msgpack==1.1.0
160
+ multidict==6.0.5
161
+ multiprocess==0.70.16
162
+ murmurhash==1.0.10
163
+ nb_conda_store_kernels==0.1.5
164
+ nbclient==0.10.0
165
+ nbconvert==7.16.4
166
+ nbformat==5.10.4
167
+ nest-asyncio==1.6.0
168
+ networkx==3.3
169
+ nodejs==0.1.1
170
+ notebook==7.2.1
171
+ notebook_shim==0.2.4
172
+ numba==0.60.0
173
+ numpy==1.26.4
174
+ nvidia-cublas-cu12==12.1.3.1
175
+ nvidia-cuda-cupti-cu12==12.1.105
176
+ nvidia-cuda-nvrtc-cu12==12.1.105
177
+ nvidia-cuda-runtime-cu12==12.1.105
178
+ nvidia-cudnn-cu12==8.9.2.26
179
+ nvidia-cudnn-cu12==9.1.0.70
180
+ nvidia-cufft-cu12==11.0.2.54
181
+ nvidia-curand-cu12==10.3.2.106
182
+ nvidia-cusolver-cu12==11.4.5.107
183
+ nvidia-cusparse-cu12==12.1.0.106
184
+ nvidia-nccl-cu12==2.19.3
185
+ nvidia-nccl-cu12==2.20.5
186
+ nvidia-nvjitlink-cu12==12.6.20
187
+ nvidia-nvtx-cu12==12.1.105
188
+ oauthlib==3.2.2
189
+ omegaconf==2.3.0
190
+ onnx==1.16.2
191
+ onnxruntime==1.19.0
192
+ opencv-python==4.10.0.84
193
+ optimum==1.21.4
194
+ optional-django==0.1.0
195
+ orjson==3.10.7
196
+ overrides==7.7.0
197
+ packaging==24.1
198
+ pamela==1.2.0
199
+ pandas==2.2.2
200
+ pandocfilters==1.5.1
201
+ parso==0.8.4
202
+ pathspec==0.12.1
203
+ peft==0.12.0
204
+ pexpect==4.9.0
205
+ pillow==10.4.0
206
+ pip==23.1.2
207
+ platformdirs==3.11.0
208
+ pooch==1.8.2
209
+ preshed==3.0.9
210
+ prometheus_client==0.20.0
211
+ prompt_toolkit==3.0.47
212
+ protobuf==5.27.3
213
+ psutil==6.0.0
214
+ ptyprocess==0.7.0
215
+ pure_eval==0.2.3
216
+ pyOpenSSL==24.2.1
217
+ pyarrow-hotfix==0.6
218
+ pyarrow==17.0.0
219
+ pycparser==2.22
220
+ pydantic==2.8.2
221
+ pydantic_core==2.20.1
222
+ pydot==3.0.1
223
+ pygame==2.6.0
224
+ pygit2==1.15.1
225
+ pygtrie==2.5.0
226
+ pyparsing==3.1.2
227
+ python-dateutil==2.9.0.post0
228
+ python-dotenv==1.0.1
229
+ python-json-logger==2.0.7
230
+ pytz==2024.1
231
+ pyzmq==26.1.0
232
+ qtconsole==5.5.2
233
+ referencing==0.35.1
234
+ regex==2024.7.24
235
+ requests==2.32.3
236
+ rfc3339-validator==0.1.4
237
+ rfc3986-validator==0.1.1
238
+ rich==13.7.1
239
+ rpds-py==0.20.0
240
+ ruamel.yaml.clib==0.2.8
241
+ ruamel.yaml==0.18.6
242
+ safetensors==0.4.4
243
+ scikit-learn==1.5.1
244
+ scipy==1.14.0
245
+ scmrepo==3.3.7
246
+ seaborn==0.13.2
247
+ semver==3.0.2
248
+ sentencepiece==0.2.0
249
+ sentry-sdk==2.14.0
250
+ seqeval==1.2.2
251
+ setproctitle==1.3.3
252
+ setuptools==65.5.0
253
+ shellingham==1.5.4
254
+ shortuuid==1.0.13
255
+ shtab==1.7.1
256
+ simplejson==3.19.2
257
+ six==1.16.0
258
+ smart-open==7.0.4
259
+ smmap==5.0.1
260
+ sniffio==1.3.1
261
+ soundfile==0.12.1
262
+ soupsieve==2.5
263
+ sox==1.5.0
264
+ soxr==0.5.0.post1
265
+ spacy-legacy==3.0.12
266
+ spacy-loggers==1.0.5
267
+ spacy==3.7.5
268
+ sqltrie==0.11.1
269
+ srsly==2.4.8
270
+ stable_baselines3==2.3.2
271
+ stack-data==0.6.3
272
+ sympy==1.13.2
273
+ tabulate==0.9.0
274
+ tensorboard-data-server==0.7.2
275
+ tensorboard==2.17.1
276
+ terminado==0.18.1
277
+ thinc==8.2.5
278
+ threadpoolctl==3.5.0
279
+ tinycss2==1.3.0
280
+ tokenizers==0.19.1
281
+ tomlkit==0.13.2
282
+ torch==2.4.1
283
+ torchaudio==2.4.1
284
+ torchvision==0.19.1
285
+ tornado==6.4.1
286
+ tqdm==4.66.5
287
+ traitlets==5.14.3
288
+ transformers==4.43.4
289
+ triton==3.0.0
290
+ trl==0.9.6
291
+ txt2tags==3.9
292
+ typer==0.12.3
293
+ types-python-dateutil==2.9.0.20240316
294
+ typing_extensions==4.12.2
295
+ tyro==0.8.6
296
+ tzdata==2024.1
297
+ uri-template==1.3.0
298
+ urllib3==2.2.2
299
+ vine==5.1.0
300
+ voluptuous==0.15.2
301
+ wandb==0.17.9
302
+ wasabi==1.1.3
303
+ wcwidth==0.2.13
304
+ weasel==0.4.1
305
+ webcolors==24.8.0
306
+ webencodings==0.5.1
307
+ websocket-client==1.8.0
308
+ widgetsnbextension==4.0.11
309
+ wrapt==1.16.0
310
+ xlrd==2.0.1
311
+ xxhash==3.4.1
312
+ yarl==1.9.4
313
+ zc.lockfile==3.0.post1
wandb/run-20240917_140929-fx32bo52/files/wandb-metadata.json ADDED
@@ -0,0 +1,281 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.18.0-513.24.1.el8_9.x86_64-x86_64-with-glibc2.28",
3
+ "python": "3.11.4",
4
+ "heartbeatAt": "2024-09-17T13:09:30.062514",
5
+ "startedAt": "2024-09-17T13:09:29.455485",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "<python with no main file>",
11
+ "codePathLocal": null,
12
+ "host": "codon-gpu-014.ebi.ac.uk",
13
+ "username": "amrufai",
14
+ "executable": "/nfs/production/literature/amina-mardiyyah/envs/llm-prompt/bin/python",
15
+ "cpu_count": 48,
16
+ "cpu_count_logical": 48,
17
+ "cpu_freq": {
18
+ "current": 2758.5387083333335,
19
+ "min": 0.0,
20
+ "max": 0.0
21
+ },
22
+ "cpu_freq_per_core": [
23
+ {
24
+ "current": 2800.0,
25
+ "min": 0.0,
26
+ "max": 0.0
27
+ },
28
+ {
29
+ "current": 2800.0,
30
+ "min": 0.0,
31
+ "max": 0.0
32
+ },
33
+ {
34
+ "current": 2800.0,
35
+ "min": 0.0,
36
+ "max": 0.0
37
+ },
38
+ {
39
+ "current": 2800.0,
40
+ "min": 0.0,
41
+ "max": 0.0
42
+ },
43
+ {
44
+ "current": 2800.0,
45
+ "min": 0.0,
46
+ "max": 0.0
47
+ },
48
+ {
49
+ "current": 2800.0,
50
+ "min": 0.0,
51
+ "max": 0.0
52
+ },
53
+ {
54
+ "current": 2800.0,
55
+ "min": 0.0,
56
+ "max": 0.0
57
+ },
58
+ {
59
+ "current": 2800.0,
60
+ "min": 0.0,
61
+ "max": 0.0
62
+ },
63
+ {
64
+ "current": 2800.0,
65
+ "min": 0.0,
66
+ "max": 0.0
67
+ },
68
+ {
69
+ "current": 2800.0,
70
+ "min": 0.0,
71
+ "max": 0.0
72
+ },
73
+ {
74
+ "current": 2800.0,
75
+ "min": 0.0,
76
+ "max": 0.0
77
+ },
78
+ {
79
+ "current": 2800.0,
80
+ "min": 0.0,
81
+ "max": 0.0
82
+ },
83
+ {
84
+ "current": 2800.0,
85
+ "min": 0.0,
86
+ "max": 0.0
87
+ },
88
+ {
89
+ "current": 2800.0,
90
+ "min": 0.0,
91
+ "max": 0.0
92
+ },
93
+ {
94
+ "current": 2800.0,
95
+ "min": 0.0,
96
+ "max": 0.0
97
+ },
98
+ {
99
+ "current": 2800.0,
100
+ "min": 0.0,
101
+ "max": 0.0
102
+ },
103
+ {
104
+ "current": 2800.0,
105
+ "min": 0.0,
106
+ "max": 0.0
107
+ },
108
+ {
109
+ "current": 2800.0,
110
+ "min": 0.0,
111
+ "max": 0.0
112
+ },
113
+ {
114
+ "current": 2800.0,
115
+ "min": 0.0,
116
+ "max": 0.0
117
+ },
118
+ {
119
+ "current": 2800.0,
120
+ "min": 0.0,
121
+ "max": 0.0
122
+ },
123
+ {
124
+ "current": 809.858,
125
+ "min": 0.0,
126
+ "max": 0.0
127
+ },
128
+ {
129
+ "current": 2800.0,
130
+ "min": 0.0,
131
+ "max": 0.0
132
+ },
133
+ {
134
+ "current": 2800.0,
135
+ "min": 0.0,
136
+ "max": 0.0
137
+ },
138
+ {
139
+ "current": 2800.0,
140
+ "min": 0.0,
141
+ "max": 0.0
142
+ },
143
+ {
144
+ "current": 2800.0,
145
+ "min": 0.0,
146
+ "max": 0.0
147
+ },
148
+ {
149
+ "current": 2800.0,
150
+ "min": 0.0,
151
+ "max": 0.0
152
+ },
153
+ {
154
+ "current": 2800.0,
155
+ "min": 0.0,
156
+ "max": 0.0
157
+ },
158
+ {
159
+ "current": 2800.0,
160
+ "min": 0.0,
161
+ "max": 0.0
162
+ },
163
+ {
164
+ "current": 2800.0,
165
+ "min": 0.0,
166
+ "max": 0.0
167
+ },
168
+ {
169
+ "current": 2800.0,
170
+ "min": 0.0,
171
+ "max": 0.0
172
+ },
173
+ {
174
+ "current": 2800.0,
175
+ "min": 0.0,
176
+ "max": 0.0
177
+ },
178
+ {
179
+ "current": 2800.0,
180
+ "min": 0.0,
181
+ "max": 0.0
182
+ },
183
+ {
184
+ "current": 2800.0,
185
+ "min": 0.0,
186
+ "max": 0.0
187
+ },
188
+ {
189
+ "current": 2800.0,
190
+ "min": 0.0,
191
+ "max": 0.0
192
+ },
193
+ {
194
+ "current": 2800.0,
195
+ "min": 0.0,
196
+ "max": 0.0
197
+ },
198
+ {
199
+ "current": 2800.0,
200
+ "min": 0.0,
201
+ "max": 0.0
202
+ },
203
+ {
204
+ "current": 2800.0,
205
+ "min": 0.0,
206
+ "max": 0.0
207
+ },
208
+ {
209
+ "current": 2800.0,
210
+ "min": 0.0,
211
+ "max": 0.0
212
+ },
213
+ {
214
+ "current": 2800.0,
215
+ "min": 0.0,
216
+ "max": 0.0
217
+ },
218
+ {
219
+ "current": 2800.0,
220
+ "min": 0.0,
221
+ "max": 0.0
222
+ },
223
+ {
224
+ "current": 2800.0,
225
+ "min": 0.0,
226
+ "max": 0.0
227
+ },
228
+ {
229
+ "current": 2800.0,
230
+ "min": 0.0,
231
+ "max": 0.0
232
+ },
233
+ {
234
+ "current": 2800.0,
235
+ "min": 0.0,
236
+ "max": 0.0
237
+ },
238
+ {
239
+ "current": 2800.0,
240
+ "min": 0.0,
241
+ "max": 0.0
242
+ },
243
+ {
244
+ "current": 2800.0,
245
+ "min": 0.0,
246
+ "max": 0.0
247
+ },
248
+ {
249
+ "current": 2800.0,
250
+ "min": 0.0,
251
+ "max": 0.0
252
+ },
253
+ {
254
+ "current": 2800.0,
255
+ "min": 0.0,
256
+ "max": 0.0
257
+ },
258
+ {
259
+ "current": 2800.0,
260
+ "min": 0.0,
261
+ "max": 0.0
262
+ }
263
+ ],
264
+ "disk": {
265
+ "/": {
266
+ "total": 47.760292053222656,
267
+ "used": 15.848091125488281
268
+ }
269
+ },
270
+ "gpu": "NVIDIA A100 80GB PCIe",
271
+ "gpu_count": 1,
272
+ "gpu_devices": [
273
+ {
274
+ "name": "NVIDIA A100 80GB PCIe",
275
+ "memory_total": 85899345920
276
+ }
277
+ ],
278
+ "memory": {
279
+ "total": 502.83758544921875
280
+ }
281
+ }
wandb/run-20240917_140929-fx32bo52/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 0.0225, "train/grad_norm": 0.9898543953895569, "train/learning_rate": 2.332778702163062e-06, "train/epoch": 8.999287241625089, "train/global_step": 6313, "_timestamp": 1726583402.385004, "_runtime": 4832.899541139603, "_step": 17, "eval/loss": 0.18012596666812897, "eval/f1": 0.8372712264569614, "eval/precision": 0.8346487347372147, "eval/recall": 0.8399102500178075, "eval/accuracy": 0.9647665678680793, "eval/runtime": 74.7153, "eval/samples_per_second": 150.197, "eval/steps_per_second": 18.778}
wandb/run-20240917_140929-fx32bo52/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240917_140929-fx32bo52/logs/debug.log ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-09-17 14:09:29,476 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Current SDK version is 0.17.9
2
+ 2024-09-17 14:09:29,476 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Configure stats pid to 1271154
3
+ 2024-09-17 14:09:29,476 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Loading settings from /homes/amrufai/.config/wandb/settings
4
+ 2024-09-17 14:09:29,476 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Loading settings from /nfs/production/literature/amina-mardiyyah/wandb/settings
5
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Loading settings from environment variables: {}
6
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Inferring run settings from compute environment: {'program': '<python with no main file>'}
7
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Applying login settings: {}
8
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_setup.py:_flush():77] Applying login settings: {}
9
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:_log_setup():524] Logging user logs to /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_140929-fx32bo52/logs/debug.log
10
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:_log_setup():525] Logging internal logs to /nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/wandb/run-20240917_140929-fx32bo52/logs/debug-internal.log
11
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:_jupyter_setup():470] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7fbb5e686b90>
12
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:init():608] calling init triggers
13
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:init():615] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:init():658] starting backend
16
+ 2024-09-17 14:09:29,477 INFO MainThread:1271154 [wandb_init.py:init():662] setting up manager
17
+ 2024-09-17 14:09:29,479 INFO MainThread:1271154 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-09-17 14:09:29,485 INFO MainThread:1271154 [wandb_init.py:init():670] backend started and connected
19
+ 2024-09-17 14:09:29,490 INFO MainThread:1271154 [wandb_run.py:_label_probe_notebook():1344] probe notebook
20
+ 2024-09-17 14:09:29,504 INFO MainThread:1271154 [wandb_run.py:_label_probe_notebook():1354] Unable to probe notebook: 'NoneType' object has no attribute 'get'
21
+ 2024-09-17 14:09:29,504 INFO MainThread:1271154 [wandb_init.py:init():768] updated telemetry
22
+ 2024-09-17 14:09:29,514 INFO MainThread:1271154 [wandb_init.py:init():801] communicating run to backend with 90.0 second timeout
23
+ 2024-09-17 14:09:29,907 INFO MainThread:1271154 [wandb_init.py:init():852] starting run threads in backend
24
+ 2024-09-17 14:09:30,803 INFO MainThread:1271154 [wandb_run.py:_console_start():2465] atexit reg
25
+ 2024-09-17 14:09:30,803 INFO MainThread:1271154 [wandb_run.py:_redirect():2311] redirect: wrap_raw
26
+ 2024-09-17 14:09:30,803 INFO MainThread:1271154 [wandb_run.py:_redirect():2376] Wrapping output streams.
27
+ 2024-09-17 14:09:30,803 INFO MainThread:1271154 [wandb_run.py:_redirect():2401] Redirects installed.
28
+ 2024-09-17 14:09:30,811 INFO MainThread:1271154 [wandb_init.py:init():895] run started, returning control to user process
29
+ 2024-09-17 14:09:30,821 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
30
+ 2024-09-17 14:09:30,821 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
31
+ 2024-09-17 14:11:21,293 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
32
+ 2024-09-17 14:11:21,353 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
33
+ 2024-09-17 14:11:21,353 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
34
+ 2024-09-17 14:11:23,083 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
35
+ 2024-09-17 14:11:23,089 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
36
+ 2024-09-17 14:11:23,090 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
37
+ 2024-09-17 14:11:28,594 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
38
+ 2024-09-17 14:11:36,241 INFO MainThread:1271154 [wandb_run.py:_config_callback():1392] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['XLMRobertaForTokenClassification'], 'finetuning_task': None, 'id2label': {0: 'O', 1: 'B-PER', 2: 'I-PER', 3: 'B-ORG', 4: 'I-ORG', 5: 'B-LOC', 6: 'I-LOC'}, 'label2id': {'O': 0, 'B-PER': 1, 'I-PER': 2, 'B-ORG': 3, 'I-ORG': 4, 'B-LOC': 5, 'I-LOC': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0', 'transformers_version': '4.43.4', 'adapters': {'adapters': {}, 'config_map': {}, 'fusion_config_map': {}, 'fusions': {}}, 'gradient_checkpointing': False, 'model_type': 'xlm-roberta', 'output_past': True, 'vocab_size': 250002, 'hidden_size': 1024, 'num_hidden_layers': 24, 'num_attention_heads': 16, 'hidden_act': 'gelu', 'intermediate_size': 4096, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 514, 'type_vocab_size': 1, 'initializer_range': 0.02, 'layer_norm_eps': 1e-05, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 2e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date/logs', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 3407, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/nfs/production/literature/amina-mardiyyah/NER/ewc_stabilised_no_date', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'eval_f1', 'greater_is_better': True, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'eval_use_gather_object': False}
39
+ 2024-09-17 14:11:36,243 INFO MainThread:1271154 [wandb_config.py:__setitem__():154] config set model/num_parameters = 558848007 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7fbad9edf090>>
40
+ 2024-09-17 14:11:36,243 INFO MainThread:1271154 [wandb_run.py:_config_callback():1392] config_cb model/num_parameters 558848007 None
41
+ 2024-09-17 15:31:39,471 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
42
+ 2024-09-17 15:31:39,472 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
43
+ 2024-09-17 15:31:41,931 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
44
+ 2024-09-17 15:31:41,938 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
45
+ 2024-09-17 15:31:41,938 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
46
+ 2024-09-17 15:31:49,020 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
47
+ 2024-09-17 15:31:49,568 INFO MainThread:1271154 [jupyter.py:save_ipynb():372] not saving jupyter notebook
48
+ 2024-09-17 15:31:49,568 INFO MainThread:1271154 [wandb_init.py:_pause_backend():435] pausing backend
49
+ 2024-09-17 15:32:22,919 INFO MainThread:1271154 [wandb_init.py:_resume_backend():440] resuming backend
wandb/run-20240917_140929-fx32bo52/run-fx32bo52.wandb ADDED
Binary file (362 kB). View file