diff --git a/mup-1b-100m-e3-full-gpt2lmmup.json b/mup-1b-100m-e3-full-gpt2lmmup.json new file mode 100644 index 0000000000000000000000000000000000000000..a78edc925629cc38586a2dd4fbe7d202aef93d99 --- /dev/null +++ b/mup-1b-100m-e3-full-gpt2lmmup.json @@ -0,0 +1 @@ +{"eval_loss": 10.503863334655762, "eval_runtime": 35656.762, "eval_samples_per_second": 4.751, "eval_steps_per_second": 0.148, "epoch": 0.01} \ No newline at end of file diff --git a/mup-200m-100m-e3-full-gpt2lmmup.json b/mup-200m-100m-e3-full-gpt2lmmup.json new file mode 100644 index 0000000000000000000000000000000000000000..f797a76e1d8bba8c1e07d8cbdb6b3a1e6f12129d --- /dev/null +++ b/mup-200m-100m-e3-full-gpt2lmmup.json @@ -0,0 +1 @@ +{"eval_loss": 10.863893508911133, "eval_runtime": 35678.3953, "eval_samples_per_second": 4.748, "eval_steps_per_second": 0.148} \ No newline at end of file diff --git a/mup-200m-100m-e3/config.json b/mup-200m-100m-e3/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3c89de94cdc339ba7df6839701d4fa04b955a065 --- /dev/null +++ b/mup-200m-100m-e3/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.1, + "intermediate_size": 12288, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 3072, + "n_head": 64, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-200m-100m-e3/pytorch_model.bin b/mup-200m-100m-e3/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..02364d485b7fc5f2cfe85a21dafb6f526be07c2a --- /dev/null +++ b/mup-200m-100m-e3/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:600250335c93538740793cc7e8a4682defb6f7e202a3641016cf4ee1754bbe75 +size 6080539805 diff --git a/mup-200m-100m-e3/training_args.bin b/mup-200m-100m-e3/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0e8deb6debfd686b602f8dfe5d2c7b05a1f165bd --- /dev/null +++ b/mup-200m-100m-e3/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a9b9d007760086d2e55ee15759f906d50d344e948eb0bca87f0e434c9a4dd63 +size 3387 diff --git a/mup-2b-100m-e3-full-gpt2lmmup.json b/mup-2b-100m-e3-full-gpt2lmmup.json new file mode 100644 index 0000000000000000000000000000000000000000..21cd2466f8af116d4bb9a8164c00e68695d24e05 --- /dev/null +++ b/mup-2b-100m-e3-full-gpt2lmmup.json @@ -0,0 +1 @@ +{"eval_loss": 9.270681381225586, "eval_runtime": 46206.7705, "eval_samples_per_second": 3.666, "eval_steps_per_second": 0.229} \ No newline at end of file diff --git a/mup-2b-100m-e3/checkpoint-100/config.json b/mup-2b-100m-e3/checkpoint-100/config.json new file mode 100644 index 0000000000000000000000000000000000000000..206494207d182569dd9eedb97a900ced82bdf150 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-100/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 13312, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 3328, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-2b-100m-e3/checkpoint-100/optimizer.pt b/mup-2b-100m-e3/checkpoint-100/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..91c515eca1ef3b95304cef2498c5529b7b08ce31 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-100/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:859a1cea6ab663b2651f6b47d6e85d9acb2a67b1ddf7200f7e5c350ab22c1f3b +size 14128712873 diff --git a/mup-2b-100m-e3/checkpoint-100/pytorch_model.bin b/mup-2b-100m-e3/checkpoint-100/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..f3f8eea86f0476580681d90ce09683d450abdcfb --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-100/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9aff050f6759cb9c9bd9fb82d5908dc746bd5fa9f0754b969286830370959df2 +size 7076932005 diff --git a/mup-2b-100m-e3/checkpoint-100/rng_state.pth b/mup-2b-100m-e3/checkpoint-100/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..213cb727666a3e0291a137e9f014e650f3be8706 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-100/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c85e1a7fcee9f0a4e6f8ebf1dc4a63a8c96bd1dd0a67ff0ff15c506aae95304 +size 14575 diff --git a/mup-2b-100m-e3/checkpoint-100/scheduler.pt b/mup-2b-100m-e3/checkpoint-100/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..4e56abbb618df6be14e8e38fac0265d36a57a011 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-100/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f41127ce2e7b57031bf845923aadf69d1d932623d3cc306538c8cffe3f95b12f +size 691 diff --git a/mup-2b-100m-e3/checkpoint-100/trainer_state.json b/mup-2b-100m-e3/checkpoint-100/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..6a2b08bfcbd1dd625bfcdbcacb742dad36486b1f --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-100/trainer_state.json @@ -0,0 +1,15 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.2631010986526346, + "global_step": 100, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [], + "max_steps": 380, + "num_train_epochs": 1, + "total_flos": 2.50936282841088e+17, + "trial_name": null, + "trial_params": null +} diff --git a/mup-2b-100m-e3/checkpoint-100/training_args.bin b/mup-2b-100m-e3/checkpoint-100/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..aa4c4b243cd83ca72ae78bb76bbca13363e894dc --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-100/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a529bfebc81b3f76599a2706a7dedaa2e159f90434d17d7113ba289ad90ba128 +size 3387 diff --git a/mup-2b-100m-e3/checkpoint-200/config.json b/mup-2b-100m-e3/checkpoint-200/config.json new file mode 100644 index 0000000000000000000000000000000000000000..206494207d182569dd9eedb97a900ced82bdf150 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-200/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 13312, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 3328, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-2b-100m-e3/checkpoint-200/optimizer.pt b/mup-2b-100m-e3/checkpoint-200/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..caba7a46c14ad8b4691520849712958256926953 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-200/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc8643fa7f495fb7235042b7a64c424947f35a3834c2e7ecaf81a5faff11ddb3 +size 14128712873 diff --git a/mup-2b-100m-e3/checkpoint-200/pytorch_model.bin b/mup-2b-100m-e3/checkpoint-200/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..238d32bb7eca9f593a3bac0b1c84b9c3e6aa3bb4 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-200/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64362990e2eaa64481e967769bacc656341aa1ab8490f6c9a62abaa928ac7f9d +size 7076932005 diff --git a/mup-2b-100m-e3/checkpoint-200/rng_state.pth b/mup-2b-100m-e3/checkpoint-200/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..9c6dbf95eb9baebca965add5859888c56c8ed597 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-200/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:850a75d03b68709ebdb3b0429ccf1c2bfbc264ba389a899cc590928049b771ed +size 14575 diff --git a/mup-2b-100m-e3/checkpoint-200/scheduler.pt b/mup-2b-100m-e3/checkpoint-200/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f51d475b1ca12195a990369a485e37ac3cfffe72 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-200/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ff11a668fd015bf8fec7ef706376a383794600ae7f4f86f3b9aeffe27fd39be +size 691 diff --git a/mup-2b-100m-e3/checkpoint-200/trainer_state.json b/mup-2b-100m-e3/checkpoint-200/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..402a8b19a70c3f0a63f4f211c680c3f5bc507c7c --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-200/trainer_state.json @@ -0,0 +1,15 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5262021973052692, + "global_step": 200, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [], + "max_steps": 380, + "num_train_epochs": 1, + "total_flos": 5.01872565682176e+17, + "trial_name": null, + "trial_params": null +} diff --git a/mup-2b-100m-e3/checkpoint-200/training_args.bin b/mup-2b-100m-e3/checkpoint-200/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..aa4c4b243cd83ca72ae78bb76bbca13363e894dc --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-200/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a529bfebc81b3f76599a2706a7dedaa2e159f90434d17d7113ba289ad90ba128 +size 3387 diff --git a/mup-2b-100m-e3/checkpoint-300/config.json b/mup-2b-100m-e3/checkpoint-300/config.json new file mode 100644 index 0000000000000000000000000000000000000000..206494207d182569dd9eedb97a900ced82bdf150 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-300/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 13312, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 3328, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-2b-100m-e3/checkpoint-300/optimizer.pt b/mup-2b-100m-e3/checkpoint-300/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7a510a1b49a2b71d0a9e735c7e5c51e4e337d82c --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-300/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83e214ea9a2859b68902931336a551a4a675fe79c227fa5b6bf330b7d270514b +size 14128712873 diff --git a/mup-2b-100m-e3/checkpoint-300/pytorch_model.bin b/mup-2b-100m-e3/checkpoint-300/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..d9af1b97b642155ed4b934411e3d2d34bcba0f5f --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-300/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0b6c22fcc540890eee68ac4649802608865e27c9584d1917a079d4034b8ab41 +size 7076932005 diff --git a/mup-2b-100m-e3/checkpoint-300/rng_state.pth b/mup-2b-100m-e3/checkpoint-300/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..7b18da519801ea41898df55ce968ba8c412bdbd3 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-300/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37467b99308898a6ecf9cd44334c15ea2bf2d409cc4222f629e0118cdb7e93e6 +size 14575 diff --git a/mup-2b-100m-e3/checkpoint-300/scheduler.pt b/mup-2b-100m-e3/checkpoint-300/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..9f8a9962b2b1866f629b67355349edba2c7229fb --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-300/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30fd152d86dafeffe2455270e22218f0b6cab42a3c9b31a01b2bc6d8ceadc7c4 +size 691 diff --git a/mup-2b-100m-e3/checkpoint-300/trainer_state.json b/mup-2b-100m-e3/checkpoint-300/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..707f1b3ed20d9adef0955189451b2e813349b388 --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-300/trainer_state.json @@ -0,0 +1,15 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.7893032959579038, + "global_step": 300, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [], + "max_steps": 380, + "num_train_epochs": 1, + "total_flos": 7.52808848523264e+17, + "trial_name": null, + "trial_params": null +} diff --git a/mup-2b-100m-e3/checkpoint-300/training_args.bin b/mup-2b-100m-e3/checkpoint-300/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..aa4c4b243cd83ca72ae78bb76bbca13363e894dc --- /dev/null +++ b/mup-2b-100m-e3/checkpoint-300/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a529bfebc81b3f76599a2706a7dedaa2e159f90434d17d7113ba289ad90ba128 +size 3387 diff --git a/mup-2b-100m-e3/config.json b/mup-2b-100m-e3/config.json new file mode 100644 index 0000000000000000000000000000000000000000..206494207d182569dd9eedb97a900ced82bdf150 --- /dev/null +++ b/mup-2b-100m-e3/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 13312, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 3328, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-2b-100m-e3/pytorch_model.bin b/mup-2b-100m-e3/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..67bf658114f0bbb790186b7f5b17e9c3b8033061 --- /dev/null +++ b/mup-2b-100m-e3/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c72ffdc020dcc3f482979a1b900d14d4832f0f3f07845f1feaba439e3d93b68 +size 7076932005 diff --git a/mup-2b-100m-e3/training_args.bin b/mup-2b-100m-e3/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..81c35996da1f93cb7bd14f67fd12d176a6fcf80c --- /dev/null +++ b/mup-2b-100m-e3/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0360c4e8f458bf1a55bfbc7c49d3cd1b51ecf6c15f5985ea0900cb0d4d27bfb +size 3387 diff --git a/mup-2b5-100m-e3-full-gpt2lmmup.json b/mup-2b5-100m-e3-full-gpt2lmmup.json new file mode 100644 index 0000000000000000000000000000000000000000..babff691199e46285cb34ffe3f1dd30da120e317 --- /dev/null +++ b/mup-2b5-100m-e3-full-gpt2lmmup.json @@ -0,0 +1 @@ +{"eval_loss": 6.969681739807129, "eval_runtime": 14651.8437, "eval_samples_per_second": 11.563, "eval_steps_per_second": 0.723, "epoch": 1.0} \ No newline at end of file diff --git a/mup-2b5-100m-e3/checkpoint-100/config.json b/mup-2b5-100m-e3/checkpoint-100/config.json new file mode 100644 index 0000000000000000000000000000000000000000..f2d4ab81464c896a920087ced88b4d7b4a8f4c41 --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 16384, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 4096, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-2b5-100m-e3/checkpoint-100/optimizer.pt b/mup-2b5-100m-e3/checkpoint-100/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5ca70dc9c55f5d864c7e09b824b2d880aa7cda8b --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:049fa73dc06e31ac0fc557bbd1880e3c30b86604acea743361e3bbbe2ef4cbe2 +size 21013034333 diff --git a/mup-2b5-100m-e3/checkpoint-100/pytorch_model-00001-of-00002.bin b/mup-2b5-100m-e3/checkpoint-100/pytorch_model-00001-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..1e1db03d8042adccbdbfc95ccda736e6be830e57 --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/pytorch_model-00001-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa17c85fa110ada3a0858e4cb2de270ecd3ee40f30e9d6684d3f2c21a11cac20 +size 9982107965 diff --git a/mup-2b5-100m-e3/checkpoint-100/pytorch_model-00002-of-00002.bin b/mup-2b5-100m-e3/checkpoint-100/pytorch_model-00002-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..ea794db18488d9b221192db5e0c61842cde11426 --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/pytorch_model-00002-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:305257ca9d4a2ccbff94906606e7f27125926dc9ce016218df1c6f2e2b6033bb +size 1360399056 diff --git a/mup-2b5-100m-e3/checkpoint-100/pytorch_model.bin.index.json b/mup-2b5-100m-e3/checkpoint-100/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..d31972f5835b2ee854ba8dba6ecb387529392231 --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/pytorch_model.bin.index.json @@ -0,0 +1,180 @@ +{ + "metadata": { + "total_size": 11342446640 + }, + "weight_map": { + "lm_head.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.0.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.2.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.ln_f.bias": "pytorch_model-00002-of-00002.bin", + "transformer.ln_f.weight": "pytorch_model-00002-of-00002.bin", + "transformer.wpe.weight": "pytorch_model-00001-of-00002.bin", + "transformer.wte.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/mup-2b5-100m-e3/checkpoint-100/rng_state.pth b/mup-2b5-100m-e3/checkpoint-100/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..7941abe409816d288b605b0753c893aecf0c2fd2 --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f851ccd36da51400a8dc63a45a85ee8ac4dcfac83932885c7b52532d228405d1 +size 14575 diff --git a/mup-2b5-100m-e3/checkpoint-100/scaler.pt b/mup-2b5-100m-e3/checkpoint-100/scaler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1e00f588d3f0176a99d362447a49f57ff6e1b1ad --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/scaler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfa44e8523f62833816d29aa6c576eaa7783e3bbdb3e132e248b1d8aaee6132b +size 557 diff --git a/mup-2b5-100m-e3/checkpoint-100/scheduler.pt b/mup-2b5-100m-e3/checkpoint-100/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..0ae0f24d9827e8ef9e6f63c35369714b47f4a72a --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f7e566314834b141fb5740c7bd281b7d032132544cc57f4aa141cad9641c3f9 +size 691 diff --git a/mup-2b5-100m-e3/checkpoint-100/trainer_state.json b/mup-2b5-100m-e3/checkpoint-100/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..46b361a371d0a43f225612131a0ba9eb6a6158b8 --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/trainer_state.json @@ -0,0 +1,15 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5262021973052692, + "global_step": 100, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [], + "max_steps": 190, + "num_train_epochs": 1, + "total_flos": 7.601860185686016e+17, + "trial_name": null, + "trial_params": null +} diff --git a/mup-2b5-100m-e3/checkpoint-100/training_args.bin b/mup-2b5-100m-e3/checkpoint-100/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0eb15111b824fe0193d821b0253d50509d62b6cb --- /dev/null +++ b/mup-2b5-100m-e3/checkpoint-100/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f53afd7a4c3b2020710831470410e571d9481019dd9692f0f5924a4abc0cf1e3 +size 3387 diff --git a/mup-2b5-100m-e3/config.json b/mup-2b5-100m-e3/config.json new file mode 100644 index 0000000000000000000000000000000000000000..f2d4ab81464c896a920087ced88b4d7b4a8f4c41 --- /dev/null +++ b/mup-2b5-100m-e3/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 16384, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 4096, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-2b5-100m-e3/pytorch_model-00001-of-00002.bin b/mup-2b5-100m-e3/pytorch_model-00001-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..54d7245f3e5bbed3cba490097e7288776c8707b4 --- /dev/null +++ b/mup-2b5-100m-e3/pytorch_model-00001-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1df6267cca6b60a123b3b9611c5ccbd8f91b5a38203bf50a124e1fa1c0363dfc +size 9982107965 diff --git a/mup-2b5-100m-e3/pytorch_model-00002-of-00002.bin b/mup-2b5-100m-e3/pytorch_model-00002-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..e475fe55919d431dbbb170d0877df6fd8b0b61fc --- /dev/null +++ b/mup-2b5-100m-e3/pytorch_model-00002-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4cde8ef624b72ede114ccb1793cf4ffa3d59c32a0688f33b13f6636542331551 +size 1360399056 diff --git a/mup-2b5-100m-e3/pytorch_model.bin.index.json b/mup-2b5-100m-e3/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..d31972f5835b2ee854ba8dba6ecb387529392231 --- /dev/null +++ b/mup-2b5-100m-e3/pytorch_model.bin.index.json @@ -0,0 +1,180 @@ +{ + "metadata": { + "total_size": 11342446640 + }, + "weight_map": { + "lm_head.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.0.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.11.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.2.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.ln_f.bias": "pytorch_model-00002-of-00002.bin", + "transformer.ln_f.weight": "pytorch_model-00002-of-00002.bin", + "transformer.wpe.weight": "pytorch_model-00001-of-00002.bin", + "transformer.wte.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/mup-2b5-100m-e3/training_args.bin b/mup-2b5-100m-e3/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..0eb15111b824fe0193d821b0253d50509d62b6cb --- /dev/null +++ b/mup-2b5-100m-e3/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f53afd7a4c3b2020710831470410e571d9481019dd9692f0f5924a4abc0cf1e3 +size 3387 diff --git a/mup-3b-100m-e3-full-gpt2lmmup.json b/mup-3b-100m-e3-full-gpt2lmmup.json new file mode 100644 index 0000000000000000000000000000000000000000..5e84b5ab3a1c1e1451bcf755e88b781825d921cc --- /dev/null +++ b/mup-3b-100m-e3-full-gpt2lmmup.json @@ -0,0 +1 @@ +{"eval_loss": 6.870556831359863, "eval_runtime": 16069.3588, "eval_samples_per_second": 10.543, "eval_steps_per_second": 0.659, "epoch": 1.0} \ No newline at end of file diff --git a/mup-3b-100m-e3/checkpoint-100/config.json b/mup-3b-100m-e3/checkpoint-100/config.json new file mode 100644 index 0000000000000000000000000000000000000000..cc3939892f1fd961b9edefe961a90d83cc981ada --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 17408, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 4352, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-3b-100m-e3/checkpoint-100/optimizer.pt b/mup-3b-100m-e3/checkpoint-100/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f1a86f76a2fc4d707f00d34ee2d52a0099f7c0af --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:590f451475b76a5bb9001c773d83e4b88c75c838fed4534df9f033eadba5470b +size 23609798041 diff --git a/mup-3b-100m-e3/checkpoint-100/pytorch_model-00001-of-00002.bin b/mup-3b-100m-e3/checkpoint-100/pytorch_model-00001-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..a50872e8be7d2335df1e709bb3ca3a44c0b589e6 --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/pytorch_model-00001-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61159c7f524a355b65044bed24541028017265e83427b9c19a9a1a75c006948f +size 9997736349 diff --git a/mup-3b-100m-e3/checkpoint-100/pytorch_model-00002-of-00002.bin b/mup-3b-100m-e3/checkpoint-100/pytorch_model-00002-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..51c827e9e420f8fd589a38a9fa44f34b85b1ade0 --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/pytorch_model-00002-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:639deb25bc2c13466f03e85aab820a2c107f03982379eaf3ce473dd5f4c38a9e +size 2694614901 diff --git a/mup-3b-100m-e3/checkpoint-100/pytorch_model.bin.index.json b/mup-3b-100m-e3/checkpoint-100/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..c25d2ada1b3f0909e5b391003284951a36fa6a5a --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/pytorch_model.bin.index.json @@ -0,0 +1,180 @@ +{ + "metadata": { + "total_size": 12692291632 + }, + "weight_map": { + "lm_head.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.0.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.masked_bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_1.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_1.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.2.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.ln_f.bias": "pytorch_model-00002-of-00002.bin", + "transformer.ln_f.weight": "pytorch_model-00002-of-00002.bin", + "transformer.wpe.weight": "pytorch_model-00001-of-00002.bin", + "transformer.wte.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/mup-3b-100m-e3/checkpoint-100/rng_state.pth b/mup-3b-100m-e3/checkpoint-100/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c365ed9430e8e47e1ee71469df63d8da2a549378 --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bccf3cda0d933f5f7bb7859163a2e83eefeef8218429908f35fc718e326740df +size 14575 diff --git a/mup-3b-100m-e3/checkpoint-100/scaler.pt b/mup-3b-100m-e3/checkpoint-100/scaler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1e00f588d3f0176a99d362447a49f57ff6e1b1ad --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/scaler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfa44e8523f62833816d29aa6c576eaa7783e3bbdb3e132e248b1d8aaee6132b +size 557 diff --git a/mup-3b-100m-e3/checkpoint-100/scheduler.pt b/mup-3b-100m-e3/checkpoint-100/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..98c9eb65a229debc77fd68453d69b24aa4d94a18 --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96cd9e71e6cb99c3e129d0a33d4fa0d5f079e1ff04df09a4e8e7d88ca6a01a02 +size 691 diff --git a/mup-3b-100m-e3/checkpoint-100/trainer_state.json b/mup-3b-100m-e3/checkpoint-100/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..2e949b9ca48b5958c0ab8a9bdb68df2951723738 --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/trainer_state.json @@ -0,0 +1,15 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5262021973052692, + "global_step": 100, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [], + "max_steps": 190, + "num_train_epochs": 1, + "total_flos": 8.581652284440576e+17, + "trial_name": null, + "trial_params": null +} diff --git a/mup-3b-100m-e3/checkpoint-100/training_args.bin b/mup-3b-100m-e3/checkpoint-100/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5fb4f25ccac324f9937e08ecab989b76f2ff85d3 --- /dev/null +++ b/mup-3b-100m-e3/checkpoint-100/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee5a7ca6b12b62d3ee6f578e53b23b4113c84732ffaa9d8d7cff6999532bf537 +size 3387 diff --git a/mup-3b-100m-e3/config.json b/mup-3b-100m-e3/config.json new file mode 100644 index 0000000000000000000000000000000000000000..cc3939892f1fd961b9edefe961a90d83cc981ada --- /dev/null +++ b/mup-3b-100m-e3/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 17408, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 4352, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-3b-100m-e3/pytorch_model-00001-of-00002.bin b/mup-3b-100m-e3/pytorch_model-00001-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..5d1e0bef7a88161161ff13ff65044d365967440b --- /dev/null +++ b/mup-3b-100m-e3/pytorch_model-00001-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b44158dadea58339110f0ee1aa454c55df872bf08d4e34206c3d8981452663d +size 9997736349 diff --git a/mup-3b-100m-e3/pytorch_model-00002-of-00002.bin b/mup-3b-100m-e3/pytorch_model-00002-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..2e09f53238f925afcdeea2003cf2db1540a2c425 --- /dev/null +++ b/mup-3b-100m-e3/pytorch_model-00002-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ca11bc105b58aa0681876ba028f4cd51006625e023093d342e8d308914b6c16 +size 2694614901 diff --git a/mup-3b-100m-e3/pytorch_model.bin.index.json b/mup-3b-100m-e3/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..c25d2ada1b3f0909e5b391003284951a36fa6a5a --- /dev/null +++ b/mup-3b-100m-e3/pytorch_model.bin.index.json @@ -0,0 +1,180 @@ +{ + "metadata": { + "total_size": 12692291632 + }, + "weight_map": { + "lm_head.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.0.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.masked_bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_1.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_1.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.2.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.9.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.ln_f.bias": "pytorch_model-00002-of-00002.bin", + "transformer.ln_f.weight": "pytorch_model-00002-of-00002.bin", + "transformer.wpe.weight": "pytorch_model-00001-of-00002.bin", + "transformer.wte.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/mup-3b-100m-e3/training_args.bin b/mup-3b-100m-e3/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..5fb4f25ccac324f9937e08ecab989b76f2ff85d3 --- /dev/null +++ b/mup-3b-100m-e3/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee5a7ca6b12b62d3ee6f578e53b23b4113c84732ffaa9d8d7cff6999532bf537 +size 3387 diff --git a/mup-3b5-100m-e3-full-gpt2lmmup.json b/mup-3b5-100m-e3-full-gpt2lmmup.json new file mode 100644 index 0000000000000000000000000000000000000000..cc5c83e15d9de6ee14c40919e630303e108b7354 --- /dev/null +++ b/mup-3b5-100m-e3-full-gpt2lmmup.json @@ -0,0 +1 @@ +{"eval_loss": 6.883823394775391, "eval_runtime": 16717.2968, "eval_samples_per_second": 10.134, "eval_steps_per_second": 0.633, "epoch": 1.0} \ No newline at end of file diff --git a/mup-3b5-100m-e3/checkpoint-100/config.json b/mup-3b5-100m-e3/checkpoint-100/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c89c4471675188d1ead6a2e0ef13de6e1ffa24d --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 18432, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 4608, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": false, + "vocab_size": 50257 +} diff --git a/mup-3b5-100m-e3/checkpoint-100/optimizer.pt b/mup-3b5-100m-e3/checkpoint-100/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..04f5e65994098df964934c67931a6237a49d8fb4 --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2a98488d55c99a2547d7fd062162a3229860ed14ad2044c855554eb3344ebc4 +size 26357556681 diff --git a/mup-3b5-100m-e3/checkpoint-100/pytorch_model-00001-of-00002.bin b/mup-3b5-100m-e3/checkpoint-100/pytorch_model-00001-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..401a9631f60e217e4b01a488184e8d2afc7acc7e --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/pytorch_model-00001-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60fbb7266991536dcb61ef1a8f8e2841b176b5a086bbd028f9f47beb55486c59 +size 9790034561 diff --git a/mup-3b5-100m-e3/checkpoint-100/pytorch_model-00002-of-00002.bin b/mup-3b5-100m-e3/checkpoint-100/pytorch_model-00002-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..28cf95cf2c7e0d1b605c3a06326766bfc9e47031 --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/pytorch_model-00002-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28ce0c918e97156f9d021dea864ab763be1761c08273bf73310804cca41f7c3e +size 4327659005 diff --git a/mup-3b5-100m-e3/checkpoint-100/pytorch_model.bin.index.json b/mup-3b5-100m-e3/checkpoint-100/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..a453a1bab2e072a7aa46eaaaba4288a5116263c9 --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/pytorch_model.bin.index.json @@ -0,0 +1,180 @@ +{ + "metadata": { + "total_size": 14117634096 + }, + "weight_map": { + "lm_head.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.0.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.masked_bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_1.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_1.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.masked_bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_1.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_1.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.2.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.8.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.masked_bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.ln_1.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.ln_1.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.ln_f.bias": "pytorch_model-00002-of-00002.bin", + "transformer.ln_f.weight": "pytorch_model-00002-of-00002.bin", + "transformer.wpe.weight": "pytorch_model-00001-of-00002.bin", + "transformer.wte.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/mup-3b5-100m-e3/checkpoint-100/rng_state.pth b/mup-3b5-100m-e3/checkpoint-100/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c365ed9430e8e47e1ee71469df63d8da2a549378 --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bccf3cda0d933f5f7bb7859163a2e83eefeef8218429908f35fc718e326740df +size 14575 diff --git a/mup-3b5-100m-e3/checkpoint-100/scaler.pt b/mup-3b5-100m-e3/checkpoint-100/scaler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1e00f588d3f0176a99d362447a49f57ff6e1b1ad --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/scaler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfa44e8523f62833816d29aa6c576eaa7783e3bbdb3e132e248b1d8aaee6132b +size 557 diff --git a/mup-3b5-100m-e3/checkpoint-100/scheduler.pt b/mup-3b5-100m-e3/checkpoint-100/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..325b9781497a3c777bc2ada8a6eb4623bc7dfb02 --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08e7910132718b7b09982cd5a6bc29605e45eafd7b09d4d1f772b55fa781189b +size 691 diff --git a/mup-3b5-100m-e3/checkpoint-100/trainer_state.json b/mup-3b5-100m-e3/checkpoint-100/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..0f7c7676c40846a58d8c85b9591799edde0c942f --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/trainer_state.json @@ -0,0 +1,15 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5262021973052692, + "global_step": 100, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [], + "max_steps": 190, + "num_train_epochs": 1, + "total_flos": 9.62081801109504e+17, + "trial_name": null, + "trial_params": null +} diff --git a/mup-3b5-100m-e3/checkpoint-100/training_args.bin b/mup-3b5-100m-e3/checkpoint-100/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..42a506cb5d02bc0d96e7a45b0f22d47e16001352 --- /dev/null +++ b/mup-3b5-100m-e3/checkpoint-100/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfa71a54b9260bfd5ca844c5e130f22f2225f6f0e829b91bc766959a5d8df96d +size 3387 diff --git a/mup-3b5-100m-e3/config.json b/mup-3b5-100m-e3/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7c89c4471675188d1ead6a2e0ef13de6e1ffa24d --- /dev/null +++ b/mup-3b5-100m-e3/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 18432, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 4608, + "n_head": 128, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": false, + "vocab_size": 50257 +} diff --git a/mup-3b5-100m-e3/pytorch_model-00001-of-00002.bin b/mup-3b5-100m-e3/pytorch_model-00001-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..a3aaf1b9948f78a273dc4ccde398c6b943fe63a1 --- /dev/null +++ b/mup-3b5-100m-e3/pytorch_model-00001-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5bee4aa1d5c3a055e9de9cbde2f0089720fd35df96b49db7996a23cc322addc +size 9790034561 diff --git a/mup-3b5-100m-e3/pytorch_model-00002-of-00002.bin b/mup-3b5-100m-e3/pytorch_model-00002-of-00002.bin new file mode 100644 index 0000000000000000000000000000000000000000..895e7e494049a4df8bbbc29e59b9c2270ba6de6a --- /dev/null +++ b/mup-3b5-100m-e3/pytorch_model-00002-of-00002.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee3c717b367c51e28ff292f655b582be9e86bd4afef4271be04cddf058b8363b +size 4327659005 diff --git a/mup-3b5-100m-e3/pytorch_model.bin.index.json b/mup-3b5-100m-e3/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..a453a1bab2e072a7aa46eaaaba4288a5116263c9 --- /dev/null +++ b/mup-3b5-100m-e3/pytorch_model.bin.index.json @@ -0,0 +1,180 @@ +{ + "metadata": { + "total_size": 14117634096 + }, + "weight_map": { + "lm_head.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.0.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.0.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.1.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.10.attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.attn.masked_bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_1.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_1.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.10.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.attn.masked_bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_1.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_1.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.11.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.2.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.2.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.3.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.4.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.5.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.6.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.7.mlp.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_attn.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.c_proj.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.attn.masked_bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_1.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.ln_2.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.bias": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_fc.weight": "pytorch_model-00001-of-00002.bin", + "transformer.h.8.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.8.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.c_attn.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.c_attn.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.attn.masked_bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.ln_1.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.ln_1.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.ln_2.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.ln_2.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.mlp.c_fc.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.mlp.c_fc.weight": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.mlp.c_proj.bias": "pytorch_model-00002-of-00002.bin", + "transformer.h.9.mlp.c_proj.weight": "pytorch_model-00002-of-00002.bin", + "transformer.ln_f.bias": "pytorch_model-00002-of-00002.bin", + "transformer.ln_f.weight": "pytorch_model-00002-of-00002.bin", + "transformer.wpe.weight": "pytorch_model-00001-of-00002.bin", + "transformer.wte.weight": "pytorch_model-00001-of-00002.bin" + } +} diff --git a/mup-3b5-100m-e3/training_args.bin b/mup-3b5-100m-e3/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..42a506cb5d02bc0d96e7a45b0f22d47e16001352 --- /dev/null +++ b/mup-3b5-100m-e3/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfa71a54b9260bfd5ca844c5e130f22f2225f6f0e829b91bc766959a5d8df96d +size 3387 diff --git a/mup-800m-100m-e3-full-gpt2lmmup.json b/mup-800m-100m-e3-full-gpt2lmmup.json new file mode 100644 index 0000000000000000000000000000000000000000..c84e580b0beb62c82ab3f2ad13f12310969b5bd8 --- /dev/null +++ b/mup-800m-100m-e3-full-gpt2lmmup.json @@ -0,0 +1 @@ +{"eval_loss": 10.501672744750977, "eval_runtime": 19631.6072, "eval_samples_per_second": 8.63, "eval_steps_per_second": 0.27, "epoch": 0.01} \ No newline at end of file diff --git a/mup-800m-100m-e3/config.json b/mup-800m-100m-e3/config.json new file mode 100644 index 0000000000000000000000000000000000000000..d47a25b9078f5cf21d1fc261182c895f660a490c --- /dev/null +++ b/mup-800m-100m-e3/config.json @@ -0,0 +1,34 @@ +{ + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_mult": 8.0, + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.01, + "intermediate_size": 8192, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_embd": 2048, + "n_head": 64, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "num_layers": 12, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50257 +} diff --git a/mup-800m-100m-e3/pytorch_model.bin b/mup-800m-100m-e3/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..59bbb9b2af55b58d7ee7a8d4758b4c8fac164a06 --- /dev/null +++ b/mup-800m-100m-e3/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc2ac8961f093b4652a4b03576267273027eefef683cfb5d1ef54a6e2846a48b +size 2849946173 diff --git a/mup-800m-100m-e3/training_args.bin b/mup-800m-100m-e3/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..e6d2375da6c37647db0e48a9ca43f87c50ce8f57 --- /dev/null +++ b/mup-800m-100m-e3/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4bd735f045412800863daa4ff54012c625b362504d08515303228d130e20641 +size 3387