bibidentuhanoi commited on
Commit
559e353
·
verified ·
1 Parent(s): bee7755

Trained with Unsloth

Browse files
README.md CHANGED
@@ -1,6 +1,9 @@
1
  ---
2
  library_name: transformers
3
- tags: []
 
 
 
4
  ---
5
 
6
  # Model Card for Model ID
 
1
  ---
2
  library_name: transformers
3
+ tags:
4
+ - unsloth
5
+ - trl
6
+ - sft
7
  ---
8
 
9
  # Model Card for Model ID
config.json CHANGED
@@ -1,11 +1,11 @@
1
  {
2
- "_name_or_path": "/kaggle/working/BMO-7B-Instruct",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
- "eos_token_id": 32000,
9
  "hidden_act": "silu",
10
  "hidden_size": 4096,
11
  "initializer_range": 0.02,
@@ -17,11 +17,11 @@
17
  "num_key_value_heads": 8,
18
  "rms_norm_eps": 1e-05,
19
  "rope_theta": 10000.0,
20
- "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
- "torch_dtype": "float16",
23
- "transformers_version": "4.39.0.dev0",
24
- "unsloth_version": "2024.3",
25
  "use_cache": false,
26
- "vocab_size": 32002
27
  }
 
1
  {
2
+ "_name_or_path": "cognitivecomputations/dolphin-2.6-mistral-7b",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
  "hidden_act": "silu",
10
  "hidden_size": 4096,
11
  "initializer_range": 0.02,
 
17
  "num_key_value_heads": 8,
18
  "rms_norm_eps": 1e-05,
19
  "rope_theta": 10000.0,
20
+ "sliding_window": null,
21
  "tie_word_embeddings": false,
22
+ "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.40.1",
24
+ "unsloth_version": "2024.4",
25
  "use_cache": false,
26
+ "vocab_size": 32001
27
  }
generation_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
- "eos_token_id": 32000,
5
- "transformers_version": "4.39.0.dev0"
6
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.40.1"
6
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f03253df2a2bb422963caea59ff6aa04f6626558770ed295f209fcfee41566ed
3
- size 4943178624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8770bd8a33c92e6d072539055af37adad65392bafe333f4322db10c11671825e
3
+ size 4943170528
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:555e47c9750cd1c865c8321cdb68be2b6c0507f13f6903c2d73e35d95335b7d0
3
- size 4999819232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6fcd3385ed1bb776f5c6871d5f34eb1b85716a4f3e969ce0a9a1244624de870
3
+ size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8b72b82bf0b6fd84d8324fbd03070a07467e90fc0e593eb2d8edd7dea541942
3
- size 4540532640
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4df32caf350d20d669dac0ed82750c417eaa2cd7430acc6d7ff4435e187b974f
3
+ size 4540524536
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 14483496960
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00003-of-00003.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 14483480576
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00003-of-00003.safetensors",