grimjim commited on
Commit
94dffda
·
verified ·
1 Parent(s): a3eaa1a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -56
README.md CHANGED
@@ -1,56 +1,55 @@
1
- ---
2
- base_model:
3
- - grimjim/llama-3-aaditya-OpenBioLLM-8B
4
- - NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
5
- - cgato/L3-TheSpice-8b-v0.8.3
6
- library_name: transformers
7
- tags:
8
- - mergekit
9
- - merge
10
- pipeline_tag: text-generation
11
- license: llama3
12
- license_link: LICENSE
13
- ---
14
- # grimjim/Llama-3-Luminurse-v0.2-OAS-8B-8bpw-exl2
15
-
16
- This repo contains an exl2 quant of [Llama-3-Luminurse-v0.2-OAS-8B](https://huggingface.co/grimjim/Llama-3-Luminurse-v0.2-OAS-8B) at 8bpw. For suggested sampler settings, refer to the model card of the original repo.
17
-
18
- This model is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
19
-
20
- Luminurse is a merge based on Lumimaid, enhanced with a biomedical model, with a dash of TheSpice thrown in to improve formatting of text generation.
21
-
22
- Built with Meta Llama 3.
23
-
24
- ## Merge Details
25
- ### Merge Method
26
-
27
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS) as a base.
28
-
29
- ### Models Merged
30
-
31
- The following models were included in the merge:
32
- * [grimjim/llama-3-aaditya-OpenBioLLM-8B](https://huggingface.co/grimjim/llama-3-aaditya-OpenBioLLM-8B)
33
- * [cgato/L3-TheSpice-8b-v0.8.3](https://huggingface.co/cgato/L3-TheSpice-8b-v0.8.3)
34
-
35
- ### Configuration
36
-
37
- The following YAML configuration was used to produce this model:
38
-
39
- ```yaml
40
- base_model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
41
- slices:
42
- - sources:
43
- - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
44
- layer_range: [0,32]
45
- - model: grimjim/llama-3-aaditya-OpenBioLLM-8B
46
- layer_range: [0,32]
47
- parameters:
48
- weight: 0.2
49
- - model: cgato/L3-TheSpice-8b-v0.8.3
50
- layer_range: [0,32]
51
- parameters:
52
- weight: 0.04
53
- merge_method: task_arithmetic
54
- dtype: bfloat16
55
-
56
- ```
 
1
+ ---
2
+ base_model:
3
+ - grimjim/llama-3-aaditya-OpenBioLLM-8B
4
+ - NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
5
+ - cgato/L3-TheSpice-8b-v0.8.3
6
+ library_name: transformers
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ pipeline_tag: text-generation
11
+ license: cc-by-nc-4.0
12
+ ---
13
+ # grimjim/Llama-3-Luminurse-v0.2-OAS-8B-8bpw-exl2
14
+
15
+ This repo contains an exl2 quant of [Llama-3-Luminurse-v0.2-OAS-8B](https://huggingface.co/grimjim/Llama-3-Luminurse-v0.2-OAS-8B) at 8bpw. For suggested sampler settings, refer to the model card of the original repo.
16
+
17
+ This model is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
+
19
+ Luminurse is a merge based on Lumimaid, enhanced with a biomedical model, with a dash of TheSpice thrown in to improve formatting of text generation.
20
+
21
+ Built with Meta Llama 3.
22
+
23
+ ## Merge Details
24
+ ### Merge Method
25
+
26
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS) as a base.
27
+
28
+ ### Models Merged
29
+
30
+ The following models were included in the merge:
31
+ * [grimjim/llama-3-aaditya-OpenBioLLM-8B](https://huggingface.co/grimjim/llama-3-aaditya-OpenBioLLM-8B)
32
+ * [cgato/L3-TheSpice-8b-v0.8.3](https://huggingface.co/cgato/L3-TheSpice-8b-v0.8.3)
33
+
34
+ ### Configuration
35
+
36
+ The following YAML configuration was used to produce this model:
37
+
38
+ ```yaml
39
+ base_model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
40
+ slices:
41
+ - sources:
42
+ - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
43
+ layer_range: [0,32]
44
+ - model: grimjim/llama-3-aaditya-OpenBioLLM-8B
45
+ layer_range: [0,32]
46
+ parameters:
47
+ weight: 0.2
48
+ - model: cgato/L3-TheSpice-8b-v0.8.3
49
+ layer_range: [0,32]
50
+ parameters:
51
+ weight: 0.04
52
+ merge_method: task_arithmetic
53
+ dtype: bfloat16
54
+
55
+ ```