|
--- |
|
license: mit |
|
datasets: |
|
- garage-bAInd/Open-Platypus |
|
pipeline_tag: text-generation |
|
--- |
|
\***drumroll please**\* |
|
|
|
**Introducing Tinypus!** |
|
|
|
![image/jpeg](/static-proxy?url=https%3A%2F%2Fcdn-uploads.huggingface.co%2Fproduction%2Fuploads%2F6324eabf05bd8a54c6eb1650%2FgJDAcOioOe0nzJLVzPfav.jpeg%3C%2Fspan%3E)%3C!-- HTML_TAG_END --> |
|
|
|
I passthrough merged base Tiny Llama Chat with itself, then fine-tuned with around 1/3 of Platypus dataset. |
|
|
|
Observations: |
|
|
|
- It's smarter (I think?) |
|
|
|
- It sometimes throws "### Instruction:" line. This could be due to the platypus dataset, or the fact that I know jackshit about programming. You can add it to "custom stopping strings" in oobaboga. |
|
|
|
- It may be possible to train very specialized mini experts and merge them??? |
|
|
|
**Template** |
|
|
|
Same with TinyLlama/TinyLlama-1.1B-Chat-v1.0 |
|
|
|
**Merge details** |
|
|
|
slices: |
|
- sources: |
|
- model: E://text-generation-webui//models//TinyLlama |
|
|
|
layer_range: [0, 12] |
|
|
|
- sources: |
|
|
|
- model: E://text-generation-webui//models//TinyLlama |
|
|
|
layer_range: [4, 22] |
|
|
|
merge_method: passthrough |
|
|
|
dtype: bfloat16 |
|
|
|
**QLoRA Details** |
|
|
|
Chunk Length: 1152 |
|
R/A: 64/128 |
|
Epoch: 1 |
|
q-k-v-o |