Daemontatox
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -1,25 +1,106 @@
|
|
1 |
---
|
2 |
-
base_model: NovaSky-AI/Sky-T1-32B-Preview
|
3 |
-
tags:
|
4 |
-
- text-generation-inference
|
5 |
-
- transformers
|
6 |
-
- unsloth
|
7 |
-
- qwen2
|
8 |
-
- trl
|
9 |
license: apache-2.0
|
10 |
language:
|
11 |
- en
|
12 |
datasets:
|
13 |
- Magpie-Align/Magpie-Reasoning-V1-150K-CoT-QwQ
|
14 |
library_name: transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
---
|
16 |
|
17 |
-
#
|
|
|
|
|
18 |
|
19 |
- **Developed by:** Daemontatox
|
20 |
-
- **
|
21 |
-
- **
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
|
23 |
-
|
24 |
|
25 |
-
|
|
|
1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
license: apache-2.0
|
3 |
language:
|
4 |
- en
|
5 |
datasets:
|
6 |
- Magpie-Align/Magpie-Reasoning-V1-150K-CoT-QwQ
|
7 |
library_name: transformers
|
8 |
+
tags:
|
9 |
+
- text-generation-inference
|
10 |
+
- transformers
|
11 |
+
- unsloth
|
12 |
+
- qwen2
|
13 |
+
- trl
|
14 |
---
|
15 |
|
16 |
+
# Sky-T1-32B-Preview Fine-Tuned Model
|
17 |
+
|
18 |
+
## Model Details
|
19 |
|
20 |
- **Developed by:** Daemontatox
|
21 |
+
- **Model type:** Text Generation
|
22 |
+
- **Language(s):** English
|
23 |
+
- **License:** Apache 2.0
|
24 |
+
- **Finetuned from model:** [NovaSky-AI/Sky-T1-32B-Preview](https://huggingface.co/NovaSky-AI/Sky-T1-32B-Preview)
|
25 |
+
- **Training dataset:** [Magpie-Reasoning-V1-150K-CoT-QwQ](https://huggingface.co/datasets/Magpie-Align/Magpie-Reasoning-V1-150K-CoT-QwQ)
|
26 |
+
- **Training framework:** [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's [TRL](https://github.com/huggingface/trl) library
|
27 |
+
|
28 |
+
## Model Description
|
29 |
+
|
30 |
+
This model is a fine-tuned version of the **NovaSky-AI/Sky-T1-32B-Preview** model, specifically optimized for text generation tasks. It was trained on the **Magpie-Reasoning-V1-150K-CoT-QwQ** dataset, which focuses on reasoning and chain-of-thought (CoT) tasks. The training process was accelerated using **Unsloth**, achieving a 2x speedup compared to traditional methods.
|
31 |
+
|
32 |
+
## Intended Use
|
33 |
+
|
34 |
+
This model is intended for **text generation** tasks, particularly those requiring reasoning and logical coherence. It can be used for:
|
35 |
+
|
36 |
+
- Chain-of-thought reasoning
|
37 |
+
- Question answering
|
38 |
+
- Content generation
|
39 |
+
- Educational tools
|
40 |
+
|
41 |
+
## Training Details
|
42 |
+
|
43 |
+
- **Training framework:** Unsloth + Huggingface TRL
|
44 |
+
- **Training speed:** 2x faster than traditional methods
|
45 |
+
- **Dataset:** Magpie-Reasoning-V1-150K-CoT-QwQ
|
46 |
+
- **Base model:** NovaSky-AI/Sky-T1-32B-Preview
|
47 |
+
|
48 |
+
## How to Use
|
49 |
+
|
50 |
+
You can use this model with the Huggingface `transformers` library:
|
51 |
+
|
52 |
+
```python
|
53 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
54 |
+
|
55 |
+
# Load the model and tokenizer
|
56 |
+
model = AutoModelForCausalLM.from_pretrained("Daemontatox/Sky-T1-32B-Preview-Finetuned")
|
57 |
+
tokenizer = AutoTokenizer.from_pretrained("Daemontatox/Sky-T1-32B-Preview-Finetuned")
|
58 |
+
|
59 |
+
# Generate text
|
60 |
+
input_text = "Explain the concept of chain-of-thought reasoning."
|
61 |
+
inputs = tokenizer(input_text, return_tensors="pt")
|
62 |
+
outputs = model.generate(**inputs, max_length=200)
|
63 |
+
|
64 |
+
# Decode and print the output
|
65 |
+
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
66 |
+
```
|
67 |
+
|
68 |
+
## Limitations
|
69 |
+
-**The model may generate incorrect or nonsensical responses if the input is ambiguous or outside its training domain.**
|
70 |
+
|
71 |
+
-**It is primarily trained on English data, so performance may degrade for other languages.**
|
72 |
+
|
73 |
+
## Ethical Considerations
|
74 |
+
-**Bias:** The model may inherit biases present in the training data. Users should be cautious when deploying it in sensitive applications.
|
75 |
+
|
76 |
+
-**Misuse:** The model should not be used for generating harmful, misleading, or unethical content.
|
77 |
+
|
78 |
+
|
79 |
+
|
80 |
+
|
81 |
+
```
|
82 |
+
@misc{novasky-sky-t1-32b-preview,
|
83 |
+
author = {NovaSky-AI},
|
84 |
+
title = {Sky-T1-32B-Preview},
|
85 |
+
year = {2023},
|
86 |
+
publisher = {Hugging Face},
|
87 |
+
howpublished = {\url{https://huggingface.co/NovaSky-AI/Sky-T1-32B-Preview}},
|
88 |
+
}
|
89 |
+
|
90 |
+
@misc{unsloth,
|
91 |
+
author = {Unsloth Team},
|
92 |
+
title = {Unsloth: Faster Training for Transformers},
|
93 |
+
year = {2023},
|
94 |
+
publisher = {GitHub},
|
95 |
+
howpublished = {\url{https://github.com/unslothai/unsloth}},
|
96 |
+
}
|
97 |
+
|
98 |
+
```
|
99 |
+
|
100 |
+
|
101 |
+
## Acknowledgements
|
102 |
+
Thanks to **NovaSky-AI** for the base model.
|
103 |
|
104 |
+
Thanks to **Unsloth** for the faster training framework.
|
105 |
|
106 |
+
Thanks to **Huggingface** for the TRL library and tools.
|