Daemontatox commited on
Commit
eee8792
·
verified ·
1 Parent(s): 77bb24e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +93 -12
README.md CHANGED
@@ -1,25 +1,106 @@
1
  ---
2
- base_model: NovaSky-AI/Sky-T1-32B-Preview
3
- tags:
4
- - text-generation-inference
5
- - transformers
6
- - unsloth
7
- - qwen2
8
- - trl
9
  license: apache-2.0
10
  language:
11
  - en
12
  datasets:
13
  - Magpie-Align/Magpie-Reasoning-V1-150K-CoT-QwQ
14
  library_name: transformers
 
 
 
 
 
 
15
  ---
16
 
17
- # Uploaded model
 
 
18
 
19
  - **Developed by:** Daemontatox
20
- - **License:** apache-2.0
21
- - **Finetuned from model :** NovaSky-AI/Sky-T1-32B-Preview
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
23
- This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
24
 
25
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
1
  ---
 
 
 
 
 
 
 
2
  license: apache-2.0
3
  language:
4
  - en
5
  datasets:
6
  - Magpie-Align/Magpie-Reasoning-V1-150K-CoT-QwQ
7
  library_name: transformers
8
+ tags:
9
+ - text-generation-inference
10
+ - transformers
11
+ - unsloth
12
+ - qwen2
13
+ - trl
14
  ---
15
 
16
+ # Sky-T1-32B-Preview Fine-Tuned Model
17
+
18
+ ## Model Details
19
 
20
  - **Developed by:** Daemontatox
21
+ - **Model type:** Text Generation
22
+ - **Language(s):** English
23
+ - **License:** Apache 2.0
24
+ - **Finetuned from model:** [NovaSky-AI/Sky-T1-32B-Preview](https://huggingface.co/NovaSky-AI/Sky-T1-32B-Preview)
25
+ - **Training dataset:** [Magpie-Reasoning-V1-150K-CoT-QwQ](https://huggingface.co/datasets/Magpie-Align/Magpie-Reasoning-V1-150K-CoT-QwQ)
26
+ - **Training framework:** [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's [TRL](https://github.com/huggingface/trl) library
27
+
28
+ ## Model Description
29
+
30
+ This model is a fine-tuned version of the **NovaSky-AI/Sky-T1-32B-Preview** model, specifically optimized for text generation tasks. It was trained on the **Magpie-Reasoning-V1-150K-CoT-QwQ** dataset, which focuses on reasoning and chain-of-thought (CoT) tasks. The training process was accelerated using **Unsloth**, achieving a 2x speedup compared to traditional methods.
31
+
32
+ ## Intended Use
33
+
34
+ This model is intended for **text generation** tasks, particularly those requiring reasoning and logical coherence. It can be used for:
35
+
36
+ - Chain-of-thought reasoning
37
+ - Question answering
38
+ - Content generation
39
+ - Educational tools
40
+
41
+ ## Training Details
42
+
43
+ - **Training framework:** Unsloth + Huggingface TRL
44
+ - **Training speed:** 2x faster than traditional methods
45
+ - **Dataset:** Magpie-Reasoning-V1-150K-CoT-QwQ
46
+ - **Base model:** NovaSky-AI/Sky-T1-32B-Preview
47
+
48
+ ## How to Use
49
+
50
+ You can use this model with the Huggingface `transformers` library:
51
+
52
+ ```python
53
+ from transformers import AutoModelForCausalLM, AutoTokenizer
54
+
55
+ # Load the model and tokenizer
56
+ model = AutoModelForCausalLM.from_pretrained("Daemontatox/Sky-T1-32B-Preview-Finetuned")
57
+ tokenizer = AutoTokenizer.from_pretrained("Daemontatox/Sky-T1-32B-Preview-Finetuned")
58
+
59
+ # Generate text
60
+ input_text = "Explain the concept of chain-of-thought reasoning."
61
+ inputs = tokenizer(input_text, return_tensors="pt")
62
+ outputs = model.generate(**inputs, max_length=200)
63
+
64
+ # Decode and print the output
65
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
66
+ ```
67
+
68
+ ## Limitations
69
+ -**The model may generate incorrect or nonsensical responses if the input is ambiguous or outside its training domain.**
70
+
71
+ -**It is primarily trained on English data, so performance may degrade for other languages.**
72
+
73
+ ## Ethical Considerations
74
+ -**Bias:** The model may inherit biases present in the training data. Users should be cautious when deploying it in sensitive applications.
75
+
76
+ -**Misuse:** The model should not be used for generating harmful, misleading, or unethical content.
77
+
78
+
79
+
80
+
81
+ ```
82
+ @misc{novasky-sky-t1-32b-preview,
83
+ author = {NovaSky-AI},
84
+ title = {Sky-T1-32B-Preview},
85
+ year = {2023},
86
+ publisher = {Hugging Face},
87
+ howpublished = {\url{https://huggingface.co/NovaSky-AI/Sky-T1-32B-Preview}},
88
+ }
89
+
90
+ @misc{unsloth,
91
+ author = {Unsloth Team},
92
+ title = {Unsloth: Faster Training for Transformers},
93
+ year = {2023},
94
+ publisher = {GitHub},
95
+ howpublished = {\url{https://github.com/unslothai/unsloth}},
96
+ }
97
+
98
+ ```
99
+
100
+
101
+ ## Acknowledgements
102
+ Thanks to **NovaSky-AI** for the base model.
103
 
104
+ Thanks to **Unsloth** for the faster training framework.
105
 
106
+ Thanks to **Huggingface** for the TRL library and tools.