--- license: apache-2.0 language: - en datasets: - Magpie-Align/Magpie-Reasoning-V1-150K-CoT-QwQ library_name: transformers tags: - text-generation-inference - transformers - unsloth - qwen2 - trl base_model: - NovaSky-AI/Sky-T1-32B-Preview --- ![image](./image.webp) # Sky-T1-32B-Preview Fine-Tuned Model ## Model Details - **Developed by:** Daemontatox - **Model type:** Text Generation - **Language(s):** English - **License:** Apache 2.0 - **Finetuned from model:** [NovaSky-AI/Sky-T1-32B-Preview](https://huggingface.co/NovaSky-AI/Sky-T1-32B-Preview) - **Training dataset:** [Magpie-Reasoning-V1-150K-CoT-QwQ](https://huggingface.co/datasets/Magpie-Align/Magpie-Reasoning-V1-150K-CoT-QwQ) - **Training framework:** [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's [TRL](https://github.com/huggingface/trl) library ## Model Description This model is a fine-tuned version of the **NovaSky-AI/Sky-T1-32B-Preview** model, specifically optimized for text generation tasks. It was trained on the **Magpie-Reasoning-V1-150K-CoT-QwQ** dataset, which focuses on reasoning and chain-of-thought (CoT) tasks. The training process was accelerated using **Unsloth**, achieving a 2x speedup compared to traditional methods. ## Intended Use This model is intended for **text generation** tasks, particularly those requiring reasoning and logical coherence. It can be used for: - Chain-of-thought reasoning - Question answering - Content generation - Educational tools ## Training Details - **Training framework:** Unsloth + Huggingface TRL - **Training speed:** 2x faster than traditional methods - **Dataset:** Magpie-Reasoning-V1-150K-CoT-QwQ - **Base model:** NovaSky-AI/Sky-T1-32B-Preview ## How to Use You can use this model with the Huggingface `transformers` library: ```python from transformers import AutoModelForCausalLM, AutoTokenizer # Load the model and tokenizer model = AutoModelForCausalLM.from_pretrained("Daemontatox/Sky-T1-32B-Preview-Finetuned") tokenizer = AutoTokenizer.from_pretrained("Daemontatox/Sky-T1-32B-Preview-Finetuned") # Generate text input_text = "Explain the concept of chain-of-thought reasoning." inputs = tokenizer(input_text, return_tensors="pt") outputs = model.generate(**inputs, max_length=200) # Decode and print the output print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Limitations -**The model may generate incorrect or nonsensical responses if the input is ambiguous or outside its training domain.** -**It is primarily trained on English data, so performance may degrade for other languages.** ## Ethical Considerations -**Bias:** The model may inherit biases present in the training data. Users should be cautious when deploying it in sensitive applications. -**Misuse:** The model should not be used for generating harmful, misleading, or unethical content. ``` @misc{novasky-sky-t1-32b-preview, author = {NovaSky-AI}, title = {Sky-T1-32B-Preview}, year = {2023}, publisher = {Hugging Face}, howpublished = {\url{https://huggingface.co/NovaSky-AI/Sky-T1-32B-Preview}}, } @misc{unsloth, author = {Unsloth Team}, title = {Unsloth: Faster Training for Transformers}, year = {2023}, publisher = {GitHub}, howpublished = {\url{https://github.com/unslothai/unsloth}}, } ``` ## Acknowledgements Thanks to **NovaSky-AI** for the base model. Thanks to **Unsloth** for the faster training framework. Thanks to **Huggingface** for the TRL library and tools.