gemma-2-2b-norobots / README.md
Zangs3011's picture
Update README.md
c8220ef verified
---
library_name: peft
tags:
- nlp
- code
- instruct
- llama
datasets:
- HuggingFaceH4/no_robots
base_model: google/gemma-2-2b-it
license: apache-2.0
---
# monsterapi/gemma-2-2b-norobots
**Base Model for Fine-tuning:** [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it)
**Service Used:** [MonsterAPI](https://monsterapi.ai)
**License:** Apache-2.0
## Overview
`monsterapi/gemma-2-2b-norobots` is a fine-tuned language model designed to improve instruction-following capabilities. The model was trained using the "No Robots" dataset, a high-quality set of 10,000 instructions and demonstrations curated by expert human annotators. This fine-tuning process enhances the base model's performance in understanding and executing single-turn instructions, similar to the goals outlined in OpenAI's InstructGPT.
### Dataset Details
**Dataset Summary:**
The "No Robots" dataset is a collection of 10,000 high-quality instructions and demonstrations created by skilled human annotators. The dataset is modeled after the instruction dataset described in OpenAI's InstructGPT paper. It mainly includes single-turn instructions across various categories, aiming to improve the instruction-following capabilities of language models during supervised fine-tuning (SFT).
## Fine-tuning Details
**Fine-tuned Model Name:** `monsterapi/gemma-2-2b-norobots`
**Training Time:** 31 minutes
**Cost:** $1.10
**Epochs:** 1
**Gradient Accumulation Steps:** 32
The model was fine-tuned using MonsterAPI's finetuning service, optimizing the base model `google/gemma-2-2b-it` to perform better on instruction-following tasks.
## Hyperparameters & Additional Details
- **Base Model:** `google/gemma-2-2b-it`
- **Dataset:** No Robots (10,000 instructions and demonstrations)
- **Training Duration:** 31 minutes
- **Cost per Epoch:** $1.10
- **Total Finetuning Cost:** $1.10
- **Gradient Accumulation Steps:** 32
## Use Cases
This model is well-suited for tasks that require improved instruction-following capabilities, such as:
- Chatbots and virtual assistants
- Content creation tools
- Automated customer support systems
- Task automation in various industries
## How to Use
You can load the model directly using the Hugging Face Transformers library:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "monsterapi/gemma-2-2b-norobots"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
# Example usage
input_text = "Explain the concept of supervised fine-tuning in simple terms."
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
## Acknowledgements
The fine-tuning process was carried out using MonsterAPI's finetuning service, which offers a seamless experience for optimizing large language models.
## Contact
For further details or queries, please contact [MonsterAPI](https://monsterapi.ai) or visit the official documentation.