|
|
|
--- |
|
library_name: peft |
|
tags: |
|
- nlp |
|
- code |
|
- instruct |
|
- llama |
|
datasets: |
|
- HuggingFaceH4/no_robots |
|
base_model: google/gemma-2-2b-it |
|
license: apache-2.0 |
|
--- |
|
|
|
|
|
# monsterapi/gemma-2-2b-norobots |
|
|
|
**Base Model for Fine-tuning:** [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it) |
|
**Service Used:** [MonsterAPI](https://monsterapi.ai) |
|
**License:** Apache-2.0 |
|
|
|
## Overview |
|
|
|
`monsterapi/gemma-2-2b-norobots` is a fine-tuned language model designed to improve instruction-following capabilities. The model was trained using the "No Robots" dataset, a high-quality set of 10,000 instructions and demonstrations curated by expert human annotators. This fine-tuning process enhances the base model's performance in understanding and executing single-turn instructions, similar to the goals outlined in OpenAI's InstructGPT. |
|
|
|
### Dataset Details |
|
|
|
**Dataset Summary:** |
|
The "No Robots" dataset is a collection of 10,000 high-quality instructions and demonstrations created by skilled human annotators. The dataset is modeled after the instruction dataset described in OpenAI's InstructGPT paper. It mainly includes single-turn instructions across various categories, aiming to improve the instruction-following capabilities of language models during supervised fine-tuning (SFT). |
|
|
|
## Fine-tuning Details |
|
|
|
**Fine-tuned Model Name:** `monsterapi/gemma-2-2b-norobots` |
|
**Training Time:** 31 minutes |
|
**Cost:** $1.10 |
|
**Epochs:** 1 |
|
**Gradient Accumulation Steps:** 32 |
|
|
|
The model was fine-tuned using MonsterAPI's finetuning service, optimizing the base model `google/gemma-2-2b-it` to perform better on instruction-following tasks. |
|
|
|
## Hyperparameters & Additional Details |
|
|
|
- **Base Model:** `google/gemma-2-2b-it` |
|
- **Dataset:** No Robots (10,000 instructions and demonstrations) |
|
- **Training Duration:** 31 minutes |
|
- **Cost per Epoch:** $1.10 |
|
- **Total Finetuning Cost:** $1.10 |
|
- **Gradient Accumulation Steps:** 32 |
|
|
|
## Use Cases |
|
|
|
This model is well-suited for tasks that require improved instruction-following capabilities, such as: |
|
- Chatbots and virtual assistants |
|
- Content creation tools |
|
- Automated customer support systems |
|
- Task automation in various industries |
|
|
|
## How to Use |
|
|
|
You can load the model directly using the Hugging Face Transformers library: |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
model_name = "monsterapi/gemma-2-2b-norobots" |
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
model = AutoModelForCausalLM.from_pretrained(model_name) |
|
|
|
# Example usage |
|
input_text = "Explain the concept of supervised fine-tuning in simple terms." |
|
inputs = tokenizer(input_text, return_tensors="pt") |
|
outputs = model.generate(**inputs) |
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
``` |
|
|
|
## Acknowledgements |
|
|
|
The fine-tuning process was carried out using MonsterAPI's finetuning service, which offers a seamless experience for optimizing large language models. |
|
|
|
## Contact |
|
|
|
For further details or queries, please contact [MonsterAPI](https://monsterapi.ai) or visit the official documentation. |
|
|