File size: 2,101 Bytes
25d2ef8 50f3e92 25d2ef8 50f3e92 25d2ef8 ccc94f4 a9b25a4 3016581 25d2ef8 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 |
---
base_model:
- unsloth/phi-4
- bunnycore/Phi-4-rp-v1-lora
library_name: transformers
tags:
- mergekit
- merge
---
Phi-4-RP-V0.2 is based on the Phi-4 architecture, which is a state-of-the-art large language model designed to handle a wide range of natural language tasks with high efficiency and performance.
## Primary Use Cases
- Interactive Storytelling : Engage users in dynamic, immersive stories where they can take on different roles and make choices that influence the narrative.
- Role-Playing Games (RPGs) : Provide rich, interactive experiences in RPGs, enhancing gameplay through intelligent character interactions.
- Virtual Assistants : Offer personalized, engaging conversations that simulate human-like interactions for customer support or entertainment purposes.
## Training Data
Phi-4-RP-V0.2 is specifically trained on role-playing datasets to ensure comprehensive understanding and versatility in various role-playing contexts. This includes but is not limited to:
- Role-playing game scripts and narratives.
- Interactive storytelling scenarios.
- Character dialogues and interactions from diverse fictional settings.
## Input Formats
Given the nature of the training data, phi-4 is best suited for prompts using the chat format as follows:
```
<|im_start|>system<|im_sep|>
You are a medieval knight and must provide explanations to modern people.<|im_end|>
<|im_start|>user<|im_sep|>
How should I explain the Internet?<|im_end|>
<|im_start|>assistant<|im_sep|>
```
## Merge Details
### Merge Method
This model was merged using the passthrough merge method using [unsloth/phi-4](https://huggingface.co/unsloth/phi-4) + [bunnycore/Phi-4-rp-v1-lora](https://huggingface.co/bunnycore/Phi-4-rp-v1-lora) as a base.
### Models Merged
The following models were included in the merge:
### Configuration
The following YAML configuration was used to produce this model:
```yaml
base_model: unsloth/phi-4+bunnycore/Phi-4-rp-v1-lora
dtype: bfloat16
merge_method: passthrough
models:
- model: unsloth/phi-4+bunnycore/Phi-4-rp-v1-lora
tokenizer_source: unsloth/phi-4
```
|