File size: 2,101 Bytes
25d2ef8
 
 
 
 
 
 
 
 
 
50f3e92
25d2ef8
50f3e92
 
 
 
25d2ef8
ccc94f4
 
 
 
 
 
 
 
a9b25a4
 
 
3016581
 
 
 
 
 
 
 
25d2ef8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
---
base_model:
- unsloth/phi-4
- bunnycore/Phi-4-rp-v1-lora
library_name: transformers
tags:
- mergekit
- merge

---
Phi-4-RP-V0.2 is based on the Phi-4 architecture, which is a state-of-the-art large language model designed to handle a wide range of natural language tasks with high efficiency and performance.

## Primary Use Cases
- Interactive Storytelling : Engage users in dynamic, immersive stories where they can take on different roles and make choices that influence the narrative.
- Role-Playing Games (RPGs) : Provide rich, interactive experiences in RPGs, enhancing gameplay through intelligent character interactions.
- Virtual Assistants : Offer personalized, engaging conversations that simulate human-like interactions for customer support or entertainment purposes.

## Training Data

Phi-4-RP-V0.2 is specifically trained on role-playing datasets to ensure comprehensive understanding and versatility in various role-playing contexts. This includes but is not limited to:

- Role-playing game scripts and narratives.
- Interactive storytelling scenarios.
- Character dialogues and interactions from diverse fictional settings.

## Input Formats

Given the nature of the training data, phi-4 is best suited for prompts using the chat format as follows:
```
<|im_start|>system<|im_sep|>
You are a medieval knight and must provide explanations to modern people.<|im_end|>
<|im_start|>user<|im_sep|>
How should I explain the Internet?<|im_end|>
<|im_start|>assistant<|im_sep|>
```

## Merge Details
### Merge Method

This model was merged using the passthrough merge method using [unsloth/phi-4](https://huggingface.co/unsloth/phi-4) + [bunnycore/Phi-4-rp-v1-lora](https://huggingface.co/bunnycore/Phi-4-rp-v1-lora) as a base.

### Models Merged

The following models were included in the merge:


### Configuration

The following YAML configuration was used to produce this model:

```yaml


base_model: unsloth/phi-4+bunnycore/Phi-4-rp-v1-lora
dtype: bfloat16
merge_method: passthrough
models:
  - model: unsloth/phi-4+bunnycore/Phi-4-rp-v1-lora
tokenizer_source: unsloth/phi-4

```