|
--- |
|
tags: |
|
- finetuned |
|
- quantized |
|
- 4-bit |
|
- AWQ |
|
- transformers |
|
- pytorch |
|
- mistral |
|
- instruct |
|
- text-generation |
|
- conversational |
|
- license:apache-2.0 |
|
- autotrain_compatible |
|
- endpoints_compatible |
|
- text-generation-inference |
|
- finetune |
|
- chatml |
|
model-index: |
|
- name: Noromaid-7B-0.4-DPO |
|
results: [] |
|
base_model: NeverSleep/Noromaid-7B-0.4-DPO |
|
datasets: |
|
- Undi95/Llama2-13B-no_robots-alpaca-lora |
|
- NobodyExistsOnTheInternet/ToxicDPOqa |
|
- Undi95/toxic-dpo-v0.1-NoWarning |
|
license: cc-by-nc-4.0 |
|
library_name: transformers |
|
model_creator: IkariDev and Undi |
|
model_name: Noromaid 7B v0.4 DPO |
|
model_type: mistral |
|
pipeline_tag: text-generation |
|
inference: false |
|
prompt_template: '<|im_start|>system |
|
|
|
{system_message}<|im_end|> |
|
|
|
<|im_start|>user |
|
|
|
{prompt}<|im_end|> |
|
|
|
<|im_start|>assistant |
|
|
|
' |
|
quantized_by: Suparious |
|
--- |
|
# Noromaid 7B v0.4 DPO - AWQ |
|
|
|
- Model creator: [IkariDev and Undi](https://huggingface.co/NeverSleep) |
|
- Original model: [Noromaid 7B v0.4 DPO](https://huggingface.co/NeverSleep/Noromaid-7B-0.4-DPO) |
|
|
|
![image/png](/static-proxy?url=https%3A%2F%2Fcdn-uploads.huggingface.co%2Fproduction%2Fuploads%2F630dfb008df86f1e5becadc3%2FVKX2Z2yjZX5J8kXzgeCYO.png%3C%2Fspan%3E)%3C!-- HTML_TAG_END --> |
|
|
|
## Model description |
|
|
|
This repo contains AWQ model files for [IkariDev and Undi's Noromaid 7B v0.4 DPO](https://huggingface.co/NeverSleep/Noromaid-7B-0.4-DPO). |
|
|
|
These files were quantised using hardware kindly provided by [SolidRusT Networks](https://solidrust.net/). |
|
|
|
### About AWQ |
|
|
|
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings. |
|
|
|
AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead. |
|
|
|
It is supported by: |
|
|
|
- [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ |
|
- [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types. |
|
- [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference) |
|
- [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers |
|
- [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code |
|
|
|
## Prompt template: ChatML |
|
|
|
```plaintext |
|
<|im_start|>system |
|
{system_message}<|im_end|> |
|
<|im_start|>user |
|
{prompt}<|im_end|> |
|
<|im_start|>assistant |
|
``` |
|
|
|
## Training data |
|
|
|
- [no_robots dataset](https://huggingface.co/Undi95/Llama2-13B-no_robots-alpaca-lora) let the model have more human behavior, enhances the output. |
|
- [Aesir Private RP dataset] New data from a new and never used before dataset, add fresh data, no LimaRP spam, this is 100% new. Thanks to the [MinvervaAI Team](https://huggingface.co/MinervaAI) and, in particular, [Gryphe](https://huggingface.co/Gryphe) for letting us use it! |
|
- [Another private Aesir dataset] |
|
- [Another private Aesir dataset] |
|
- [limarp](https://huggingface.co/datasets/lemonilia/LimaRP) |
|
|
|
## DPO training data |
|
|
|
- [Intel/orca_dpo_pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs) |
|
- [NobodyExistsOnTheInternet/ToxicDPOqa](https://huggingface.co/datasets/NobodyExistsOnTheInternet/ToxicDPOqa) |
|
- [Undi95/toxic-dpo-v0.1-NoWarning](https://huggingface.co/datasets/Undi95/toxic-dpo-v0.1-NoWarning) |
|
|