Commit
·
1192270
1
Parent(s):
8f7bc52
Update README.md
Browse files
README.md
CHANGED
@@ -15,8 +15,8 @@ datasets:
|
|
15 |
- alisawuffles/WANLI
|
16 |
#pipeline_tag:
|
17 |
#- text-classification
|
18 |
-
widget:
|
19 |
-
|
20 |
|
21 |
model-index: # info: https://github.com/huggingface/hub-docs/blame/main/modelcard.md
|
22 |
- name: DeBERTa-v3-large-mnli-fever-anli-ling-wanli
|
@@ -102,7 +102,7 @@ model-index: # info: https://github.com/huggingface/hub-docs/blame/main/modelca
|
|
102 |
## Model description
|
103 |
This model was fine-tuned on the [MultiNLI](https://huggingface.co/datasets/multi_nli), [Fever-NLI](https://github.com/easonnie/combine-FEVER-NSMN/blob/master/other_resources/nli_fever.md), Adversarial-NLI ([ANLI](https://huggingface.co/datasets/anli)), [LingNLI](https://arxiv.org/pdf/2104.07179.pdf) and [WANLI](https://huggingface.co/datasets/alisawuffles/WANLI) datasets, which comprise 885 242 NLI hypothesis-premise pairs. This model is the best performing NLI model on the Hugging Face Hub as of 06.06.22 and can be used for zero-shot classification. It significantly outperforms all other large models on the [ANLI benchmark](https://github.com/facebookresearch/anli).
|
104 |
|
105 |
-
The foundation model is [DeBERTa-v3-large from Microsoft](https://huggingface.co/microsoft/deberta-v3-large).
|
106 |
|
107 |
## Intended uses & limitations
|
108 |
#### How to use the model
|
|
|
15 |
- alisawuffles/WANLI
|
16 |
#pipeline_tag:
|
17 |
#- text-classification
|
18 |
+
#widget:
|
19 |
+
#- text: "I first thought that I really liked the movie, but upon second thought it was actually disappointing. [SEP] The movie was not good."
|
20 |
|
21 |
model-index: # info: https://github.com/huggingface/hub-docs/blame/main/modelcard.md
|
22 |
- name: DeBERTa-v3-large-mnli-fever-anli-ling-wanli
|
|
|
102 |
## Model description
|
103 |
This model was fine-tuned on the [MultiNLI](https://huggingface.co/datasets/multi_nli), [Fever-NLI](https://github.com/easonnie/combine-FEVER-NSMN/blob/master/other_resources/nli_fever.md), Adversarial-NLI ([ANLI](https://huggingface.co/datasets/anli)), [LingNLI](https://arxiv.org/pdf/2104.07179.pdf) and [WANLI](https://huggingface.co/datasets/alisawuffles/WANLI) datasets, which comprise 885 242 NLI hypothesis-premise pairs. This model is the best performing NLI model on the Hugging Face Hub as of 06.06.22 and can be used for zero-shot classification. It significantly outperforms all other large models on the [ANLI benchmark](https://github.com/facebookresearch/anli).
|
104 |
|
105 |
+
The foundation model is [DeBERTa-v3-large from Microsoft](https://huggingface.co/microsoft/deberta-v3-large). DeBERTa-v3 combines several recent innovations compared to classical Masked Language Models like BERT, RoBERTa etc., see the [paper](https://arxiv.org/abs/2111.09543)
|
106 |
|
107 |
## Intended uses & limitations
|
108 |
#### How to use the model
|