Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
datasets:
|
3 |
+
- cuad
|
4 |
+
- theatticusproject/cuad
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
pipeline_tag: question-answering
|
8 |
+
---
|
9 |
+
|
10 |
+
|
11 |
+
# BERT-large fine-tuned on CUAD
|
12 |
+
|
13 |
+
This is a **BERT-large** model ([`bert-large-uncased-whole-word-masking`][2]) fine-tuned on the [**CUAD**][3] dataset
|
14 |
+
from [*CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review* (Hendrycks et al., 2021)][1], with the **BertforQuestionAnswering** model architecture.
|
15 |
+
|
16 |
+
The questions ask for information often found in contracts;
|
17 |
+
the model would return the relevant text string and its starting index in the given document if the information exists.
|
18 |
+
The CUAD dataset is in SQuAD 2.0 format.
|
19 |
+
|
20 |
+
For details of the dataset and usage of the relevant training/testing scripts, check out the paper and their [Github repo][4].
|
21 |
+
|
22 |
+
[1]: https://arxiv.org/abs/2103.06268
|
23 |
+
[2]: https://huggingface.co/bert-large-uncased-whole-word-masking
|
24 |
+
[3]: https://www.atticusprojectai.org/cuad
|
25 |
+
[4]: https://github.com/TheAtticusProject/cuad
|