Update README.md
Browse files
README.md
CHANGED
@@ -222,10 +222,22 @@ print("Generated Translations:", results_detokenized)
|
|
222 |
|
223 |
|
224 |
|
225 |
-
|
226 |
## Evaluation
|
227 |
|
228 |
-
Below are the evaluation results on Flores-200 dev and devtest compared to NLLB-3.3 ([Costa-jussà et al., 2022](https://arxiv.org/abs/2207.04672)) for CA-XX and XX-CA directions.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
229 |
|
230 |
|
231 |
#### Flores200-dev
|
|
|
222 |
|
223 |
|
224 |
|
|
|
225 |
## Evaluation
|
226 |
|
227 |
+
Below are the evaluation results on Flores-200 dev and devtest compared to NLLB-3.3 ([Costa-jussà et al., 2022](https://arxiv.org/abs/2207.04672)) for CA-XX and XX-CA directions. The metrics have been computed excluding Asturian, Aranese, and Aragonese as we report them separately. The evaluation was conducted using [MT Lens](https://github.com/langtech-bsc/mt-evaluation). We report the following metrics:
|
228 |
+
|
229 |
+
<details>
|
230 |
+
<summary>Click to show metrics details</summary>
|
231 |
+
</details>
|
232 |
+
|
233 |
+
- `BLEU`: Sacrebleu implementation. Signature: nrefs:1— case:mixed— eff:no— tok:13a— smooth:exp—version:2.3.1
|
234 |
+
- `TER`: Sacrebleu implementation.
|
235 |
+
- `ChrF`: Sacrebleu implementation.
|
236 |
+
- `Comet`: Model checkpoint: "Unbabel/wmt22-comet-da".
|
237 |
+
- `Comet-kiwi`: Model checkpoint: "Unbabel/wmt22-cometkiwi-da".
|
238 |
+
- `Bleurt`: Model checkpoint: "lucadiliello/BLEURT-20".
|
239 |
+
|
240 |
+
</details>
|
241 |
|
242 |
|
243 |
#### Flores200-dev
|