dariolopez commited on
Commit
1aa8f09
·
1 Parent(s): c8fc3ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -1
README.md CHANGED
@@ -6,4 +6,39 @@ language:
6
  - es
7
  library_name: sentence-transformers
8
  pipeline_tag: question-answering
9
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  - es
7
  library_name: sentence-transformers
8
  pipeline_tag: question-answering
9
+ ---
10
+
11
+ # Model Description
12
+
13
+ The trained model is a fine-tuned version of [PlanTL-GOB-ES/roberta-base-bne](https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne) focused on question/answer using [MS-MARCO dataset translated into Spanish](https://huggingface.co/datasets/dariolopez/ms-marco-es-500k).
14
+
15
+ # How to use
16
+
17
+ In progress
18
+
19
+ # Training
20
+
21
+ * [Base Model](https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne)
22
+ * [Config used to train](https://huggingface.co/dariolopez/roberta-base-bne-finetuned-msmarco-qa-es/blob/main/train_config.json)
23
+ * Dataset: [dariolopez/ms-marco-es](https://huggingface.co/datasets/dariolopez/ms-marco-es) (query - positive - negative)
24
+ * Loss: TripletLoss
25
+
26
+ ## Config
27
+
28
+ ```
29
+ {
30
+ "model_name": "PlanTL-GOB-ES/roberta-base-bne",
31
+ "max_seq_length": 512,
32
+ "epochs": 10,
33
+ "warmup_steps": 1000,
34
+ "batch_size": 16,
35
+ "optimizer_params": {
36
+ "lr": 2e-05
37
+ },
38
+ "loss": "tl",
39
+ "dataset_train_size": 500000,
40
+ "dataset_name": "dariolopez/ms-marco-es-500k",
41
+ "seed": 42,
42
+ "length_embedding": 768
43
+ }
44
+ ```