Update README.md
Browse files
README.md
CHANGED
@@ -10,41 +10,51 @@ This is a pre-trained language translation model that aims to create a translati
|
|
10 |
|
11 |
## Model Details
|
12 |
|
13 |
-
- Transformer architecture used
|
14 |
-
- Trained on a 210000 corpus pairs
|
15 |
-
- Pre-trained Helsinki-NLP/opus-mt-en-swc
|
16 |
-
- 2 models to enforce biderectional translation
|
17 |
### Model Description
|
18 |
|
19 |
<!-- Provide a longer summary of what this model is. -->
|
20 |
|
21 |
|
22 |
|
23 |
-
- **Developed by:**
|
24 |
-
- **
|
25 |
-
- **
|
26 |
-
- **
|
27 |
-
- **
|
28 |
-
- **License:** [More Information Needed]
|
29 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
30 |
|
31 |
### Model Sources [optional]
|
32 |
|
33 |
<!-- Provide the basic links for the model. -->
|
34 |
|
35 |
-
- **Repository:** [
|
36 |
-
- **Paper [optional]:**
|
37 |
-
- **Demo [optional]:**
|
38 |
|
39 |
## Uses
|
40 |
|
41 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
|
|
42 |
|
43 |
### Direct Use
|
44 |
|
45 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
46 |
|
47 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
48 |
|
49 |
### Downstream Use [optional]
|
50 |
|
|
|
10 |
|
11 |
## Model Details
|
12 |
|
13 |
+
- Transformer architecture used
|
14 |
+
- Trained on a 210000 corpus pairs
|
15 |
+
- Pre-trained Helsinki-NLP/opus-mt-en-swc
|
16 |
+
- 2 models to enforce biderectional translation
|
17 |
### Model Description
|
18 |
|
19 |
<!-- Provide a longer summary of what this model is. -->
|
20 |
|
21 |
|
22 |
|
23 |
+
- **Developed by:** Peter Rogendo
|
24 |
+
- **Model type:** Transformer
|
25 |
+
- **Language(s) (NLP):** Transformer, Pandas, Numpy
|
26 |
+
- **License:** Distributed under the MIT License
|
27 |
+
- **Finetuned from model [Helsinki-NLP/opus-mt-en-swc]:** [This pre-trained model was re-trained on a swahili-english sentence pairs that were collected across Kenya. Swahili is the national language and is among the top three of the most spoken language in Africa. The sentences that were used to train this model were 210000 in total.]
|
|
|
|
|
28 |
|
29 |
### Model Sources [optional]
|
30 |
|
31 |
<!-- Provide the basic links for the model. -->
|
32 |
|
33 |
+
- **Repository:** [https://github.com/Rogendo/Eng-Swa-Translator]
|
34 |
+
- **Paper [optional]:**
|
35 |
+
- **Demo [optional]:**
|
36 |
|
37 |
## Uses
|
38 |
|
39 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
40 |
+
This translation model is intended to be used in many cases, from language translators, screen assistants, to even in official cases such as translating legal documents.
|
41 |
|
42 |
### Direct Use
|
43 |
|
44 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
45 |
|
46 |
+
# Use a pipeline as a high-level helper
|
47 |
+
|
48 |
+
from transformers import pipeline
|
49 |
+
|
50 |
+
pipe = pipeline("text2text-generation", model="Rogendo/sw-en")
|
51 |
+
|
52 |
+
# Load model directly
|
53 |
+
|
54 |
+
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
55 |
+
|
56 |
+
tokenizer = AutoTokenizer.from_pretrained("Rogendo/sw-en")
|
57 |
+
model = AutoModelForSeq2SeqLM.from_pretrained("Rogendo/sw-en")
|
58 |
|
59 |
### Downstream Use [optional]
|
60 |
|