Rogendo commited on
Commit
89b776a
·
verified ·
1 Parent(s): b1a2588

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -15
README.md CHANGED
@@ -10,41 +10,51 @@ This is a pre-trained language translation model that aims to create a translati
10
 
11
  ## Model Details
12
 
13
- - Transformer architecture used.
14
- - Trained on a 210000 corpus pairs.
15
- - Pre-trained Helsinki-NLP/opus-mt-en-swc.
16
- - 2 models to enforce biderectional translation.
17
  ### Model Description
18
 
19
  <!-- Provide a longer summary of what this model is. -->
20
 
21
 
22
 
23
- - **Developed by:** [Peter Rogendo]
24
- - **Funded by [optional]:** [More Information Needed]
25
- - **Shared by [optional]:** [More Information Needed]
26
- - **Model type:** [More Information Needed]
27
- - **Language(s) (NLP):** [More Information Needed]
28
- - **License:** [More Information Needed]
29
- - **Finetuned from model [optional]:** [More Information Needed]
30
 
31
  ### Model Sources [optional]
32
 
33
  <!-- Provide the basic links for the model. -->
34
 
35
- - **Repository:** [More Information Needed]
36
- - **Paper [optional]:** [More Information Needed]
37
- - **Demo [optional]:** [More Information Needed]
38
 
39
  ## Uses
40
 
41
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
 
42
 
43
  ### Direct Use
44
 
45
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
46
 
47
- [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
48
 
49
  ### Downstream Use [optional]
50
 
 
10
 
11
  ## Model Details
12
 
13
+ - Transformer architecture used
14
+ - Trained on a 210000 corpus pairs
15
+ - Pre-trained Helsinki-NLP/opus-mt-en-swc
16
+ - 2 models to enforce biderectional translation
17
  ### Model Description
18
 
19
  <!-- Provide a longer summary of what this model is. -->
20
 
21
 
22
 
23
+ - **Developed by:** Peter Rogendo
24
+ - **Model type:** Transformer
25
+ - **Language(s) (NLP):** Transformer, Pandas, Numpy
26
+ - **License:** Distributed under the MIT License
27
+ - **Finetuned from model [Helsinki-NLP/opus-mt-en-swc]:** [This pre-trained model was re-trained on a swahili-english sentence pairs that were collected across Kenya. Swahili is the national language and is among the top three of the most spoken language in Africa. The sentences that were used to train this model were 210000 in total.]
 
 
28
 
29
  ### Model Sources [optional]
30
 
31
  <!-- Provide the basic links for the model. -->
32
 
33
+ - **Repository:** [https://github.com/Rogendo/Eng-Swa-Translator]
34
+ - **Paper [optional]:**
35
+ - **Demo [optional]:**
36
 
37
  ## Uses
38
 
39
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
40
+ This translation model is intended to be used in many cases, from language translators, screen assistants, to even in official cases such as translating legal documents.
41
 
42
  ### Direct Use
43
 
44
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
45
 
46
+ # Use a pipeline as a high-level helper
47
+
48
+ from transformers import pipeline
49
+
50
+ pipe = pipeline("text2text-generation", model="Rogendo/sw-en")
51
+
52
+ # Load model directly
53
+
54
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
55
+
56
+ tokenizer = AutoTokenizer.from_pretrained("Rogendo/sw-en")
57
+ model = AutoModelForSeq2SeqLM.from_pretrained("Rogendo/sw-en")
58
 
59
  ### Downstream Use [optional]
60