Upload README.md
Browse files
README.md
CHANGED
@@ -22241,7 +22241,7 @@ Notably, we do not use the popular MS-MARCO retrieval dataset in our training co
|
|
22241 |
We train Granite Embedding Models using IBM's computing cluster, Cognitive Compute Cluster, which is outfitted with NVIDIA A100 80gb GPUs. This cluster provides a scalable and efficient infrastructure for training our models over multiple GPUs.
|
22242 |
|
22243 |
**Ethical Considerations and Limitations:**
|
22244 |
-
The data used to train the base language model was filtered to remove text containing hate, abuse, and profanity. Granite-Embedding-278m-Multilingual is
|
22245 |
|
22246 |
**Resources**
|
22247 |
- ⭐️ Learn about the latest updates with Granite: https://www.ibm.com/granite
|
|
|
22241 |
We train Granite Embedding Models using IBM's computing cluster, Cognitive Compute Cluster, which is outfitted with NVIDIA A100 80gb GPUs. This cluster provides a scalable and efficient infrastructure for training our models over multiple GPUs.
|
22242 |
|
22243 |
**Ethical Considerations and Limitations:**
|
22244 |
+
The data used to train the base language model was filtered to remove text containing hate, abuse, and profanity. Granite-Embedding-278m-Multilingual is finetuned on 12 languages, and has a context length of 512 tokens (longer texts will be truncated to this size).
|
22245 |
|
22246 |
**Resources**
|
22247 |
- ⭐️ Learn about the latest updates with Granite: https://www.ibm.com/granite
|