michaelfeil
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ This model extends LLama-3 8B's context length from 8k to > 160K, developed by G
|
|
19 |
|
20 |
**Infra:**
|
21 |
|
22 |
-
We build on top of the EasyContext Blockwise RingAttention library [3] to scalably and efficiently train on contexts up to
|
23 |
|
24 |
**Data:**
|
25 |
|
|
|
19 |
|
20 |
**Infra:**
|
21 |
|
22 |
+
We build on top of the EasyContext Blockwise RingAttention library [3] to scalably and efficiently train on contexts up to 262144 tokens on [Crusoe Energy](https://huggingface.co/crusoeai) high performance L40S cluster.
|
23 |
|
24 |
**Data:**
|
25 |
|