juliehunter
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -66,7 +66,7 @@ The model architecture and hyperparameters are the same as for [Lucie-7B](https:
|
|
66 |
* max learning rate: 3e-5
|
67 |
* min learning rate: 3e-6
|
68 |
|
69 |
-
|
70 |
|
71 |
## Testing the model
|
72 |
|
@@ -150,7 +150,8 @@ Lucie-7B LLM and its training dataset
|
|
150 |
|
151 |
## Acknowledgements
|
152 |
|
153 |
-
This work was performed using HPC resources from GENCI–IDRIS (Grant 2024-GC011015444).
|
|
|
154 |
|
155 |
Lucie-7B was created by members of [LINAGORA](https://labs.linagora.com/) and the [OpenLLM-France](https://www.openllm-france.fr/) community, including in alphabetical order:
|
156 |
Olivier Gouvert (LINAGORA),
|
@@ -173,6 +174,8 @@ and
|
|
173 |
Olivier Ferret (CEA)
|
174 |
for their helpful input.
|
175 |
|
|
|
|
|
176 |
## Contact
|
177 |
|
178 |
|
|
66 |
* max learning rate: 3e-5
|
67 |
* min learning rate: 3e-6
|
68 |
|
69 |
+
<sup>*</sup>As noted above, while Lucie-7B-Instruct is trained on sequences of 4096 tokens, it maintains the capacity of the base model, Lucie-7B, to handle context sizes of up to 32K tokens.
|
70 |
|
71 |
## Testing the model
|
72 |
|
|
|
150 |
|
151 |
## Acknowledgements
|
152 |
|
153 |
+
This work was performed using HPC resources from GENCI–IDRIS (Grant 2024-GC011015444). We gratefully acknowledge support from GENCI and IDRIS and from Pierre-François Lavallée (IDRIS) and Stephane Requena (GENCI) in particular.
|
154 |
+
|
155 |
|
156 |
Lucie-7B was created by members of [LINAGORA](https://labs.linagora.com/) and the [OpenLLM-France](https://www.openllm-france.fr/) community, including in alphabetical order:
|
157 |
Olivier Gouvert (LINAGORA),
|
|
|
174 |
Olivier Ferret (CEA)
|
175 |
for their helpful input.
|
176 |
|
177 |
+
Finally, we thank the entire OpenLLM-France community, whose members have helped in diverse ways.
|
178 |
+
|
179 |
## Contact
|
180 |
|
181 |