Text Generation
llama
pretrained
llama-3
openllm-france
Jeronymous commited on
Commit
93da255
·
verified ·
1 Parent(s): a511b4f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -0
README.md CHANGED
@@ -1,6 +1,7 @@
1
  ---
2
  license: apache-2.0
3
  pipeline_tag: text-generation
 
4
  language:
5
  - fr
6
  - en
@@ -26,3 +27,19 @@ widget:
26
  # top_k: null
27
  # max_new_tokens: null
28
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  pipeline_tag: text-generation
4
+ base_model: OpenLLM-France/Lucie-7B
5
  language:
6
  - fr
7
  - en
 
27
  # top_k: null
28
  # max_new_tokens: null
29
  ---
30
+
31
+ # Model Card
32
+
33
+ This repository contains checkpoints (splitted for 512 GPUs) in DeepSpeed format for the [Lucie-7B model](https://huggingface.co/OpenLLM-France/Lucie-7B),
34
+ which was trained using [this repository of code](https://github.com/OpenLLM-France/Lucie-Training)
35
+ based on [a fork of `Megatron-Deepspeed`](https://github.com/OpenLLM-France/Megatron-DeepSpeed).
36
+
37
+ Each checkpoint is in a subbranch (revision), which names specifies the number of training steps.
38
+ For instance `step0400000` corresponds to the checkpoint after 4M training steps.
39
+
40
+ Those checkpoints are provided so that the model can be retrained from a given point.
41
+
42
+ ## Contact
43
+
44
45
+