Lucie LLM
Collection
Open weights LLM for French, English, German, Spanish and Italian
•
6 items
•
Updated
•
13
This repository contains checkpoints (splitted for 512 GPUs) in DeepSpeed format for the Lucie-7B model,
which was trained using this repository of code
based on a fork of Megatron-Deepspeed
.
Each checkpoint is in a subbranch (revision), which names specifies the number of training steps.
For instance step0400000
corresponds to the checkpoint after 4M training steps.
Those checkpoints are provided so that the model can be retrained from a given point.