BigScience Reddit AMA on Thursday, March 24th, 2022

Come ask questions about everything on the new HuggingFace BigScience language modeldata, the licenses, and the cluster!

We have started training a 176B parameter multilingual language model on the French supercomputer Jean Zay – literally out in the open. This is not only the first time a multilingual LLM (46 languages!) at this scale will be fully accessible to the ML research community, but the whole decision, engineering and training process is transparent and open. We'll be training for several months and the community can follow along, engage and ask questions, etc. For regular training updates follow @BigScienceLLM.

The model, compute and training

More info

For the AMA we’re joined by

Model / Engineering