Come ask questions about everything on the new HuggingFace BigScience language model, data, the licenses, and the cluster!
We have started training a 176B parameter multilingual language model on the French supercomputer Jean Zay – literally out in the open. This is not only the first time a multilingual LLM (46 languages!) at this scale will be fully accessible to the ML research community, but the whole decision, engineering and training process is transparent and open. We'll be training for several months and the community can follow along, engage and ask questions, etc. For regular training updates follow @BigScienceLLM.
The model, compute and training
More info
Model / Engineering