The most popular and comprehensive Open Source ECM platform
BLOOM: The Open-Source Alternative to GPT-3-like Large Language Models
OpenAI’s GPT-3 may have been first, or at least first to generate a widespread ‘wow’ from people once they saw what AI Large learning Models can do. Since then there have been competitors, like Google’s LaMDA, China’s Wu Dao 2.0, MUM, MetaAI’s OPT, and other models, trained on massive data sets to have billions and even trillions of parameters.
And now there is BLOOM (BigScience Large Open-science Open-access Multilingual Language Model), an Open Source alternative to GPT-3. It was created by the BigScience research project, a public-private collaboration with more than 1000 contributors, that was launched in 2021. The BLOOM model was trained on the new Jean Zay supercomputer in France and funded by a 3 million euro budget from a French research agency grant.
Thomas Wolf, the BigScience co-lead and Hugging Face co-founder, said that “large ML models have changed the world of AI research over the last two years but the huge compute cost necessary to train them resulted in very few teams actually having the ability to train and research them.”
Teven Le Scao, AI engineer at Hugging Face, told VentureBeat that while “GPT-3 is monolingual, BLOOM was designed from the start to be multilingual so it was trained on several languages, and also to incorporate a significant amount of programming language data. BLOOM supports 46 human languages and 13 programming languages — so that’s a very sizable difference.”
Alberto Romero, CambrianAI analyst, wrote that “BigScience and BLOOM are, without a doubt, the most notable attempt at bringing down all the barriers that big tech has erected — willingly or unwillingly — throughout the last decade in the AI field. And BLOOM is also the most sincere and honest undertaking to building AI (large language models in particular) that benefits everyone.”