ETHZ and EPFL announced the release of a Large Language Model (LLM) developed on public infrastructure: Trained on the “Alps” supercomputer at the Swiss National Supercomputing Centre (CSCS) in 8B and 70B parameters configurations, using open-source training data, respecting web crawling opt-outs during data acquisition, and natively fluent in over 1000 languages. Quoting: "The model will be fully open: source code and weights will be publicly available, and the training data will be transparent and reproducible".

I don't know how good it's going to be, but if true for me this is the real definition of "open-source" in AI (not the ridiculous, corporate-promiscuous definition by the Open Source Initiative).

ethz.ch/en/news-and-events/eth

0
0
0

If you have a fediverse account, you can quote this note from your own instance. Search https://infosec.exchange/users/tomgag/statuses/114836639270348675 on your instance and quote it. (Note that quoting is not supported in Mastodon.)