Researchers at EPFL, ETH Zurich and CSCS have developed Apertus, a fully open Large Language Model (LLM) – one of the largest of its kind. As a foundational technology, Apertus enables innovation and strengthens AI expertise across research, society and industry by allowing others to build upon it. Apertus is currently available through strategic partner Swisscom, the AI platform Hugging Face, and the Public AI network.
Notes: 6.74 · 10^24 FLOPs - reported 6 FLOP / parameter / token * 70 * 10^9 parameters * 15 * 10^12 tokens = 6.3e+24 FLOP 989500000000000 FLOP / GPU / sec [GH200, bf16 reported] * 6*10^6 GPU-hours * 3600 sec / hour * 0.3 [assumed utilization] = 6.41196e+24 FLOP
Size Notes: "Trained on 15 trillion tokens across more than 1,000 languages – 40% of the data is non-English"
Notes: 70B