>
FLOPs6.9e+24
Notes: 6ND: 6 * 115B * 10T = 6.9e24
Training Code AccessibilityApache 2.0 https://huggingface.co/Tele-AI/TeleChat2-115B "TeleChat model supports commercial use if you plan to treat TeleChat The model or its derivatives are used for commercial purposes, you need to contact the mailbox below tele_ai@chinatelecom.cn" no clear license but same disclaimer as above https://github.com/Tele-AI/TeleChat2/ this is seems to be pre-training code: https://github.com/Tele-AI/TeleChat2/tree/main/deepspeed
Size Notes: The open source TeleChat2-115B model is trained using 10 trillion tokens of high-quality Chinese and English corpus
Parameters115000000000