>
Notes: 6ND = 2.7e9 * 1.3e12 * 6 = 2,106E+22 "stable-code-3b is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. "
Size Notes: 1.3T tokens "stable-code-3b is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. "
Notes: 2796431360 from https://huggingface.co/stabilityai/stable-code-3b#model-architecture "stable-code-3b is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. "