This is Gemma 1.1 7B (IT), an update over the original instruction-tuned Gemma release. Gemma 1.1 was trained using a novel RLHF method, leading to substantial gains on quality, coding capabilities, factuality, instruction following and multi-turn conversation quality. We also fixed a bug in multi-turn conversations, and made sure that model responses don't always start with "Sure,". We believe this release represents an improvement for most use cases, but we encourage users to test in their particular applications. The previous model will continue to be available in the same repo. We appreciate the enthusiastic adoption of Gemma, and we continue to welcome all feedback from the community.
Notes: 6ND = 6*6000000000000*8540000000=3.0744e+23
Size Notes: "These models were trained on a dataset of text data that includes a wide variety of sources, totaling 6 trillion tokens. "
Notes: Safetensors Model size 8.54B params