This is Gemma 1.1 7B (IT), an update over the original instruction-tuned Gemma release. Gemma 1.1 was trained using a novel RLHF method, leading to substantial gains on quality, coding capabilities, factuality, instruction following and multi-turn conversation quality. We also fixed a bug in multi-turn conversations, and made sure that model responses don't always start with "Sure,". We believe this release represents an improvement for most use cases, but we encourage users to test in their particular applications. The previous model will continue to be available in the same repo. We appreciate the enthusiastic adoption of Gemma, and we continue to welcome all feedback from the community.
FLOPs3.07e+23
Notes: 6ND = 6*6000000000000*8540000000=3.0744e+23
Training Code Accessibilityhttps://huggingface.co/google/gemma-1.1-7b-it "This repository is publicly accessible, but you have to accept the conditions to access its files and content."
HardwareGoogle TPU v5e
Size Notes: "These models were trained on a dataset of text data that includes a wide variety of sources, totaling 6 trillion tokens. "
Parameters8540000000
Notes: Safetensors Model size 8.54B params