In this technical report, we propose YAYI 2, including both base and chat models, with 30 billion parameters. YAYI 2 is pre-trained from scratch on a multilingual corpus which contains 2.65 trillion tokens filtered by our pre-training data processing pipeline.
FLOPs4.77e+23
Notes: 1000 A800 GPUs
Training Code Accessibilityunclear license https://huggingface.co/wenge-research/yayi2-30b this license file https://github.com/wenge-research/YAYI2/blob/main/COMMERCIAL_LICENSE prohibits particular usage (i.e. military, direct competitors) To use YAYI2 models commercially, you must apply for a commercial license. apache 2.0 for code
HardwareNVIDIA A800 PCIe 40 GB
Hardware Quantity1000
Parameters30000000000