As opposed to scaling-up protein language models (PLMs), we seek improving performance via protein-specific optimization. Although the proportionality between the language model size and the richness of its learned representations is validated, we prioritize accessibility and pursue a path of data-efficient, cost-reduced, and knowledge-guided optimization. Through over twenty experiments ranging from masking, architecture, and pre-training data, we derive insights from protein-specific experimentation into building a model that interprets the language of life, optimally. We present Ankh, the first general-purpose PLM trained on Google’s TPU-v4 surpassing the state-of-the-art performance with fewer parameters (<10% for pre-training, <7% for inference, and <30% for the embedding dimension). We provide a representative range of structure and function benchmarks where Ankh excels. We further provide a protein variant generation analysis on High-N and One-N input data scales where Ankh succeeds in learning protein evolutionary conservation-mutation trends and introducing functional diversity while retaining key structural-functional characteristics. We dedicate our work to promoting accessibility to research innovation via attainable resources.
Notes: Table 9 from here: https://www.biorxiv.org/content/10.1101/2023.07.05.547496v1.full.pdf
Size Notes: Pretrained over UniRef50; 45M proteins and 14B amino acids, per Table 2 952B tokens from Table 9 at: https://www.biorxiv.org/content/10.1101/2023.07.05.547496v1 (This is total tokens over multiple epochs)
Notes: Figure 1 indicates 450M, but the model on huggingface https://huggingface.co/ElnaggarLab/ankh-base as well as Table 9 from https://www.biorxiv.org/content/10.1101/2023.07.05.547496v1.full.pdf each indicate 740M parameters. Notebook for counting params: https://colab.research.google.com/drive/1EGI5_vDl4pOBUukJexMHQR16BFKJe4a5?usp=sharing