--- library_name: transformers license: apache-2.0 metrics: - perplexity base_model: - facebook/esm1b_t33_650M_UR50S --- ## **Pretraining on Combined Phosphosite Data** ESM-1b is trained from scratch by Masked Language Modeling objective. The data is combination of phosphosite data which are used to train **isikz/esm1b_msa_mlm_pt_phosphosite** and **isikz/esm1b_mlm_pt_phosphosite**. The total number of data is 1055221.