xTrimoPGLM: Unified 100B-Scale Pre-trained Transformer for Deciphering the Language of Protein

Study explores the compatibility and potential for joint optimization of autoencoding and autoregressive pretraining objectives. They train a model at [what I think is] a huge scale and get results suggesting that it significantly outperforms other baselines and facilitates an atomic-resolution view of protein structures. This is another step towards protein foundation models.

3