Leveraging a Small Language Model to Achieve Full-Length mRNA Analysis with Improved Prediction of Stability & Translation Efficiency

Time: 12:25 pm
day: Day One

Details:

  • CodonBERT is pre-trained on 10 million mRNA sequences to capture codon dependencies and interactions across different species
  • mRNA-LM extends CodonBERT from the CDS to full-length mRNA by integrating two pre-trained BERT models on UTR regions
  • Both CodonBERT and mRNA-LM show significant improvements in predicting mRNA properties compared to existing methods

Speakers: