Contrastive learning for enzyme class prediction improves accuracy, reliability, and sensitivity and identifies promiscuous activity that is then verified in the lab!
Tianhao Yu, @OceanHCui @luoyunan @HuiminZhaoLab
| Website | https://yangkky.github.io/ |
Contrastive learning for enzyme class prediction improves accuracy, reliability, and sensitivity and identifies promiscuous activity that is then verified in the lab!
Tianhao Yu, @OceanHCui @luoyunan @HuiminZhaoLab
RT @ml4proteins
Next week on 4/11 @ 4 pm EST, we'll have @NotinPascal talk about hybrid protein language models for fitness prediction!
Using human-readible text as input and output is incredibly dumb and tedious
---
RT @Abebab
aii, let's do this... AI/ML version
what’s your critical hot take on AI/ML that would have you in this position
https://twitter.com/Abebab/status/1642550494265589761
Machine learning to predict the binding energy between an enzyme and its ligand.
Carlos Ramírez-Palacios @CG_Martini
Finetune a masked language model to make edits that improve some function of a sequence in order to find sequences that are better than anything in the training set.
@vishakh_pk Richard Yuanzhe Pang @hhexiy @ank_parikh
We study the problem of extrapolative controlled generation, i.e., generating sequences with attribute values beyond the range seen in training. This task is of significant importance in automated design, especially drug discovery, where the goal is to design novel proteins that are \textit{better} (e.g., more stable) than existing sequences. Thus, by definition, the target sequences and their attribute values are out of the training distribution, posing challenges to existing methods that aim to directly generate the target sequence. Instead, in this work, we propose Iterative Controlled Extrapolation (ICE) which iteratively makes local edits to a sequence to enable extrapolation. We train the model on synthetically generated sequence pairs that demonstrate small improvement in the attribute value. Results on one natural language task (sentiment analysis) and two protein engineering tasks (ACE2 stability and AAV fitness) show that ICE considerably outperforms state-of-the-art approaches despite its simplicity. Our code and models are available at: https://github.com/vishakhpk/iter-extrapolation.
RT @ChelseaParlett
3 Lectures
118 Slides
♾ hours making graphics
A peek at the graphics I use to explain Recurrent Neural Networks and related topics😅