Comparative performance of SGPO methods under low-throughput wet-lab constraints
Determine how existing Steered Generation for Protein Optimization methods—including classifier guidance and posterior sampling for diffusion models as well as reinforcement learning fine-tuning of protein language models—perform and compare to each other in real-world protein optimization campaigns where fitness is measured via low-throughput wet-lab assays with only hundreds of labeled sequence–fitness pairs, in order to establish practical best practices for protein fitness optimization with limited experimental feedback.
References
However, by and large, past studies have optimized surrogate rewards and/or utilized large amounts of labeled data for steering, making it unclear how well existing methods perform and compare to each other in real-world optimization campaigns where fitness is measured by low-throughput wet-lab assays.