# PP489 Optimization Variants Iter1 x TIGIT (YM_0988) ## Overview YM_0988 includes ABC001 against 2 TIGIT homologs. We explored several model hypothesis: (i) Does pre-training aid predicitivity and (ii) does the featurization of the input sequences matter. To test pretraining, we refer to `mata_descriptions` with the term **warm** to include pretraining, and **cold** to start from a randomly initialized seed. For featurization, we explored **label-encoded** sequences with a one-hot-encoder of amino acid identities, versus an **ESM-featurized** embedding to represent each sequence in the PPI. Optimization was performed on the human ortholog. ## Experimental details We studied the efficacy of generating binders with different model hyperparameters. This dataset includes 26726 unique scFvs and 2 unique target sequences. A more extensive methods section can be found in our publication [here](https://pmc.ncbi.nlm.nih.gov/articles/PMC12296056/). ## Misc dataset details We define the following binders: ### A-library (scFvs) There are several terms you can filter by: - `ABC001_WT_`: These are WT replicates. - `ABC001_label_encoded_cold`: Label encoded sequences with no pretraining - `ABC001_label_encoded_warm`: Label encoded sequences with pretraining - `ABC001_esm_cold`: ESM featurized sequences with no pretraining - `ABC001_esm_warm`: ESM featurized sequences with pretraining ### Alpha-library - `TIGIT_22-137_POI-AGA2`: Human TIGIT - `TIGIT_Mouse`: Mouse TIGIT