Enhancing selectivity using Wasserstein distance based reweighing (2401.11562v2)
Abstract: Given two labeled data-sets $\mathcal{S}$ and $\mathcal{T}$, we design a simple and efficient greedy algorithm to reweigh the loss function such that the limiting distribution of the neural network weights that result from training on $\mathcal{S}$ approaches the limiting distribution that would have resulted by training on $\mathcal{T}$. On the theoretical side, we prove that when the metric entropy of the input datasets is bounded, our greedy algorithm outputs a close to optimal reweighing, i.e., the two invariant distributions of network weights will be provably close in total variation distance. Moreover, the algorithm is simple and scalable, and we prove bounds on the efficiency of the algorithm as well. As a motivating application, we train a neural net to recognize small molecule binders to MNK2 (a MAP Kinase, responsible for cell signaling) which are non-binders to MNK1 (a highly similar protein). In our example dataset, of the 43 distinct small molecules predicted to be most selective from the enamine catalog, 2 small molecules were experimentally verified to be selective, i.e., they reduced the enzyme activity of MNK2 below 50\% but not MNK1, at 10$\mu$M -- a 5\% success rate.
- Efficient sketches for earth-mover distance, with applications. In 2009 50th Annual IEEE Symposium on Foundations of Computer Science, pages 324–330, 2009.
- C. Bianca and Christian Dogbe. On the existence and uniqueness of invariant measure for multidimensional stochastic processes. Nonlinear Studies - The International Journal, 2017.
- Dirichlet-enhanced spam filtering based on biased samples. In Advances in Neural Information Processing Systems, 2007.
- Quantitative concentration inequalities for empirical measures on non-compact spaces. Probability Theory and Related Fields, 137:541–593, 2007.
- Weighted csiszár-kullback-pinsker inequalities and applications to transportation inequalities. Annales de la Faculté des sciences de Toulouse : Mathématiques, 14(3):331–352, 2005.
- Stochastic gradient descent performs variational inference, converges to limit cycles for deep networks. In International Conference on Learning Representations, 2018.
- Sample selection bias correction theory. In Proceedings of the International Conference on Algorithmic Learning Theory, 2008.
- Joint distribution optimal transportation for domain adaptation. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.
- Large Deviations Techniques and Applications. Springer-Verlag Berlin Heidelberg, 2010.
- Correcting sample selection bias in maximum entropy density estimation. In Advances in Neural Information Processing Systems, 2005.
- Faster scaling algorithms for general graph matching problems. Journal of the ACM, 38(4):815–853, 1991.
- Neural message passing for quantum chemistry. In Doina Precup and Yee Whye Teh, editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 1263–1272. PMLR, 06–11 Aug 2017.
- Correcting sample selection bias by unlabeled data. In Advances in Neural Information Processing Systems, 2007.
- A generalized hypergreedy algorithm for weighted perfect matching. In BIT, 1993.
- Molecular graph convolutions: moving beyond fingerprints. In Journal of Computer-Aided Molecular Design, volume 30, page 595–608, 2016.
- On label shift in domain adaptation via wasserstein distance. CoRR, abs/2110.15520, 2021.
- Machine learning on dna-encoded libraries: A new paradigm for hit finding. Journal of Medicinal Chemistry, 63(16):8857–8866, 2020.
- Asher Mullard. DNA tags help the hunt for drugs. Nature, 530:367–369, 2016.
- Dataset Shift in Machine Learning. The MIT Press, 2009.
- On a greedy heuristic for complete matching. In Siam Journal of Computing, pages 676–681, 1981.
- The central role of the propensity score in observational studies for causal effects. In Biometrika, 1983.
- Dna-encoded chemical libraries. Nature Review Methods Primers, 2(3), 2022.
- R. Sharathkumar and Pankaj K. Agarwal. Algorithms for the transportation problem in geometric settings. In Proceedings of the 2012 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 306–317, 2012.
- Hidetoshi Shimodaira. Improving predictive inference under covariate shift by weighting the log-likelihood function. Journal of Statistical Planning and Inference, 90(2):227–244, 2000.
- Pravin Vaidya. Geometry helps in matching. Siam J. of Computing, 18(6):1201–1225, 1989.
- Optimization of molecules via deep reinforcement learning. Scientific Reports, 9(10752), 2019.
Sponsor
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.