Beyond Squared Error: Exploring Loss Design for Enhanced Training of Generative Flow Networks (2410.02596v1)
Abstract: Generative Flow Networks (GFlowNets) are a novel class of generative models designed to sample from unnormalized distributions and have found applications in various important tasks, attracting great research interest in their training algorithms. In general, GFlowNets are trained by fitting the forward flow to the backward flow on sampled training objects. Prior work focused on the choice of training objects, parameterizations, sampling and resampling strategies, and backward policies, aiming to enhance credit assignment, exploration, or exploitation of the training process. However, the choice of regression loss, which can highly influence the exploration and exploitation behavior of the under-training policy, has been overlooked. Due to the lack of theoretical understanding for choosing an appropriate regression loss, most existing algorithms train the flow network by minimizing the squared error of the forward and backward flows in log-space, i.e., using the quadratic regression loss. In this work, we rigorously prove that distinct regression losses correspond to specific divergence measures, enabling us to design and analyze regression losses according to the desired properties of the corresponding divergence measures. Specifically, we examine two key properties: zero-forcing and zero-avoiding, where the former promotes exploitation and higher rewards, and the latter encourages exploration and enhances diversity. Based on our theoretical framework, we propose three novel regression losses, namely, Shifted-Cosh, Linex(1/2), and Linex(1). We evaluate them across three benchmarks: hyper-grid, bit-sequence generation, and molecule generation. Our proposed losses are compatible with most existing training algorithms, and significantly improve the performances of the algorithms concerning convergence speed, sample diversity, and robustness.
- Shun-ichi Amari. Differential-geometrical methods in statistics, volume 28. Springer Science & Business Media, 2012.
- Wasserstein generative adversarial networks. In International conference on machine learning, pp. 214–223. PMLR, 2017.
- Dyngfn: Bayesian dynamic causal discovery using generative flow networks. arXiv preprint arXiv:2302.04178, 2023.
- Flow network based generative models for non-iterative diverse candidate generation. Advances in Neural Information Processing Systems, 34:27381–27394, 2021a.
- Gflownet foundations. arXiv preprint arXiv:2111.09266, 2021b.
- A theory of non-acyclic generative flow networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pp. 11124–11131, 2024.
- Order-preserving gflownets. arXiv preprint arXiv:2310.00386, 2023.
- Bayesian structure learning with generative flow networks. In Uncertainty in Artificial Intelligence, pp. 518–528. PMLR, 2022.
- Joint bayesian inference of graphical structure and parameters with a single generative flow network. Advances in Neural Information Processing Systems, 36, 2024.
- Variational inference via χ𝜒\chiitalic_χ upper bound minimization. Advances in Neural Information Processing Systems, 30, 2017.
- Extreme q-learning: Maxent rl without entropy. arXiv preprint arXiv:2301.02328, 2023.
- Generative flow networks assisted biological sequence editing. In NeurIPS 2023 Generative AI and Biology (GenBio) Workshop, 2023.
- Aligning language models with preferences through f-divergence minimization. arXiv preprint arXiv:2302.08215, 2023.
- Amortizing intractable inference in large language models. arXiv preprint arXiv:2310.04363, 2023.
- Biological sequence design with gflownets. In International Conference on Machine Learning, pp. 9786–9801. PMLR, 2022.
- Gflownets for ai-driven scientific discovery. Digital Discovery, 2(3):557–577, 2023a.
- Multi-objective gflownets. In International conference on machine learning, pp. 14631–14653. PMLR, 2023b.
- Learning energy decompositions for partial inference of gflownets. arXiv preprint arXiv:2310.03301, 2023.
- Pessimistic backward policy for gflownets. arXiv preprint arXiv:2405.16012, 2024.
- Junction tree variational autoencoder for molecular graph generation. In International conference on machine learning, pp. 2323–2332. PMLR, 2018.
- Learning to scale logits for temperature-conditional gflownets. arXiv preprint arXiv:2310.02823, 2023a.
- Local search gflownets. arXiv preprint arXiv:2310.02710, 2023b.
- Ant colony sampling with gflownets for combinatorial optimization. arXiv preprint arXiv:2403.07041, 2024.
- A theory of continuous generative flow networks. In International Conference on Machine Learning, pp. 18269–18300. PMLR, 2023.
- Qgfn: Controllable greediness with action values. arXiv preprint arXiv:2402.05234, 2024.
- Rényi divergence variational inference. Advances in neural information processing systems, 29, 2016.
- On divergences and informations in statistics and information theory. IEEE Transactions on Information Theory, 52(10):4394–4412, 2006.
- Learning gflownets from partial episodes for improved convergence and stability. In International Conference on Machine Learning, pp. 23467–23483. PMLR, 2023.
- Trajectory balance: Improved credit assignment in gflownets. Advances in Neural Information Processing Systems, 35:5955–5967, 2022a.
- Gflownets and variational inference. arXiv preprint arXiv:2210.00580, 2022b.
- Tom Minka et al. Divergence measures and message passing. Technical report, Technical report, Microsoft Research, 2005.
- Maximum entropy gflownets with soft q-learning. In International Conference on Artificial Intelligence and Statistics, pp. 2593–2601. PMLR, 2024.
- f-gan: Training generative neural samplers using variational divergence minimization. Advances in neural information processing systems, 29, 2016.
- Generative augmented flow networks. arXiv preprint arXiv:2210.03308, 2022.
- Better training of gflownets with local credit and incomplete trajectories. In International Conference on Machine Learning, pp. 26878–26890. PMLR, 2023a.
- Stochastic generative flow networks. In Uncertainty in Artificial Intelligence, pp. 1628–1638. PMLR, 2023b.
- Gflownet pretraining with inexpensive rewards. arXiv preprint arXiv:2409.09702, 2024.
- Yury Polyanskiy, 2019. URL: https://people.lids.mit.edu/yp/homepage/data/LN_fdiv.pdf. Last visited on 2024/09/23.
- Thompson sampling for improved exploration in gflownets. arXiv preprint arXiv:2306.17693, 2023.
- Towards understanding and improving gflownet training. In International Conference on Machine Learning, pp. 30956–30975. PMLR, 2023.
- Latent logic tree extraction for event sequence explanation from llms. arXiv preprint arXiv:2406.01124, 2024.
- Generative flow networks as entropy-regularized rl. In International Conference on Artificial Intelligence and Statistics, pp. 4213–4221. PMLR, 2024.
- Amortizing intractable inference in diffusion models for vision, language, and control. arXiv preprint arXiv:2405.20971, 2024.
- Beyond reverse kl: Generalizing direct preference optimization with diverse divergence constraints. arXiv preprint arXiv:2309.16240, 2023.
- Generative flow networks for discrete probabilistic modeling. In International Conference on Machine Learning, pp. 26412–26428. PMLR, 2022.
- Diffusion generative flow samplers: Improving learning signals through partial trajectory optimization. arXiv preprint arXiv:2310.02679, 2023a.
- Let the flows tell: Solving graph combinatorial problems with gflownets. Advances in neural information processing systems, 36:11952–11969, 2023b.
- Distributional gflownets with quantile flows. arXiv preprint arXiv:2302.05793, 2023c.
- Variational f-divergence minimization. arXiv preprint arXiv:1907.11891, 2019.
- Information geometric measurements of generalisation. 1995.
- Sample-efficient multi-objective molecular optimization with gflownets. Advances in Neural Information Processing Systems, 36, 2024.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.