Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
140 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Maximum Discrepancy Generative Regularization and Non-Negative Matrix Factorization for Single Channel Source Separation (2404.15296v1)

Published 26 Mar 2024 in math.NA, cs.LG, cs.NA, eess.SP, and stat.ML

Abstract: The idea of adversarial learning of regularization functionals has recently been introduced in the wider context of inverse problems. The intuition behind this method is the realization that it is not only necessary to learn the basic features that make up a class of signals one wants to represent, but also, or even more so, which features to avoid in the representation. In this paper, we will apply this approach to the training of generative models, leading to what we call Maximum Discrepancy Generative Regularization. In particular, we apply this to problem of source separation by means of Non-negative Matrix Factorization (NMF) and present a new method for the adversarial training of NMF bases. We show in numerical experiments, both for image and audio separation, that this leads to a clear improvement of the reconstructed signals, in particular in the case where little or no strong supervision data is available.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Image decomposition into a bounded variation component and an oscillating component. J. Math. Imaging Vision, 22(1):71–88, 2005.
  2. Random search for hyper-parameter optimization. J. Mach. Learn. Res., 13:281–305, feb 2012.
  3. Simultaneous structure and texture image inpainting. IEEE Trans. Image Process., 12(8):882–889, 2003.
  4. Li Deng. The MNIST database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6):141–142, 2012.
  5. Regularising inverse problems with generative machine learning models. Journal of Mathematical Imaging and Vision, 66:37–56, 2024.
  6. Single-channel audio source separation with NMF: divergences, constraints and algorithms. Audio Source Separation, pages 1–24, 2018.
  7. Multi-parameter approaches in image processing. In Handbook of mathematical models and algorithms in computer vision and imaging—mathematical imaging and vision, pages 943–967. Springer, Cham, [2023] ©2023.
  8. A kernel two-sample test. The Journal of Machine Learning Research, 13(1):723–773, 2012.
  9. Array programming with NumPy. Nature, 585(7825):357–362, September 2020.
  10. The Elements of Statistical Learning. Springer Series in Statistics. Springer New York Inc., New York, NY, USA, 2001.
  11. Source separation with deep generative priors. In International Conference on Machine Learning, pages 4724–4735. PMLR, 2020.
  12. Deep convolutional neural network for inverse problems in imaging. IEEE Transactions on Image Processing, 26(9):4509–4522, 2017.
  13. Sparse NMF–half-baked or well done? Mitsubishi Electric Research Labs (MERL), Cambridge, MA, USA, Tech. Rep., no. TR2015-023, 11:13–15, 2015.
  14. SDR–half-baked or well done? In ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 626–630. IEEE, 2019.
  15. Algorithms for non-negative matrix factorization. Advances in neural information processing systems, 13, 2000.
  16. Learning the parts of objects by non-negative matrix factorization. Nature, 401(6755):788–791, 1999.
  17. Blind audio source separation with minimum-volume beta-divergence nmf. IEEE Transactions on Signal Processing, 68:3400–3410, 2020.
  18. A systematic survey of regularization and normalization in gans. ACM Computing Surveys, 55(11):1–37, 2023.
  19. Adversarial regularizers in inverse problems. Advances in neural information processing systems, 31, 2018.
  20. librosa: Audio and music signal analysis in python. In Proceedings of the 14th python in science conference, volume 8, pages 18–25, 2015.
  21. Learned reconstruction methods with convergence guarantees: a survey of concepts and applications. IEEE Signal Processing Magazine, 40(1):164–182, 2023.
  22. Alfred Müller. Integral probability metrics and their generating classes of functions. Advances in applied probability, 29(2):429–443, 1997.
  23. Librispeech: An ASR corpus based on public domain audio books. In 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 5206–5210, 2015.
  24. Mini-batch stochastic approaches for accelerated multiplicative updates in nonnegative matrix factorisation with beta-divergence. In 2016 IEEE 26th International Workshop on Machine Learning for Signal Processing (MLSP), pages 1–6. IEEE, 2016.
  25. Musan: A music, speech, and noise corpus. arXiv preprint arXiv:1510.08484, 2015.
  26. Generative adversarial source separation. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 26–30. IEEE, 2018.
  27. Modeling textures with total variation minimization and oscillating patterns in image processing. J. Sci. Comput., 19(1-3):553–572, 2003. Special issue in honor of the sixtieth birthday of Stanley Osher.
  28. Audio source separation and speech enhancement. John Wiley & Sons, 2018.
  29. Supervised speech separation based on deep learning: An overview. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 26(10):1702–1726, 2018.
  30. Discriminative NMF and its application to single-channel source separation. In Interspeech, pages 865–869, 2014.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets