MM-BCD-SCA Algorithm in IVA Optimization
- The MM-BCD-SCA algorithm is a method for optimizing IVA via majorization-minimization and block coordinate descent that alternates updates of separation matrices and auxiliary variables.
- It integrates iterative projection (IP) and iterative source steering (ISS) strategies to update the separation matrix with closed-form solutions for small block sizes.
- The framework transforms the cost into a quadratic form using surrogate functions, enabling practical convergence and posing challenges for extending to higher block sizes (d ≥ 3).
The MM-BCD-SCA algorithm refers to a class of methods for majorization-minimization (MM)-based optimization in independent vector analysis (IVA), employing block coordinate descent (BCD) strategies. Within this context, the algorithmic family leverages surrogates for negative log-likelihood, alternating coordinate minimization of the separation matrix and auxiliary variables, using iterative block updates. MM-BCD-SCA (Majorization-Minimization, Block Coordinate Descent, Source Column Algorithm) specifically encompasses approaches such as iterative projection (IP) and iterative source steering (ISS), including their higher-block generalizations such as ISS.
1. Surrogate Function and the Majorization-Minimization Principle
In MM-based IVA, optimization is carried out on a separation matrix by minimizing a surrogate objective given by:
where each is a positive definite matrix computed at each MM iteration from the current separated signals. These matrices are constructed as , with an auxiliary weight matrix updated at each outer iteration. The surrogate is obtained via the super-Gaussian majorization (tight at ), rendering the cost function quadratic in the variables of at each iteration (Ikeshita et al., 2022). This allows for analytical updates in block coordinates.
2. Block Coordinate Descent Methodologies
Block coordinate descent (BCD) reduces by partitioning (or its inverse ) and optimally updating blocks in closed form. Two principal strategies are prominent:
- Iterative Projection (IP): Operates directly on , updating one (IP) or two (IP) rows at a time by solving low-dimensional subproblems, each yielding closed-form updates. For IP, each update computes with . IP solves a subproblem (Ikeshita et al., 2022).
- Iterative Source Steering (ISS): Operates on , updating its columns with the constraint . ISS updates one column at a time, and ISS extends this to updating two columns simultaneously.
For , global minimization of remains intractable; therefore, BCD strategies offer practical convergence and reliable performance empirically.
3. Closed-Form Update Formulas for ISS
Let be the block size (e.g., for ISS, for ISS). The update proceeds by partitioning into blocks of columns. In each BCD step, the update is represented as with , a block-matrix characterized by , , .
The cost decomposes as:
where , .
The Q-step (for ) involves , followed by an update to . The P-step depends on the block size:
- For , .
- For , the update uses eigendecomposition: set , then compute roots , and associated eigenvectors , , normalized as .
For , a closed-form solution for the P-step is not available, which remains an open problem and direction for potential further generalizations.
4. Unified Framework for ISS Methods
The MM-BCD-SCA framework allows systematic development of block-update source steering methods for general block sizes dividing . The unified recipe is as follows (Ikeshita et al., 2022):
- Partition the mixing matrix into blocks of columns.
- For each block, optionally permute so the active block is leading.
- For each block, update by first solving the Q-step (closed-form for any ), then the P-step (analytical for , open for ).
- Update the separated signals and auxiliary matrix accordingly.
- Iterate to convergence of the MM surrogate.
ISS and ISS can be directly obtained via this systematic process, and the framework is extensible, conditional on analytical tractability of the P-step for higher .
5. Algorithmic Workflow and Pseudocode
A typical majorization-minimization cycle for ISS proceeds as follows:
- Initialize (optionally via data whitening), set for each dataset .
- Outer MM loop: update .
- For each block (inner BCD cycle):
- For each and , compute and , update by subtracting .
- Compute , build , eigenvalues , eigenvectors , set .
- Update .
- Permute rows of and by the block-cycle permutation . This cycle is iterated to MM convergence, and outputs the separated .
6. Computational Complexity and Empirical Performance
The computational overhead per MM iteration for ISS methods depends on block size :
- ISS (one column):
- ISS (two columns):
- IP/IP (one/two rows):
Here, is the number of datasets, is the number of sensors, the signal length. The dominant cost for ISS methods is the computation of and updates.
Numerical experiments on reverberant speech mixtures (for ) indicate that IP and ISS converge in approximately $20$ to $30$ MM iterations to a target source-to-distortion ratio (SDR), while ISS and IP require more iterations (). ISS achieves the convergence rate of IP while retaining the $1/m$ computational complexity advantage characteristic of ISS-type updates (Ikeshita et al., 2022).
7. Extensions and Open Problems
The ISS framework, as outlined, immediately yields practical and efficient algorithms for and . For , the P-step does not admit an analytical solution, which delineates a notable open problem for extending the family to higher block-sizes. This suggests future research directions aimed at either characterizing closed-form solutions for or developing efficient numerical subroutines for high-dimensional block updates. The MM-BCD-SCA paradigm offers a systematic path, unifying disparate algorithms under a common MM and BCD scaffold and facilitating informed tradeoffs between convergence speed and per-iteration cost (Ikeshita et al., 2022).