Max-Min-Max Submodular Optimization
- Max-min-max submodular optimization is a framework for selecting a subset that maximizes the worst-case value across multiple monotone submodular functions under a budget constraint.
- The approach employs innovative discrete algorithms that iteratively use greedy selection, linear programming, and Monte Carlo evaluations to achieve scalable, near-optimal results.
- Applications span robust experimental design, fair influence maximization, and novel fair centrality maximization, ensuring balanced performance across diverse objectives.
Max-min-max submodular optimization, often referred to in the literature as multiobjective submodular maximization under a cardinality constraint, considers selecting a subset of elements from a finite ground set so as to simultaneously maximize the worst-case value across multiple monotone submodular objective functions. Formally, given submodular functions indexed by and a budget , the problem is to find such that is maximized. This formulation is central to robust combinatorial optimization, encompassing applications in fair influence maximization, robust experimental design, and (as newly introduced) fair centrality maximization, where ensuring good performance under each objective is essential.
1. Formal Problem Definition and Representative Applications
Let be a finite ground set of elements, and let be an index set of size . For each , is a monotone submodular function: for all and , and . The objective is: This framework arises in:
- Robust experimental design: Simultaneously maximizing a family over uncertain parameters .
- Fair influence maximization: Each color denotes a demographic group, with measuring expected influence spread in group .
- Fair centrality maximization: The new application introduced, optimizing groupwise harmonic centrality in graphs after adding up to edges.
2. Continuous Relaxation, Multilinear Extension, and Practical Limitations
Theoretical approaches to this problem have explored continuous relaxations via the multilinear extension. For , define as a random subset of containing independently with probability ; then the multilinear extension is . The relaxed problem is: However, exactly evaluating involves summing over sets and is thus intractable. Practical approaches rely on Monte Carlo estimation or continuous-greedy methods (e.g., Frank–Wolfe), but these require repeated estimation of and its gradients, leading to significant computational overhead especially as and increase.
3. Discrete (Greedy-Style) Asymptotically Optimal Algorithm
A new scalable, discrete algorithm attains a approximation with high probability, avoiding the multilinear extension and relying solely on standard submodular oracle calls. The method constructs iteratively via rounds, where in each round it solves a linear program (LP) over the simplex to select an element to add:
Algorithm (sketch):
- Run independent trials.
- For each trial: a. Initialize . b. For to , i. Let . ii. Solve the LP over and : maximize subject to for every , , . iii. Sample and add to . c. If , update .
- Return .
Key performance results are as follows:
- In expectation over the random process, for each , .
- With high probability (martingale concentration argument, Theorem 6), for all , provided , where .
The algorithm relies only on computing and via submodular oracles. The LP can be efficiently approximated via a multiplicative-weights (MWU) subroutine and lazy evaluations in oracle calls per outer iteration.
4. Algorithmic Rounding and Ensuring Integral Solutions
Since the main greedy step maintains as an integral set at all times, explicit rounding is unnecessary. To remove the technical requirement , a preprocessing phase identifies and includes up to elements of highest marginal gain (across colors), forming a set . Modified objectives are constructed, now with all singleton marginals . A continuous relaxation is then run on budget , yielding a fractional solution that is rounded via "swap rounding" to an integral set . Lemma 9 and swap rounding analysis ensure a final approximation guarantee for all .
5. Computational Complexity and Scalability Features
The algorithm achieves practical scalability under the following resource bounds:
- Submodular oracle calls: .
- Total running time: .
Crucial speed-ups include:
- The MWU approach to LP solving with rounds and each round using lazy marginal gain bounding.
- Preprocessing to reduce the impact of large-gain elements and further control the dependence.
- Lazy evaluations of marginal gains (maintaining upper bounds ).
Empirically, the LP can be solved via standard solvers (e.g., Gurobi) or by MWU. The introduction of a "tilt" parameter in the LP objective biases the allocation toward colors currently yielding the minimum value, which improves practical convergence.
6. Applications: Fair Centrality Maximization
A significant new application is groupwise harmonic centrality in networks. For a node in a directed graph , classical harmonic centrality is ; adding an edge to increases this quantity submodularly. The fair variant seeks to maximize
Selecting up to edges to add to defines as the post-edit, groupwise normalized harmonic sum. Each is nonnegative monotone submodular, and the resulting task is . Standard continuous methods fail to scale to graphs with tens of thousands of nodes, whereas the new discrete method retains its theoretical and practical guarantees in this regime.
7. Empirical Performance and Comparative Analysis
Experiments were performed on:
- Max--cover instances (, ) from stochastic Kronecker, Barabási–Albert, and Erdős–Rényi models.
- 20 Amazon co-purchase networks (up to nodes) for fair centrality with .
- Simulated Antelope Valley social networks (, up to 13) for fair influence maximization.
Compared algorithms included the LP Greedy method (with both Gurobi and MWU linear program solving, plus lazy updates), round-robin greedy (Udwani-style), Saturate (bi-criteria method), Udwani’s MWU ((1-1/e)-approximation), and Frank–Wolfe continuous method (for influence). Major findings:
- LP Greedy achieves the highest min-cover on max--cover, with fewer oracle calls than MWU.
- On fair centrality, LP Greedy outperforms Saturate and MWU both in objective and running time, solving up to 10,000-node graphs in minutes.
- For fair influence, LP Greedy matches or outperforms Frank–Wolfe for nontrivial budgets and is more broadly applicable.
- Ablation studies reveal that 20 outer repetitions suffice and a tilt factor optimizes practical performance.
Overall, this algorithm bridges the prior theoretical-practical gap, attaining the asymptotically optimal approximation via efficient, scalable discrete algorithms deployable on large-scale real-world tasks in fair optimization of submodular objectives.