Submodular Information Measures (SIM)
- Submodular Information Measures (SIM) are a combinatorial generalization of classical information metrics that replace Shannon entropy with a monotone submodular set function.
- SIMs offer strong theoretical guarantees with efficient greedy algorithms that achieve near-optimal (1-1/e) approximations for tasks like summarization and active learning.
- Applications of SIMs span active learning, privacy filtering, causal inference, and representation learning, often yielding significant performance gains in empirical studies.
A submodular information measure (SIM) is a combinatorial generalization of classical information-theoretic measures, such as entropy, mutual information, and conditional mutual information, in which the foundational role of Shannon entropy is replaced with a general monotone submodular set function. SIMs provide an abstract algebraic framework for modeling information, relevance, coverage, independence, and diversity on arbitrary ground sets, extending beyond random variables to structured data, feature sets, and combinatorial objects. SIMs have deep implications across data subset selection, active learning, summarization, privacy, representation learning, causal inference, and extremal combinatorics.
1. Formal Definitions and Mathematical Structure
Let be a finite ground set and a normalized, monotone, submodular set function: , whenever , and for all , .
The key submodular information measures are:
- Submodular Conditional Gain (SCG):
Intuition: The incremental "utility" provided by beyond .
0
Intuition: The amount of “shared information” or representativeness of 1 with respect to 2.
3
Equivalently, 4. Intuition: Relevance of 5 to 6 penalized by overlap with 7.
These extend immediately to multi-set analogues, total correlation, and composite objectives. For example, the total correlation of 8 disjoint sets is 9 (Majee et al., 2023).
When 0 is the entropy of a collection of random variables, these recover classical Shannon-information measures. For canonical submodular functions like coverage, facility-location, concave-over-modular, or certain graph-cut-type objectives, SIMs coincide exactly with entropic mutual information under explicit constructions (Iyer, 19 Jan 2026).
2. Theoretical Properties: Axioms and Independence
SIMs inherit critical properties from submodularity (Asnani et al., 2021, Iyer et al., 2020):
- Nonnegativity: 1 and 2 for normalized, monotone 3.
- Symmetry: 4.
- Monotonicity: 5 is non-decreasing for fixed 6; 7 is monotone in 8.
- Submodularity in One Argument: 9 is submodular in 0 when 1's third discrete derivatives are non-negative; this holds for facility-location, set cover, concave-over-modular, and some graph-cut functions (Iyer et al., 2020, Kothawade et al., 2021).
- Chain Rule: 2.
Independence concepts are generalized:
- Joint Independence: 3.
- Pairwise Independence: 4 if for all 5, 6.
- Multi-set Independence: 7 (Asnani et al., 2021).
These fundamental axioms enable the use of SIMs in combinatorial optimization, privacy, summarization, and learning tasks that require formal guarantees.
3. Canonical Submodular Function Classes and Entropic Correspondence
The most widely used SIMs are grounded in the following classes (Iyer, 19 Jan 2026, Iyer et al., 2020):
| Function family | 8 definition | Typical use cases |
|---|---|---|
| Coverage/set-cover | 9 | Diversity, coverage |
| Facility-location | 0 | Representation, information overlap |
| Graph-cut-type | 1 | Redundancy, separation, clustering |
| Concave-over-mod | 2, 3 concave nondecreasing | Robustness, budgeted diversity |
| Log-determinant | 4 (5 kernel) | Volume, diversity, uncertainty |
Recent work demonstrates exact entropic constructions: given any of these 6, there exists a random vector 7 so that 8, and all submodular information measures reduce to their classical Shannon counterparts (Iyer, 19 Jan 2026).
4. Optimization Algorithms and Greedy Guarantees
Maximization of any nonnegative, monotone SIM (e.g., 9, 0, 1) under a cardinality or matroid constraint admits a 2-approximation via the greedy algorithm (Kothawade et al., 2022, Kothawade et al., 2021, Kothawade et al., 2021):
- Initialize 3.
- For 4 to 5:
- For each 6 not in 7, compute marginal gain: e.g., 8.
- Add 9 with maximal 0 to 1.
Lazy-greedy and partitioning reduce computational cost, particularly for SMI based on facility-location (requiring only 2 similarity evaluations) (Kothawade et al., 2022, Kothawade et al., 2021).
Curvature bounds ([curvature 3]) further tighten approximation ratios. In practice, facility-location, graph-cut, and log-determinant functions exhibit low curvature, making greedy nearly optimal (Kothawade et al., 2022).
5. Applications: Data Selection, Summarization, and Learning
SIMs constitute core objectives in broad machine learning settings:
- Active Learning and Data Discovery: SCG and SMI are used to mine rare or unknown classes by rewarding dissimilarity from labeled sets (SCG) and then intensifying discovery by targeting known hits (SMI/SCMI). Empirically, these approaches dominate baselines on rare-class and OOD selection in image classification and object detection, with 10–15% absolute gains in accuracy for unknowns (Kothawade et al., 2022, Kothawade et al., 2021, Kothawade et al., 2022).
- Targeted Subset Selection: SMI and variants (facility-location, log-det, graph-cut, COM) select samples that optimally trade off query relevance and target coverage. Theoretical bounds guarantee that maximizing SMI under realistic similarity-separation assumptions ensures high query relevance and coverage (Beck et al., 2024, Kothawade et al., 2021).
- Privacy and Fairness: SCMI and its constraints operationalize privacy by enforcing independence from a sensitive set under a user-defined threshold (Asnani et al., 2021, Kaushal et al., 2020). Privacy filters and marginal-independence filters compose efficiently with submodular maximization objectives.
- Summarization and Representation Learning: SIMs unify generic, query-focused, privacy- and update-aware summarization as direct maximizations of SMI, SCG, or CSMI, generalizing models such as ROUGE, DPPs, and graph-cut methods (Kaushal et al., 2020, Kothawade et al., 2021). In representation learning, submodular total correlation losses (e.g., SCoRe framework) simultaneously minimize intra-class variance and inter-class bias, outperforming standard contrastive methods for imbalanced data (Majee et al., 2023).
A selection of empirical results:
| Application | SIM Instantiation | Typical Gain over Baselines | Reference |
|---|---|---|---|
| Active Data Discovery | Fl_cg+mi, Logdet_cg+mi | 10–15% higher accuracy on unknowns | (Kothawade et al., 2022) |
| OOD Avoidance | Fl-CMI, LogDet-CMI | 4–7% accuracy lift | (Kothawade et al., 2022) |
| Targeted TSS | LogdetMI, FL2MI | ~20–30% absolute improvement on rare classes | (Kothawade et al., 2021) |
| Summarization | FL-SMI, GraphCut-SMI, LogDet-SMI | Near human-level V-ROUGE | (Kaushal et al., 2020) |
| Representation Learning | FL-/GC -C_f | 1–9% boost in class-imblanced recognition | (Majee et al., 2023) |
6. Extensions: Causal Inference, Information Inequalities, and Advanced Properties
SIMs extend classical independence, conditional independence, and causal Markov properties to non-entropic settings, unifying information-theoretic and combinatorial perspectives. The generalized causal Markov condition for SIMs matches the standard DAG-based independence structure, independent of the choice of submodular 4 (Steudel et al., 2010).
Unified derivations of information inequalities (Han’s, Shearer’s, monotonicity sequences, total correlation bounds) follow broadly from submodularity. These yield refined combinatorial bounds, e.g., on projection sizes, Boolean influences, and extremal graph properties (Sason, 2022).
Recent developments include the study of SIMs for weak submodularity in quadratic estimation and optimal experimental design (alphabetic optimality criteria), where closed-form utility functions (log-det, trace, min-eigenvalue) are directly submodular or enjoy quantifiable approximation via greedy (Hashemi et al., 2019).
7. Modeling Flexibility, Parameterizations, and Practical Considerations
Modern extensions, such as PRISM (Kothawade et al., 2021), introduce multi-parameterized SIMs to interpolate between relevance, diversity, privacy, and coverage. Typical parameters:
- 5 (graph-cut): relevance vs. diversity
- 6 (facility-location, COM): similarity-to-query trade-off
- 7 (conditional gain): strength of avoidance/penalty to a private set
By tuning these, SIMs adapt to a wide regime of problems: rare-class mining, guided summarization, OOD filtering, distributed and scalable optimization.
Several concrete choices are supported with efficient greedy algorithms (Kothawade et al., 2021, Kaushal et al., 2020, Beck et al., 2024), and the entire framework is modality-agnostic—applicable to images, video, text, sensor sets, and gradient embeddings.
Summary Table of Core SIM Formulae
| Name | Formula | Typical Use |
|---|---|---|
| SCG | 8 | Dissimilarity, novelty |
| SMI | 9 | Relevance, coverage, overlap |
| SCMI | 0 | Targeting under exclusion |
Through their algebraic generality and foundational approximation guarantees, submodular information measures constitute a principled, tractable, and highly expressive toolkit for information-centric decision-making in structured data systems (Kothawade et al., 2022, Asnani et al., 2021, Iyer et al., 2020, Beck et al., 2024, Kothawade et al., 2021).