Indirect Cost Functions Overview
- Indirect cost functions are defined as the minimal cumulative cost over sequential experiments that replicate a target distribution of posteriors in information acquisition.
- They satisfy key axioms like monotonicity and sub-additivity, ensuring that less informative or sequentially composed experiments incur lower costs.
- Posterior-separable forms connect these costs to statistical divergences and entropy-based penalties, with significant implications for dynamic learning and cost-sharing frameworks.
An indirect cost function quantifies the minimum expected cost of generating an information structure when the design is unconstrained except by the cost assigned to individual direct experiments and the possibility to sequentially combine such experiments. This concept is central to the economic theory of information acquisition, where information is modeled as a stochastic reduction in uncertainty over a finite state space, and the costs of different acquisition strategies are analyzed according to their efficiency and compositional properties.
1. Formal Model and Definition
The state space is finite. An information structure or "experiment" is formalized as a distribution over posterior beliefs , with mean (expected value) equal to the prior . The cost of producing such a structure in one step is given by a direct cost function . This function is required to be bounded and (piecewise) continuous, convex under mixture (i.e., mixing experiments does not incur extra cost), and to satisfy for all deterministic (degenerate) posteriors .
The indirect cost of generating a distribution is then defined as the minimal expected cumulative direct cost over all finite-horizon sequential designs replicating in law, allowing for "information-disposal" (steps that forget information at no cost). Formally,
where the infimum is over all Markov chains on beliefs with the required boundary conditions and martingale property. This definition enables the modeling of the cheapest way to implement given a direct cost structure and is robust to sequential composition and probabilistic mixing of experiments (Zhong, 2018).
2. Fundamental Axioms: Monotonicity and Sub-additivity
All indirect cost functions derived as above satisfy two key properties:
- Monotonicity: If is a Blackwell-garbling of (i.e., — is less informative than ), then
This reflects the principle that less informative structures can be implemented at lower or equal cost.
- Sub-additivity: For any two-stage experiment where is obtained by drawing from then drawing from (that is, sequentially implementing and then, conditionally, ),
This encodes the cost advantage of breaking an acquisition task into sequential sub-experiments.
These axioms are both necessary and sufficient: a function is an indirect cost for some valid if and only if it satisfies monotonicity and sub-additivity. Furthermore, if already satisfies these it is the indirect cost function induced by itself, i.e., it is "optimized" and further indirectification leaves it unchanged (Zhong, 2018).
3. Posterior-separable and Uniformly Posterior-separable Indirect Costs
A cost function is posterior-separable if there exists a divergence such that
It is uniformly posterior-separable if there is a convex potential with
Uniform posterior-separability is equivalent to monotonicity plus the stricter additivity property: for all two-stage joins. This class captures costs whereby the expected acquisition cost only depends on the dispersion of posteriors around the prior, evaluated through a convex potential—directly linking information costs to statistical divergences and entropy-based penalties.
If a direct cost "favors incremental evidence"—locally approximating, up to , a quadratic form with positive semi-definite , and admitting a convex such that and —then the induced indirect cost is uniformly posterior-separable. Conversely, uniform posterior-separability of the indirect cost requires that favors incremental evidence in this sense.
4. Prior-independence and Indirect Cost Separability
A direct cost function is prior-independent if it depends only on the conditional distributions (likelihood matrices) and not on the prior over states. For (the binary-state case), any prior-independent direct cost can be written in terms of the joint signal-state law and its marginals. The induced indirect cost is uniformly posterior-separable (i.e., admits a potential representation) if and only if there exists such that
with denoting mutual information, the Kullback–Leibler divergence, and the signal law conditional on . In this regime, the indirect cost is
For , there is no nontrivial prior-independent direct cost which generates a posterior-separable indirect cost. This delineates a sharp boundary between the binary and multi-state settings in terms of the admissible cost functionals.
5. Representative Cost Function Examples and Operational Insights
Common choices for direct cost functions and their induced indirect forms include:
| Cost Type | definition | Indirect Cost Property |
|---|---|---|
| Mutual information | Additive, posterior-separable | |
| Quadratic variance | Additive, quadratic potential |
- For mutual information, the corresponding potential is (negative Shannon entropy), and both direct and indirect costs are equal.
- For quadratic variance, when and again the cost is additive, i.e., direct and indirect costs coincide.
A corollary is the dilution property: for a mixture , , which means "diluting" an experiment in time or mixtures scales its cost linearly.
In continuous time limits, the most efficient implementation of over leverages a compound Poisson signal, jumping to at rate .
An application is in dynamic learning: when a decision-maker faces a per-period flow cost based on and selects an optimal stopping rule, sub-additivity ensures that optimal information delivery can be characterized by a static allocation and a Poissonian event structure, with equal-cost smoothing across periods.
6. Indirect Cost Functions in Statistical Cost Sharing
In cost-sharing problems where the cost function is not directly accessible but instead estimated from empirical data, “statistical cost sharing” approaches assess solution concepts (such as the core and the Shapley value) based on observed tuples and their distribution (Balkanski et al., 2017). Here, the indirect cost corresponds to the statistical framework that estimates cooperative cost functionals using data, sample complexity analysis, and probabilistic guarantees (e.g., uniform convergence based on VC-dimension).
For monotone submodular cost functions with curvature , the statistical Shapley value can be approximated within a factor of , and this bound is information-theoretically tight. Statistical analogues of the Shapley axioms (balance, symmetry, zero-element, additivity) uniquely specify the data-driven Shapley value in this context.
7. Implications, Boundaries, and Research Directions
The paper of indirect cost functions clarifies which functional forms over distributions of posteriors are compatible with compositional, implementable, and economically meaningful cost measures. The equivalence of indirect cost functions with monotonicity and sub-additivity axioms provides a rigorous operational characterization, while the theory of posterior-separability establishes a bridge to potential-based representations widely used in information theory and learning theory.
The sharp distinction between the binary and multi-state cases for prior-independent, posterior-separable indirect costs restricts the general applicability of certain statistical measures as cost surrogates. The explicit representation of indirect cost functions in dynamic and statistical learning continues to motivate both theoretical and applied research in information acquisition, mechanism design, and data-driven cooperative game theory.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free