Composite Function Vectors
- Composite function vectors are functions formed by composing smooth mapping with scalar or vector-valued functions, characterized by structured rules for derivatives.
- Their analysis employs multivariate Bell polynomials and FaĆ di Bruno formulas to precisely enumerate derivative contributions and combinatorial structures.
- Applications span derivative-free and stochastic optimization, finite-difference schemes, and composition operators in analytic function spaces.
A composite function vector refers to the broad class of functions formed by composing a (possibly vector-valued) smooth mapping with another (scalar- or vector-valued) function, as well as the associated mathematical structures and rules for their derivatives, finite differences, and analytical properties. The study of composite function vectors is central to areas such as multivariate calculus, derivative-free optimization, stochastic optimization, and the theory of analytic function spaces, with special attention to the structure of their derivativesāencapsulated by multivariate generalizations of the FaĆ di Bruno and Bell polynomial formulasāas well as their discrete and functional-analytic generalizations.
1. Multivariate Bell Polynomials and the FaĆ di Bruno Formula
The most precise symbolic description of repeated derivatives of a composite function vector , where and , is given by the multivariate FaĆ di Bruno formula, expressed in terms of multivariate Bell polynomials(Schumann, 2019). For multi-indices and , the partial multivariate Bell polynomial packages all combinatorial contributions of higher-order partials of , such that
where are multi-indexed derivatives of , and those of . This generalization recovers the classical (single-variable) FaĆ di Bruno and Bell polynomial structure when , but extends to all orders and mixed partials in higher dimensions. Combinatorially, Bell polynomials enumerate all ways to partition derivatives among the argumentsāa necessity given the tensor character of and .
2. Discrete Composite Function Vector: Finite-Difference Structure
The discrete analog of composite function vectors is governed by a finite-difference version of the FaĆ di Bruno formula(Duarte et al., 2008). For and , and for multi-step directions , the repeated difference operator applied to is given by
where the inner sum is over all partitions of the binary multi-index into nonzero multi-indices , and is the -fold directional difference. This formula is entirely algebraic and applies to any abelian group, with no smoothness assumptions, providing a direct combinatorial correspondence to the continuous case.
3. Analytical Properties and Fully Linear Model Approximations
Composite function vectors arise as core objects in derivative-free optimization (DFO) when the objective function is a composition with convex lsc and smooth(Hare, 2016). When each component can be approximated by fully linear models satisfying
the error in the composite model is
with a local Lipschitz constant. Subdifferential proximity at a focal point can be guaranteed with
where is the Lipschitz modulus of at . These error bounds justify the use of composite models in derivative-free trust region algorithms even when is nonsmooth but retains composite structure.
4. Stochastic Composite Function Vector Optimization
The composite function vector framework is central in modern optimization, required for problems where the objective takes the form , with or , smooth, and a regularizer(Zhang et al., 2019). The chain rule for the exact gradient reads
but naive stochastic estimators for are biased due to the nonlinear action of on the inner estimate . The Composite Incremental Variance-Reduced (CIVR) algorithm introduces independent SARAH/SPIDER-style variance reduction for and , yielding total sample complexity for expectation-form and for finite-sum form, matching the best known first-order methods for such nonconvex composite structures. The approach retains practical efficiency for a broad range of machine learning and reinforcement learning problems.
5. Composition Operators on Vector-Valued Analytic Function Spaces
In functional analysis, composite function vectors appear as composition operators acting on Banach spaces of -valued analytic functions via (Laitila et al., 2015). The boundedness, compactness, and weak compactness of depend on both the operator's action on scalar-valued spaces and the structure of . Key facts include:
- is bounded iff is bounded.
- For infinite dimensional, is never compact on strong spaces , but can be weakly compact under conditions derived from the scalar case (e.g., Shapiro's condition for ).
- For reflexive , weak compactness of on is equivalent to compactness on for large classes of spaces, such as , weighted Bergman, Bloch, and BMOA(X).
This theory connects composition of vector-valued functions to operator theory and function space geometry, exposing deep links between vectorial analytic structure and function-theoretic operator properties.
6. Concrete Examples and Special Cases
Explicit calculation in low dimension clarifies the structure of the composite function vector theory. For , , , the mixed partial contains contributions from both second-order partials of (via ) and products of first partials (via ), with the multivariate Bell polynomial structure systematically packaging all terms(Schumann, 2019). For finite differences, the combinatorics are determined by partitions of the multi-index, with the formula reducing to familiar forms for , (Duarte et al., 2008).
7. Extensions, Assumptions, and Open Problems
The theory is sensitive to the assumptions on and . For convex composite optimization, interior point assumptions and model interpolation are critical(Hare, 2016). In analytic function spaces, key open questions concern the extension of Hilbert-Schmidt boundedness and weakly compact composition to non-classical spaces and operator-weighted settings(Laitila et al., 2015). In stochastic optimization, extensions to multilevel compositional problems are tractable by similar variance-reduction techniques, facilitating application to hierarchical models in machine learning(Zhang et al., 2019).
References
- "Multivariate Bell Polynomials and Derivatives of Composed Functions" (Schumann, 2019)
- "A discrete Faa di Bruno's formula" (Duarte et al., 2008)
- "Compositions of Convex Functions and Fully Linear Models" (Hare, 2016)
- "Composition operators on vector-valued analytic function spaces: a survey" (Laitila et al., 2015)
- "A Stochastic Composite Gradient Method with Incremental Variance Reduction" (Zhang et al., 2019)