Compute-Law Compositionality
- Compute-law compositionality is a framework that formalizes how meanings of complex expressions are systematically built from their parts using computable functions.
- It integrates classical semantic theories, probabilistic algebras, and neural architectures, employing metrics like the modulus of continuity and variance-decline score to quantify composition.
- The framework bridges linguistic, cognitive, and computational models, advancing our understanding of systematic generalization in both human cognition and artificial intelligence.
Compute-law compositionality concerns the existence, formalization, and empirical manifestation of explicit, quantitative principles—‘computable laws’—governing how the meanings of complex structures are determined by their parts and composition rules. This domain encompasses classical semantic theories, categorical models, symbolic and subsymbolic computational architectures, as well as neural and probabilistic systems. The study of compute-law compositionality aims to bridge linguistic, cognitive, and algorithmic models by making precise the mappings that underlie systematic generalization and emergent combinatorial capacity in both human cognition and artificial intelligence.
1. Formal Foundations of Compositional Law
Classical compositionality, in the Fregean tradition, asserts that for any complex expression constructed from sub-expressions , by combination rule , the meaning is entirely determined by the meanings of , , and the operation : where is a computable semantic-composition function. Category-theoretic generalizations formalize grammar as a category with objects (syntactic types) and morphisms (composition rules), with semantics as a functor into a semantic category , enforcing
thereby ensuring that syntactic and semantic compositions commute (Russin et al., 2024).
In formal computational modeling, compositionality is specified by inductively defined sets and functions. For natural language, let (lexicon) and (strings) be as follows:
- If , then
Constituent structures are defined by recursive Merge operations. At each level , a (partial) semantic function computes higher-level structure meaning, preserving the meanings of subcomponents (Kaushik et al., 2020).
In probabilistic process algebras, the compositionality of an operator is quantified via its modulus of continuity : where is a bisimulation pseudometric and is determined by the fixed-point on rule multiplicities (Gebler et al., 2014).
2. Core Set-theoretic and Functional Laws
Compute-law compositionality is captured by explicit set-theoretic and functional lemmas:
- Non-isomorphism of domains: The linear domain of strings () is not isomorphic to constituent-time patterns (), ensuring that meaning is not directly recoverable from the string alone. Each semantic function constructs higher-level meanings via partial, incremental application along structurally defined domains, culminating in a final interpretation : for with levels of merge (Kaushik et al., 2020).
- Weak Compositionality: Because each is partial, meaning is built in stages, ensuring that only complete structures (those with at the root) project uniquely to valid syntax trees.
- Probabilistic Modulus of Continuity: For algebraic operators, the impact on process distance is bounded above by a law computed from rule multiplicities via a fixed-point: where is the least solution to (with defined from SOS rules) (Gebler et al., 2014).
3. Cognitive and Neural Constraints
Cognitive-plausible models impose architectural requirements and computational constraints stemming from these laws:
- Constituents must be represented incrementally, indexed by string positions or time.
- Semantic functions operate in a strict bottom-up (and optionally top-down) manner, with operations applied in an online, left-to-right fashion.
- Partial constituency representations are maintained; only at the root is full structure recovered.
- Structure-dependent transformations are computed over constituent domains (), whereas rigid (structure-independent) operations can act only on strings ().
In the DORA model (a symbolic-connectionist role–filler binding architecture), these requirements are met via layered dynamical systems:
- Four layers (Propositions, Role-Binder, Predicate/Object, Semantics)
- Timing-based role–filler binding—temporal asynchrony encodes relations
- Hebbian mapping and competitive inhibition realize incremental structure mapping
- Empirical evaluations with word embeddings demonstrate that structural constraints prune mapping errors and optimize relational learning (Kaushik et al., 2020).
4. Compositionality in Deep Neural and Probabilistic Systems
Recent advances have produced explicit compute-law characterizations in deep learning and probabilistic scene parsing:
- Neural Models: Compositionality may be “built in” using architectural inductive biases (role–filler separation, tensor-product binding, relational bottlenecks) or “learned” through meta-learning regimes. Metalearning discovers compositional algorithms through exposure to multiple tasks, resulting in an inner-loop implementation that generalizes systematically (Russin et al., 2024).
- CLAP (Compositional Law Parsing): In visual reasoning, each concept is mapped by a latent random function , instantiated as a Neural Process. The compositional law of a scene bundles . Laws may be composed across scenes by recombining latent variables corresponding to each concept, enabling exchange and novel law formation. Compositionality is diagnosed by the variance-decline score, quantifying the match between learned latents and ground-truth sub-laws (Shi et al., 2022).
| Method | Law Formalization | Empirical Domain |
|---|---|---|
| Symbolic (Kaushik) | Partial semantics via | Natural language syntax/semantics |
| Probabilistic (Desharnais et al.) | Modulus of continuity via fixed points | Probabilistic processes |
| Deep learning (CLAP) | Latent random functions + variance-decline | Scene/physics parsing |
| DNNs, LLMs | Token-level composition + ICL | NLP, code, reasoning |
5. Quantitative Evaluation and Empirical Metrics
Compute-law compositionality enables quantitative assessment:
- Geometric compositionality score: For phrase in context, embedding aligns with context subspace ; compositionality is measured by
Values near 1 indicate literal (compositional) usage; scores near 0 identify idiomatic or metaphorical usage (Gong et al., 2016).
- Variance-decline score (CLAP): For each concept–law, the score compares variance in with and without fixing law : implies that encodes law (Shi et al., 2022).
- Precision in mapping structure: In DORA, predicted mapping matrices are compared to ground-truth to compute precision (TP/(TP+FP)); performance increases when embedding structure better reflects syntax (Kaushik et al., 2020).
6. Emergent Laws in Large-scale Neural Systems
Empirical work demonstrates robust, quantitative “computable laws” connecting architecture, learning strategy, and compositional generalization:
- Out-of-distribution generalization error decays exponentially in training examples under strong inductive biases: where is the compositional generalization rate for bias .
- Under metalearning, the inner-loop error on new composition tasks scales as in prompt-length and meta-training set size .
- Increasing model scale and corpus size in pre-trained DNNs leads to strengthening 'emergent' compositionality (Russin et al., 2024).
Instruction-tuned, large pre-trained models display compositional behaviors across NLP and medical reasoning, and code generation benchmarks. These findings position compute-law compositionality as the theoretical bedrock and practical instrument for both measuring and inducing systematic generalization in artificial and biological systems.
7. Worked Examples in Formal Compositionality
The application of these laws is exemplified in both symbolic and neural systems:
- Natural language syntax: For ,
- assigns categories and position indices
- constructs NP, VP constituents
- merges to full S
is thus a composition of meanings from atomic terms and three merges (Kaushik et al., 2020).
- Probabilistic operators: For non-deterministic choice , the modulus is
and for probabilistic prefix , (Gebler et al., 2014).
- CLAP law composition: By swapping global latent variables between sample scenes A and B for concept sets , new composed laws are instantiated, and decoded to generate scenes with combined sub-law behaviors (Shi et al., 2022).
These formal examples and empirical systems collectively instantiate compute-law compositionality as a rigorous, quantitative, and implementable principle across symbolic, probabilistic, and connectionist models.