Papers
Topics
Authors
Recent
Search
2000 character limit reached

Compute-Law Compositionality

Updated 22 December 2025
  • Compute-law compositionality is a framework that formalizes how meanings of complex expressions are systematically built from their parts using computable functions.
  • It integrates classical semantic theories, probabilistic algebras, and neural architectures, employing metrics like the modulus of continuity and variance-decline score to quantify composition.
  • The framework bridges linguistic, cognitive, and computational models, advancing our understanding of systematic generalization in both human cognition and artificial intelligence.

Compute-law compositionality concerns the existence, formalization, and empirical manifestation of explicit, quantitative principles—‘computable laws’—governing how the meanings of complex structures are determined by their parts and composition rules. This domain encompasses classical semantic theories, categorical models, symbolic and subsymbolic computational architectures, as well as neural and probabilistic systems. The study of compute-law compositionality aims to bridge linguistic, cognitive, and algorithmic models by making precise the mappings that underlie systematic generalization and emergent combinatorial capacity in both human cognition and artificial intelligence.

1. Formal Foundations of Compositional Law

Classical compositionality, in the Fregean tradition, asserts that for any complex expression EE constructed from sub-expressions AA, BB by combination rule CC, the meaning E\llbracket E \rrbracket is entirely determined by the meanings of AA, BB, and the operation CC: E=F(A,B,C)\llbracket E \rrbracket = F\left(\llbracket A \rrbracket,\, \llbracket B \rrbracket,\, C\right) where FF is a computable semantic-composition function. Category-theoretic generalizations formalize grammar as a category C\mathcal{C} with objects (syntactic types) and morphisms (composition rules), with semantics as a functor M:CS\mathbb{M}:\mathcal{C}\to \mathcal{S} into a semantic category S\mathcal{S}, enforcing

M(fg)=M(f)M(g)\mathbb{M}(f \circ g) = \mathbb{M}(f) \circ \mathbb{M}(g)

thereby ensuring that syntactic and semantic compositions commute (Russin et al., 2024).

In formal computational modeling, compositionality is specified by inductively defined sets and functions. For natural language, let L0L_0 (lexicon) and SS (strings) be as follows:

  • SL0S \supseteq L_0
  • If w,vSw, v \in S, then (w+v)S(w + v) \in S

Constituent structures LiL_i are defined by recursive Merge operations. At each level ii, a (partial) semantic function fif_i computes higher-level structure meaning, preserving the meanings of subcomponents (Kaushik et al., 2020).

In probabilistic process algebras, the compositionality of an operator is quantified via its modulus of continuity zz: d0(f(t1,,tn),f(t1,,tn))z(d0(t1,t1),,d0(tn,tn))d_0(f(t_1,\ldots,t_n), f(t'_1,\ldots,t'_n)) \leq z(d_0(t_1, t'_1), \ldots, d_0(t_n, t'_n)) where d0d_0 is a bisimulation pseudometric and zz is determined by the fixed-point on rule multiplicities (Gebler et al., 2014).

2. Core Set-theoretic and Functional Laws

Compute-law compositionality is captured by explicit set-theoretic and functional lemmas:

  • Non-isomorphism of domains: The linear domain of strings (SS) is not isomorphic to constituent-time patterns (TiT_i^*), ensuring that meaning is not directly recoverable from the string alone. Each semantic function fif_i constructs higher-level meanings via partial, incremental application along structurally defined domains, culminating in a final interpretation hh: m(e)=hnfn(e,fn1(e,,f1(e)))m(e) = h_n \circ f_n(e, f_{n-1}(e, \ldots, f_1(e))) for ee with nn levels of merge (Kaushik et al., 2020).
  • Weak Compositionality: Because each fif_i is partial, meaning is built in stages, ensuring that only complete structures (those with θ=1\theta=1 at the root) project uniquely to valid syntax trees.
  • Probabilistic Modulus of Continuity: For algebraic operators, the impact on process distance is bounded above by a law zxz_x computed from rule multiplicities via a fixed-point: zx(ε1,,εn)=min{iMf(i)εi,1}z_x(\varepsilon_1,\ldots,\varepsilon_n) = \min\left\{ \sum_i M_f(i)\varepsilon_i, 1 \right\} where MfM_f is the least solution to M=suprRfHr(M)M = \sup_{r \in R_f} H_r(M) (with HrH_r defined from SOS rules) (Gebler et al., 2014).

3. Cognitive and Neural Constraints

Cognitive-plausible models impose architectural requirements and computational constraints stemming from these laws:

  • Constituents must be represented incrementally, indexed by string positions or time.
  • Semantic functions fif_i operate in a strict bottom-up (and optionally top-down) manner, with operations applied in an online, left-to-right fashion.
  • Partial constituency representations are maintained; only at the root is full structure recovered.
  • Structure-dependent transformations are computed over constituent domains (TiT_i^*), whereas rigid (structure-independent) operations can act only on strings (SS).

In the DORA model (a symbolic-connectionist role–filler binding architecture), these requirements are met via layered dynamical systems:

  • Four layers (Propositions, Role-Binder, Predicate/Object, Semantics)
  • Timing-based role–filler binding—temporal asynchrony encodes relations
  • Hebbian mapping and competitive inhibition realize incremental structure mapping
  • Empirical evaluations with word embeddings demonstrate that structural constraints prune mapping errors and optimize relational learning (Kaushik et al., 2020).

4. Compositionality in Deep Neural and Probabilistic Systems

Recent advances have produced explicit compute-law characterizations in deep learning and probabilistic scene parsing:

  • Neural Models: Compositionality may be “built in” using architectural inductive biases (role–filler separation, tensor-product binding, relational bottlenecks) or “learned” through meta-learning regimes. Metalearning discovers compositional algorithms through exposure to multiple tasks, resulting in an inner-loop implementation that generalizes systematically (Russin et al., 2024).
  • CLAP (Compositional Law Parsing): In visual reasoning, each concept cc is mapped by a latent random function fc:XRdzf^c:{\mathcal X}\rightarrow\mathbb{R}^{d_z}, instantiated as a Neural Process. The compositional law of a scene bundles f1,,fKf^1,\ldots,f^K. Laws may be composed across scenes by recombining latent variables gcg^c corresponding to each concept, enabling exchange and novel law formation. Compositionality is diagnosed by the variance-decline score, quantifying the match between learned latents and ground-truth sub-laws (Shi et al., 2022).
Method Law Formalization Empirical Domain
Symbolic (Kaushik) Partial semantics via fi,sf_i,s Natural language syntax/semantics
Probabilistic (Desharnais et al.) Modulus of continuity zz via fixed points Probabilistic processes
Deep learning (CLAP) Latent random functions + variance-decline Scene/physics parsing
DNNs, LLMs Token-level composition + ICL NLP, code, reasoning

5. Quantitative Evaluation and Empirical Metrics

Compute-law compositionality enables quantitative assessment:

  • Geometric compositionality score: For phrase pp in context, embedding vpv_p aligns with context subspace ScS_c; compositionality is measured by

C(p,context)=cos(vp,ProjSc(vp))C(p,\,\text{context}) = \cos(v_p,\,\mathrm{Proj}_{S_c}(v_p))

Values near 1 indicate literal (compositional) usage; scores near 0 identify idiomatic or metaphorical usage (Gong et al., 2016).

  • Variance-decline score (CLAP): For each concept–law, the score sl,cs_{l,c} compares variance in gcg^c with and without fixing law ll: sl,c=1NBb=1NBvˉcvlc,bvˉcs_{l,c} = \frac{1}{N_B}\sum_{b=1}^{N_B} \frac{\bar v^c - v^{c, b}_l}{\bar v^c} sl,c1s_{l,c} \approx 1 implies that gcg^c encodes law ll (Shi et al., 2022).
  • Precision in mapping structure: In DORA, predicted mapping matrices are compared to ground-truth to compute precision (TP/(TP+FP)); performance increases when embedding structure better reflects syntax (Kaushik et al., 2020).

6. Emergent Laws in Large-scale Neural Systems

Empirical work demonstrates robust, quantitative “computable laws” connecting architecture, learning strategy, and compositional generalization:

  • Out-of-distribution generalization error decays exponentially in training examples under strong inductive biases: Erro.o.d.(n)c(B)exp(λ(B)n)\mathrm{Err}_{o.o.d.}(n) \leq c(B) \exp(-\lambda(B)n) where λ(B)\lambda(B) is the compositional generalization rate for bias BB.
  • Under metalearning, the inner-loop error on new composition tasks scales as O(Kα+Mβ)O(K^{-\alpha} + M^{-\beta}) in prompt-length KK and meta-training set size MM.
  • Increasing model scale and corpus size in pre-trained DNNs leads to strengthening 'emergent' compositionality (Russin et al., 2024).

Instruction-tuned, large pre-trained models display compositional behaviors across NLP and medical reasoning, and code generation benchmarks. These findings position compute-law compositionality as the theoretical bedrock and practical instrument for both measuring and inducing systematic generalization in artificial and biological systems.

7. Worked Examples in Formal Compositionality

The application of these laws is exemplified in both symbolic and neural systems:

  • Natural language syntax: For s=“Big dogs bite men”s=\text{“Big dogs bite men”},

    1. f1(s)=f_1(s) = assigns categories and position indices
    2. f2(s,f1(s))=f_2(s,f_1(s)) = constructs NP, VP constituents
    3. f3(s,...)=f_3(s,...) = merges to full S

    m(s)=h3(f3(s,f2(s),f1(s)))m(s) = h_3( f_3( s, f_2(s), f_1(s) ) )

    m(s)m(s) is thus a composition of meanings from atomic terms and three merges (Kaushik et al., 2020).

  • Probabilistic operators: For non-deterministic choice g(x,y)=x+yg(x, y) = x + y, the modulus zgz_g is

    zg(ε1,ε2)=max{ε1,ε2}z_g(\varepsilon_1, \varepsilon_2) = \max\{\varepsilon_1, \varepsilon_2\}

    and for probabilistic prefix f(x)=a.([q]x[1q]0)f(x)=a.([q]x \oplus [1-q]0), zf(ε)=qεz_f(\varepsilon)=q\varepsilon (Gebler et al., 2014).

  • CLAP law composition: By swapping global latent variables gcg^c between sample scenes A and B for concept sets CA,CBC_A, C_B, new composed laws are instantiated, and decoded to generate scenes with combined sub-law behaviors (Shi et al., 2022).

These formal examples and empirical systems collectively instantiate compute-law compositionality as a rigorous, quantitative, and implementable principle across symbolic, probabilistic, and connectionist models.

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Compute-Law Compositionality.