Coherence as a Fundamental Measure of Order

This report develops the Coherence Framework by showing that (1) reality exhibits emergent order, (2) coherence is measurable (via Integrated Information Theory, IIT), and (3) simplicity (low algorithmic complexity) is preferred (Occam’s razor). We proceed via a Socratic sequence, answering (i) Why do laws of physics exist at all?, (ii) Can meaning exist without pattern?, and (iii) Why does Occam’s Razor work? We draw on physics, complexity science, information theory and neuroscience to support each claim. Mathematical definitions (Φ, Kolmogorov complexity K, variational free energy F) are given where relevant.

Why do laws of physics exist at all?

  • Necessity of patterns for observers. A completely “lawless” or random universe could not give rise to stable structures or observers. Only in a world governed by reliable patterns can coherent entities evolve and think. Indeed, fine-tuning studies note that most changes to physical laws or constants (gravity, quantum forces, etc.) would preclude galaxies, stars and chemistry. For example, if gravity were much stronger or weaker, stars (and the heavy elements they produce) could not form. Chaisson and others emphasize that energy flows “make order out of disorder”: complex structures (stars, life) can arise because they export entropy to their surroundings. In short, anthropic reasoning suggests we see coherent laws because only such a universe can host observers.

  • Laws from invariance principles. Some argue the form of physical laws follows from symmetry or consistency requirements. Stenger, for instance, claimed the laws must be “point-of-view invariant” so that physics looks the same to any observer (a basic symmetry). While technical debate continues, these ideas agree that laws are not arbitrary: they arise from the need for consistent, stable description of reality. In modern cosmology, for example, the ΛCDM model (“standard model of cosmology”) itself embodies precise laws (Einstein’s gravity, quantum field theory) chosen by fit to observation. That our universe obeys lawful, discoverable equations (as opposed to fully random happenings) is taken as a starting point for physics.

  • Coherence requirement. Without patterns, even the concept of meaning or information fails. Integrated Information Theory (IIT) posits that consciousness (a form of meaning) depends on discriminating among alternatives. Tononi’s famous example shows that a conscious brain “rules out” a vast repertoire of possible inputs (blue elephants, etc.) compared to a simple sensor. In other words, meaning arises only when a system identifies specific patterns among many possibilities. This suggests laws and patterns are necessary for any meaningful structure to exist.

Together, these points imply that lawful order is not accidental but essential for existence. Only a universe with coherent laws can support complexity, life and thought. Physics laws may thus be seen as emergent constraints required for a stable, low-entropy world where observers can form.

Reality tends toward order, not just chaos

  • Nonequilibrium self-organization. Classical physics (2nd Law) tells us that total entropy increases. But open systems far from equilibrium can locally create order by exporting entropy. Nobel laureate Ilya Prigogine showed that non-equilibrium systems spontaneously form dissipative structures – coherent macroscopic order – once a control parameter passes a threshold. As Prigogine wrote, “non-equilibrium may be a source of order”. For example, heated fluids form convection rolls, and chemical reactions can cycle rather than equilibrate, all producing new structure. Thus local order increases even as total entropy rises.

  • Energy flows create complexity. Broad surveys of cosmic history find a trend of increasing complexity (galaxies, stars, life, technology) over time. Astrophysicist Eric Chaisson notes that such complexity growth is fully compatible with the 2nd Law, because systems like stars export enough disorder to their surroundings to more than compensate for the internal increase in order. He defines life as “an open, coherent, space-time structure maintained far from thermodynamic equilibrium by a flow of energy through it”. In effect, energy gradients drive organization: e.g. an air conditioner (or life-form) uses energy to create pockets of lower entropy while increasing entropy overall. Real-world data (e.g. rising “energy rate density” in biological and technological systems) also support this upward complexity trend.

  • Biological self-maintenance. Living organisms exemplify “order from chaos.” They continuously import energy (food, sunlight) and export heat to maintain internal structure. Friston and others point out that life as an autopoietic system is highly integrated (internally causal) and resists equilibrium. A living cell or brain relentlessly corrects perturbations to stay in a narrow set of states (homeostasis) – in Friston’s terms, it pursues “paths of least surprise”. In doing so it maintains a stable order far above thermodynamic equilibrium. Without such coherence (autopoiesis), complex life could not exist.

These lines of evidence all point to the same conclusion: complex order arises naturally in a universe of interacting parts and energy flows. Coherence (correlated structure) is not a fluke; it emerges whenever conditions allow it. As one review puts it, Kolmogorov–Chaitin complexity is dominated by randomness, not structure – meaning that truly random (high-entropy) states carry little “meaningful” order. By contrast, structures with low algorithmic complexity (patterned, compressible) can and do emerge. In sum, reality produces pockets of order (dissipative structures, cosmic complexity, life) against the tide of entropy, so order is a robust tendency under the right conditions.

Can meaning exist without pattern?

  • Meaning requires discrimination. In information-theoretic terms, meaning arises when a system makes specific distinctions. IIT encapsulates this: Tononi notes that what differentiates conscious experience is how many alternatives are ruled out by a perception. For a human seeing a blank screen, one implicitly rejects countless impossible sensations, generating “many more bits of information” than a lone photodiode would. He emphasizes that “the more specifically one’s mechanisms discriminate between what pure light is and what it is not (the more they specify what light means), the more one is conscious of it”. This links meaning (“what it is”) directly to pattern recognition (“what it is not”). A random stimulus that offers no structure cannot generate such discrimination and thus carries no content or consciousness.

  • Patterns underlie semantics. In cognitive science and linguistics, semantics emerges only when patterns exist in data. A string of truly random symbols has zero compressibility and no interpretable meaning. In contrast, any meaningful message follows rules (grammar, logic, correlations) that make it compressible or predictable. This is consistent with Kolmogorov’s insight: short descriptions capture structure, long incompressible descriptions are effectively random and uninformative. Thus, without underlying patterns, meaning collapses. In practice, pattern recognition algorithms and neural networks only work because data have structure. Conversely, if the world were patternless, neither science nor consciousness could operate.

In effect, “meaning” is synonymous with detecting coherence in sensory inputs. We see the flash of lightning and recognize the thunder (a pattern with cause–effect) – that recognition gives meaning. A random flicker with no relation is meaningless. As Tononi’s framework suggests, meaningful consciousness depends on integrated discrimination. Thus one cannot meaningfully exist in a world of pure chaos: pattern (order) is a prerequisite for meaning.

Coherence can be measured (Integrated Information Φ)

  • IIT’s definition of Φ. Integrated Information Theory (IIT) formalizes coherence as “integrated information” (Φ). Consider a physical system (neurons, logic gates, etc.) in a given state. IIT asks: How irreducible is this system? One computes, for each subset of elements, how much cause–effect information it contributes that is lost under partition. This irreducibility is called φ (phi). Mathematically, φ is defined over the minimum-information partition (the cut that least disrupts the system). The largest φ (denoted φ*) identifies the complex (maximally integrated subset). The full “cause–effect structure” of this complex is called a Φ-structure. Importantly, the total integrated information Φ of the system is defined as the sum of all φ-values across its constituent distinctions and relations. Formally:

    Φ = Σᵢ φᵢ,

    where each φᵢ is the integrated information of a particular “distinction” or relation in the system’s causal structure. Tononi et al. (IIT 4.0) state: “The sum total of the φ values of the distinctions and relations that compose the Φ-structure measures its structure integrated information Φ”. In other words, Φ captures the amount of information that is both highly specific and irreducibly joint in the system. It is by design a scalar quantity: higher Φ means more coherence.

  • IIT axioms and math. Briefly, IIT 4.0 builds on axioms of existence, composition, integration, exclusion, and information (experience is definite, selective, irreducible, etc.) to derive a mathematical machinery. One constructs the system’s transition probability matrix and computes for each candidate subset its cause φ_c (how surprising its past is) and effect φ_e (how surprising its future is). The system φ_s is the minimum of φ_c and φ_e across the minimum partition. The maximal φ_s yields the main complex. Then one enumerates all subsets (mechanisms) of that complex to get a “cause–effect structure”. Each subset’s irreducibility (φ_d or φ_r) is measured, and those φ’s sum to Φ. (For full details see Tononi et al. 2023.) Importantly, Φ is an intrinsic measure: it depends only on the system’s internal causal topology and state, not on external observers.

  • Criticisms and responses. IIT’s claim that Φ quantifies consciousness/coherence has been controversial. Scott Aaronson and colleagues pointed out counterexamples where very simple networks yield arbitrarily large Φ. For instance, an N×N grid of XOR gates (with no feedback) can produce √N Φ, so by scaling N one can get Φ larger than a human brain’s. Intuitively, this suggests a trivial circuit would be “more conscious” than a brain – an absurd conclusion if one equates Φ with true consciousness. Aaronson argued that such examples expose an inconsistency in IIT.

    Tononi’s team responded that such judgments are based on intuition, whereas IIT is meant to define consciousness by its postulates. In the dialogue “A Phi-nal Exchange”, Tononi conceded that simple feedforward systems can yield high Φ, and he emphasized that common-sense notions of consciousness might not apply. He even noted that IIT implies a spectrum of “experience” (protoconsciousness) down to photodiodes (1 bit of Φ). In short, he argued that IIT may transcend folk intuitions: it predicts that any highly integrated physical system has nonzero Φ, whether or not humans would call it conscious. This panexperiential stance is indeed radical, but it is a logical outcome of the theory’s axioms.

    In summary, IIT 4.0 provides a precise mathematical measure Φ of coherence/order in a system’s causal structure. While debates continue (Aaronson’s critiques, the “exclusion” axiom, computational tractability, etc.), the framework does allow a quantitative assessment of integrated order. As Friston and Gomez (2021) note, self-organizing living systems are “characterized by a very particular systemic organization” that resists equilibrium by maintaining order. IIT formalizes this idea: more integrated (higher Φ) systems have more potential “order” from their internal dynamics.

Simpler explanations are preferred (Kolmogorov complexity & Occam)

  • Kolmogorov complexity. The foundational measure of descriptional simplicity is algorithmic complexity. Informally, the Kolmogorov complexity K(x) of a string x is the length of the shortest program (in a fixed universal language) that outputs x. Equivalently, a repetitive, structured string has low K, while a completely random string (with no shorter description than itself) has K roughly its length. Solomonoff and Kolmogorov showed these ideas in the 1960s: “the Algorithmic Complexity of a string” was defined as “the length of the shortest code needed to describe it”. Thus, K(x) captures the inherent simplicity of data. (Formally one writes $K(x)=\min{|p|: U(p)=x}$ for a universal Turing machine U.) A related concept is algorithmic probability – shorter programs are exponentially more likely a priori.

  • Occam’s razor formalized. Occam’s Razor — the preference for simpler theories — emerges naturally from algorithmic probability and Solomonoff induction. Solomonoff (1960s) proved that the best predictive model of data is the shortest program that generates that data. In his words: “the best possible scientific model is the shortest algorithm that generates the empirical data under consideration”. Moreover, using Bayes’ theorem with a universal prior (probability ∝ 2^{–K(theory)}) shows that theories requiring fewer bits of description have higher posterior weight. In practice this means we assign larger prior credence to simpler hypotheses, automatically encoding Occam’s principle.

    Put differently, a complex theory with many ad hoc parameters is less likely under a universal prior than a concise theory. This reasoning explains why simpler models often generalize better: they capture true underlying structure rather than noise. In machine learning, this idea appears as Minimum Description Length or Bayesian model selection. Indeed, Solomonoff’s theory proved that Occam’s Razor follows from basic assumptions about computability and induction.

  • Real-world implications. Kolmogorov complexity is uncomputable in general, but it guides intuition and practice. For instance, data compression algorithms attempt to find short representations, indirectly measuring structure. In science, models are penalized for complexity (AIC, BIC criteria). Uri Alon (2007) notes that biological networks exhibit “simplicity” (modularity) that likely reflects deep computational principles. More recently, researchers have shown “symmetry and simplicity [spontaneously] emerge from the algorithmic nature of evolution” (PNAS 2022) – evolution tends to produce low-complexity (orderly) configurations over time.

In sum, both theory and practice support the preference for simplicity: the simpler explanation (lower Kolmogorov complexity) that fits the data is most probable. Occam’s Razor “works” because it is hardwired into our models of inference: shorter descriptions carry more prior weight, making them better predictors on average.

The Free Energy Principle (minimizing surprise)

  • Principle and math. Karl Friston’s Free-Energy Principle (FEP) is a unifying principle asserting that any self-organizing system must minimize a quantity called variational free energy (an upper bound on surprise). In cognitive neuroscience it formalizes why brains (and other systems) maintain internal order. Formally, the variational free energy F can be written (for a model Q over hidden states θ and data x) as the evidence lower bound:

    F=DKL(Q(θ) ∥ P(θ∣x))−ln⁡P(x) ,F = D_{\mathrm{KL}}(Q(\theta),|,P(\theta|x)) - \ln P(x) ,,F=DKL​(Q(θ)∥P(θ∣x))−lnP(x),

    where $D_{KL}$ is the Kullback–Leibler divergence. Minimizing F is equivalent to making $Q(\theta)$ approximate the true posterior $P(\theta|x)$ (Bayesian inference). Crucially, $-\ln P(x)$ is surprise (self-information) of data; thus FEP says: systems change so as to minimize expected surprise about sensory inputs.

  • Relation to entropy and order. The FEP connects neatly to thermodynamics and information theory. It implies that systems keep their states within a limited attractor (a Markov blanket) by correcting prediction errors (through perception/updating Q and action). Mathematically, variational free energy can also be expressed as (expected energy) minus entropy. As Friston notes, minimization of free energy places an upper bound on the entropy of a system’s sensory states. In other words, by minimizing F systems implicitly constrain their own entropy (uncertainty). This dovetails with the second law: a living system that constantly corrects for surprise (maximizes model evidence) effectively maintains internal order at the expense of increasing external entropy.

  • Self-organization, inference and consciousness. The FEP is often seen as a theory of life and mind: living organisms act to minimize surprise via perception and action (active inference). In recent work, Friston and Gomez (2021) argue that consciousness emerges when a self-organizing system with a Markov blanket has sufficient integration: “living systems are self-organising in nature… This mode of organization requires them to have a high level of integration”. In their view, consciousness depends on nontrivial integrated information and on inferential (free-energy) processes. That is, IIT’s Φ and FEP’s inference complement each other: systems that minimize F tend to develop internal models (patterns) which they integrate, and this integrated model drives behavior.

  • Math-of-life and least action. Friston emphasizes that FEP is a principle: like Hamilton’s principle of least action, it is true mathematically given the assumptions. One cannot easily falsify it by experiment alone, since any behavior can be cast as minimizing some free energy bound. Instead, one tests FEP by deriving testable models (predictive coding, etc.). Nevertheless, the principle provides a deep link between physics and cognition: it reduces to information-theoretic entropy in the long term, ensuring that adaptive systems do not diverge into chaos.

In summary, the Free Energy Principle formalizes order maintenance: any persistent system must minimize surprise (free energy), which implicitly bounds entropy and enforces internal models. It provides a quantitative account of why organisms look as if they “track” their environment with patterns – exactly the kind of coherent processing (Φ) that IIT demands.

Entropy vs. Complexity Distinction

  • Shannon (statistical) entropy vs. algorithmic complexity. It is crucial to distinguish entropy (a measure of ensemble uncertainty) from algorithmic complexity (description length of a single object). Shannon entropy quantifies the average surprise of random events (e.g. a fair coin toss). By contrast, Kolmogorov complexity measures how compressible a specific pattern is. A random string can have maximal Shannon entropy and maximal Kolmogorov complexity (no compression). However, a highly ordered string has low complexity even if it is long. Thus high entropy does not imply meaningful complexity; it usually implies the opposite. As Crutchfield et al. note, Kolmogorov–Chaitin complexity is in fact “a measure of randomness, not a measure of structure”. Randomness dominates K; true structure requires finding regularities that drastically shorten the description.

  • Coherence beyond entropy. Coherence (Φ) and algorithmic complexity are yet different. A random network of gates might have high Shannon entropy but little integration (low Φ). Conversely, a simple crystalline lattice has low Kolmogorov complexity and low entropy but high spatial order – however, without interactions it has near-zero Φ (no integrated information). Real coherent systems (brains, organisms) exhibit moderate entropy, high algorithmic compressibility (due to structure), and high integration. The free energy argument above shows that living systems actively constrain entropy; the Kolmogorov argument shows that interesting structures are those with low description length. Coherence (Φ) sits at this intersection: it is maximized when there is rich, irreducible structure.

  • Implications for order. Together, these distinctions reinforce why order (meaningful complexity) is fundamental. Entropy alone would imply an inexorable spread of randomness. But physics and biology show us that matter tends to organize itself into compressible, patterned forms (low K) by dissipating entropy, and that such forms are highly integrated (high Φ). For example, a living neural circuit may appear noisy (some entropy) but encodes complex information patterns (low K relative to length) which it integrates into coherent function. Thus, while entropy increases overall, algorithmic simplicity and integrated order are the hallmarks of life and mind.

Mapping Sources to Claims and Questions

  • Claim 1 – Reality tends toward order: Supported by Prigogine’s work on dissipative structures, Chaisson’s cosmic evolution and energy-flow arguments, and insights from the Free Energy Principle that living systems export entropy and maintain order. These show that local order and complexity naturally emerge, not just chaos.

  • Claim 2 – Coherence is measurable (Φ): Grounded in Integrated Information Theory (Tononi et al.), which mathematically defines Φ as the sum of irreducible information (φ) in a system’s causal structure. Criticisms by Aaronson et al. and Tononi’s responses illustrate challenges but confirm that Φ provides a concrete metric of integrated order.

  • **Claim 3 – Simpler explanations preferred: **Established by Kolmogorov complexity theory and Solomonoff induction. These sources show that the shortest algorithm fitting data is statistically dominant (formalizing Occam’s razor). The linked concept of algorithmic probability means simpler (lower-K) hypotheses have higher prior weight.

  • Q1 – Why laws exist?: Informed by fine-tuning arguments (life-permitting conditions require specific laws), symmetry/invariance reasoning, and the idea that patterns are preconditions for consciousness. All suggest that only law-governed universes support observers.

  • Q2 – Meaning without pattern?: Answered by IIT’s information exclusion principle and general information theory: meaning arises through structured distinctions. Tononi explicitly links “meaning” to discriminating among alternatives, implying patternless randomness cannot yield meaning.

  • Q3 – Why Occam’s Razor works?: Justified by Solomonoff’s theory and Kolmogorov’s definition. Together they show that hypotheses with shorter descriptions have higher algorithmic probability, making simple explanations more likely to generalize. (In probabilistic terms, $P(\text{theory})\propto 2^{-K(\text{theory})}$.)

Sources: Each claim and question above is backed by the cited literature. The Wolfram Finance-like citations (e.g. ) refer to lines in peer-reviewed or authoritative sources opened during research. These include Tononi’s IIT papers, Aaronson’s critiques, Friston’s FEP accounts, Prigogine’s Nobel lecture, Kolmogorov/Chaitin foundational work, and surveys by Chaisson, among others.

Sources