Structural Information Theory
The Structural Information Theory (Teoría de la Información Estructural, TIE, or SIT in english) offers an alternative approach to understanding data, moving away from measuring statistical uncertainty, entropy (Shannon's theory) towards measuring the inherent structure, geometry, and generative laws of a system. SIT views information deterministically as structure.
Here is a simplified discussion of SIT's core concepts, ensuring the critical mathematical definitions are included:
I. The Formalism: Deconstructing Data into Structure
SIT begins by dissecting a sequence of data into its fundamental components to define structural relationships.
1. The Presence Map ( )
This is the basic building block of SIT.
-
For every symbol (
) in an alphabet, the Presence Map ( ) is a binary vector (a sequence of 1s and 0s) that acts as the "fingerprint" of that symbol in time. A '1' indicates the symbol appeared at that specific time position. -
These maps obey the Fundamental Law of Exclusion—if you add up all the maps for all symbols, you get a vector of all ones, confirming that one and only one symbol occupies every position.
2. The Language of Structure: Transformations ( )
Structure is defined by how these Presence Maps relate to one another.
-
These relationships are described by a finite set of invertible transformations (
), which are geometric "rules" applied to the binary vectors. -
Basic rules in
include Cyclic Shift ( ), Reflection ( ), and Negation (:). In a physical context, the rules governing the universe could be seen as the transformations within .
3. Defining Relationships: Structural Adjacency
Two Presence Maps (
-
This approximation is measured using the Hamming distance (
), which counts the number of positions that differ. This distance must be below a tolerance threshold ( ) to account for "noise" or exceptions. -
These adjacencies create the Symbolic Dependency Graph (
), which visually maps the file’s internal structure.
II. Core Principles: Finding the Irreducible Information
These principles determine which parts of the data are genuinely new information and which are redundant.
1. The Principle of Structural Partition
The Dependency Graph (
-
Consequence: The true informational basis of the file is not the original
symbols, but only the Base Maps, one derived from each independent family. The remaining maps are dependent and therefore considered redundant. -
The Conjecture of Redundancy suggests that
is often much smaller than , potentially bounded by .
2. The Principle of Minimum Structural Description (Causal Information)
This defines the primary measure of SIT: Structural Information (
-
is not the file’s statistical randomness (entropy), but the length of the shortest computer program required to generate the file using the SIT structural rules. -
For systems generated by simple rules (
), this cost is tiny, bounded by $$I_S(F) = O(r \cdot \log N)$$, making it independent of the overall file length ( ).
III. The Definitive Metric and the Role of Time
SIT provides a specific, computable metric for information and offers a unique perspective on randomness.
1. The Structural Information Metric ( )
The total structural information is the sum of costs for describing the irreducible parts, plus the costs for describing how the dependent parts are generated:
This formula quantifies the information as the sum of three elements:
-
The Algorithmic Complexity (
) of the fundamentally independent Base Maps (the truly irreducible parts). -
The cost of describing the structural relations (the transformations,
) used to link the other maps. -
The cost of the Exceptions/Errors (
) where the rules fail.
2. Structural Entropy ( )
Structural Entropy (
-
is defined by the group of symmetries ( ) that leave the file structure unchanged. -
-
A file with strong symmetries (e.g., a highly repetitive texture) will have a large symmetry group (
), resulting in a drastically lower Structural Entropy ( ) compared to its Shannon Entropy.
3. Temporal Randomness
SIT proposes that much of what is perceived as randomness is not due to the choice of the symbol itself, but rather the instant of its appearance. Randomness is often an "illusion causal" (causal illusion), resulting from observing the superposition and desynchronisation of multiple underlying deterministic processes.
-
Since each Presence Map is a symbol's informational trajectory in time, plotting the intervals between symbol appearances yields its stroboscopic curve.
-
For structured systems, these curves are contained within an Evolution Envelope, which defines the limits of causal possibilities for that symbol. The size of this envelope measures the symbol's temporal "freedom".
source: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5374335
(in Spanish)
#theory #information #statistics #structure #entropy #Shannon #geometry