Structural Information Theory

The Structural Information Theory (Teoría de la Información Estructural, TIE, or SIT in english) offers an alternative approach to understanding data, moving away from measuring statistical uncertainty, entropy (Shannon's theory) towards measuring the inherent structure, geometry, and generative laws of a system. SIT views information deterministically as structure.

Here is a simplified discussion of SIT's core concepts, ensuring the critical mathematical definitions are included:

I. The Formalism: Deconstructing Data into Structure

SIT begins by dissecting a sequence of data into its fundamental components to define structural relationships.

1. The Presence Map (MΦ)

This is the basic building block of SIT.

2. The Language of Structure: Transformations (τ)

Structure is defined by how these Presence Maps relate to one another.

3. Defining Relationships: Structural Adjacency

Two Presence Maps (MΦ and MΦ) are structurally adjacent if one map can be closely generated by applying a transformation (T) from the defined space to the other map.

II. Core Principles: Finding the Irreducible Information

These principles determine which parts of the data are genuinely new information and which are redundant.

1. The Principle of Structural Partition

The Dependency Graph (GΣ) naturally breaks down into m unconnected components (families).

2. The Principle of Minimum Structural Description (Causal Information)

This defines the primary measure of SIT: Structural Information (IS(F)).

III. The Definitive Metric and the Role of Time

SIT provides a specific, computable metric for information and offers a unique perspective on randomness.

1. The Structural Information Metric (IS(F))

The total structural information is the sum of costs for describing the irreducible parts, plus the costs for describing how the dependent parts are generated:

IS(F)=j=1mCalg(Mbasej)+i=m+1k(log2(|Tj|)+Calg(Ei))

This formula quantifies the information as the sum of three elements:

  1. The Algorithmic Complexity (Calg) of the fundamentally independent Base Maps (the truly irreducible parts).

  2. The cost of describing the structural relations (the transformations, T) used to link the other maps.

  3. The cost of the Exceptions/Errors (E) where the rules fail.

2. Structural Entropy (HS)

Structural Entropy (HS) is a global geometric measure, unlike Shannon's local statistical measure.

3. Temporal Randomness

SIT proposes that much of what is perceived as randomness is not due to the choice of the symbol itself, but rather the instant of its appearance. Randomness is often an "illusion causal" (causal illusion), resulting from observing the superposition and desynchronisation of multiple underlying deterministic processes.


source: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5374335
(in Spanish)

#theory #information #statistics #structure #entropy #Shannon #geometry