Neuro-Semiotic Reasoning: Computational Semiosis for Observational Causal Inference
Introducing our first research paper on computational semiosis and observational causal inference
Today we're sharing our first research paper: "Neuro-Semiotic Reasoning: Computational Semiosis for Observational Causal Inference."
This paper introduces the theoretical foundation and architectural design of o-machine's reasoning engine. It addresses what we call the Causal Reasoning Gap: the inability of current AI systems to derive verified causation from real-world observational signals.
The Core Problem
Large language models excel at statistical pattern matching but fail systematically at causal reasoning. They can tell you what happened, but not why it happened. They optimize for fluency, not truth. In high-stakes decision environments—investment due diligence, supply chain risk assessment, geopolitical analysis—this limitation produces what we call "hallucinations of expertise": confident summaries that reflect training data consensus while remaining blind to actual causal mechanisms.
The gap exists because statistical systems are built on a fundamentally different foundation: they optimize for token probability rather than causal understanding from observational evidence.
Our Approach: Computational Semiosis
We present the Neuro-Semiotic Reasoning Engine, an architecture that constructs verified causal understanding through semiotic interpretation of differential relationships. Drawing on Charles Sanders Peirce's theory of unlimited semiosis, we demonstrate that recursive interpretation of facts through multiple contextual lenses enables capabilities difficult to achieve in statistical systems.
The key insight: causal meaning emerges through recursive interpretation of differential relationships across four dimensions:
- Temporal: What changed over time (velocity, acceleration, phase transitions)
- Spatial: What's connected (causal chains, dependencies)
- Conceptual: What phase or pattern (industry transitions, strategic pivots)
- Absence: What's missing (expected-but-absent events as diagnostic signals)
When interpretations across all four dimensions converge on a consistent causal structure, this provides verification without ground truth—verification through triangulation rather than experimental validation.
Five Architectural Pillars
The system is built on five complementary pillars, each addressing a specific computational requirement:
- Temporal Knowledge Graphs: Atomic facts as the unit of storage, not documents. Every fact is timestamped with full provenance, enabling velocity analysis and absence detection.
- Concept-Mediated Ontology: Entities connect through higher-order concepts that evolve their meaning through use. This enables cross-domain inference and concept-level pattern recognition.
- Adversarial Verification (The Socratic Protocol): Every inference undergoes systematic falsification through a Generator/Critic debate before acceptance. Interpretations must survive adversarial scrutiny across all four dimensions.
- Behavioral Modeling: The system maintains models of expected behavior, making absence a first-class signal. "Company X stopped hiring perception engineers" becomes meaningful only when compared to learned baselines.
- Evolutionary Optimization: The system's reasoning strategies evolve through selection pressure. It autonomously improves detection patterns and verification protocols over time.
Emergent Capabilities
These architectural choices enable six capabilities that are structurally difficult in statistical systems:
- Shadow Mapping: Inferring hidden strategies from weak, cross-domain signals
- Absence Detection: Reasoning about what isn't happening as a diagnostic tool
- Multi-Hop Causal Chains: Tracing implications across domains with temporal lag prediction
- Counterfactual Reasoning: Identifying causal dependencies through expected-but-absent outcomes
- Concept-Mediated Discovery: Finding unknown actors through behavioral patterns, not keywords
- Complete Explainability: Every inference traceable back through its full interpretive chain to atomic facts and source documents
The Paradigm Shift
We argue this represents a paradigm shift analogous to pre-Google search engines to Google. Before Google, search engines ranked by keyword frequency. Google derived relevance from the structure of relationships (links between pages). Similarly, we derive causation from the structure of differential relationships between facts, not from token co-occurrence frequencies.
Just as Google's graph-based architecture couldn't be approximated by better keyword matching, observational causal reasoning cannot be approximated by larger language models or more training data. The required primitives—temporal differentials, absence detection, concept-mediated inference, adversarial verification—are structurally different from token probability computation.
What's Next
The architecture described in the paper has been implemented and is under active development. We're preparing a closed beta launch for Q2 2026 with design partners from private equity, venture capital, and corporate development teams.
Comprehensive empirical results across 15 benchmark scenarios will be reported following the closed beta period. This paper establishes the theoretical foundation and architectural design as our primary contribution.
Read the Paper
The full paper is available for download. We welcome collaboration with researchers and domain experts interested in extending this framework to additional domains.
Citation:
Trajkow, M. (2026). Neuro-Semiotic Reasoning: Computational Semiosis for Observational Causal Inference.
Questions or interested in collaborating? Reach out at martin@o-machine.com