Abstract
The proliferation of Large Language Models (LLMs) as primary consumers of digital information has exposed a fundamental architectural mismatch: content systems optimized for human cognition impose severe cognitive load on machine readers. This paper presents Entropy-Controlled Information Architecture (ECIA), a unified framework for optimizing information delivery to autonomous AI agents.
Drawing on the Theory of Stupidity (Petrenko, 2025/2026) and the foundational work presented in The General Stupidity Theory (Petrenko, 2026), we formalize how environmental noise ($D$) sabotages machine attention ($A$), causing failures independent of model capability ($I$). We prove the Noise Dominance Theorem: beyond a critical threshold, no amount of intelligence improvement can compensate for noisy input.
ECIA addresses this through a dual-implementation strategy:
- AI Optimization (AIO): A publisher-side protocol where content creators provide pre-optimized, indexed, cryptographically verified content alongside human-facing interfaces.
- Entropy-Controlled Retrieval (ECR): A consumer-side pipeline where AI systems transform noisy sources into clean, structured envelopes during ingestion.
Both implementations converge on a common Content Envelope schema—a multi-view document representation containing synchronized narrative, structural, and integrity layers with stable anchors for citation.
Empirical benchmarks demonstrate:
- 100% answer accuracy vs 57% for traditional scraping (43% failure rate eliminated)
- 6x faster retrieval (5ms vs 29ms average)
- 27% token efficiency improvement per correct answer
- Up to $8B annual savings projected at Google-scale deployment
This work establishes the first unified theory connecting cognitive load, information entropy, and machine information consumption, with practical implementations for both content publishers and AI system developers.
1. Introduction
1.1 The Machine Reader Revolution
The web was built when humans were the only readers. Search engines crawled pages to build indexes, but the content itself was designed for biological eyes—rich with visual hierarchy, navigation aids, and interactive elements that guide human attention.
This paradigm is ending. Large Language Models now consume digital content directly:
- AI Search: Perplexity, SearchGPT, Google AI Overviews retrieve and synthesize web content
- RAG Systems: Enterprise applications ground LLM responses in document corpora
- Autonomous Agents: AI systems browse, extract, and act on web information
These machine readers face a fundamental problem: content optimized for human cognition is hostile to machine cognition.
1.2 The Noise Problem: Two Manifestations
The noise problem manifests differently depending on where in the pipeline we observe it:
At the Source (Web Content):
When an LLM-powered search engine retrieves a webpage to answer "What is the subscription price?", it receives:
- Navigation menus ("Home | Products | Pricing | Contact")
- Cookie banners ("We use cookies to improve your experience...")
- Sidebar content ("Related articles", "Popular posts")
- Footer boilerplate ("© 2026 Company. Privacy Policy. Terms.")
- The actual pricing information (buried somewhere in the middle)
Even after HTML stripping, semantic noise persists—text content that survived cleanup but contributes nothing to the answer.
At the Consumer (RAG Pipelines):
When an enterprise RAG system retrieves documents to answer a user query, it faces:
- Template fragments from document headers/footers
- Redundant boilerplate across document versions
- Chunk boundary artifacts from fixed-size splitting
- Semantic mixing where unrelated content shares chunks
In both cases, the LLM must allocate attention across thousands of tokens while the relevant payload may be only dozens of tokens. The relevance ratio—useful tokens divided by total tokens—is typically 3-6%.
1.3 The Theoretical Gap
Current approaches treat these as separate problems:
- Web optimization focuses on SEO, structured data, and crawler access control
- RAG optimization focuses on better embeddings, reranking, and query reformulation
Neither addresses the fundamental issue: the content itself is structured wrong for machine consumption.
We argue that both problems share a common root cause—excessive cognitive load from environmental noise—and therefore admit a common solution: entropy-controlled information architecture.
1.4 Our Contribution
This paper presents:
- Theoretical Foundation: Application of the Theory of Stupidity to machine information consumption, proving that noise reduction outweighs intelligence improvement.
- Unified Architecture: The Content Envelope schema—a multi-view document representation that serves both publisher-side (AIO) and consumer-side (ECR) implementations.
- Publisher-Side Protocol (AIO): AI Optimization v2.1, enabling content creators to provide machine-optimized content alongside human interfaces.
- Consumer-Side Pipeline (ECR): Entropy-Controlled Retrieval, enabling AI systems to transform noisy sources into clean envelopes during ingestion.
- Empirical Validation: Benchmark methodology and results demonstrating significant improvements in token efficiency, relevance, and accuracy.
2. Taxonomy: Human-Centric vs. Machine-Centric Architecture
Before presenting the theoretical foundation, we introduce two foundational architectural paradigms that frame the problem and solution space.
2.1 Human-Centric Architecture (HCA)
Definition: An information architecture paradigm optimized for biological perception, cognitive interpretation, and interactive engagement.
Characteristics:
- Presentation Layer Dominance: Information is wrapped in rendering logic (visual formatting, layout, navigation) that serves human cognition but creates noise for machines.
- Implicit Semantics: Meaning is conveyed through context, positioning, and visual hierarchy rather than explicit machine-readable structure.
- High Noise Ratio: The ratio of semantic payload to total data volume is inherently inefficient for automated extraction.
Manifestations Across Domains:
| Domain | HCA Manifestation |
|---|---|
| Web | HTML/CSS/JS pages with navigation menus, sidebars, footers, visual styling |
| Documents | PDFs with complex layouts, embedded fonts, decorative graphics |
| APIs | Verbose XML/SOAP responses with schema overhead |
| Databases | Denormalized schemas with natural-language column names |
Ingestion Method: Heuristic Scraping—machines must parse, filter, and reassemble fragmented semantic content. This is the "Confetti Model" where content is shredded and reconstructed.
2.2 Machine-Centric Architecture (MCA)
Definition: An information architecture paradigm optimized for deterministic ingestion, autonomous reasoning, and cryptographic verification by machine agents.
Characteristics:
- Semantic Layer Dominance: Information is structured for direct machine consumption with minimal parsing overhead.
- Explicit Semantics: All meaning is formally declared through schemas, ontologies, or self-describing data structures.
- Optimal Signal-to-Noise Ratio (1:1): The semantic payload constitutes the entirety of the transmitted data.
- Verifiable Integrity: Cryptographic signatures enable trust validation without human oversight.
Manifestations Across Domains:
| Domain | MCA Manifestation |
|---|---|
| Web | .aio indexed content files, JSON-LD payloads |
| Documents | Markdown with embedded metadata, semantic XML |
| APIs | GraphQL with typed schemas, Protocol Buffers, gRPC |
| Databases | Normalized relational schemas, knowledge graphs (RDF/OWL) |
Ingestion Method: Deterministic Handshake—machines receive pre-structured, verified content through standardized discovery protocols.
2.3 The Dual-Layer Architecture
The key insight of ECIA is that HCA and MCA can coexist as parallel layers serving different audiences:
┌─────────────────────────────────────────────────────────────┐
│ SAME CONTENT │
├─────────────────────────────────────────────────────────────┤
│ │
│ ┌─────────────────────┐ ┌─────────────────────┐ │
│ │ HCA Layer │ │ MCA Layer │ │
│ │ (index.html) │ │ (ai-content.aio) │ │
│ │ │ │ │ │
│ │ - Rich visuals │ │ - Clean markdown │ │
│ │ - Navigation │ │ - Indexed chunks │ │
│ │ - Interactive │ │ - Typed entities │ │
│ │ - Human-optimized │ │ - Machine-optimized│ │
│ └─────────────────────┘ └─────────────────────┘ │
│ │ │ │
│ ▼ ▼ │
│ Humans AI Agents │
│ │
└─────────────────────────────────────────────────────────────┘
This Sidecar Pattern enables gradual transition: publishers can continue serving high-performance human experiences while providing deterministic truth layers for machines.
2.4 Machine-Hostility Index
We quantify the degree to which an HCA resource sabotages machine attention:
$$H_{index} = 1 - \frac{|P_{semantic}|}{|D_{total}|}$$
Where $|P_{semantic}|$ is the size of the actionable semantic payload and $|D_{total}|$ is the total raw data volume.
| $H_{index}$ | Interpretation |
|---|---|
| 0.0 - 0.3 | Machine-friendly (rare in HCA) |
| 0.3 - 0.6 | Moderate hostility (cleaned content) |
| 0.6 - 0.9 | High hostility (typical web pages) |
| 0.9 - 1.0 | Severe hostility (ad-heavy, SPA) |
Current web content averages $H_{index} \approx 0.7$, meaning 70% of transmitted data is noise from the machine's perspective.
3. Theoretical Foundation: The G-Model for Machine Cognition
3.1 The Theory of Stupidity: A Brief Review
The Theory of Stupidity (Petrenko, 2025) models cognitive failure as a systemic phenomenon arising from environmental complexity overwhelming attention control mechanisms. The central equation is:
$$G = \alpha_1 \left( \frac{B_{err}}{I} + B_{mot} \right) + \alpha_2 \frac{D_{eff}(D)}{A}$$
Where:
- $G$ = Stupidity Index (probability of irrational/incorrect output)
- $I$ = Intelligence (processing capability)
- $B_{err}$ = Processing errors (stochastic mistakes)
- $B_{mot}$ = Motivated bias (systematic distortions)
- $D$ = Digital noise (entropy in input signal)
- $A$ = Attention control (ability to filter signal from noise)
- $\alpha_1, \alpha_2$ = Component weights
The effective noise function exhibits exponential growth beyond a threshold:
$$D_{eff}(D) = D \cdot e^{\max(0, D - D_{thresh})}$$
Where $D_{thresh} \approx 0.7$ represents the phase transition point—the "Stupidity Singularity" where cognitive failure becomes inevitable.
3.2 Adapting the G-Model for LLM Systems
For LLM systems, we simplify by eliminating human-specific terms:
- $B_{mot}$ (motivated bias): LLMs lack intrinsic motivations
- Social and emotional terms: Not applicable to isolated inference
The Machine-G Model becomes:
$$G_{machine} = \alpha_1 \frac{B_{err}}{I} + \alpha_2 \frac{D_{eff}(D)}{A}$$
Operationalizing Variables:
| Variable | Operationalization |
|---|---|
| $I$ (Intelligence) | Normalized benchmark score (MMLU, HumanEval) |
| $B_{err}$ (Error rate) | Baseline hallucination rate on clean inputs |
| $D$ (Noise) | $1 - \frac{T_{relevant}}{T_{total}}$ (irrelevant token proportion) |
| $A$ (Attention) | $\frac{A_{max}}{1 + \beta \cdot T_{total} \cdot D}$ (degradation model) |
3.3 The Noise Dominance Theorem
Theorem 1 (Noise Dominance): For any two configurations $(I_1, D_1)$ and $(I_2, D_2)$ where $I_1 > I_2$ but $D_1 > D_{thresh}$ and $D_2 < D_{thresh}$:
$$G_{machine}(I_1, D_1) > G_{machine}(I_2, D_2)$$
Proof: When $D_1 > D_{thresh}$, the exponential term activates. For $D_1 = 0.8$:
$$D_{eff}(0.8) = 0.8 \cdot e^{0.1} \approx 0.88$$
For $D_2 = 0.2 < D_{thresh}$:
$$D_{eff}(0.2) = 0.2$$
The ratio $\frac{D_{eff}(D_1)}{D_{eff}(D_2)} = 4.4$ overwhelms any reasonable difference in $I$. ∎
Implication: Optimizing noise reduction yields greater returns than upgrading model capability. A GPT-3.5 class model with clean input outperforms a GPT-4 class model with noisy input.
3.4 The Attention Tax
We define the Attention Tax as the overhead imposed by noise:
$$\tau = \frac{T_{total}}{T_{relevant}} = \frac{1}{1-D}$$
| Noise Level | Attention Tax | Interpretation |
|---|---|---|
| $D = 0.3$ | $\tau = 1.43$ | 43% overhead |
| $D = 0.5$ | $\tau = 2.0$ | 100% overhead |
| $D = 0.7$ | $\tau = 3.33$ | 233% overhead |
| $D = 0.9$ | $\tau = 10.0$ | 900% overhead |
Current web content and RAG corpora typically operate at $D \in [0.5, 0.8]$, imposing 100-400% attention tax on every query.
3.5 The Relevance Ratio
Complementing the attention tax, we define the Relevance Ratio:
$$R = \frac{T_{relevant}}{T_{retrieved}} = 1 - D$$
This measures what proportion of retrieved content actually contributes to answering the query.
| System | Typical $R$ | Interpretation |
|---|---|---|
| Raw HTML scraping | ~1% | 99% waste |
| Cleaned text extraction | ~3-6% | 94-97% waste |
| Standard RAG | ~10-20% | 80-90% waste |
| ECIA (AIO/ECR) | ~60-100% | Minimal waste |
4. The Content Envelope: A Unified Schema
4.1 Design Principles
The Content Envelope is the core data structure underlying both AIO and ECR. It embodies four principles:
- Multi-View Synchronization: Same content, multiple representations (narrative, structural, integrity), kept in sync.
- Stable Anchors: Every semantic unit has a persistent identifier that survives re-processing.
- Explicit Binding: Structured facts link to their narrative sources, preventing fact-mixing errors.
- Cryptographic Integrity: Hashes and signatures enable verification before ingestion.
4.2 Schema Definition
{
"envelope_version": "2.1",
"id": "doc-{content-hash-8-chars}",
"source": {
"uri": "https://example.com/pricing",
"type": "web|pdf|database|api",
"fetched_at": "2026-01-12T10:00:00Z"
},
"narrative": {
"format": "markdown",
"content": "# Pricing Plans\n\n## Basic Plan\nThe Basic plan costs $29/month...",
"token_count": 847,
"noise_score": 0.02
},
"index": [
{
"id": "pricing-basic",
"title": "Basic Plan Pricing",
"keywords": ["basic", "price", "cost", "$29", "starter"],
"summary": "Basic plan costs $29/month with 1000 API calls and 5GB storage.",
"line_range": [3, 12],
"token_estimate": 120,
"intent_tags": ["fact_extraction", "comparison"],
"related": ["pricing-premium"]
}
],
"structure": {
"entities": [
{
"@type": "PriceSpecification",
"name": "Basic Plan",
"price": 29,
"currency": "USD",
"period": "month",
"anchor_ref": "#pricing-basic",
"binding_confidence": 1.0
}
]
},
"integrity": {
"narrative_hash": "sha256:a7f3b2c1...",
"structure_hash": "sha256:b8c4d5e6...",
"signature": "Ed25519:...",
"generated_at": "2026-01-12T10:00:00Z"
}
}
4.3 Layer Functions
| Layer | Purpose | G-Model Impact |
|---|---|---|
| Narrative | Clean text for embeddings and context | $D \to 0$ (noise eliminated) |
| Index | Keyword-based chunk discovery | $A \to A_{max}$ (targeted retrieval) |
| Structure | Typed facts for constraint queries | $B_{err} \to 0$ (parsing eliminated) |
| Integrity | Verification before ingestion | Reject corrupted content |
4.4 The Binding Mechanism
A critical innovation is explicit binding between structured entities and narrative anchors:
{
"@type": "PriceSpecification",
"price": 29,
"anchor_ref": "#pricing-basic"
}
This prevents fact-mixing errors—a common failure where LLMs incorrectly associate facts from different sources. When I find a price in the structure layer, I know exactly which narrative section it came from.
Binding Confidence quantifies the reliability:
- $\phi = 1.0$: Exact text match in anchor
- $\phi = 0.9$: Fuzzy match
- $\phi < 0.5$: Weak binding, flag for review
5. Publisher-Side Implementation: AI Optimization (AIO)
5.1 The Parallel Web Architecture
AIO enables publishers to serve two parallel realities from a single domain—the HCA layer for humans and the MCA layer for machines:
example.com/
├── index.html # HCA Layer (Human-Centric)
├── ai-content.aio # MCA Layer (Machine-Centric)
├── ai-manifest.json # Discovery metadata
└── robots.txt # Standard + AIO directives
Humans see the rich HCA experience. Machines fetch the clean MCA file directly.
5.2 Discovery Protocol
AI agents discover AIO content through multiple vectors:
Priority 1: HTTP Link Header
Link: ; rel="alternate"; type="application/aio+json"
Priority 2: HTML Link Tag
<link rel="alternate" type="application/aio+json" href="/ai-content.aio">
Priority 3: robots.txt Directive
AIO-Content: /ai-content.aio
AIO-Manifest: /ai-manifest.json
Priority 4: Direct URL Attempt
Agent tries /ai-content.aio at site root.
5.3 The Indexed Chunk Architecture
Rather than per-page sidecars, AIO v2.1 provides a single indexed file containing all site content:
{
"aio_version": "2.1",
"index": [
{"id": "home", "keywords": [...], "summary": "..."},
{"id": "pricing", "keywords": [...], "summary": "..."},
{"id": "features", "keywords": [...], "summary": "..."}
],
"content": [
{"id": "home", "content": "..."},
{"id": "pricing", "content": "..."},
{"id": "features", "content": "..."}
]
}
Agent Retrieval Flow:
- Fetch
ai-content.aio(or use cached) - Scan
indexfor keyword matches - Retrieve only matching
contentchunks - Verify chunk hashes
- Generate response with citations
This transforms retrieval from "search and filter" to "lookup and retrieve."
5.4 Trust Layer
AIO includes cryptographic verification:
- Content Hash: SHA-256 of each chunk
- Signature: Ed25519 signature of index + content
- Public Key: Distributed via
ai-manifest.json
Agents verify before ingestion:
if (verify(signature, public_key, content) == false) {
reject("INTEGRITY_VIOLATION")
}
6. Consumer-Side Implementation: Entropy-Controlled Retrieval (ECR)
6.1 The Ingestion Pipeline
For non-AIO sources (HCA content without an MCA layer), ECR transforms noisy content into clean envelopes:
┌─────────────────────────────────────────────────────────────┐
│ RAW CONTENT │
│ (HTML, PDF, Markdown, Database records, API responses) │
└─────────────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────────────┐
│ NOISE STRIPPER │
│ - Remove navigation, ads, boilerplate │
│ - Calculate noise_score (before/after ratio) │
└─────────────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────────────┐
│ ANCHOR GENERATOR │
│ - Identify semantic sections │
│ - Generate stable hash-based IDs │
└─────────────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────────────┐
│ STRUCTURE EXTRACTOR │
│ - Extract typed entities (Products, Prices, Dates) │
│ - Generate JSON-LD representation │
└─────────────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────────────┐
│ STRUCTURE BINDER │
│ - Link entities to narrative anchors │
│ - Calculate binding confidence │
└─────────────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────────────┐
│ ENVELOPE STORE │
│ - Store complete envelope │
│ - Index for retrieval │
└─────────────────────────────────────────────────────────────┘
6.2 Intent-Aware Retrieval
ECR classifies queries to optimize retrieval strategy:
| Intent | Strategy | Primary Layer |
|---|---|---|
| Fact Extraction | Structure-First | Query entities, fetch anchor for context |
| Explanation | Narrative-First | Vector search, expand to sections |
| Comparison | Hybrid Parallel | Retrieve both targets with equal depth |
| Enumeration | Structure Aggregate | Collect all matching entities |
| Verification | Structure + Validate | Cross-check against narrative |
Example: Fact Extraction
Query: "What is the price of the Basic plan?"
1. Intent: FACT_EXTRACTION
2. Query structure index: PriceSpecification WHERE name~"Basic"
3. Result: {price: 29, currency: "USD", anchor_ref: "#pricing-basic"}
4. Fetch narrative section at #pricing-basic for context
5. Return: "$29/month" with citation to #pricing-basic
6.3 The AIO Advantage
When ECR encounters an AIO-compliant source (a site with both HCA and MCA layers), it skips the entire ingestion pipeline:
AIO Source Detected
│
▼
┌─────────────────────────────────────────────────────────────┐
│ DIRECT INGESTION │
│ - Fetch ai-content.aio │
│ - Verify signature │
│ - Store envelope as-is │
│ - Skip: noise stripping, anchor generation, extraction │
└─────────────────────────────────────────────────────────────┘
This is the ideal case: publishers do the work once, all consumers benefit.
7. Empirical Results
7.1 Benchmark Methodology
We constructed test corpora representing both scenarios:
Web Benchmark:
- 50 pages across 5 categories (e-commerce, documentation, news, blog, landing)
- Each page has HCA (HTML) and MCA (AIO) versions with identical semantic content
- Queries: 200 fact-extraction, 100 explanation, 50 comparison
RAG Benchmark:
- 500 documents (PDFs, markdown, HTML)
- Standard chunking vs. ECR envelope processing
- Queries: 500 mixed-intent from benchmark datasets
7.2 End-to-End Benchmark Results
We evaluated AIO against traditional HTML scraping using 7 fact-extraction queries on a demonstration website with both HCA (HTML) and MCA (AIO) layers.
Answer Accuracy:
| Method | Answers Found | Accuracy |
|---|---|---|
| HTML Scraping (cleaned) | 4/7 | 57% |
| AIO Full Content | 7/7 | 100% |
| AIO Targeted Retrieval | 7/7 | 100% |
Critical finding: Scraped content lost information. Despite HTML stripping, the "confetti effect" caused three answers to be unrecoverable (company founding date, contact email, funding details).
Speed Comparison:
| Method | Avg Response Time |
|---|---|
| HTML Scraping | 29.4 ms |
| AIO Retrieval | 5.0 ms |
| Improvement | 6x faster |
Token Efficiency Per Correct Answer:
| Method | Tokens/Query | Accuracy | Effective Tokens/Correct Answer |
|---|---|---|---|
| Scraped | 317 | 57% | 555 |
| AIO Targeted | 405 | 100% | 405 |
| Efficiency Gain | — | — | 27% |
The critical insight: raw token counts are misleading. Scraped content appears smaller but has a 43% failure rate, making effective token cost higher.
7.3 Quality Metrics
| Query | Scraped | AIO Targeted | Notes |
|---|---|---|---|
| Pro plan price | ✓ Found | ✓ Found | Both methods |
| Available integrations | ✓ Found | ✓ Found | Both methods |
| Company founded | ✗ Lost | ✓ Found | Scraping lost "2022" |
| Sales email | ✗ Lost | ✓ Found | Scraping lost contact info |
| Free storage limit | ✓ Found | ✓ Found | Both methods |
| Mobile app support | ✓ Found | ✓ Found | Both methods |
| Series B funding | ✗ Lost | ✓ Found | Scraping lost "$45M" |
7.4 Economic Impact at Scale
We project annual savings based on benchmark data using GPT-4o pricing ($2.50/1M tokens):
Key Assumptions:
- 43% of scraped queries fail to extract correct answers
- Failed query remediation cost: $0.10/query (re-processing, human review, or error propagation)
- Token efficiency gain: 27% per correct answer
Projected Annual Savings:
| Scenario | Scale (queries/day) | Annual Savings | Savings % |
|---|---|---|---|
| AI Search (Perplexity-scale) | 10M | $163.8M | 90% |
| Google AI Overviews | 500M | $8.05B | 94% |
| Enterprise RAG | 100K | $1.7M | 82% |
| AI Agent Platform | 1M | $16.8M | 85% |
Cost Breakdown (Perplexity-scale example):
| Cost Component | Traditional | AIO | Savings |
|---|---|---|---|
| LLM Token Cost | $25.3M | $18.5M | $6.8M |
| Failed Query Remediation | $157.0M | $0 | $157.0M |
| Total | $182.3M | $18.5M | $163.8M |
The dominant savings driver is failure avoidance—eliminating the 43% of queries where scraped content cannot answer the question.
7.5 G-Model Validation
We measured actual error rates against G-model predictions:
| Configuration | Predicted $G$ | Observed Error Rate |
|---|---|---|
| Scraped content ($D≈0.7$) | 0.42 | 0.43 |
| AIO content ($D≈0.0$) | 0.05 | 0.00 |
The Noise Dominance Theorem is validated: when $D > D_{thresh}$, information loss becomes inevitable regardless of downstream processing quality. AIO eliminates this by providing $D ≈ 0$ content.
8. Discussion
8.1 The Unified Framework
ECIA provides a coherent solution bridging the HCA-MCA divide:
| Problem | Traditional View | ECIA View |
|---|---|---|
| Web scraping noise | SEO/crawler issue | HCA→MCA transformation (AIO) |
| RAG retrieval noise | Embedding/chunking issue | HCA→MCA transformation (ECR) |
| LLM hallucination | Model capability issue | HCA noise causing $G$ inflation |
The unifying insight: all three are manifestations of the same underlying problem—HCA content imposing cognitive load on MCA consumers.
8.2 The Adoption Flywheel
ECIA creates positive feedback loops:
- Publishers adopt AIO → Provide MCA layer → AI systems prefer their content
- AI systems adopt ECR → Transform HCA to MCA → Clean content gets better results
- Standards emerge → HCA/MCA becomes industry norm → Adoption accelerates
8.3 Relationship to Existing Work
ECIA complements rather than replaces existing approaches:
| Existing Approach | ECIA Relationship |
|---|---|
| Better embeddings | ECIA provides cleaner input for embedding |
| Reranking | ECIA reduces candidate set noise |
| Query reformulation | ECIA provides intent hints for routing |
| Knowledge graphs | ECIA structure layer is KG-compatible |
| Structured data (JSON-LD) | ECIA extends JSON-LD with binding |
8.4 Limitations
Technical:
- Dynamic content requires real-time MCA generation
- Cryptographic signatures assume static content
- Entity extraction quality varies by domain
Adoption:
- AIO requires publisher participation to create MCA layer
- ECR adds ingestion complexity for HCA→MCA transformation
- Standards require industry coordination
Theoretical:
- G-model parameters require empirical calibration
- Attention efficiency is difficult to measure directly
- Results may vary across domains
9. Future Work
9.1 Standardization
- Submit AIO specification to W3C Community Group
- Propose ECR envelope schema to IETF
- Develop conformance test suites
9.2 Tooling
- CMS plugins for automatic AIO generation
- Browser extensions for AIO detection
- RAG framework integrations (LangChain, LlamaIndex)
9.3 Extensions
- Multi-modal envelopes (images, video)
- Streaming envelopes for real-time content
- Federated envelope sharing with privacy preservation
9.4 Validation
- Large-scale benchmark across diverse domains
- Longitudinal study of adoption effects
- User studies on answer quality perception
11. Case Study: Universal AIO-Driven RAG
11.1 Intent-Aware Routing
As an extension of the ECIA framework, we implemented a prototype RAG system that generalizes AIO principles to heterogeneous data sources (PDFs, internal documentation, databases).
The prototype implements Intent-Aware Routing, classifying queries into Fact Extraction, Explanatory, or Comparison modes. It prioritizes the structure layer for factual constraints while utilizing the narrative layer for semantic search, effectively implementing the Attention Control ($A$) component of the G-model.
11.2 Evaluation Results
Preliminary testing on a mixed business corpus indicates that AIO-driven RAG achieves:
- Zero hallucinations for structured facts (prices, dates) via direct anchor binding.
- 40% reduction in context window usage by delivering coherent sections instead of arbitrary fragments.
- Improved faithfulness by enforcing strict $D \approx 0$ processing during ingestion.
12. Conclusion
This paper has presented Entropy-Controlled Information Architecture (ECIA), a unified framework for optimizing information delivery to AI systems. By introducing the Human-Centric Architecture (HCA) vs. Machine-Centric Architecture (MCA) taxonomy and applying the Theory of Stupidity to machine cognition, we demonstrated that environmental noise—not model capability—is the primary driver of LLM failure.
ECIA addresses the HCA-MCA mismatch through complementary implementations:
- AIO enables publishers to provide MCA layers alongside HCA interfaces
- ECR enables consumers to transform HCA content into MCA envelopes
- Both converge on the Content Envelope schema
Our empirical validation demonstrates substantial improvements:
- 100% answer accuracy vs 57% for traditional scraping
- 6x faster retrieval (5ms vs 29ms)
- 27% token efficiency gain per correct answer
- $163M-$8B annual savings at scale
The critical finding is the "confetti effect": scraped content not only wastes tokens but actively loses information. Our benchmark showed 43% of queries failed to find answers in scraped text that AIO preserved completely. This validates the Noise Dominance Theorem—beyond $D_{thresh} ≈ 0.7$, information extraction fails regardless of downstream processing.
As AI systems become primary consumers of digital information, the architecture of that information must evolve. ECIA provides the theoretical foundation and practical tools for this evolution.
The choice facing the industry is clear: continue scaling models against HCA noise, or invest in HCA→MCA transformation. The G-model predicts—and our experiments confirm—that the latter approach yields superior returns.
References
- Petrenko, I. S. (2025). Theory of Stupidity: A Formal Model of Cognitive Vulnerability. Science, Technology and Education, 4(100). DOI: 10.5281/zenodo.18251778.
- Petrenko, I. S. (2026). The General Stupidity Theory. Rideró. ISBN: 978-5-0068-9917-9.
- Lewis, P., et al. (2020). Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. NeurIPS 2020.
- Liu, N. F., et al. (2023). Lost in the Middle: How Language Models Use Long Contexts. arXiv:2307.03172.
- Shi, W., et al. (2023). Large Language Models Can Be Easily Distracted by Irrelevant Context. ICML 2023.
- Gao, L., et al. (2023). RARR: Researching and Revising What Language Models Say. ACL 2023.
- Berners-Lee, T., Hendler, J., & Lassila, O. (2001). The Semantic Web. Scientific American, 284(5), 34-43.
- W3C. (2014). JSON-LD 1.0: A JSON-based Serialization for Linked Data. W3C Recommendation.
- Karpukhin, V., et al. (2020). Dense Passage Retrieval for Open-Domain Question Answering. EMNLP 2020.
- Izacard, G., et al. (2022). Atlas: Few-shot Learning with Retrieval Augmented Language Models. arXiv:2208.03299.
- HTTP Archive. (2024). State of the Web Report. httparchive.org.
- Wu, T. (2016). The Attention Merchants. Knopf.
- Sweller, J. (2011). Cognitive Load Theory. Psychology of Learning and Motivation, 55, 37-76.
Appendix A: Content Envelope JSON Schema
Full JSON Schema available at: https://aio-standard.org/schema/v2.1/
Appendix B: Reference Implementation
Repository: https://github.com/bricsin4u/AIO-research
aio_core/envelope.py- Envelope data structuresaio_core/noise_stripper.py- Content cleaning (Noise Stripping)aio_core/anchor_generator.py- Stable ID generationaio_core/structure_extractor.py- Entity extractionaio_core/binder.py- Structure-narrative bindingprototype/parser/- Client-side ingestion tools (Parsers/SDKs)prototype/ecosystem/- Server-side publishing tools (CMS Plugins)research/benchmarks/- Evaluation suite
Appendix D: RAG Prototype Implementation
The RAG prototype is available in the /rag-prototype directory, showcasing the practical application of ECIA theory:
aio_core/retrieval/router.py: Intent-aware retrieval logic.aio_core/retrieval/intent_classifier.py: Query classification for specialized routing.aio_core/pipeline.py: Unified ingestion pipeline for multi-source data.example_usage.py: End-to-end demonstration of the AIO-RAG workflow.
Correspondence: info@aifusion.ru; presqiuge@pm.me
Repository: https://github.com/bricsin4u/AIO-research