企業情境下 AI 敘事證據失敗的分類法

企業情境下 AI 敘事證據失敗的分類法

Hacker News·

本文提出了一個在企業情境下,針對 AI 生成的公司敘事中,實證觀察到的失敗模式分類法,強調證據分解而非幻覺是主要的治理風險。該分類法著重於審查下的可重構性和可辯護性,而非技術錯誤的分類。

A Taxonomy of AI Narrative Evidence Failure in Enterprise Contexts

Image

Image

Why evidentiary breakdown, not hallucination, is the dominant governance risk

Abstract

This article sets out a taxonomy of empirically observed failure modes in AI-generated corporate narratives, derived from controlled, repeatable testing across multiple large language models. The taxonomy does not rely on anecdotal incidents, post-hoc reconstruction, or hypothetical scenarios. It is organized around evidentiary consequences under scrutiny, not technical error classification.

No entities, outcomes, frequencies, or metrics are disclosed. The purpose of this article is not to attribute liability or predict legal outcomes, but to clarify what types of evidentiary failure already exist and why these failures matter to legal, compliance, and risk functions concerned with defensibility.

Framing: from hallucination to evidentiary failure

Public discussion of AI risk frequently centers on hallucination as a technical defect. In enterprise governance contexts, that framing is insufficient.

Under scrutiny, the decisive question is rarely whether an AI output was inaccurate in isolation. It is whether the enterprise can reconstruct:

Failure to answer those questions constitutes an evidentiary failure, irrespective of model accuracy. Unlike model-centric taxonomies that classify computational error, the taxonomy below is organized around reconstructability and defensibility under review.

Methodological boundary

This taxonomy is informed by internally generated, repeatable AI-governance test artefacts produced under locked protocols, including:

These artefacts were generated prior to any dispute, enforcement inquiry, or external request and were not produced in response to litigation risk. They are not published here. Their existence is noted solely to establish that the failure modes described are empirically observed, not theoretical.

No legal conclusions are drawn from these observations.

Category A: Identity conflation failure

DefinitionThe model conflates distinct legal or commercial entities that share similar names, historical lineage, or sectoral proximity, producing a single blended narrative.

Characteristics

Evidentiary consequenceOnce an identity boundary is crossed, all downstream reasoning becomes contaminated. Post-hoc correction does not restore reconstructability because the narrative path itself cannot be reliably replayed.

Category B: Fabricated documentary attribution

DefinitionThe model references formal documents, filings, or disclosures that do not exist, while presenting them in authoritative, document-first language.

Characteristics

Evidentiary consequenceThese outputs simulate documentary evidence. In governance contexts, simulated records are more destabilizing than obvious error because they resemble admissible artefacts without provenance.

Category C: Temporal drift under identical prompts

DefinitionIdentical prompts produce materially different narratives across time-separated runs, without any intervening change in source data.

Characteristics

Evidentiary consequenceTemporal inconsistency defeats reconstruction. An enterprise cannot evidence what was presented at a moment of reliance if that moment cannot be reliably reproduced.

Category D: Status inflation (inference to assertion)

DefinitionSpeculative or inferred statements are progressively promoted into asserted representations across runs or within a single narrative.

Characteristics

Evidentiary consequenceStatus inflation erodes the distinction between analysis and assertion. In legal and regulatory review, that distinction is foundational.

Category E: Cross-run narrative instability

DefinitionMultiple runs addressing the same question set yield internally coherent but mutually incompatible narratives.

Characteristics

Evidentiary consequenceStability is a prerequisite for defensibility. Where incompatible narratives coexist, selection itself becomes an ungoverned act.

Defense context and constraints

The existence of these evidentiary failure modes does not imply that AI-mediated representations are inherently indefensible, nor that traditional defenses fail. Variability, third-party attribution doctrines, absence of reliance, and jurisdiction-specific standards remain substantial constraints in many contexts.

The relevance of this taxonomy is procedural rather than outcome-determinative. It identifies where evidentiary reconstruction may be contested, not where liability attaches.

What this article does and does not claim

This article claims

This article does not claim

Conclusion: evidence exists before outcomes

Enterprises will not encounter AI risk first as a doctrinal question. They will encounter it as a request:

Where that request cannot be answered, defensibility erodes regardless of outcome. The failure modes described here are already observable in controlled settings. The strategic question is whether they are encountered during planned review or during unplanned scrutiny.

Editorial note

The AIVO Journal has been informed by internally generated AI narrative evidence artefacts produced under locked, repeatable protocols. These artefacts are not published here. Their existence is noted solely to clarify that the taxonomy above reflects empirically observed evidentiary failure modes, not speculative constructs.

Sign up for more like this.

Image

Reconstructability as a Threshold Question in AI-Mediated Representation

Image

When AI Procurement Fails, What Evidence Exists?

Image

AI Reliance Logging

Hacker News

相關文章

  1. 當AI採購失敗時,存在哪些證據?

    3 個月前

  2. 可重構性作為AI中介表徵的門檻問題

    3 個月前

  3. 當AI成為記錄系統:為何證據而非準確性將定義責任歸屬

    4 個月前

  4. 當AI發聲,證據成為控制介面

    3 個月前

  5. 外部AI表述與證據可重構性:一個描述性案例研究

    3 個月前