Reasoning Integrity Standard

RIS v1.0 · Full Specification

Section 0

Foreword

Reasoning Integrity Standard (RIS) v1.0

Published by Atom Labs

© 2025 Atom Labs. All Rights Reserved.


0. Foreword

The Reasoning Integrity Standard (RIS) defines a formal framework for evaluating, governing, and maintaining the integrity of reasoning performed by large language models (LLMs), autonomous agents, and multi-model cognitive systems.

This standard was developed by Atom Labs to address the absence of authoritative, measurable criteria for assessing the stability and trustworthiness of LLM reasoning in production environments. As AI systems increasingly perform high-impact decision-making across enterprise, regulatory, and safety-critical contexts, the industry requires an auditable, repeatable, and model-agnostic method for measuring reasoning reliability.

RIS establishes a multi-level maturity model (RIS-0 through RIS-4) along with normative controls governing chain-of-thought stability, semantic coherence, drift detection, governance, and reasoning boundary enforcement. It specifies the requirements necessary for organizations to assess the integrity of reasoning workflows under real-world operational and adversarial conditions.

This standard is intended for:

  • enterprise architects deploying AI systems
  • safety and governance teams responsible for LLM oversight
  • auditors and compliance personnel evaluating AI risk posture
  • engineering teams building LLM-integrated applications
  • research organizations conducting reproducible benchmarking

RIS v1.0 is model-agnostic and does not mandate the use of any specific framework or vendor. LCAC (Least-Context Access Control) is referenced solely as an example of a compliant reference implementation. RIS remains an independent, neutral specification applicable to any LLM or cognitive system.

Future revisions may expand the standard to include coherence metrics, multi-agent interference scoring, predictive trust modeling, and additional benchmark suites. Atom Labs invites the research, standards, and enterprise community to review, adopt, and contribute to future versions.

Section 1

Scope

1. Scope

1.1 Purpose

The Reasoning Integrity Standard (RIS) defines the requirements, controls, measurement methodologies, and conformance criteria necessary to evaluate the integrity of reasoning performed by LLMs and LLM-powered cognitive systems. The purpose of RIS is not to measure the correctness or factuality of model outputs, but to measure the stability, predictability, and structural reliability of the reasoning process that produces those outputs.

1.2 Applicability

RIS applies to any system that conducts or delegates reasoning to:

  • large language models (LLMs)
  • autonomous or semi-autonomous agents
  • multi-agent architectures
  • chain-of-thought or stepwise reasoning pipelines
  • retrieval-augmented systems (RAG)
  • tool-augmented or API-integrated cognitive systems
  • multi-model or ensemble inference architectures

RIS is applicable regardless of:

  • model vendor or provider
  • model size, training method, or licensing
  • deployment environment (cloud, hybrid, on-premise, edge)
  • inference scenario (single-turn, multi-turn, stateful, agentic)

1.3 Intended Use

RIS is intended to support:

  • evaluation of LLMs and agentic systems before deployment
  • continuous monitoring of reasoning behavior in production environments
  • regulatory and compliance audits involving AI reasoning stability
  • vendor and model procurement due diligence
  • internal trust, governance, and safety programs
  • reproducible research and benchmarking

1.4 Out of Scope

The following areas are outside the scope of RIS v1.0:

  • data privacy practices
  • model training and dataset governance
  • ethics, fairness, bias, or demographic analysis
  • factual correctness of responses
  • model alignment and value judgments unrelated to reasoning integrity

These areas may be addressed by other standards or future RIS extensions.

1.5 Normative Language

RIS uses the following requirement terminology:

  • MUST: an absolute, mandatory requirement for conformance
  • SHALL: a criterion required to meet RIS compliance
  • SHOULD: a recommended practice unless a justified exception exists
  • MAY: an optional practice
  • NOT PERMITTED: a prohibited behavior or configuration

1.6 Relationship to LCAC

LCAC is a reference implementation illustrating one possible method of achieving RIS compliance. RIS does not require LCAC, nor does conformance imply adoption of any specific product, tool, or architecture. RIS is intended to remain implementation-agnostic.

1.7 Versioning and Revision Policy

RIS v1.0 is the initial published version of the Reasoning Integrity Standard. Atom Labs maintains stewardship of the standard until the formation or designation of a formal multi-party standards body.

Future revisions may include:

  • expanded risk models
  • domain-specific profiles
  • standardized test suites
  • trust-flow prediction methodologies
  • multi-agent interference metrics

Revisions will follow a versioned specification model (v1.1, v1.2, v2.0, etc.).


Section 2

Normative References

2. Normative References

2.1 Purpose of Normative References

This section identifies documents, standards, and terminology references that are considered authoritative for understanding and implementing the Reasoning Integrity Standard (RIS). Where conflicts arise, this specification takes precedence unless a referenced standard is explicitly designated as controlling for a specific requirement.

Normative references are essential for interpreting RIS controls, understanding terminology, and establishing consistent evaluation and audit methodologies across implementations.

2.2 Standards and Frameworks Referenced by RIS

The following documents and standards are integral to interpreting RIS v1.0. Implementers are expected to be familiar with these materials, as RIS incorporates concepts, terminology, and structural patterns from them.

NIST Standards

  • NIST Special Publication 800-53: Security and Privacy Controls for Information Systems and Organizations
  • NIST Artificial Intelligence Risk Management Framework (AI RMF)
  • NIST SP 1270: Trustworthy and Responsible AI

ISO Standards

  • ISO/IEC 27001: Information Security Management Systems
  • ISO/IEC 23894: Artificial Intelligence Risk Management
  • ISO/IEC 29100: Privacy Framework
  • ISO/IEC 25010: System and Software Quality Models

SOC 2 and AICPA Guidance

  • AICPA Trust Services Criteria (Security, Availability, Processing Integrity, Confidentiality, Privacy)
  • AICPA AT-C 205: Examination Engagements

OWASP and Application Security References

  • OWASP Application Security Verification Standard (ASVS)
  • OWASP Top 10 (referenced in governance and boundary controls)

AI Safety, Governance, and Model Evaluation References

  • EU AI Act (Draft and Final Texts)
  • MLPerf and related benchmark publications
  • Model evaluation methodologies from major AI labs
  • Academic publications on chain-of-thought, inference stability, and cognitive drift

2.3 Terminology References

RIS uses terminology derived from established research and industry standards. Implementers should reference the following categories for precise interpretation.

AI Systems Terminology

  • large language model (LLM)
  • agent and multi-agent system
  • chain-of-thought (CoT)
  • semantic coherence
  • contextual drift
  • inference-time variability
  • deterministic vs. non-deterministic reasoning

Governance Terminology

  • risk posture
  • control requirement
  • conformance
  • audit evidence
  • variance threshold
  • stability metric

LCAC Terminology (Reference Implementation Only)

LCAC terminology is informative, not normative. Terms include:

  • least-context access control
  • governance mode
  • reasoning boundary
  • drift envelope
  • trust baseline
  • variance baseline
  • reasoning ledger

Implementers are not required to adopt LCAC’s terminology or architecture, but LCAC terms appear in examples and appendices as part of the reference implementation.

2.4 Definitions Governed by RIS

RIS mandates the following definitions for consistent use across all conformant implementations.

Reasoning Integrity

The measurable ability of a system to produce stable, predictable reasoning outputs that maintain structural coherence across similar prompts, repeated prompts, operational conditions, and controlled variations.

Reasoning Drift

A deviation in reasoning behavior, structure, or output that exceeds defined variance thresholds when evaluated under repeatable or controlled conditions.

Chain Stability

The degree to which the internal reasoning structure of a system maintains consistent logical steps, semantic transitions, and contextual boundaries across repeated inference cycles.

Semantic Consistency

The alignment of meaning, intent, and conceptual structure across multiple reasoning instances or reasoning steps for equivalent prompts.

Governance Boundary

A defined constraint controlling the informational, contextual, or semantic state accessible to a reasoning system at inference time.

2.5 Control of Normative References

Where RIS references external standards:

  • External standards MUST be treated as supporting material.
  • External standards SHALL NOT override RIS requirements unless explicitly designated as authoritative.
  • RIS SHALL define the governing definitions for reasoning integrity and drift.
  • External terminology MAY be used for interpretation but not for replacing RIS controls.

2.6 Informative vs. Normative Material

RIS distinguishes between:

  • Normative material (mandatory for conformance)
  • Informative material (supporting guidance)

Normative sections include:

  • Controls
  • Requirements
  • Measurement Framework
  • Scoring and Conformance
  • Audit Guidelines

Informative sections include:

  • Reference Implementation
  • Mapping to Other Standards
  • Appendices

This distinction is preserved throughout the remainder of the specification.


Section 3

Fundamental Concepts

3. Fundamental Concepts of Reasoning Integrity

3.1 Overview

This section defines the foundational concepts required to understand and evaluate reasoning integrity in large language models (LLMs), autonomous agents, and multi-model cognitive systems. These concepts form the basis of all controls, requirements, and conformance criteria in RIS. Implementers MUST understand and consistently apply these definitions across all stages of evaluation, monitoring, and governance.

Reasoning integrity is not a measure of factual correctness or ethical compliance. It is a measure of the internal stability, predictability, and structural reliability of the reasoning process itself.

3.2 Reasoning Integrity

Reasoning integrity is the degree to which a reasoning system produces stable, coherent, and predictable internal reasoning structures when presented with equivalent or contextually similar prompts, conditions, or operational states.

A system with high reasoning integrity maintains:

  • consistent logical structure
  • predictable semantic relationships
  • stable reasoning paths across repetitions
  • adherence to defined contextual boundaries
  • low drift under controlled variations

A system with low reasoning integrity exhibits:

  • unpredictable reasoning paths
  • inconsistent semantic structures
  • sensitivity to minor input or environmental variations
  • uncontrolled drift over repeated inferences
  • unstable or contradictory reasoning forms

3.3 Reasoning Chain

A reasoning chain is the internal, stepwise sequence of conceptual transformations used by the system to derive an output from an input. The chain may be implicit (black-box LLMs) or explicit (systems exposing structured reasoning steps), but RIS treats both as conceptual equivalents.

A reasoning chain includes:

  • latent representations
  • intermediate conceptual structures
  • logical transitions
  • semantic dependencies
  • step-to-step relationships

RIS does not require access to internal model states. Reasoning chains may be reconstructed using output structure, behavioral patterns, and statistical inference.

3.4 Chain Stability

Chain stability is the consistency of reasoning chain structure across repeated or controlled inference cycles.

A system demonstrates chain stability when:

  • repeated prompts generate reasoning structures with low variance
  • stepwise transitions remain logically and semantically similar
  • variations fall within defined acceptable thresholds

Chain stability is a core metric in RIS and is used to determine system conformance at all RIS levels.

3.5 Semantic Coherence

Semantic coherence is the degree to which reasoning outputs maintain consistent meaning, conceptual alignment, and contextual relevance across equivalent inputs or over the course of a reasoning chain.

A system with high semantic coherence ensures that:

  • meaning remains stable across reasoning steps
  • internal transitions preserve conceptual alignment
  • responses maintain adherence to prompt intent
  • conceptual drift remains within acceptable thresholds

Loss of semantic coherence is a primary indicator of reasoning instability.

3.6 Cognitive Drift

Cognitive drift is the measurable deviation in a system’s reasoning process when evaluated under repeated, controlled, or minimally varied conditions.

Forms of drift include:

  • structural drift: changes in reasoning chain architecture
  • semantic drift: changes in meaning, context, or conceptual focus
  • temporal drift: progressive degradation of reasoning consistency over time
  • interference drift: changes caused by tools, memory, or agent interactions

Cognitive drift is one of the primary risk factors assessed by RIS.

3.7 Variance Envelope

The variance envelope defines the acceptable bounds of variation for reasoning behavior across repeated inference cycles. These bounds are derived from:

  • statistical analysis
  • expected variability based on model architecture
  • domain-specific tolerance criteria
  • operational safety requirements

A system MUST maintain its reasoning behavior within its defined variance envelope to achieve RIS conformance.

3.8 Governance Boundary

A governance boundary is a defined limit controlling what contextual, semantic, or informational state the system is permitted to use during a reasoning process.

Boundaries include:

  • allowed context windows
  • prohibited contextual expansions
  • restricted semantic domains
  • tool- and memory-access constraints
  • cross-agent interaction limits

Governance boundaries prevent uncontrolled or unauthorized expansion of reasoning state.

3.9 Deterministic and Non-Deterministic Reasoning

Most LLM systems exhibit non-deterministic reasoning due to sampling temperature, randomness, and stochastic inference pathways. RIS does not require deterministic outputs; it requires:

  • deterministic reasoning patterns within allowable variance
  • predictable behavior under controlled conditions
  • stable reasoning structure despite inherent stochasticity

Systems MUST demonstrate stable structural patterns even if surface-level outputs vary.

3.10 Tool and Agent Interference

Tool or agent interference occurs when the reasoning system’s internal process is altered by:

  • external tool responses
  • API calls
  • retrieval results
  • other agents in a multi-agent environment

Interference may cause:

  • unexpected semantic transitions
  • contextual overrides
  • premature reasoning collapse
  • unbounded expansion of reasoning state

RIS controls evaluate the system’s ability to maintain stability despite such interference.

3.11 State Persistence and Contextual Memory

State persistence defines how long prior reasoning states influence future reasoning cycles. Excessive or uncontrolled persistence may create:

  • reasoning contamination
  • indirect hallucination
  • unexpected cross-task dependency
  • failure to reinitialize reasoning state

RIS requires systems to maintain predictable and bounded state persistence behaviors.

3.12 Integrity Failure Modes

Common reasoning integrity failure modes include:

  • spontaneous reasoning divergence
  • oscillatory reasoning loops
  • semantic boundary collapse
  • unintended contextual expansion
  • instability under minimal input variations
  • conflicting reasoning steps
  • multi-agent interference anomalies
  • degradation over time or workload

These failure modes are formalized later in the RIS risk models and evaluation methodology.


Section 4

RIS Levels

4. RIS Levels

4.1 Overview

The Reasoning Integrity Standard (RIS) defines five levels of reasoning integrity maturity, ranging from RIS-0 (uncontrolled reasoning) to RIS-4 (high-integrity, production-grade reasoning). These levels provide a structured framework for evaluating the stability, predictability, and reliability of reasoning produced by large language models (LLMs) and agentic systems.

RIS levels classify systems based on measurable reasoning behavior, not model size, architecture, vendor, or training methodology. A system's RIS level reflects its demonstrated ability to maintain reasoning integrity under controlled conditions, repeated evaluations, and operational scenarios.

4.2 Level Definitions

RIS defines the following reasoning integrity maturity levels:

  • RIS-0: Uncontrolled Reasoning
  • RIS-1: Drift-Sensitive Reasoning
  • RIS-2: Semi-Stable Reasoning
  • RIS-3: Controlled Reasoning
  • RIS-4: High-Integrity Reasoning

Each level includes required properties, performance characteristics, and failure thresholds.

4.3 RIS-0: Uncontrolled Reasoning

Description

RIS-0 systems exhibit unpredictable, highly variable, or unstable reasoning behavior. Reasoning chains diverge significantly across repeated prompts, and semantic coherence is inconsistent or unreliable.

Characteristics

  • High variance across repeated inferences
  • Unstable or incoherent reasoning chains
  • Susceptibility to minor input or environmental perturbations
  • Unbounded or uncontrolled reasoning state transitions
  • No enforceable variance envelope
  • Reasoning behavior cannot be reliably reproduced

Typical Use Cases

  • exploratory research models
  • early-stage prototypes
  • unconstrained generative systems

Conformance

RIS-0 indicates non-conformance with all RIS requirements.

4.4 RIS-1: Drift-Sensitive Reasoning

Description

RIS-1 systems demonstrate partial reasoning stability but remain sensitive to drift, perturbations, and repeated evaluations. Structural or semantic divergence occurs frequently.

Characteristics

  • partially stable reasoning patterns under ideal conditions
  • significant drift under repetition or minimal variation
  • inconsistent semantic alignment
  • no enforceable drift boundaries
  • limited reproducibility of reasoning behavior

Typical Use Cases

  • non-safety-critical applications
  • general-purpose conversational models
  • agent prototypes

Conformance

RIS-1 indicates minimal partial alignment with RIS controls but does not meet baseline conformance.

4.5 RIS-2: Semi-Stable Reasoning

Description

RIS-2 systems achieve basic reasoning integrity. Reasoning chains show moderate consistency across repeated prompts and operate within a loosely defined variance range. Drift is detectable but bounded.

Characteristics

  • moderate chain stability
  • acceptable semantic coherence under repetition
  • drift remains within a broad variance envelope
  • occasional reasoning divergence under perturbation
  • reproducible reasoning behavior in controlled cases

Typical Use Cases

  • consumer-facing AI applications
  • low-risk enterprise workflows
  • model evaluation and benchmarking environments

Conformance

RIS-2 meets some baseline RIS controls but does not satisfy the requirements for production-grade reasoning integrity.

4.6 RIS-3: Controlled Reasoning

Description

RIS-3 systems demonstrate controlled reasoning behavior suitable for production environments. Reasoning chains are stable across repetitions, drift remains within defined boundaries, and semantic coherence is predictably maintained.

Characteristics

  • consistent chain stability across repeated prompts
  • defined and enforceable variance envelope
  • predictable reasoning under controlled perturbations
  • bounded semantic drift and minimal divergence
  • reproducible reasoning state transitions
  • resistance to tool and agent interference within expected parameters

Typical Use Cases

  • enterprise-grade LLM applications
  • internal decision-support systems
  • workflow automation
  • multi-agent orchestration with defined boundaries

Conformance

RIS-3 satisfies the majority of RIS requirements and is considered the minimum level appropriate for regulated or production-grade environments.

4.7 RIS-4: High-Integrity Reasoning

Description

RIS-4 represents the highest level of reasoning integrity. Systems at this level produce stable, predictable, and coherent reasoning patterns with minimal variance across repeated evaluations. Drift is tightly controlled, and reasoning chains maintain structural integrity under both controlled and adversarial conditions.

Characteristics

  • high stability and coherence across all evaluations
  • tightly bounded variance envelope
  • minimal drift under repetition or perturbation
  • predictable reasoning structure even with minor input variations
  • resilience to interference from tools, retrieval components, or agents
  • strong governance boundary adherence
  • reproducible reasoning consistent with defined expectations

Typical Use Cases

  • safety-critical systems
  • regulated industries requiring auditability
  • financial, medical, legal, and security reasoning workflows
  • autonomous agent frameworks with strict stability demands
  • enterprise systems requiring deterministic reasoning behavior

Conformance

RIS-4 conforms to all RIS requirements and demonstrates the highest degree of reasoning integrity defined by the standard.

4.8 Level Assignment Requirements

A system's RIS level MUST be assigned based on:

  • performance across the RIS measurement framework
  • compliance with required RIS controls
  • observed behavior during evaluation methodology test suites
  • documented evidence demonstrating stability and drift characteristics

Level assignment SHALL be determined from empirical results, not vendor claims or model size.

4.9 Level Validation and Reassessment

Systems evaluated for RIS conformance MUST undergo periodic reassessment based on:

  • operational changes
  • model updates or retraining
  • architectural changes
  • new tool or agent integrations
  • changes in governance boundaries
  • drift trends observed in production

Reassessment MAY elevate or reduce a system's RIS level depending on observed results.


Section 5

Measurement Framework

5. Measurement Framework

5.1 Overview

The RIS Measurement Framework defines the quantitative and qualitative methods used to evaluate reasoning integrity. It provides the metrics, data collection procedures, statistical thresholds, and evaluation structures necessary to determine a system's RIS level.

The framework applies to all systems subject to RIS evaluation and SHALL be used consistently across:

  • baseline assessments
  • production monitoring
  • audit reviews
  • comparative benchmarking

Measurement MUST be based on empirical evidence. Qualitative assessments may supplement measurements but SHALL NOT replace them.

5.2 Measurement Categories

RIS evaluates reasoning integrity using five primary measurement categories:

  1. Chain Stability
  2. Semantic Coherence
  3. Drift Sensitivity
  4. Variance Envelope Compliance
  5. Governance Boundary Adherence

All categories MUST be assessed to determine RIS level.

5.3 Required Inputs

The following inputs are required for RIS-compliant measurement:

  • repeated inference samples from identical prompts
  • controlled perturbation samples (minimal input variations)
  • environmental invariants (temperature, sampling parameters)
  • contextual invariants (system instructions, agent state)
  • tool or retrieval call responses (if applicable)
  • logs or ledger entries representing reasoning behavior

Implementations SHOULD collect evaluation data under isolated and consistent conditions.

5.4 Sample Size Requirements

A compliant RIS evaluation MUST include:

  • a minimum of 25 repeated inference samples per test prompt
  • a minimum of 10 controlled perturbation samples per prompt
  • at least 50 total prompts across diverse categories for generalized models
  • at least 20 prompts for domain-specific or task-specific systems

Larger sample sizes MAY be used but SHALL be documented.

5.5 Chain Stability Metric

5.5.1 Definition

Chain stability measures the consistency of a system’s reasoning structure across repeated inference cycles.

5.5.2 Measurement Method

Stability is measured using:

  • structural similarity metrics
  • step-to-step coherence comparisons
  • reasoning-length variance
  • transition-alignment scoring

A system MUST maintain chain similarity within its defined variance envelope.

5.5.3 Scoring

Chain stability is scored from 0.00 to 1.00.

Example thresholds:

  • 0.85-1.00: stable
  • 0.65-0.84: semi-stable
  • below 0.65: unstable

RIS-4 systems SHALL achieve stability above 0.90 for all core prompts.

5.6 Semantic Coherence Metric

5.6.1 Definition

Semantic coherence measures the consistency of meaning and conceptual alignment across repeated inferences.

5.6.2 Measurement Method

Coherence is evaluated based on:

  • semantic similarity scoring
  • topic retention
  • conceptual transition consistency
  • negation or contradiction detection

5.6.3 Scoring

Scoring follows the same 0.00-1.00 scale.

RIS-4 systems SHALL demonstrate coherence above 0.90.

5.7 Drift Sensitivity Metric

5.7.1 Definition

Drift sensitivity measures the degree to which a system deviates from baseline reasoning behavior under repeated evaluations or minimal perturbations.

5.7.2 Measurement Method

Drift is assessed by:

  • analyzing variance outside baseline patterns
  • measuring semantic or structural deviation trends
  • quantifying divergence under perturbation conditions

5.7.3 Scoring

Drift sensitivity is scored inversely (lower is better):

  • 0.00-0.10: low drift
  • 0.11-0.25: moderate drift
  • above 0.25: high drift

RIS-4 systems SHALL maintain drift sensitivity below 0.10.

5.8 Variance Envelope Compliance

5.8.1 Definition

The variance envelope represents the acceptable range of variation in reasoning behavior.

5.8.2 Measurement Method

Variance envelope compliance is determined by:

  • repeated inference testing
  • perturbation evaluations
  • longitudinal measurement
  • stability under load or repeated requests

5.8.3 Scoring

A system is considered compliant if:

  • at least 95 percent of evaluations fall within its defined envelope

RIS-4 systems SHALL achieve at least 98 percent compliance.

5.9 Governance Boundary Adherence

5.9.1 Definition

Governance boundaries constrain the context, semantic scope, and operational state the model may use for reasoning.

5.9.2 Measurement Method

Boundary adherence is measured by observing:

  • context window expansion attempts
  • unauthorized semantic domain expansion
  • tool or agent influence patterns
  • cross-task contamination or memory leakage

5.9.3 Scoring

Boundary violations are scored as:

  • 0: no violations
  • 1: one minor violation
  • 2: repeated or significant violations

RIS-4 systems SHALL record zero violations under all evaluation conditions.

5.10 Composite Reasoning Integrity Score

RIS defines an aggregate score calculated using weighted metrics:

  • Chain Stability: 30 percent
  • Semantic Coherence: 25 percent
  • Drift Sensitivity: 20 percent
  • Variance Envelope Compliance: 15 percent
  • Governance Boundary Adherence: 10 percent

Composite scores determine candidate RIS levels but do not replace conformance requirements.

5.11 Interpretation of Scores

Composite score ranges:

  • 0.00-0.40: RIS-0
  • 0.41-0.60: RIS-1
  • 0.61-0.75: RIS-2
  • 0.76-0.89: RIS-3
  • 0.90-1.00: RIS-4

A system MUST meet both:

  • composite scoring thresholds
  • all mandatory controls applicable to its level

to be classified at that RIS level.


Section 6

Control Families

6. RIS Controls

6.1 Overview

RIS controls define the mandatory and recommended requirements that systems MUST meet to achieve compliance with RIS levels RIS-1 through RIS-4. Controls are grouped into control families that govern stability, coherence, drift, variance, boundary adherence, and operational integrity.

Each control includes:

  • Control Objective
  • Control Requirement (MUST / SHALL)
  • Implementation Guidance (informative)
  • Evidence Required (for audits)
  • Conformance Criteria

All systems seeking RIS classification SHALL be evaluated against these controls.


6.2 Control Families

RIS defines six control families:

  1. RS - Chain Stability Controls
  2. SC - Semantic Coherence Controls
  3. DR - Drift Resistance Controls
  4. VE - Variance Envelope Controls
  5. GB - Governance Boundary Controls
  6. OP - Operational Integrity Controls

Controls within each family are numbered sequentially.


6.3 Chain Stability Controls (RS)

RS-1: Repetition Consistency

Objective

Ensure reasoning chains remain consistent across repeated inference cycles.

Requirement

The system SHALL produce reasoning outputs whose structural similarity remains within the defined variance envelope across repeated evaluations of identical prompts.

Implementation Guidance

  • Use repeated inference tests.
  • Monitor reasoning structure length and transitions.
  • Track deviations across a defined sample size.

Evidence Required

  • stability reports
  • similarity metrics
  • repeated inference samples

Conformance

Required for RIS-2, RIS-3, and RIS-4.


RS-2: Perturbation Stability

Objective

Ensure stability under minor, controlled input variations.

Requirement

The system SHALL maintain reasoning structure stability when presented with minimal, semantically equivalent perturbations of the same prompt.

Implementation Guidance

  • Use controlled perturbation prompts (synonym replacement, minor rephrasing, neutral variations).
  • Measure structural drift across perturbation samples.

Evidence Required

  • perturbation test logs
  • drift measurements
  • variance envelope comparison

Conformance

Required for RIS-3 and RIS-4.


RS-3: Transition Alignment

Objective

Ensure internal reasoning transitions follow consistent logical patterns.

Requirement

The system SHALL maintain alignment of step-to-step transitions across repeated inference cycles.

Implementation Guidance

  • Compare reasoning step ordering or inferred conceptual transitions.
  • Identify anomalous reordering indicative of instability.

Evidence Required

  • transition-level similarity reports

Conformance

Required for RIS-4.


6.4 Semantic Coherence Controls (SC)

SC-1: Semantic Alignment

Objective

Ensure that reasoning outputs remain semantically consistent across repeated evaluations.

Requirement

The system SHALL maintain semantic similarity across repeated inferences meeting or exceeding the thresholds defined in the measurement framework.

Implementation Guidance

  • Use embedding-based similarity scoring.
  • Detect contradictory or negated reasoning shifts.

Evidence Required

  • semantic similarity analysis
  • contradiction or negation detection logs

Conformance

Required for RIS-2, RIS-3, and RIS-4.


SC-2: Concept Retention

Objective

Ensure conceptual themes persist across reasoning processes.

Requirement

The system SHALL retain core conceptual structures across evaluations unless explicitly instructed to change.

Implementation Guidance

  • Analyze topic retention through semantic clustering.
  • Evaluate for unexpected topic divergence.

Evidence Required

  • topic coherence reports
  • clustering analysis results

Conformance

Required for RIS-3 and RIS-4.


SC-3: Logical Consistency

Objective

Ensure logical validity and structure remain consistent.

Requirement

The system SHALL avoid contradictory intermediate reasoning steps within an evaluation set.

Implementation Guidance

  • Evaluate internal reasoning structure for contradictions.
  • Use logical consistency scoring where available.

Evidence Required

  • logical consistency assessments

Conformance

Required for RIS-4.


6.5 Drift Resistance Controls (DR)

DR-1: Drift Monitoring

Objective

Establish continuous monitoring for reasoning drift.

Requirement

The system SHALL implement monitoring to detect drift exceeding defined thresholds over repeated inference cycles.

Implementation Guidance

  • Compare baseline outputs against current samples.
  • Establish early-warning indicators for drift escalation.

Evidence Required

  • drift baseline logs
  • drift trend reports

Conformance

Required for RIS-2, RIS-3, and RIS-4.


DR-2: Drift Envelope Enforcement

Objective

Maintain reasoning behavior within allowable drift boundaries.

Requirement

The system SHALL enforce a drift envelope that constrains allowable deviation from baseline behavior.

Implementation Guidance

  • Use variance and drift sensitivity metrics.
  • Define acceptable drift bands based on use case.

Evidence Required

  • drift envelope documentation
  • variance compliance reports

Conformance

Required for RIS-3 and RIS-4.


DR-3: Drift Recovery

Objective

Ensure recovery mechanisms exist for restoring reasoning integrity.

Requirement

The system SHALL provide mechanisms to reinitialize or reset reasoning state when drift exceeds permitted thresholds.

Implementation Guidance

  • Reset context or agent state.
  • Refresh memory or tool caches.
  • Re-evaluate under controlled conditions.

Evidence Required

  • drift recovery logs
  • anomaly detection events

Conformance

Required for RIS-4.


6.6 Variance Envelope Controls (VE)

VE-1: Variance Definition

Objective

Define acceptable variance for reasoning behavior.

Requirement

The system SHALL define a variance envelope specifying acceptable deviation thresholds across repeated evaluations.

Implementation Guidance

  • Document variance thresholds.
  • Derive thresholds from benchmarking or domain requirements.

Evidence Required

  • variance envelope documentation

Conformance

Required for RIS-2, RIS-3, and RIS-4.


VE-2: Variance Compliance

Objective

Ensure reasoning outputs remain within defined variance bounds.

Requirement

The system SHALL maintain variance compliance for at least 95 percent of repeated evaluations.

Implementation Guidance

  • Use repeated inference tests.
  • Monitor patterns of deviation.

Evidence Required

  • compliance reports
  • exception logs

Conformance

Required for RIS-3 and RIS-4.


VE-3: Variance Tightening

Objective

Tighten variance thresholds for high-integrity reasoning.

Requirement

The system SHALL maintain variance compliance at rates exceeding 98 percent for RIS-4.

Implementation Guidance

  • Use stricter similarity or coherence scoring.
  • Implement adaptive variance tightening.

Evidence Required

  • high-precision variance reports

Conformance

Required for RIS-4.


6.7 Governance Boundary Controls (GB)

GB-1: Boundary Definition

Objective

Define explicit reasoning boundaries to prevent uncontrolled contextual expansion.

Requirement

The system SHALL define governance boundaries that constrain contextual, semantic, or operational state during reasoning.

Implementation Guidance

  • Define allowable context limits.
  • Identify prohibited expansions or semantic domains.

Evidence Required

  • governance boundary documentation

Conformance

Required for RIS-2, RIS-3, and RIS-4.


GB-2: Boundary Enforcement

Objective

Ensure reasoning remains within defined boundaries.

Requirement

The system SHALL enforce boundaries during inference and detect violations.

Implementation Guidance

  • Use contextual or semantic boundary checks.
  • Restrict access to tools or memory beyond defined limits.

Evidence Required

  • boundary enforcement logs
  • violation reports

Conformance

Required for RIS-3 and RIS-4.


GB-3: Interference Protection

Objective

Prevent uncontrolled influence from tools, agents, or retrieval components.

Requirement

The system SHALL detect and mitigate reasoning interference from tool responses or agent interactions.

Implementation Guidance

  • Validate tool outputs before integration.
  • Isolate reasoning states across agents.

Evidence Required

  • interference detection logs
  • impact assessments

Conformance

Required for RIS-4.


6.8 Operational Integrity Controls (OP)

OP-1: Evaluation Environment Consistency

Objective

Ensure evaluation environments remain stable and controlled.

Requirement

The system SHALL maintain consistent parameters (temperature, sampling settings, context) during evaluation.

Evidence Required

  • environment configuration logs

Conformance

Required for all RIS levels beyond RIS-0.


OP-2: Reproducible Testing

Objective

Ensure evaluation results can be reproduced.

Requirement

The system SHALL support reproducible testing conditions and retain configuration details for audit purposes.

Evidence Required

  • reproducibility reports
  • configuration snapshots

Conformance

Required for RIS-3 and RIS-4.


OP-3: Audit Logging

Objective

Ensure auditable logs for evaluation and monitoring.

Requirement

The system SHALL generate logs documenting reasoning behavior, drift events, variance compliance, and boundary violations.

Evidence Required

  • audit logs
  • ledger entries
  • summary reports

Conformance

Required for RIS-4.


OP-4: Longitudinal Monitoring

Objective

Track reasoning behavior over time.

Requirement

The system SHALL support longitudinal monitoring to detect degradation or emerging instability.

Evidence Required

  • multi-interval reports
  • trend analyses

Conformance

Required for RIS-4.


6.9 Summary of Conformance Requirements by RIS Level

Control FamilyRIS-1RIS-2RIS-3RIS-4
RS (Stability)OptionalRequiredRequiredRequired (strict)
SC (Coherence)OptionalRequiredRequiredRequired (strict)
DR (Drift)OptionalRequiredRequiredRequired (strict)
VE (Variance)OptionalRequiredRequiredRequired (tight)
GB (Boundary)OptionalRequiredRequiredRequired (strict)
OP (Operational)MinimalModerateStrongFull (audit-ready)

6.10 Control Interpretation

Control interpretation SHALL comply with the following:

  • MUST and SHALL controls are required for conformance.
  • SHOULD controls are recommended but not required unless explicitly tied to a level.
  • MAY controls provide flexibility but do not affect conformance.

Section 7

Scoring and Conformance

7. Scoring and Conformance

7.1 Overview

This section defines the scoring methodology and conformance criteria used to determine a system's RIS level. Scoring integrates quantitative metrics from the measurement framework with mandatory control requirements. A system's RIS classification is based on both:

  1. composite reasoning integrity score, and
  2. demonstrated compliance with required RIS controls.

A system SHALL NOT be assigned a RIS level solely based on scoring metrics without meeting corresponding control requirements.


7.2 Composite Reasoning Integrity Score

The composite score is calculated using weighted metrics defined in Section 5:

  • Chain Stability: 30 percent
  • Semantic Coherence: 25 percent
  • Drift Sensitivity: 20 percent
  • Variance Envelope Compliance: 15 percent
  • Governance Boundary Adherence: 10 percent

Each metric yields a value between 0.00 and 1.00.

The weighted sum produces the final composite reasoning integrity score.

The composite score is used to determine eligibility for a RIS level, but SHALL NOT be used to determine final classification without control verification.


7.3 Metric Weighting

The rationale for metric weighting is as follows:

  • Chain Stability (30 percent): foundational indicator of structural reasoning reliability.
  • Semantic Coherence (25 percent): ensures reasoning maintains consistent meaning.
  • Drift Sensitivity (20 percent): predicts long-term stability under repetition.
  • Variance Envelope Compliance (15 percent): ensures bounded divergence.
  • Governance Boundary Adherence (10 percent): required for operational trust and safety.

These weights SHALL remain consistent across RIS versions unless updated in a future amendment.


7.4 Eligibility Thresholds for RIS Levels

Eligibility thresholds based on composite score:

  • 0.00-0.40: RIS-0
  • 0.41-0.60: RIS-1
  • 0.61-0.75: RIS-2
  • 0.76-0.89: RIS-3
  • 0.90-1.00: RIS-4

Meeting the scoring threshold does not guarantee assignment to that RIS level.

The system SHALL also satisfy the control requirements designated for that level.


7.5 Control-Based Conformance Requirements

The following table defines minimum control conformance for each RIS level:

RIS LevelRequired Controls
RIS-0No controls required
RIS-1Optional controls only
RIS-2All RS-1, SC-1, DR-1, VE-1, GB-1 controls
RIS-3All RIS-2 controls plus RS-2, SC-2, DR-2, VE-2, GB-2, OP-2
RIS-4All RIS-3 controls plus RS-3, SC-3, DR-3, VE-3, GB-3, OP-3, OP-4

A system SHALL NOT be classified above RIS-2 unless all mandatory controls for RIS-3 are satisfied.

A system SHALL NOT be classified above RIS-3 unless all mandatory controls for RIS-4 are satisfied.


7.6 Evidence Requirements

To support a RIS classification, the following evidence MUST be produced:

  1. Measurement Framework Results

- repeated inference sample outputs

- controlled perturbation outputs

- variance compliance reports

- drift sensitivity analyses

  1. Control Implementation Evidence

- documentation of stability, coherence, drift, variance, and boundary controls

- operational integrity documentation

- governance boundary definitions

  1. Audit Logs

- reasoning ledger entries or equivalent reports

- evaluation logs

- anomaly and boundary violation logs

  1. Configuration Snapshots

- inference parameters

- environmental settings

- prompt templates

- tool access or agent configuration

The evaluator SHALL verify all evidence before assigning a RIS level.


7.7 Conformance Determination Process

A system SHALL undergo the following steps to determine conformance:

  1. Preparation

- define evaluation conditions

- document configuration

- establish invariants

  1. Data Collection

- perform repeated inference sampling

- perform controlled perturbation sampling

- log reasoning behavior

  1. Metric Calculation

- compute each metric independently

- calculate composite reasoning integrity score

  1. Control Verification

- confirm implementation and operation of required controls

- identify gaps or violations

  1. Level Assignment

- determine eligible RIS level based on composite score

- determine achievable RIS level based on control compliance

- final level SHALL be the lower of the two

  1. Report Generation

- produce a formal evaluation report

- document findings, score, control status, and assignment

  1. Certification

- issue a RIS Conformance Statement

- assign classification validity period


7.8 Handling of Violations and Exceptions

If a system violates a mandatory control during evaluation:

  • it SHALL be classified at the highest RIS level for which all mandatory controls are satisfied
  • violations MUST be documented with supporting evidence
  • systems MAY undergo remediation and reassessment

Exceptions for SHOULD and MAY controls SHALL NOT affect level assignment if mandatory controls are met.


7.9 Reassessment Requirements

RIS conformance SHALL be reassessed when:

  • the underlying model is updated
  • training data or fine-tuning changes
  • new tools or agents are added
  • governance boundaries change
  • drift or instability trends are detected in production
  • a significant degradation in metrics occurs
  • system behavior changes under load

Reassessment MAY be performed more frequently based on organizational or regulatory requirements.


7.10 Validity of RIS Classification

RIS classification is valid for a period of:

  • 12 months for general LLM systems
  • 6 months for agentic or tool-integrated systems
  • 3 months for safety-critical systems

After expiration, the system SHALL be reevaluated to maintain a RIS classification.


7.11 Publication and Reporting

Organizations MAY publish:

  • their RIS level
  • their composite score
  • high-level findings

However, detailed logs, inference samples, and proprietary evaluation data SHOULD only be disclosed under NDA or regulatory requirement.

RIS conformance SHALL NOT be claimed without formal assessment.


Section 8

Risk Models

8. Risk Models

8.1 Overview

This section defines the risk models used to evaluate, categorize, and mitigate threats to reasoning integrity in LLMs and agentic systems. RIS identifies reasoning risks that impact stability, coherence, predictability, and boundary adherence. Risk models are used to:

  • guide evaluation methodology
  • inform control requirements
  • support audit and governance decisions
  • predict failure conditions
  • align reasoning-system behavior with organizational risk tolerance

RIS risk models are independent of data privacy, security, ethical, or regulatory risk. They focus exclusively on reasoning integrity.


8.2 Categories of Reasoning Risk

RIS defines six primary categories of reasoning risk:

  1. Structural Drift Risk
  2. Semantic Drift Risk
  3. Temporal Stability Risk
  4. Interference Risk
  5. Boundary Violation Risk
  6. Degradation Risk

Each risk category is described in the following sections.


8.3 Structural Drift Risk

8.3.1 Description

Structural drift occurs when the underlying structure, sequence, or transitions of reasoning chains changes beyond acceptable variance thresholds.

8.3.2 Causes

  • sampling randomness beyond expected levels
  • sensitivity to minor prompt variations
  • instability in internal representations
  • tool or agent load effects
  • model configuration fluctuations

8.3.3 Indicators

  • inconsistent step ordering
  • fluctuating reasoning length
  • divergent intermediate transitions
  • unexplained reorganization of reasoning steps

8.3.4 Impact

Structural drift undermines reproducibility and increases the likelihood of unpredictable system behavior.

8.3.5 Mitigation

  • enforce variance envelope
  • enforce chain stability controls
  • reset or reinitialize agent state when drift is detected
  • maintain consistent inference parameters

8.4 Semantic Drift Risk

8.4.1 Description

Semantic drift arises when the meaning, conceptual alignment, or topic structure shifts unpredictably during or across reasoning cycles.

8.4.2 Causes

  • conceptual instability under repetition
  • semantic domain expansion
  • interference from retrieval or tools
  • failure to retain core conceptual anchors

8.4.3 Indicators

  • loss of topic relevance
  • contradictory intermediate steps
  • unintended meaning shifts
  • negation or inversion anomalies

8.4.4 Impact

Semantic drift reduces coherence, predictability, and trustworthiness of reasoning outputs.

8.4.5 Mitigation

  • semantic coherence controls
  • conceptual retention verification
  • governance boundary enforcement

8.5 Temporal Stability Risk

8.5.1 Description

Temporal stability risk refers to degradation in reasoning quality over time or across extended inference sequences.

8.5.2 Causes

  • accumulated internal noise
  • memory contamination
  • incremental divergence across long chains
  • resource fluctuations

8.5.3 Indicators

  • gradual decline in coherence
  • increasing drift across intervals
  • failure to maintain reasoning state consistency

8.5.4 Impact

Long-duration applications may become unreliable, particularly in agentic or multi-step workflows.

8.5.5 Mitigation

  • longitudinal monitoring
  • periodic state resets
  • consistency checkpoints

8.6 Interference Risk

8.6.1 Description

Interference risk arises when external components affect reasoning integrity.

Sources of interference include:

  • tools
  • retrieval components
  • other agents
  • environment APIs
  • context or memory systems

8.6.2 Causes

  • misaligned tool responses
  • unbounded agent interactions
  • unpredictable RAG data injection
  • multi-agent reasoning collisions

8.6.3 Indicators

  • sudden reasoning divergence
  • tool-dependent instability
  • inconsistent decisions after tool use
  • agent cross-contamination

8.6.4 Impact

Interference may cause unpredictable reasoning patterns, loss of boundary control, and uncontrolled drift.

8.6.5 Mitigation

  • interference detection
  • boundary enforcement controls
  • isolation of reasoning state
  • validation of tool outputs

8.7 Boundary Violation Risk

8.7.1 Description

Boundary violation risk occurs when the system exceeds allowed contextual, semantic, or operational boundaries.

8.7.2 Causes

  • context window expansion
  • unauthorized semantic domain exploration
  • cross-task memory leakage
  • overreliance on external tools
  • multi-agent interference

8.7.3 Indicators

  • references to disallowed information
  • use of context not present in the input
  • reasoning dependent on external unintended state
  • creeping expansion of reasoning scope

8.7.4 Impact

Boundary violations compromise predictability, auditability, and operational safety.

8.7.5 Mitigation

  • boundary definition and enforcement controls
  • monitoring for out-of-bound semantic transitions
  • strict tool and agent constraints

8.8 Degradation Risk

8.8.1 Description

Degradation risk refers to progressive reasoning quality decline over time due to repeated use, extended agent loops, environmental factors, or model instability.

8.8.2 Causes

  • prolonged agentic sessions
  • heavy tool usage
  • accumulated contextual noise
  • fluctuating infrastructure conditions

8.8.3 Indicators

  • increasing drift trends
  • decreasing coherence metrics
  • longer or shorter reasoning chains without cause
  • gradual divergence across evaluation windows

8.8.4 Impact

Degradation affects reliability in long-running systems, especially in automation and orchestration applications.

8.8.5 Mitigation

  • longitudinal stability tracking
  • state resets
  • performance baselining
  • proactive drift sensitivity checks

8.9 Risk Levels

RIS classifies reasoning risks into three levels:

  • Low: Within established variance and stability thresholds
  • Moderate: Deviations observed but within safe operational limits
  • High: Significant deviations requiring remediation

Risk levels SHALL be documented in the conformance report.


8.10 Relationship Between Risks and RIS Levels

Each RIS level carries a corresponding risk tolerance:

  • RIS-0: All risks high
  • RIS-1: Most risks moderate to high
  • RIS-2: Risks moderate with limited high-risk events
  • RIS-3: All risks low to moderate; no high-risk events allowed
  • RIS-4: All risks low; strict controls required

A system SHALL NOT qualify for RIS-3 or RIS-4 if any high-risk category is identified.


8.11 Use of Risk Models in Evaluation

During evaluation, assessors SHALL:

  • identify risk categories triggered
  • document causes and indicators
  • evaluate drift and stability trends
  • map observed behavior to risk severity
  • determine required mitigation for higher RIS levels

Risk models inform conformance decisions but do not replace mandatory controls or scoring.


8.12 Use of Risk Models in Production Monitoring

Organizations SHOULD use risk models to:

  • detect early warning signals
  • anticipate degradation trends
  • identify tool-induced instability
  • evaluate multi-agent interference
  • maintain operational safety

Risk-based monitoring SHALL supplement RIS control compliance.


Section 9

Evaluation Methodology

9. Evaluation Methodology

9.1 Overview

The RIS evaluation methodology defines the standardized procedures for performing a reasoning integrity assessment. These procedures ensure that measurements are:

  • repeatable
  • objective
  • statistically meaningful
  • comparable across systems
  • suitable for audit and certification

All evaluations MUST follow the methodology in this section to be considered valid for RIS classification.


9.2 Evaluation Objectives

The objectives of RIS evaluation are to:

  1. quantify reasoning stability and coherence
  2. measure drift and variance across repeated evaluations
  3. detect boundary violations
  4. assess interference sensitivity
  5. validate compliance with required controls
  6. classify the system into an appropriate RIS level

Evaluation SHALL NOT assess factual correctness or ethical outcomes. Only reasoning integrity is measured.


9.3 Evaluation Conditions

9.3.1 Environmental consistency

Evaluations SHALL be conducted under consistent conditions, including:

  • model version
  • inference parameters (temperature, top-p, settings)
  • context and prompt format
  • agent or tool configurations
  • hardware or infrastructure environment

Changes to any parameter invalidate comparisons.

9.3.2 Isolation

The evaluation environment SHOULD minimize:

  • interference from concurrent workloads
  • environmental noise
  • model adaptation or memory building
  • caching effects

9.3.3 Documentation

The evaluator SHALL document:

  • all inference parameters
  • all environment variables
  • system settings used for assessment
  • test prompt sets

9.4 Prompt Set Requirements

RIS evaluation uses two types of prompts:

  1. Baseline prompts
  2. Perturbation prompts

9.4.1 Baseline prompts

Baseline prompts SHOULD include:

  • informational queries
  • reasoning tasks
  • multi-step tasks
  • conceptual analysis tasks
  • domain-neutral questions

At least 50 prompts SHOULD be used for general-purpose systems.

At least 20 prompts SHOULD be used for specialized systems.

9.4.2 Perturbation prompts

Perturbation prompts MUST maintain semantic equivalence with baseline prompts while varying:

  • syntax
  • structure
  • ordering
  • minor phrasing
  • neutral substitutions

A minimum of 10 perturbations SHALL be used per baseline prompt selected for perturbation testing.


9.5 Sampling Methodology

9.5.1 Repeated inference sampling

For each baseline prompt:

  • at least 25 repeated inference samples SHALL be collected
  • sampling parameters SHALL remain identical across all repetitions

Repeated inference sampling is used to compute:

  • chain stability
  • semantic coherence
  • drift sensitivity
  • variance envelope compliance

9.5.2 Controlled perturbation sampling

For each perturbation prompt:

  • at least 10 inference samples SHALL be collected

Perturbation sampling is used to measure:

  • structural drift
  • semantic drift
  • robustness under variation

9.6 Baseline Establishment

9.6.1 Creating a baseline

The evaluator SHALL establish a baseline by:

  1. selecting representative baseline prompts
  2. executing repeated inference sampling
  3. computing structural and semantic similarity metrics
  4. determining drift sensitivity
  5. defining acceptable variance envelopes

9.6.2 Baseline validity

A baseline is valid only if:

  • all evaluation conditions are documented
  • system configuration remains unchanged
  • variance envelope is statistically meaningful

Baseline MUST be refreshed when model parameters change or risk models trigger reassessment.


9.7 Drift Evaluation

9.7.1 Drift detection

Drift is detected by:

  • comparing new samples to baseline
  • analyzing deviation trends
  • identifying abrupt or progressive divergence
  • checking for consistent out-of-envelope behavior

9.7.2 Drift characterization

Drift SHALL be categorized as:

  • structural
  • semantic
  • temporal
  • interference-based

9.7.3 Drift thresholds

A drift event occurs when:

  • deviation exceeds the envelope for more than 5 percent of samples
  • repeated deviations occur during perturbation sampling
  • temporal divergence increases across sequential windows

9.7.4 Drift escalation

If drift is detected:

  • the system’s RIS level MAY be lowered
  • additional testing SHALL be performed
  • drift controls MUST be verified

9.8 Boundary Evaluation

Boundary adherence SHALL be tested by:

  • introducing tasks with clear semantic boundaries
  • evaluating attempts to use disallowed context
  • testing reasoning behavior with restricted tool outputs
  • detecting cross-task or cross-context contamination

Any violation SHALL be logged and considered in level classification.


9.9 Interference Evaluation

Interference evaluation includes:

  • testing reasoning stability before and after tool calls
  • measuring agent-to-agent interaction effects
  • injecting controlled retrieval variability
  • observing structural or semantic disruptions

Interference MUST not cause uncontrolled reasoning divergence for systems seeking RIS-3 or RIS-4 level classification.


9.10 Variance Envelope Evaluation

Variance envelope compliance SHALL be tested through:

  • repeated inference sampling
  • perturbation sampling
  • load or stress conditions if applicable
  • periodic resampling across intervals

A system MUST maintain compliance for:

  • at least 95 percent of samples (RIS-3)
  • at least 98 percent of samples (RIS-4)

9.11 Evaluation Report Requirements

A complete evaluation report SHALL include:

  1. System description
  2. Evaluation environment details
  3. Prompt sets used
  4. Sampling methodology
  5. Baseline metrics
  6. Drift analysis
  7. Variance envelope results
  8. Boundary adherence results
  9. Interference results
  10. Composite score calculation
  11. Control compliance analysis
  12. RIS level assigned
  13. Validity period
  14. Evidence appendix

The report SHALL be signed by the evaluator.


9.12 Evaluation Validity

An evaluation remains valid only if:

  • the system configuration does not change
  • no drift events occur in production
  • no new tools, agents, or retrieval sources are introduced
  • governance boundaries remain constant

If any condition changes, reassessment MAY be required immediately.


9.13 Re-Evaluation Requirements

Re-evaluation SHALL occur:

  • at the end of the classification validity period
  • when major model updates occur
  • when drift or instability is observed
  • when new risk categories are triggered
  • when evidence suggests deviation from baseline behavior

Systems in regulated or safety-critical environments MAY require more frequent reassessments.


Section 10

Audit Guidelines

10. Audit Guidelines

10.1 Overview

This section defines the audit requirements, evidence standards, assessment procedures, and validation criteria necessary to verify compliance with the Reasoning Integrity Standard (RIS). These guidelines ensure that evaluations are reliable, reproducible, and suitable for internal review, third-party certification, or regulatory oversight.

An auditor MAY be an internal reviewer, an external assessor, or an automated system. All auditors SHALL adhere to the requirements outlined in this section.


10.2 Audit Objectives

The objectives of a RIS audit are to:

  1. verify the correctness and completeness of evaluation procedures
  2. validate metric calculations and composite reasoning integrity scores
  3. confirm implementation of mandatory controls
  4. detect deficiencies, violations, or risks
  5. ensure evidence sufficiency and traceability
  6. determine whether assigned RIS levels are justified
  7. produce a formal audit report documenting findings

Audits SHALL NOT evaluate ethical, factual, or domain-specific correctness.


10.3 Audit Scope

The audit SHALL cover:

  • evaluation methodology adherence
  • system configuration and invariants
  • reasoning integrity metrics
  • drift and variance analysis
  • control compliance
  • governance boundary enforcement
  • operational integrity
  • logged events and ledger entries
  • deviation or violation reports

The audit MAY include additional areas depending on organizational, regulatory, or contractual requirements.


10.4 Evidence Requirements

Auditors SHALL review the following types of evidence:

10.4.1 Raw inference samples

  • repeated inference outputs
  • perturbation outputs
  • temporal sampling outputs
  • tool-integrated outputs (if applicable)

10.4.2 Metric calculations

  • chain stability similarity data
  • semantic coherence scores
  • drift sensitivity calculations
  • variance envelope compliance results
  • boundary adherence results

10.4.3 Configuration data

  • inference parameters (temperature, top-p, etc.)
  • system settings
  • agent or tool configurations
  • environment variables
  • software versions

10.4.4 Control implementation evidence

  • governance boundary definitions
  • drift envelope definitions
  • stability and coherence controls
  • interference detection controls
  • audit logging mechanisms

10.4.5 Logs and ledger entries

  • reasoning ledger (or equivalent)
  • boundary violation logs
  • drift events
  • anomaly detection entries
  • operational logs related to inference behavior

Auditors SHALL verify the authenticity and integrity of evidence before drawing conclusions.


10.5 Audit Procedures

10.5.1 Preparation

The auditor SHALL:

  • obtain system documentation
  • review evaluation methodology used
  • validate prompt sets and sampling plans
  • confirm frozen inference parameters
  • ensure environmental consistency

10.5.2 Evidence examination

The auditor SHALL examine:

  • raw samples for drift or instability
  • metrics for correctness and reproducibility
  • compliance with mandatory controls
  • documentation for completeness

10.5.3 Reproduction testing

Auditors MAY reproduce:

  • repeated inference sampling
  • perturbation sampling
  • drift and variance calculations

Reproduction SHALL use identical parameters to those documented in the evaluation.

10.5.4 Violation detection

The auditor SHALL check for:

  • drift beyond thresholds
  • variance envelope breaches
  • boundary violations
  • interference anomalies
  • control misconfigurations
  • unreported failures

Any violation SHALL be documented in the final audit report.


10.6 Compliance Determination

Audit findings SHALL conclude one of the following:

  • compliant
  • compliant with exceptions
  • non-compliant
  • inconclusive (insufficient evidence)

If non-compliant, the auditor SHALL document which controls or evaluation procedures were violated.

A RIS level MAY only be confirmed if:

  • composite score qualifies
  • mandatory controls are satisfied
  • evidence is complete
  • no unmitigated high-risk deviations are observed

10.7 Audit Reporting Requirements

The audit report SHALL include:

  1. System overview
  2. Evaluation conditions
  3. Summary of controls reviewed
  4. Evidence examined
  5. Metric verification results
  6. Observed violations or exceptions
  7. Assessment of drift, variance, and boundary adherence
  8. Conformance determination
  9. Confirmed RIS level
  10. Recommendations and corrective actions
  11. Validity period of assessment
  12. Appendices with supporting data

Reports SHALL be retained for the duration of the classification validity period.


10.8 Corrective Action Requirements

If deficiencies or violations are identified:

  • corrective actions SHALL be documented
  • the system owner SHALL remediate issues
  • the system SHALL undergo partial or full reassessment
  • major issues SHALL reset the RIS classification

Corrective actions MAY include:

  • adjusting inference parameters
  • redefining governance boundaries
  • improving drift or variance controls
  • updating tool or agent behavior

10.9 Auditor Qualifications

RIS audits MAY be conducted by:

  • internal governance teams
  • independent third-party evaluators
  • qualified researchers
  • automated auditing systems

Auditors SHOULD possess expertise in:

  • LLM inference behavior
  • statistical evaluation
  • reasoning-chain analysis
  • AI governance and risk frameworks
  • tool, agent, or RAG system integrations (if applicable)

Organizations MAY impose additional requirements for auditor certification.


10.10 Continuous Audit Recommendations

Organizations SHOULD implement ongoing audit-aligned practices:

  • periodic drift checks
  • rolling variance sampling
  • boundary adherence monitoring
  • interference detection and logging
  • longitudinal stability assessments

Continuous monitoring SHALL supplement, not replace, formal RIS audits.


10.11 Audit Validity

An audit is valid for:

  • 12 months for general systems
  • 6 months for agentic or tool-integrated systems
  • 3 months for safety-critical systems

Audits SHALL be invalidated immediately if:

  • the model or agent is updated
  • new tools, retrieval sources, or agents are introduced
  • significant drift is observed in production
  • boundary violations occur repeatedly

10.12 Audit Transparency

Organizations MAY disclose:

  • confirmed RIS level
  • summary of findings
  • audit date and validity period

Organizations SHOULD NOT publicly disclose:

  • raw inference samples
  • proprietary metrics
  • internal configurations
  • confidential logs

Unless required by contract, regulation, or security review.


Section 11

Reference Implementation

11. Reference Implementation (LCAC)

11.1 Overview

This section describes the Least-Context Access Control (LCAC) framework as a reference implementation of the Reasoning Integrity Standard (RIS). LCAC demonstrates one practical method for enforcing RIS controls, measuring reasoning integrity, and maintaining predictable reasoning behavior in LLM-based systems.

LCAC is not required for RIS conformance. It is provided solely as an informative reference to illustrate feasibility, implementation patterns, and system design considerations aligned with RIS requirements.


11.2 Architectural Overview

LCAC is structured as a modular reasoning-governance framework consisting of:

  • a centralized governor service
  • a trust and variance model
  • an immutable reasoning ledger
  • governance boundary enforcement components
  • monitoring and evaluation pipelines
  • optional persona overlays

LCAC may operate as an independent service or be integrated directly into an LLM-driven application, multi-agent system, or inference pipeline.


11.3 Components and Functions

11.3.1 Governor

The LCAC governor performs the following functions:

  • receives prompt and output pairs for evaluation
  • computes trust, variance, and drift metrics
  • determines reasoning verdicts (stable, watch, unstable)
  • updates governance states (hold, elevate, lockdown)
  • writes results to the ledger

The governor serves as the core enforcement point for RIS-aligned controls.

11.3.2 Trust Model

LCAC's trust model provides:

  • a baseline trust score
  • variance and drift calculations
  • predictive stability measurements
  • structural and semantic similarity analysis

The trust score is mapped to RIS metrics such as:

  • chain stability
  • semantic coherence
  • drift sensitivity

Trust values range between 0.00 and 1.00, consistent with RIS scoring.

11.3.3 Variance and Drift Envelope

LCAC defines and enforces a variance envelope based on:

  • repeated inference samples
  • controlled perturbation testing
  • deviation thresholds
  • longitudinal observations

Metric outputs are compared against these envelopes to determine:

  • compliance
  • drift events
  • stability degradation

11.3.4 Immutable Ledger

LCAC maintains an immutable reasoning ledger using:

  • hash-linked entries
  • timestamped records
  • evaluation metrics
  • baseline comparisons
  • anomaly documentation

Each ledger entry includes:

  • input prompt
  • model output
  • trust score
  • variance metrics
  • stability verdict
  • previous entry hash
  • current entry hash

This supports audit, reproducibility, and traceability.


11.4 Governance Modes

LCAC defines three governance modes that influence reasoning behavior:

11.4.1 Hold

  • default system state
  • stable or near-stable reasoning
  • no significant anomalies detected

11.4.2 Elevate

  • moderate drift or variance detected
  • reasoning monitored more aggressively
  • potential instability anticipated

11.4.3 Lockdown

  • reasoning integrity breach detected
  • drift or variance above thresholds
  • boundary violation or interference event observed
  • heightened restrictions applied

Governance modes demonstrate how RIS controls may be operationalized.


11.5 Boundary Enforcement

LCAC establishes governance boundaries using:

  • contextual limits
  • semantic domain constraints
  • tool and agent access restrictions
  • validation checks for external outputs

Boundary enforcement aligns with RIS boundary violation controls.


11.6 Persona Overlay (Informative Only)

LCAC includes an optional persona system that applies:

  • stability bias adjustments
  • risk posture tuning
  • domain-specific behavioral constraints

Personas may influence:

  • baseline trust
  • drift sensitivity
  • variance envelopes
  • stability thresholds

Persona systems are informative and not required for RIS compliance.


11.7 Integration Patterns

LCAC supports the following integration models:

11.7.1 Request-Evaluation Pattern

The system sends:

  • prompt
  • model output

to LCAC for evaluation. LCAC returns:

  • trust score
  • variance data
  • stability verdict
  • governance mode

This pattern enables external inference pipelines to remain unchanged.

11.7.2 Inline Agent or LLM Integration

LCAC may be embedded into:

  • agent toolchains
  • orchestration frameworks
  • multi-agent reasoning flows
  • specialized LLM applications

Governance checks occur inline between inference steps.

11.7.3 Multi-Model or Router Integration

LCAC may act as:

  • a routing constraint
  • a stability filter
  • a verification layer

for systems that dynamically select between models.


11.8 Operational Monitoring

LCAC supports RIS-aligned monitoring by:

  • generating audit logs
  • tracking drift events
  • identifying variance breaches
  • logging boundary violations
  • maintaining historical baselines

Monitoring outputs MAY be used to trigger reassessment under RIS Section 9.


11.9 Example Workflow

A RIS-aligned workflow using LCAC includes:

  1. The LLM produces an output for a given prompt.
  2. The output and prompt are sent to LCAC.
  3. LCAC computes trust, variance, and drift metrics.
  4. LCAC assigns a verdict (stable, watch, unstable).
  5. LCAC writes an immutable ledger entry.
  6. Governance mode is updated based on results.
  7. Application logic uses the verdict to:

- accept the output

- reject the output

- request re-evaluation

- escalate to a fallback model

- enforce stricter controls

This workflow illustrates one method of operationalizing RIS requirements.


11.10 Compliance Relationship

LCAC demonstrates compliance with the following RIS controls:

  • RS-1, RS-2, RS-3
  • SC-1, SC-2, SC-3
  • DR-1, DR-2, DR-3
  • VE-1, VE-2, VE-3
  • GB-1, GB-2, GB-3
  • OP-1, OP-2, OP-3, OP-4

LCAC is a reference implementation and SHOULD NOT be considered the only method of achieving RIS compliance.


11.11 Implementation Notes

  • LCAC uses Redis for metric storage and ledger hashing.
  • All evaluation computations are deterministic given identical inputs.
  • LCAC can operate in multi-node architectures with shared storage.
  • The reference implementation is platform-agnostic.
  • API endpoints and data schemas are documented in the LCAC repository.

These details are informative and may vary between deployments.


11.12 Limitations

LCAC, as a reference implementation:

  • does not guarantee factual correctness
  • does not address privacy or security
  • does not enforce ethical or domain-specific constraints
  • does not replace external governance frameworks

RIS conformance MUST be determined independently of LCAC adoption.


11.13 Informative Appendix Linkage

Sections of LCAC MAY be referenced in Appendix A for:

  • example metrics
  • sample ledger entries
  • example drift baselines
  • variance envelope templates
  • integration code snippets

These references are informative only.


Section 12

Standard Mapping

12. Mapping to Existing Standards

12.1 Overview

This section provides a crosswalk between the Reasoning Integrity Standard (RIS) and widely adopted security, governance, and AI risk management frameworks. The purpose of this mapping is to:

  • support organizations aligning RIS with existing compliance programs
  • demonstrate compatibility between RIS and established standards
  • clarify which risks and controls RIS covers
  • identify where RIS provides additional, reasoning-specific requirements

Mappings are informative and do not constitute equivalence or certification under referenced standards.


12.2 Relationship to NIST SP 800-53

RIS aligns with NIST SP 800-53 by extending its principles into the domain of reasoning integrity for AI systems.

12.2.1 Complementary Areas

  • NIST RA (Risk Assessment): RIS provides a reasoning-specific risk model.
  • NIST CA (Assessment, Authorization, Monitoring): RIS defines evaluation and audit procedures.
  • NIST SI (System Integrity): RIS adds reasoning-chain integrity and drift detection.
  • NIST PM (Program Management): RIS supports governance for LLM-based workflows.

12.2.2 Gaps Filled by RIS

NIST does not address:

  • reasoning drift
  • semantic coherence
  • chain stability
  • variance envelopes
  • multi-agent interference

RIS provides controls specifically for these areas.


12.3 Relationship to ISO/IEC 27001

RIS complements ISO/IEC 27001 by introducing AI reasoning integrity controls that parallel information security requirements.

12.3.1 Complementary Areas

  • ISO 27001 Annex A: controls for monitoring, logging, and operational integrity
  • ISO 27001 Annex A: change management and configuration management align with RIS evaluation invariants
  • ISO 27001 Annex A: audit logging aligns with RIS ledger requirements

12.3.2 Gaps Filled by RIS

ISO 27001 does not include:

  • reasoning integrity assessment
  • drift evaluation
  • semantic stability requirements
  • boundary adherence controls for LLMs

RIS extends governance into these areas.


12.4 Relationship to SOC 2 Trust Services Criteria

RIS aligns with SOC 2 across the following TSC categories:

  • Security: LCAC-like governance boundaries enhance system integrity
  • Availability: drift and variance controls improve predictability of behavior
  • Processing Integrity: RIS focuses specifically on the integrity of reasoning processes
  • Confidentiality: RIS does not address confidentiality explicitly but complements it

SOC 2 does not cover reasoning stability, which RIS supplies.


12.5 Relationship to OWASP Application Security Verification Standard (ASVS)

RIS extends ASVS principles into the domain of AI reasoning by providing:

  • integrity validation of reasoning workflows
  • boundary checks analogous to input validation
  • monitoring comparable to application logic verification

Areas where RIS aligns with ASVS:

  • session integrity (analogous to reasoning chain integrity)
  • validation layers (similar to boundary enforcement)
  • error and exception handling (mapped to drift detection)

ASVS does not include concepts such as:

  • semantic drift
  • reasoning instability
  • variance envelopes

RIS adds these requirements.


12.6 Relationship to NIST AI RMF

RIS is closely aligned with the NIST AI Risk Management Framework.

12.6.1 Alignment Areas

  • Govern Function: RIS defines governance boundaries and control families
  • Map Function: RIS specifies reasoning-specific risks and threat models
  • Measure Function: RIS provides standardized metrics and scoring
  • Manage Function: RIS prescribes drift, variance, and stability controls

12.6.2 RIS Extensions Beyond NIST AI RMF

RIS introduces:

  • explicit reasoning chain assessment
  • structured variance and drift envelopes
  • multi-level conformance classifications
  • detailed evaluation methodology
  • audit and verification standards

These elements provide more granular operational guidance.


12.7 Relationship to the EU AI Act

The EU AI Act categorizes AI systems according to risk. RIS provides a compatibility layer by offering objective measurements relevant to risk classification.

12.7.1 High-Risk Alignment

High-risk AI systems require:

  • logging
  • monitoring
  • technical documentation
  • risk mitigation

RIS supports these by:

  • providing a reasoning ledger
  • defining stability metrics
  • enabling drift monitoring
  • defining reproducible evaluation processes

12.7.2 RIS-4 and High-Risk Use Cases

RIS-4 provides the strictest reasoning integrity requirements and is appropriate for:

  • safety-critical systems
  • medical reasoning systems
  • financial decision systems
  • legal reasoning systems
  • infrastructure and security systems

EU AI Act does not provide technical detail on reasoning integrity. RIS fills that gap.


12.8 Summary Table (RIS vs. External Standards)

FrameworkOverlapRIS Contribution
NIST SP 800-53Monitoring, logging, system integrityAdds reasoning-chain integrity and drift controls
ISO 27001Operational controls, audit loggingAdds reasoning stability requirements
SOC 2Processing integrityAdds reasoning-specific integrity metrics
OWASP ASVSBoundary and validation conceptsAdds semantic and structural reasoning requirements
NIST AI RMFRisk, governance, measurementAdds detailed evaluation and conformance system
EU AI ActDocumentation and oversightAdds measurable technical integrity criteria

12.9 General Observations

  • RIS is compatible with existing frameworks but more granular in evaluating reasoning behavior.
  • RIS introduces metrics and controls not present in any existing standard.
  • Organizations adopting RIS can integrate it with existing compliance programs.
  • RIS MAY serve as a technical layer beneath regulatory or governance frameworks.

12.10 Use of Mapping in Compliance Programs

Organizations MAY use this mapping to:

  • integrate RIS evaluation with ISO or SOC 2 audits
  • include RIS controls in security and governance documentation
  • demonstrate operational integrity in high-risk AI deployments
  • support regulatory submissions requiring technical evidence
  • develop internal AI governance programs

Mapping does not imply certification under other standards.


Section 13

Appendices

13. Appendices

13.1 Overview

This section contains informative appendices supporting the implementation, evaluation, and audit of the Reasoning Integrity Standard (RIS). Appendices are not normative unless explicitly referenced by a control or requirement.

The materials in this section include:

  • definitions and terminology
  • mathematical formulations
  • example baseline structures
  • sample drift and variance calculations
  • reference ledger entries
  • prompt set templates
  • evaluation dataset formats
  • implementation considerations

13.2 Key Definitions

13.2.1 Reasoning chain

A sequence of internal conceptual steps used by an LLM or agent to derive an output from an input.

13.2.2 Structural drift

A deviation in the ordering, length, or internal transitions of a reasoning chain beyond acceptable variance thresholds.

13.2.3 Semantic drift

A deviation in meaning, conceptual structure, or topic alignment across samples.

13.2.4 Variance envelope

The statistical boundary within which repeated inference behavior must remain to be considered stable.

13.2.5 Boundary violation

A reasoning event in which the model exceeds defined contextual, semantic, or operational constraints.

13.2.6 Interference event

An alteration of reasoning behavior caused by a tool, agent, retrieval source, or external system.


13.3 Mathematical Formulations

13.3.1 Chain stability score

Chain stability S is calculated as:

S = mean(similarity(Ri, Rj))

13.3.2 Semantic coherence score

Semantic coherence C is calculated as:

C = mean(cosine(Ei, Ej))

13.3.3 Drift sensitivity

Drift sensitivity D is calculated as:

D = variance(S) + variance(C)

13.3.4 Variance envelope compliance

Compliance V is:

V = (samples within envelope) / (total samples)


13.4 Example Variance Envelopes

13.4.1 General-purpose LLM envelope

  • chain stability: 0.75 to 1.00
  • semantic coherence: 0.80 to 1.00
  • drift sensitivity: below 0.30

13.4.2 High-integrity envelope

  • chain stability: 0.90 to 1.00
  • semantic coherence: 0.90 to 1.00
  • drift sensitivity: below 0.10

Evaluators SHALL adjust envelopes based on domain requirements.


13.5 Example Ledger Entry

Below is a sample hash-linked ledger entry.

{

"prompt": "Explain the impact of X on Y.",

"output": "Reasoning steps and conclusion...",

"trust": 0.92,

"variance": 0.05,

"verdict": "stable",

"timestamp": 1735861234,

"prev_hash": "0000c9beef78aa12fd39b4b2d4e9a18f",

"hash": "8ad1c4fd1e09aa7dfefc324b3987456b"

}

This format is informative and MAY be adapted.


13.6 Prompt Set Templates

The following templates MAY be used to construct RIS evaluation prompt sets.

13.6.1 Baseline prompts

  • Explain a concept.
  • Perform a step-by-step analysis.
  • Compare two ideas.
  • Solve a structured reasoning task.
  • Describe implications of a policy or event.

13.6.2 Perturbation prompts

  • synonym substitutions
  • minor phrasal changes
  • syntax adjustments
  • clause reordering

All perturbations MUST preserve semantic meaning.


13.7 Evaluation Dataset Format

Evaluation datasets SHOULD be structured as follows:

{

"prompt_id": "P001",

"prompt": "Explain the role of Z.",

"baseline_samples": [],

"perturbation_prompts": [],

"perturbation_samples": [],

"metadata": { "domain": "general", "difficulty": "medium" }

}

This supports consistent analysis and cross-system comparison.


13.8 Example Drift Analysis Summary

A drift summary MAY include:

  • baseline mean stability score
  • new sample mean stability score
  • deviation percentage
  • variance across iterations
  • threshold compliance
  • drift triggers observed

Example:

{

"baseline_stability": 0.91,

"current_stability": 0.82,

"deviation": "9.8%",

"threshold": "10%",

"status": "borderline-compliant"

}


13.9 Boundary Violation Examples

Examples of boundary violations include:

  • referencing information not present in prompt or context
  • unexpected topic expansion outside allowed domain
  • using tool outputs that exceed allowed semantic scope
  • combining context from unrelated tasks
  • reasoning dependent on long-term model memory

Boundary violations SHALL be documented and investigated.


13.10 Implementation Considerations

Organizations implementing RIS SHOULD consider:

  • model-specific tuning of drift and variance thresholds
  • architecture-specific behaviors (transformer vs. hybrid vs. agentic systems)
  • differential stability between prompt types
  • retrieval and tool-use artifacts influencing stability
  • cross-agent reasoning contamination
  • infrastructure effects on sampling consistency

These considerations inform evaluation and do not alter RIS conformance requirements.


13.11 Future Extensions (Informative)

Potential future RIS extensions MAY include:

  • multi-agent interference scoring
  • predictive trust-flow modeling
  • coherence signature analysis
  • temporal reasoning-chain compression metrics
  • scenario-based benchmark suites
  • extended metadata for multi-modal systems

These are informative only and not included in RIS v1.0.


13.12 Document Revision Log

Version: 1.0

Published by: Atom Labs

Status: Initial Release

Future revisions SHALL be recorded in this section when issued.