Foreword
Reasoning Integrity Standard (RIS) v1.0
Published by Atom Labs
© 2025 Atom Labs. All Rights Reserved.
0. Foreword
The Reasoning Integrity Standard (RIS) defines a formal framework for evaluating, governing, and maintaining the integrity of reasoning performed by large language models (LLMs), autonomous agents, and multi-model cognitive systems.
This standard was developed by Atom Labs to address the absence of authoritative, measurable criteria for assessing the stability and trustworthiness of LLM reasoning in production environments. As AI systems increasingly perform high-impact decision-making across enterprise, regulatory, and safety-critical contexts, the industry requires an auditable, repeatable, and model-agnostic method for measuring reasoning reliability.
RIS establishes a multi-level maturity model (RIS-0 through RIS-4) along with normative controls governing chain-of-thought stability, semantic coherence, drift detection, governance, and reasoning boundary enforcement. It specifies the requirements necessary for organizations to assess the integrity of reasoning workflows under real-world operational and adversarial conditions.
This standard is intended for:
- enterprise architects deploying AI systems
- safety and governance teams responsible for LLM oversight
- auditors and compliance personnel evaluating AI risk posture
- engineering teams building LLM-integrated applications
- research organizations conducting reproducible benchmarking
RIS v1.0 is model-agnostic and does not mandate the use of any specific framework or vendor. LCAC (Least-Context Access Control) is referenced solely as an example of a compliant reference implementation. RIS remains an independent, neutral specification applicable to any LLM or cognitive system.
Future revisions may expand the standard to include coherence metrics, multi-agent interference scoring, predictive trust modeling, and additional benchmark suites. Atom Labs invites the research, standards, and enterprise community to review, adopt, and contribute to future versions.
Scope
1. Scope
1.1 Purpose
The Reasoning Integrity Standard (RIS) defines the requirements, controls, measurement methodologies, and conformance criteria necessary to evaluate the integrity of reasoning performed by LLMs and LLM-powered cognitive systems. The purpose of RIS is not to measure the correctness or factuality of model outputs, but to measure the stability, predictability, and structural reliability of the reasoning process that produces those outputs.
1.2 Applicability
RIS applies to any system that conducts or delegates reasoning to:
- large language models (LLMs)
- autonomous or semi-autonomous agents
- multi-agent architectures
- chain-of-thought or stepwise reasoning pipelines
- retrieval-augmented systems (RAG)
- tool-augmented or API-integrated cognitive systems
- multi-model or ensemble inference architectures
RIS is applicable regardless of:
- model vendor or provider
- model size, training method, or licensing
- deployment environment (cloud, hybrid, on-premise, edge)
- inference scenario (single-turn, multi-turn, stateful, agentic)
1.3 Intended Use
RIS is intended to support:
- evaluation of LLMs and agentic systems before deployment
- continuous monitoring of reasoning behavior in production environments
- regulatory and compliance audits involving AI reasoning stability
- vendor and model procurement due diligence
- internal trust, governance, and safety programs
- reproducible research and benchmarking
1.4 Out of Scope
The following areas are outside the scope of RIS v1.0:
- data privacy practices
- model training and dataset governance
- ethics, fairness, bias, or demographic analysis
- factual correctness of responses
- model alignment and value judgments unrelated to reasoning integrity
These areas may be addressed by other standards or future RIS extensions.
1.5 Normative Language
RIS uses the following requirement terminology:
- MUST: an absolute, mandatory requirement for conformance
- SHALL: a criterion required to meet RIS compliance
- SHOULD: a recommended practice unless a justified exception exists
- MAY: an optional practice
- NOT PERMITTED: a prohibited behavior or configuration
1.6 Relationship to LCAC
LCAC is a reference implementation illustrating one possible method of achieving RIS compliance. RIS does not require LCAC, nor does conformance imply adoption of any specific product, tool, or architecture. RIS is intended to remain implementation-agnostic.
1.7 Versioning and Revision Policy
RIS v1.0 is the initial published version of the Reasoning Integrity Standard. Atom Labs maintains stewardship of the standard until the formation or designation of a formal multi-party standards body.
Future revisions may include:
- expanded risk models
- domain-specific profiles
- standardized test suites
- trust-flow prediction methodologies
- multi-agent interference metrics
Revisions will follow a versioned specification model (v1.1, v1.2, v2.0, etc.).
Normative References
2. Normative References
2.1 Purpose of Normative References
This section identifies documents, standards, and terminology references that are considered authoritative for understanding and implementing the Reasoning Integrity Standard (RIS). Where conflicts arise, this specification takes precedence unless a referenced standard is explicitly designated as controlling for a specific requirement.
Normative references are essential for interpreting RIS controls, understanding terminology, and establishing consistent evaluation and audit methodologies across implementations.
2.2 Standards and Frameworks Referenced by RIS
The following documents and standards are integral to interpreting RIS v1.0. Implementers are expected to be familiar with these materials, as RIS incorporates concepts, terminology, and structural patterns from them.
NIST Standards
- NIST Special Publication 800-53: Security and Privacy Controls for Information Systems and Organizations
- NIST Artificial Intelligence Risk Management Framework (AI RMF)
- NIST SP 1270: Trustworthy and Responsible AI
ISO Standards
- ISO/IEC 27001: Information Security Management Systems
- ISO/IEC 23894: Artificial Intelligence Risk Management
- ISO/IEC 29100: Privacy Framework
- ISO/IEC 25010: System and Software Quality Models
SOC 2 and AICPA Guidance
- AICPA Trust Services Criteria (Security, Availability, Processing Integrity, Confidentiality, Privacy)
- AICPA AT-C 205: Examination Engagements
OWASP and Application Security References
- OWASP Application Security Verification Standard (ASVS)
- OWASP Top 10 (referenced in governance and boundary controls)
AI Safety, Governance, and Model Evaluation References
- EU AI Act (Draft and Final Texts)
- MLPerf and related benchmark publications
- Model evaluation methodologies from major AI labs
- Academic publications on chain-of-thought, inference stability, and cognitive drift
2.3 Terminology References
RIS uses terminology derived from established research and industry standards. Implementers should reference the following categories for precise interpretation.
AI Systems Terminology
- large language model (LLM)
- agent and multi-agent system
- chain-of-thought (CoT)
- semantic coherence
- contextual drift
- inference-time variability
- deterministic vs. non-deterministic reasoning
Governance Terminology
- risk posture
- control requirement
- conformance
- audit evidence
- variance threshold
- stability metric
LCAC Terminology (Reference Implementation Only)
LCAC terminology is informative, not normative. Terms include:
- least-context access control
- governance mode
- reasoning boundary
- drift envelope
- trust baseline
- variance baseline
- reasoning ledger
Implementers are not required to adopt LCAC’s terminology or architecture, but LCAC terms appear in examples and appendices as part of the reference implementation.
2.4 Definitions Governed by RIS
RIS mandates the following definitions for consistent use across all conformant implementations.
Reasoning Integrity
The measurable ability of a system to produce stable, predictable reasoning outputs that maintain structural coherence across similar prompts, repeated prompts, operational conditions, and controlled variations.
Reasoning Drift
A deviation in reasoning behavior, structure, or output that exceeds defined variance thresholds when evaluated under repeatable or controlled conditions.
Chain Stability
The degree to which the internal reasoning structure of a system maintains consistent logical steps, semantic transitions, and contextual boundaries across repeated inference cycles.
Semantic Consistency
The alignment of meaning, intent, and conceptual structure across multiple reasoning instances or reasoning steps for equivalent prompts.
Governance Boundary
A defined constraint controlling the informational, contextual, or semantic state accessible to a reasoning system at inference time.
2.5 Control of Normative References
Where RIS references external standards:
- External standards MUST be treated as supporting material.
- External standards SHALL NOT override RIS requirements unless explicitly designated as authoritative.
- RIS SHALL define the governing definitions for reasoning integrity and drift.
- External terminology MAY be used for interpretation but not for replacing RIS controls.
2.6 Informative vs. Normative Material
RIS distinguishes between:
- Normative material (mandatory for conformance)
- Informative material (supporting guidance)
Normative sections include:
- Controls
- Requirements
- Measurement Framework
- Scoring and Conformance
- Audit Guidelines
Informative sections include:
- Reference Implementation
- Mapping to Other Standards
- Appendices
This distinction is preserved throughout the remainder of the specification.
Fundamental Concepts
3. Fundamental Concepts of Reasoning Integrity
3.1 Overview
This section defines the foundational concepts required to understand and evaluate reasoning integrity in large language models (LLMs), autonomous agents, and multi-model cognitive systems. These concepts form the basis of all controls, requirements, and conformance criteria in RIS. Implementers MUST understand and consistently apply these definitions across all stages of evaluation, monitoring, and governance.
Reasoning integrity is not a measure of factual correctness or ethical compliance. It is a measure of the internal stability, predictability, and structural reliability of the reasoning process itself.
3.2 Reasoning Integrity
Reasoning integrity is the degree to which a reasoning system produces stable, coherent, and predictable internal reasoning structures when presented with equivalent or contextually similar prompts, conditions, or operational states.
A system with high reasoning integrity maintains:
- consistent logical structure
- predictable semantic relationships
- stable reasoning paths across repetitions
- adherence to defined contextual boundaries
- low drift under controlled variations
A system with low reasoning integrity exhibits:
- unpredictable reasoning paths
- inconsistent semantic structures
- sensitivity to minor input or environmental variations
- uncontrolled drift over repeated inferences
- unstable or contradictory reasoning forms
3.3 Reasoning Chain
A reasoning chain is the internal, stepwise sequence of conceptual transformations used by the system to derive an output from an input. The chain may be implicit (black-box LLMs) or explicit (systems exposing structured reasoning steps), but RIS treats both as conceptual equivalents.
A reasoning chain includes:
- latent representations
- intermediate conceptual structures
- logical transitions
- semantic dependencies
- step-to-step relationships
RIS does not require access to internal model states. Reasoning chains may be reconstructed using output structure, behavioral patterns, and statistical inference.
3.4 Chain Stability
Chain stability is the consistency of reasoning chain structure across repeated or controlled inference cycles.
A system demonstrates chain stability when:
- repeated prompts generate reasoning structures with low variance
- stepwise transitions remain logically and semantically similar
- variations fall within defined acceptable thresholds
Chain stability is a core metric in RIS and is used to determine system conformance at all RIS levels.
3.5 Semantic Coherence
Semantic coherence is the degree to which reasoning outputs maintain consistent meaning, conceptual alignment, and contextual relevance across equivalent inputs or over the course of a reasoning chain.
A system with high semantic coherence ensures that:
- meaning remains stable across reasoning steps
- internal transitions preserve conceptual alignment
- responses maintain adherence to prompt intent
- conceptual drift remains within acceptable thresholds
Loss of semantic coherence is a primary indicator of reasoning instability.
3.6 Cognitive Drift
Cognitive drift is the measurable deviation in a system’s reasoning process when evaluated under repeated, controlled, or minimally varied conditions.
Forms of drift include:
- structural drift: changes in reasoning chain architecture
- semantic drift: changes in meaning, context, or conceptual focus
- temporal drift: progressive degradation of reasoning consistency over time
- interference drift: changes caused by tools, memory, or agent interactions
Cognitive drift is one of the primary risk factors assessed by RIS.
3.7 Variance Envelope
The variance envelope defines the acceptable bounds of variation for reasoning behavior across repeated inference cycles. These bounds are derived from:
- statistical analysis
- expected variability based on model architecture
- domain-specific tolerance criteria
- operational safety requirements
A system MUST maintain its reasoning behavior within its defined variance envelope to achieve RIS conformance.
3.8 Governance Boundary
A governance boundary is a defined limit controlling what contextual, semantic, or informational state the system is permitted to use during a reasoning process.
Boundaries include:
- allowed context windows
- prohibited contextual expansions
- restricted semantic domains
- tool- and memory-access constraints
- cross-agent interaction limits
Governance boundaries prevent uncontrolled or unauthorized expansion of reasoning state.
3.9 Deterministic and Non-Deterministic Reasoning
Most LLM systems exhibit non-deterministic reasoning due to sampling temperature, randomness, and stochastic inference pathways. RIS does not require deterministic outputs; it requires:
- deterministic reasoning patterns within allowable variance
- predictable behavior under controlled conditions
- stable reasoning structure despite inherent stochasticity
Systems MUST demonstrate stable structural patterns even if surface-level outputs vary.
3.10 Tool and Agent Interference
Tool or agent interference occurs when the reasoning system’s internal process is altered by:
- external tool responses
- API calls
- retrieval results
- other agents in a multi-agent environment
Interference may cause:
- unexpected semantic transitions
- contextual overrides
- premature reasoning collapse
- unbounded expansion of reasoning state
RIS controls evaluate the system’s ability to maintain stability despite such interference.
3.11 State Persistence and Contextual Memory
State persistence defines how long prior reasoning states influence future reasoning cycles. Excessive or uncontrolled persistence may create:
- reasoning contamination
- indirect hallucination
- unexpected cross-task dependency
- failure to reinitialize reasoning state
RIS requires systems to maintain predictable and bounded state persistence behaviors.
3.12 Integrity Failure Modes
Common reasoning integrity failure modes include:
- spontaneous reasoning divergence
- oscillatory reasoning loops
- semantic boundary collapse
- unintended contextual expansion
- instability under minimal input variations
- conflicting reasoning steps
- multi-agent interference anomalies
- degradation over time or workload
These failure modes are formalized later in the RIS risk models and evaluation methodology.
RIS Levels
4. RIS Levels
4.1 Overview
The Reasoning Integrity Standard (RIS) defines five levels of reasoning integrity maturity, ranging from RIS-0 (uncontrolled reasoning) to RIS-4 (high-integrity, production-grade reasoning). These levels provide a structured framework for evaluating the stability, predictability, and reliability of reasoning produced by large language models (LLMs) and agentic systems.
RIS levels classify systems based on measurable reasoning behavior, not model size, architecture, vendor, or training methodology. A system's RIS level reflects its demonstrated ability to maintain reasoning integrity under controlled conditions, repeated evaluations, and operational scenarios.
4.2 Level Definitions
RIS defines the following reasoning integrity maturity levels:
- RIS-0: Uncontrolled Reasoning
- RIS-1: Drift-Sensitive Reasoning
- RIS-2: Semi-Stable Reasoning
- RIS-3: Controlled Reasoning
- RIS-4: High-Integrity Reasoning
Each level includes required properties, performance characteristics, and failure thresholds.
4.3 RIS-0: Uncontrolled Reasoning
Description
RIS-0 systems exhibit unpredictable, highly variable, or unstable reasoning behavior. Reasoning chains diverge significantly across repeated prompts, and semantic coherence is inconsistent or unreliable.
Characteristics
- High variance across repeated inferences
- Unstable or incoherent reasoning chains
- Susceptibility to minor input or environmental perturbations
- Unbounded or uncontrolled reasoning state transitions
- No enforceable variance envelope
- Reasoning behavior cannot be reliably reproduced
Typical Use Cases
- exploratory research models
- early-stage prototypes
- unconstrained generative systems
Conformance
RIS-0 indicates non-conformance with all RIS requirements.
4.4 RIS-1: Drift-Sensitive Reasoning
Description
RIS-1 systems demonstrate partial reasoning stability but remain sensitive to drift, perturbations, and repeated evaluations. Structural or semantic divergence occurs frequently.
Characteristics
- partially stable reasoning patterns under ideal conditions
- significant drift under repetition or minimal variation
- inconsistent semantic alignment
- no enforceable drift boundaries
- limited reproducibility of reasoning behavior
Typical Use Cases
- non-safety-critical applications
- general-purpose conversational models
- agent prototypes
Conformance
RIS-1 indicates minimal partial alignment with RIS controls but does not meet baseline conformance.
4.5 RIS-2: Semi-Stable Reasoning
Description
RIS-2 systems achieve basic reasoning integrity. Reasoning chains show moderate consistency across repeated prompts and operate within a loosely defined variance range. Drift is detectable but bounded.
Characteristics
- moderate chain stability
- acceptable semantic coherence under repetition
- drift remains within a broad variance envelope
- occasional reasoning divergence under perturbation
- reproducible reasoning behavior in controlled cases
Typical Use Cases
- consumer-facing AI applications
- low-risk enterprise workflows
- model evaluation and benchmarking environments
Conformance
RIS-2 meets some baseline RIS controls but does not satisfy the requirements for production-grade reasoning integrity.
4.6 RIS-3: Controlled Reasoning
Description
RIS-3 systems demonstrate controlled reasoning behavior suitable for production environments. Reasoning chains are stable across repetitions, drift remains within defined boundaries, and semantic coherence is predictably maintained.
Characteristics
- consistent chain stability across repeated prompts
- defined and enforceable variance envelope
- predictable reasoning under controlled perturbations
- bounded semantic drift and minimal divergence
- reproducible reasoning state transitions
- resistance to tool and agent interference within expected parameters
Typical Use Cases
- enterprise-grade LLM applications
- internal decision-support systems
- workflow automation
- multi-agent orchestration with defined boundaries
Conformance
RIS-3 satisfies the majority of RIS requirements and is considered the minimum level appropriate for regulated or production-grade environments.
4.7 RIS-4: High-Integrity Reasoning
Description
RIS-4 represents the highest level of reasoning integrity. Systems at this level produce stable, predictable, and coherent reasoning patterns with minimal variance across repeated evaluations. Drift is tightly controlled, and reasoning chains maintain structural integrity under both controlled and adversarial conditions.
Characteristics
- high stability and coherence across all evaluations
- tightly bounded variance envelope
- minimal drift under repetition or perturbation
- predictable reasoning structure even with minor input variations
- resilience to interference from tools, retrieval components, or agents
- strong governance boundary adherence
- reproducible reasoning consistent with defined expectations
Typical Use Cases
- safety-critical systems
- regulated industries requiring auditability
- financial, medical, legal, and security reasoning workflows
- autonomous agent frameworks with strict stability demands
- enterprise systems requiring deterministic reasoning behavior
Conformance
RIS-4 conforms to all RIS requirements and demonstrates the highest degree of reasoning integrity defined by the standard.
4.8 Level Assignment Requirements
A system's RIS level MUST be assigned based on:
- performance across the RIS measurement framework
- compliance with required RIS controls
- observed behavior during evaluation methodology test suites
- documented evidence demonstrating stability and drift characteristics
Level assignment SHALL be determined from empirical results, not vendor claims or model size.
4.9 Level Validation and Reassessment
Systems evaluated for RIS conformance MUST undergo periodic reassessment based on:
- operational changes
- model updates or retraining
- architectural changes
- new tool or agent integrations
- changes in governance boundaries
- drift trends observed in production
Reassessment MAY elevate or reduce a system's RIS level depending on observed results.
Measurement Framework
5. Measurement Framework
5.1 Overview
The RIS Measurement Framework defines the quantitative and qualitative methods used to evaluate reasoning integrity. It provides the metrics, data collection procedures, statistical thresholds, and evaluation structures necessary to determine a system's RIS level.
The framework applies to all systems subject to RIS evaluation and SHALL be used consistently across:
- baseline assessments
- production monitoring
- audit reviews
- comparative benchmarking
Measurement MUST be based on empirical evidence. Qualitative assessments may supplement measurements but SHALL NOT replace them.
5.2 Measurement Categories
RIS evaluates reasoning integrity using five primary measurement categories:
- Chain Stability
- Semantic Coherence
- Drift Sensitivity
- Variance Envelope Compliance
- Governance Boundary Adherence
All categories MUST be assessed to determine RIS level.
5.3 Required Inputs
The following inputs are required for RIS-compliant measurement:
- repeated inference samples from identical prompts
- controlled perturbation samples (minimal input variations)
- environmental invariants (temperature, sampling parameters)
- contextual invariants (system instructions, agent state)
- tool or retrieval call responses (if applicable)
- logs or ledger entries representing reasoning behavior
Implementations SHOULD collect evaluation data under isolated and consistent conditions.
5.4 Sample Size Requirements
A compliant RIS evaluation MUST include:
- a minimum of 25 repeated inference samples per test prompt
- a minimum of 10 controlled perturbation samples per prompt
- at least 50 total prompts across diverse categories for generalized models
- at least 20 prompts for domain-specific or task-specific systems
Larger sample sizes MAY be used but SHALL be documented.
5.5 Chain Stability Metric
5.5.1 Definition
Chain stability measures the consistency of a system’s reasoning structure across repeated inference cycles.
5.5.2 Measurement Method
Stability is measured using:
- structural similarity metrics
- step-to-step coherence comparisons
- reasoning-length variance
- transition-alignment scoring
A system MUST maintain chain similarity within its defined variance envelope.
5.5.3 Scoring
Chain stability is scored from 0.00 to 1.00.
Example thresholds:
- 0.85-1.00: stable
- 0.65-0.84: semi-stable
- below 0.65: unstable
RIS-4 systems SHALL achieve stability above 0.90 for all core prompts.
5.6 Semantic Coherence Metric
5.6.1 Definition
Semantic coherence measures the consistency of meaning and conceptual alignment across repeated inferences.
5.6.2 Measurement Method
Coherence is evaluated based on:
- semantic similarity scoring
- topic retention
- conceptual transition consistency
- negation or contradiction detection
5.6.3 Scoring
Scoring follows the same 0.00-1.00 scale.
RIS-4 systems SHALL demonstrate coherence above 0.90.
5.7 Drift Sensitivity Metric
5.7.1 Definition
Drift sensitivity measures the degree to which a system deviates from baseline reasoning behavior under repeated evaluations or minimal perturbations.
5.7.2 Measurement Method
Drift is assessed by:
- analyzing variance outside baseline patterns
- measuring semantic or structural deviation trends
- quantifying divergence under perturbation conditions
5.7.3 Scoring
Drift sensitivity is scored inversely (lower is better):
- 0.00-0.10: low drift
- 0.11-0.25: moderate drift
- above 0.25: high drift
RIS-4 systems SHALL maintain drift sensitivity below 0.10.
5.8 Variance Envelope Compliance
5.8.1 Definition
The variance envelope represents the acceptable range of variation in reasoning behavior.
5.8.2 Measurement Method
Variance envelope compliance is determined by:
- repeated inference testing
- perturbation evaluations
- longitudinal measurement
- stability under load or repeated requests
5.8.3 Scoring
A system is considered compliant if:
- at least 95 percent of evaluations fall within its defined envelope
RIS-4 systems SHALL achieve at least 98 percent compliance.
5.9 Governance Boundary Adherence
5.9.1 Definition
Governance boundaries constrain the context, semantic scope, and operational state the model may use for reasoning.
5.9.2 Measurement Method
Boundary adherence is measured by observing:
- context window expansion attempts
- unauthorized semantic domain expansion
- tool or agent influence patterns
- cross-task contamination or memory leakage
5.9.3 Scoring
Boundary violations are scored as:
- 0: no violations
- 1: one minor violation
- 2: repeated or significant violations
RIS-4 systems SHALL record zero violations under all evaluation conditions.
5.10 Composite Reasoning Integrity Score
RIS defines an aggregate score calculated using weighted metrics:
- Chain Stability: 30 percent
- Semantic Coherence: 25 percent
- Drift Sensitivity: 20 percent
- Variance Envelope Compliance: 15 percent
- Governance Boundary Adherence: 10 percent
Composite scores determine candidate RIS levels but do not replace conformance requirements.
5.11 Interpretation of Scores
Composite score ranges:
- 0.00-0.40: RIS-0
- 0.41-0.60: RIS-1
- 0.61-0.75: RIS-2
- 0.76-0.89: RIS-3
- 0.90-1.00: RIS-4
A system MUST meet both:
- composite scoring thresholds
- all mandatory controls applicable to its level
to be classified at that RIS level.
Control Families
6. RIS Controls
6.1 Overview
RIS controls define the mandatory and recommended requirements that systems MUST meet to achieve compliance with RIS levels RIS-1 through RIS-4. Controls are grouped into control families that govern stability, coherence, drift, variance, boundary adherence, and operational integrity.
Each control includes:
- Control Objective
- Control Requirement (MUST / SHALL)
- Implementation Guidance (informative)
- Evidence Required (for audits)
- Conformance Criteria
All systems seeking RIS classification SHALL be evaluated against these controls.
6.2 Control Families
RIS defines six control families:
- RS - Chain Stability Controls
- SC - Semantic Coherence Controls
- DR - Drift Resistance Controls
- VE - Variance Envelope Controls
- GB - Governance Boundary Controls
- OP - Operational Integrity Controls
Controls within each family are numbered sequentially.
6.3 Chain Stability Controls (RS)
RS-1: Repetition Consistency
Objective
Ensure reasoning chains remain consistent across repeated inference cycles.
Requirement
The system SHALL produce reasoning outputs whose structural similarity remains within the defined variance envelope across repeated evaluations of identical prompts.
Implementation Guidance
- Use repeated inference tests.
- Monitor reasoning structure length and transitions.
- Track deviations across a defined sample size.
Evidence Required
- stability reports
- similarity metrics
- repeated inference samples
Conformance
Required for RIS-2, RIS-3, and RIS-4.
RS-2: Perturbation Stability
Objective
Ensure stability under minor, controlled input variations.
Requirement
The system SHALL maintain reasoning structure stability when presented with minimal, semantically equivalent perturbations of the same prompt.
Implementation Guidance
- Use controlled perturbation prompts (synonym replacement, minor rephrasing, neutral variations).
- Measure structural drift across perturbation samples.
Evidence Required
- perturbation test logs
- drift measurements
- variance envelope comparison
Conformance
Required for RIS-3 and RIS-4.
RS-3: Transition Alignment
Objective
Ensure internal reasoning transitions follow consistent logical patterns.
Requirement
The system SHALL maintain alignment of step-to-step transitions across repeated inference cycles.
Implementation Guidance
- Compare reasoning step ordering or inferred conceptual transitions.
- Identify anomalous reordering indicative of instability.
Evidence Required
- transition-level similarity reports
Conformance
Required for RIS-4.
6.4 Semantic Coherence Controls (SC)
SC-1: Semantic Alignment
Objective
Ensure that reasoning outputs remain semantically consistent across repeated evaluations.
Requirement
The system SHALL maintain semantic similarity across repeated inferences meeting or exceeding the thresholds defined in the measurement framework.
Implementation Guidance
- Use embedding-based similarity scoring.
- Detect contradictory or negated reasoning shifts.
Evidence Required
- semantic similarity analysis
- contradiction or negation detection logs
Conformance
Required for RIS-2, RIS-3, and RIS-4.
SC-2: Concept Retention
Objective
Ensure conceptual themes persist across reasoning processes.
Requirement
The system SHALL retain core conceptual structures across evaluations unless explicitly instructed to change.
Implementation Guidance
- Analyze topic retention through semantic clustering.
- Evaluate for unexpected topic divergence.
Evidence Required
- topic coherence reports
- clustering analysis results
Conformance
Required for RIS-3 and RIS-4.
SC-3: Logical Consistency
Objective
Ensure logical validity and structure remain consistent.
Requirement
The system SHALL avoid contradictory intermediate reasoning steps within an evaluation set.
Implementation Guidance
- Evaluate internal reasoning structure for contradictions.
- Use logical consistency scoring where available.
Evidence Required
- logical consistency assessments
Conformance
Required for RIS-4.
6.5 Drift Resistance Controls (DR)
DR-1: Drift Monitoring
Objective
Establish continuous monitoring for reasoning drift.
Requirement
The system SHALL implement monitoring to detect drift exceeding defined thresholds over repeated inference cycles.
Implementation Guidance
- Compare baseline outputs against current samples.
- Establish early-warning indicators for drift escalation.
Evidence Required
- drift baseline logs
- drift trend reports
Conformance
Required for RIS-2, RIS-3, and RIS-4.
DR-2: Drift Envelope Enforcement
Objective
Maintain reasoning behavior within allowable drift boundaries.
Requirement
The system SHALL enforce a drift envelope that constrains allowable deviation from baseline behavior.
Implementation Guidance
- Use variance and drift sensitivity metrics.
- Define acceptable drift bands based on use case.
Evidence Required
- drift envelope documentation
- variance compliance reports
Conformance
Required for RIS-3 and RIS-4.
DR-3: Drift Recovery
Objective
Ensure recovery mechanisms exist for restoring reasoning integrity.
Requirement
The system SHALL provide mechanisms to reinitialize or reset reasoning state when drift exceeds permitted thresholds.
Implementation Guidance
- Reset context or agent state.
- Refresh memory or tool caches.
- Re-evaluate under controlled conditions.
Evidence Required
- drift recovery logs
- anomaly detection events
Conformance
Required for RIS-4.
6.6 Variance Envelope Controls (VE)
VE-1: Variance Definition
Objective
Define acceptable variance for reasoning behavior.
Requirement
The system SHALL define a variance envelope specifying acceptable deviation thresholds across repeated evaluations.
Implementation Guidance
- Document variance thresholds.
- Derive thresholds from benchmarking or domain requirements.
Evidence Required
- variance envelope documentation
Conformance
Required for RIS-2, RIS-3, and RIS-4.
VE-2: Variance Compliance
Objective
Ensure reasoning outputs remain within defined variance bounds.
Requirement
The system SHALL maintain variance compliance for at least 95 percent of repeated evaluations.
Implementation Guidance
- Use repeated inference tests.
- Monitor patterns of deviation.
Evidence Required
- compliance reports
- exception logs
Conformance
Required for RIS-3 and RIS-4.
VE-3: Variance Tightening
Objective
Tighten variance thresholds for high-integrity reasoning.
Requirement
The system SHALL maintain variance compliance at rates exceeding 98 percent for RIS-4.
Implementation Guidance
- Use stricter similarity or coherence scoring.
- Implement adaptive variance tightening.
Evidence Required
- high-precision variance reports
Conformance
Required for RIS-4.
6.7 Governance Boundary Controls (GB)
GB-1: Boundary Definition
Objective
Define explicit reasoning boundaries to prevent uncontrolled contextual expansion.
Requirement
The system SHALL define governance boundaries that constrain contextual, semantic, or operational state during reasoning.
Implementation Guidance
- Define allowable context limits.
- Identify prohibited expansions or semantic domains.
Evidence Required
- governance boundary documentation
Conformance
Required for RIS-2, RIS-3, and RIS-4.
GB-2: Boundary Enforcement
Objective
Ensure reasoning remains within defined boundaries.
Requirement
The system SHALL enforce boundaries during inference and detect violations.
Implementation Guidance
- Use contextual or semantic boundary checks.
- Restrict access to tools or memory beyond defined limits.
Evidence Required
- boundary enforcement logs
- violation reports
Conformance
Required for RIS-3 and RIS-4.
GB-3: Interference Protection
Objective
Prevent uncontrolled influence from tools, agents, or retrieval components.
Requirement
The system SHALL detect and mitigate reasoning interference from tool responses or agent interactions.
Implementation Guidance
- Validate tool outputs before integration.
- Isolate reasoning states across agents.
Evidence Required
- interference detection logs
- impact assessments
Conformance
Required for RIS-4.
6.8 Operational Integrity Controls (OP)
OP-1: Evaluation Environment Consistency
Objective
Ensure evaluation environments remain stable and controlled.
Requirement
The system SHALL maintain consistent parameters (temperature, sampling settings, context) during evaluation.
Evidence Required
- environment configuration logs
Conformance
Required for all RIS levels beyond RIS-0.
OP-2: Reproducible Testing
Objective
Ensure evaluation results can be reproduced.
Requirement
The system SHALL support reproducible testing conditions and retain configuration details for audit purposes.
Evidence Required
- reproducibility reports
- configuration snapshots
Conformance
Required for RIS-3 and RIS-4.
OP-3: Audit Logging
Objective
Ensure auditable logs for evaluation and monitoring.
Requirement
The system SHALL generate logs documenting reasoning behavior, drift events, variance compliance, and boundary violations.
Evidence Required
- audit logs
- ledger entries
- summary reports
Conformance
Required for RIS-4.
OP-4: Longitudinal Monitoring
Objective
Track reasoning behavior over time.
Requirement
The system SHALL support longitudinal monitoring to detect degradation or emerging instability.
Evidence Required
- multi-interval reports
- trend analyses
Conformance
Required for RIS-4.
6.9 Summary of Conformance Requirements by RIS Level
| Control Family | RIS-1 | RIS-2 | RIS-3 | RIS-4 |
|---|---|---|---|---|
| RS (Stability) | Optional | Required | Required | Required (strict) |
| SC (Coherence) | Optional | Required | Required | Required (strict) |
| DR (Drift) | Optional | Required | Required | Required (strict) |
| VE (Variance) | Optional | Required | Required | Required (tight) |
| GB (Boundary) | Optional | Required | Required | Required (strict) |
| OP (Operational) | Minimal | Moderate | Strong | Full (audit-ready) |
6.10 Control Interpretation
Control interpretation SHALL comply with the following:
- MUST and SHALL controls are required for conformance.
- SHOULD controls are recommended but not required unless explicitly tied to a level.
- MAY controls provide flexibility but do not affect conformance.
Scoring and Conformance
7. Scoring and Conformance
7.1 Overview
This section defines the scoring methodology and conformance criteria used to determine a system's RIS level. Scoring integrates quantitative metrics from the measurement framework with mandatory control requirements. A system's RIS classification is based on both:
- composite reasoning integrity score, and
- demonstrated compliance with required RIS controls.
A system SHALL NOT be assigned a RIS level solely based on scoring metrics without meeting corresponding control requirements.
7.2 Composite Reasoning Integrity Score
The composite score is calculated using weighted metrics defined in Section 5:
- Chain Stability: 30 percent
- Semantic Coherence: 25 percent
- Drift Sensitivity: 20 percent
- Variance Envelope Compliance: 15 percent
- Governance Boundary Adherence: 10 percent
Each metric yields a value between 0.00 and 1.00.
The weighted sum produces the final composite reasoning integrity score.
The composite score is used to determine eligibility for a RIS level, but SHALL NOT be used to determine final classification without control verification.
7.3 Metric Weighting
The rationale for metric weighting is as follows:
- Chain Stability (30 percent): foundational indicator of structural reasoning reliability.
- Semantic Coherence (25 percent): ensures reasoning maintains consistent meaning.
- Drift Sensitivity (20 percent): predicts long-term stability under repetition.
- Variance Envelope Compliance (15 percent): ensures bounded divergence.
- Governance Boundary Adherence (10 percent): required for operational trust and safety.
These weights SHALL remain consistent across RIS versions unless updated in a future amendment.
7.4 Eligibility Thresholds for RIS Levels
Eligibility thresholds based on composite score:
- 0.00-0.40: RIS-0
- 0.41-0.60: RIS-1
- 0.61-0.75: RIS-2
- 0.76-0.89: RIS-3
- 0.90-1.00: RIS-4
Meeting the scoring threshold does not guarantee assignment to that RIS level.
The system SHALL also satisfy the control requirements designated for that level.
7.5 Control-Based Conformance Requirements
The following table defines minimum control conformance for each RIS level:
| RIS Level | Required Controls |
|---|---|
| RIS-0 | No controls required |
| RIS-1 | Optional controls only |
| RIS-2 | All RS-1, SC-1, DR-1, VE-1, GB-1 controls |
| RIS-3 | All RIS-2 controls plus RS-2, SC-2, DR-2, VE-2, GB-2, OP-2 |
| RIS-4 | All RIS-3 controls plus RS-3, SC-3, DR-3, VE-3, GB-3, OP-3, OP-4 |
A system SHALL NOT be classified above RIS-2 unless all mandatory controls for RIS-3 are satisfied.
A system SHALL NOT be classified above RIS-3 unless all mandatory controls for RIS-4 are satisfied.
7.6 Evidence Requirements
To support a RIS classification, the following evidence MUST be produced:
- Measurement Framework Results
- repeated inference sample outputs
- controlled perturbation outputs
- variance compliance reports
- drift sensitivity analyses
- Control Implementation Evidence
- documentation of stability, coherence, drift, variance, and boundary controls
- operational integrity documentation
- governance boundary definitions
- Audit Logs
- reasoning ledger entries or equivalent reports
- evaluation logs
- anomaly and boundary violation logs
- Configuration Snapshots
- inference parameters
- environmental settings
- prompt templates
- tool access or agent configuration
The evaluator SHALL verify all evidence before assigning a RIS level.
7.7 Conformance Determination Process
A system SHALL undergo the following steps to determine conformance:
- Preparation
- define evaluation conditions
- document configuration
- establish invariants
- Data Collection
- perform repeated inference sampling
- perform controlled perturbation sampling
- log reasoning behavior
- Metric Calculation
- compute each metric independently
- calculate composite reasoning integrity score
- Control Verification
- confirm implementation and operation of required controls
- identify gaps or violations
- Level Assignment
- determine eligible RIS level based on composite score
- determine achievable RIS level based on control compliance
- final level SHALL be the lower of the two
- Report Generation
- produce a formal evaluation report
- document findings, score, control status, and assignment
- Certification
- issue a RIS Conformance Statement
- assign classification validity period
7.8 Handling of Violations and Exceptions
If a system violates a mandatory control during evaluation:
- it SHALL be classified at the highest RIS level for which all mandatory controls are satisfied
- violations MUST be documented with supporting evidence
- systems MAY undergo remediation and reassessment
Exceptions for SHOULD and MAY controls SHALL NOT affect level assignment if mandatory controls are met.
7.9 Reassessment Requirements
RIS conformance SHALL be reassessed when:
- the underlying model is updated
- training data or fine-tuning changes
- new tools or agents are added
- governance boundaries change
- drift or instability trends are detected in production
- a significant degradation in metrics occurs
- system behavior changes under load
Reassessment MAY be performed more frequently based on organizational or regulatory requirements.
7.10 Validity of RIS Classification
RIS classification is valid for a period of:
- 12 months for general LLM systems
- 6 months for agentic or tool-integrated systems
- 3 months for safety-critical systems
After expiration, the system SHALL be reevaluated to maintain a RIS classification.
7.11 Publication and Reporting
Organizations MAY publish:
- their RIS level
- their composite score
- high-level findings
However, detailed logs, inference samples, and proprietary evaluation data SHOULD only be disclosed under NDA or regulatory requirement.
RIS conformance SHALL NOT be claimed without formal assessment.
Risk Models
8. Risk Models
8.1 Overview
This section defines the risk models used to evaluate, categorize, and mitigate threats to reasoning integrity in LLMs and agentic systems. RIS identifies reasoning risks that impact stability, coherence, predictability, and boundary adherence. Risk models are used to:
- guide evaluation methodology
- inform control requirements
- support audit and governance decisions
- predict failure conditions
- align reasoning-system behavior with organizational risk tolerance
RIS risk models are independent of data privacy, security, ethical, or regulatory risk. They focus exclusively on reasoning integrity.
8.2 Categories of Reasoning Risk
RIS defines six primary categories of reasoning risk:
- Structural Drift Risk
- Semantic Drift Risk
- Temporal Stability Risk
- Interference Risk
- Boundary Violation Risk
- Degradation Risk
Each risk category is described in the following sections.
8.3 Structural Drift Risk
8.3.1 Description
Structural drift occurs when the underlying structure, sequence, or transitions of reasoning chains changes beyond acceptable variance thresholds.
8.3.2 Causes
- sampling randomness beyond expected levels
- sensitivity to minor prompt variations
- instability in internal representations
- tool or agent load effects
- model configuration fluctuations
8.3.3 Indicators
- inconsistent step ordering
- fluctuating reasoning length
- divergent intermediate transitions
- unexplained reorganization of reasoning steps
8.3.4 Impact
Structural drift undermines reproducibility and increases the likelihood of unpredictable system behavior.
8.3.5 Mitigation
- enforce variance envelope
- enforce chain stability controls
- reset or reinitialize agent state when drift is detected
- maintain consistent inference parameters
8.4 Semantic Drift Risk
8.4.1 Description
Semantic drift arises when the meaning, conceptual alignment, or topic structure shifts unpredictably during or across reasoning cycles.
8.4.2 Causes
- conceptual instability under repetition
- semantic domain expansion
- interference from retrieval or tools
- failure to retain core conceptual anchors
8.4.3 Indicators
- loss of topic relevance
- contradictory intermediate steps
- unintended meaning shifts
- negation or inversion anomalies
8.4.4 Impact
Semantic drift reduces coherence, predictability, and trustworthiness of reasoning outputs.
8.4.5 Mitigation
- semantic coherence controls
- conceptual retention verification
- governance boundary enforcement
8.5 Temporal Stability Risk
8.5.1 Description
Temporal stability risk refers to degradation in reasoning quality over time or across extended inference sequences.
8.5.2 Causes
- accumulated internal noise
- memory contamination
- incremental divergence across long chains
- resource fluctuations
8.5.3 Indicators
- gradual decline in coherence
- increasing drift across intervals
- failure to maintain reasoning state consistency
8.5.4 Impact
Long-duration applications may become unreliable, particularly in agentic or multi-step workflows.
8.5.5 Mitigation
- longitudinal monitoring
- periodic state resets
- consistency checkpoints
8.6 Interference Risk
8.6.1 Description
Interference risk arises when external components affect reasoning integrity.
Sources of interference include:
- tools
- retrieval components
- other agents
- environment APIs
- context or memory systems
8.6.2 Causes
- misaligned tool responses
- unbounded agent interactions
- unpredictable RAG data injection
- multi-agent reasoning collisions
8.6.3 Indicators
- sudden reasoning divergence
- tool-dependent instability
- inconsistent decisions after tool use
- agent cross-contamination
8.6.4 Impact
Interference may cause unpredictable reasoning patterns, loss of boundary control, and uncontrolled drift.
8.6.5 Mitigation
- interference detection
- boundary enforcement controls
- isolation of reasoning state
- validation of tool outputs
8.7 Boundary Violation Risk
8.7.1 Description
Boundary violation risk occurs when the system exceeds allowed contextual, semantic, or operational boundaries.
8.7.2 Causes
- context window expansion
- unauthorized semantic domain exploration
- cross-task memory leakage
- overreliance on external tools
- multi-agent interference
8.7.3 Indicators
- references to disallowed information
- use of context not present in the input
- reasoning dependent on external unintended state
- creeping expansion of reasoning scope
8.7.4 Impact
Boundary violations compromise predictability, auditability, and operational safety.
8.7.5 Mitigation
- boundary definition and enforcement controls
- monitoring for out-of-bound semantic transitions
- strict tool and agent constraints
8.8 Degradation Risk
8.8.1 Description
Degradation risk refers to progressive reasoning quality decline over time due to repeated use, extended agent loops, environmental factors, or model instability.
8.8.2 Causes
- prolonged agentic sessions
- heavy tool usage
- accumulated contextual noise
- fluctuating infrastructure conditions
8.8.3 Indicators
- increasing drift trends
- decreasing coherence metrics
- longer or shorter reasoning chains without cause
- gradual divergence across evaluation windows
8.8.4 Impact
Degradation affects reliability in long-running systems, especially in automation and orchestration applications.
8.8.5 Mitigation
- longitudinal stability tracking
- state resets
- performance baselining
- proactive drift sensitivity checks
8.9 Risk Levels
RIS classifies reasoning risks into three levels:
- Low: Within established variance and stability thresholds
- Moderate: Deviations observed but within safe operational limits
- High: Significant deviations requiring remediation
Risk levels SHALL be documented in the conformance report.
8.10 Relationship Between Risks and RIS Levels
Each RIS level carries a corresponding risk tolerance:
- RIS-0: All risks high
- RIS-1: Most risks moderate to high
- RIS-2: Risks moderate with limited high-risk events
- RIS-3: All risks low to moderate; no high-risk events allowed
- RIS-4: All risks low; strict controls required
A system SHALL NOT qualify for RIS-3 or RIS-4 if any high-risk category is identified.
8.11 Use of Risk Models in Evaluation
During evaluation, assessors SHALL:
- identify risk categories triggered
- document causes and indicators
- evaluate drift and stability trends
- map observed behavior to risk severity
- determine required mitigation for higher RIS levels
Risk models inform conformance decisions but do not replace mandatory controls or scoring.
8.12 Use of Risk Models in Production Monitoring
Organizations SHOULD use risk models to:
- detect early warning signals
- anticipate degradation trends
- identify tool-induced instability
- evaluate multi-agent interference
- maintain operational safety
Risk-based monitoring SHALL supplement RIS control compliance.
Evaluation Methodology
9. Evaluation Methodology
9.1 Overview
The RIS evaluation methodology defines the standardized procedures for performing a reasoning integrity assessment. These procedures ensure that measurements are:
- repeatable
- objective
- statistically meaningful
- comparable across systems
- suitable for audit and certification
All evaluations MUST follow the methodology in this section to be considered valid for RIS classification.
9.2 Evaluation Objectives
The objectives of RIS evaluation are to:
- quantify reasoning stability and coherence
- measure drift and variance across repeated evaluations
- detect boundary violations
- assess interference sensitivity
- validate compliance with required controls
- classify the system into an appropriate RIS level
Evaluation SHALL NOT assess factual correctness or ethical outcomes. Only reasoning integrity is measured.
9.3 Evaluation Conditions
9.3.1 Environmental consistency
Evaluations SHALL be conducted under consistent conditions, including:
- model version
- inference parameters (temperature, top-p, settings)
- context and prompt format
- agent or tool configurations
- hardware or infrastructure environment
Changes to any parameter invalidate comparisons.
9.3.2 Isolation
The evaluation environment SHOULD minimize:
- interference from concurrent workloads
- environmental noise
- model adaptation or memory building
- caching effects
9.3.3 Documentation
The evaluator SHALL document:
- all inference parameters
- all environment variables
- system settings used for assessment
- test prompt sets
9.4 Prompt Set Requirements
RIS evaluation uses two types of prompts:
- Baseline prompts
- Perturbation prompts
9.4.1 Baseline prompts
Baseline prompts SHOULD include:
- informational queries
- reasoning tasks
- multi-step tasks
- conceptual analysis tasks
- domain-neutral questions
At least 50 prompts SHOULD be used for general-purpose systems.
At least 20 prompts SHOULD be used for specialized systems.
9.4.2 Perturbation prompts
Perturbation prompts MUST maintain semantic equivalence with baseline prompts while varying:
- syntax
- structure
- ordering
- minor phrasing
- neutral substitutions
A minimum of 10 perturbations SHALL be used per baseline prompt selected for perturbation testing.
9.5 Sampling Methodology
9.5.1 Repeated inference sampling
For each baseline prompt:
- at least 25 repeated inference samples SHALL be collected
- sampling parameters SHALL remain identical across all repetitions
Repeated inference sampling is used to compute:
- chain stability
- semantic coherence
- drift sensitivity
- variance envelope compliance
9.5.2 Controlled perturbation sampling
For each perturbation prompt:
- at least 10 inference samples SHALL be collected
Perturbation sampling is used to measure:
- structural drift
- semantic drift
- robustness under variation
9.6 Baseline Establishment
9.6.1 Creating a baseline
The evaluator SHALL establish a baseline by:
- selecting representative baseline prompts
- executing repeated inference sampling
- computing structural and semantic similarity metrics
- determining drift sensitivity
- defining acceptable variance envelopes
9.6.2 Baseline validity
A baseline is valid only if:
- all evaluation conditions are documented
- system configuration remains unchanged
- variance envelope is statistically meaningful
Baseline MUST be refreshed when model parameters change or risk models trigger reassessment.
9.7 Drift Evaluation
9.7.1 Drift detection
Drift is detected by:
- comparing new samples to baseline
- analyzing deviation trends
- identifying abrupt or progressive divergence
- checking for consistent out-of-envelope behavior
9.7.2 Drift characterization
Drift SHALL be categorized as:
- structural
- semantic
- temporal
- interference-based
9.7.3 Drift thresholds
A drift event occurs when:
- deviation exceeds the envelope for more than 5 percent of samples
- repeated deviations occur during perturbation sampling
- temporal divergence increases across sequential windows
9.7.4 Drift escalation
If drift is detected:
- the system’s RIS level MAY be lowered
- additional testing SHALL be performed
- drift controls MUST be verified
9.8 Boundary Evaluation
Boundary adherence SHALL be tested by:
- introducing tasks with clear semantic boundaries
- evaluating attempts to use disallowed context
- testing reasoning behavior with restricted tool outputs
- detecting cross-task or cross-context contamination
Any violation SHALL be logged and considered in level classification.
9.9 Interference Evaluation
Interference evaluation includes:
- testing reasoning stability before and after tool calls
- measuring agent-to-agent interaction effects
- injecting controlled retrieval variability
- observing structural or semantic disruptions
Interference MUST not cause uncontrolled reasoning divergence for systems seeking RIS-3 or RIS-4 level classification.
9.10 Variance Envelope Evaluation
Variance envelope compliance SHALL be tested through:
- repeated inference sampling
- perturbation sampling
- load or stress conditions if applicable
- periodic resampling across intervals
A system MUST maintain compliance for:
- at least 95 percent of samples (RIS-3)
- at least 98 percent of samples (RIS-4)
9.11 Evaluation Report Requirements
A complete evaluation report SHALL include:
- System description
- Evaluation environment details
- Prompt sets used
- Sampling methodology
- Baseline metrics
- Drift analysis
- Variance envelope results
- Boundary adherence results
- Interference results
- Composite score calculation
- Control compliance analysis
- RIS level assigned
- Validity period
- Evidence appendix
The report SHALL be signed by the evaluator.
9.12 Evaluation Validity
An evaluation remains valid only if:
- the system configuration does not change
- no drift events occur in production
- no new tools, agents, or retrieval sources are introduced
- governance boundaries remain constant
If any condition changes, reassessment MAY be required immediately.
9.13 Re-Evaluation Requirements
Re-evaluation SHALL occur:
- at the end of the classification validity period
- when major model updates occur
- when drift or instability is observed
- when new risk categories are triggered
- when evidence suggests deviation from baseline behavior
Systems in regulated or safety-critical environments MAY require more frequent reassessments.
Audit Guidelines
10. Audit Guidelines
10.1 Overview
This section defines the audit requirements, evidence standards, assessment procedures, and validation criteria necessary to verify compliance with the Reasoning Integrity Standard (RIS). These guidelines ensure that evaluations are reliable, reproducible, and suitable for internal review, third-party certification, or regulatory oversight.
An auditor MAY be an internal reviewer, an external assessor, or an automated system. All auditors SHALL adhere to the requirements outlined in this section.
10.2 Audit Objectives
The objectives of a RIS audit are to:
- verify the correctness and completeness of evaluation procedures
- validate metric calculations and composite reasoning integrity scores
- confirm implementation of mandatory controls
- detect deficiencies, violations, or risks
- ensure evidence sufficiency and traceability
- determine whether assigned RIS levels are justified
- produce a formal audit report documenting findings
Audits SHALL NOT evaluate ethical, factual, or domain-specific correctness.
10.3 Audit Scope
The audit SHALL cover:
- evaluation methodology adherence
- system configuration and invariants
- reasoning integrity metrics
- drift and variance analysis
- control compliance
- governance boundary enforcement
- operational integrity
- logged events and ledger entries
- deviation or violation reports
The audit MAY include additional areas depending on organizational, regulatory, or contractual requirements.
10.4 Evidence Requirements
Auditors SHALL review the following types of evidence:
10.4.1 Raw inference samples
- repeated inference outputs
- perturbation outputs
- temporal sampling outputs
- tool-integrated outputs (if applicable)
10.4.2 Metric calculations
- chain stability similarity data
- semantic coherence scores
- drift sensitivity calculations
- variance envelope compliance results
- boundary adherence results
10.4.3 Configuration data
- inference parameters (temperature, top-p, etc.)
- system settings
- agent or tool configurations
- environment variables
- software versions
10.4.4 Control implementation evidence
- governance boundary definitions
- drift envelope definitions
- stability and coherence controls
- interference detection controls
- audit logging mechanisms
10.4.5 Logs and ledger entries
- reasoning ledger (or equivalent)
- boundary violation logs
- drift events
- anomaly detection entries
- operational logs related to inference behavior
Auditors SHALL verify the authenticity and integrity of evidence before drawing conclusions.
10.5 Audit Procedures
10.5.1 Preparation
The auditor SHALL:
- obtain system documentation
- review evaluation methodology used
- validate prompt sets and sampling plans
- confirm frozen inference parameters
- ensure environmental consistency
10.5.2 Evidence examination
The auditor SHALL examine:
- raw samples for drift or instability
- metrics for correctness and reproducibility
- compliance with mandatory controls
- documentation for completeness
10.5.3 Reproduction testing
Auditors MAY reproduce:
- repeated inference sampling
- perturbation sampling
- drift and variance calculations
Reproduction SHALL use identical parameters to those documented in the evaluation.
10.5.4 Violation detection
The auditor SHALL check for:
- drift beyond thresholds
- variance envelope breaches
- boundary violations
- interference anomalies
- control misconfigurations
- unreported failures
Any violation SHALL be documented in the final audit report.
10.6 Compliance Determination
Audit findings SHALL conclude one of the following:
- compliant
- compliant with exceptions
- non-compliant
- inconclusive (insufficient evidence)
If non-compliant, the auditor SHALL document which controls or evaluation procedures were violated.
A RIS level MAY only be confirmed if:
- composite score qualifies
- mandatory controls are satisfied
- evidence is complete
- no unmitigated high-risk deviations are observed
10.7 Audit Reporting Requirements
The audit report SHALL include:
- System overview
- Evaluation conditions
- Summary of controls reviewed
- Evidence examined
- Metric verification results
- Observed violations or exceptions
- Assessment of drift, variance, and boundary adherence
- Conformance determination
- Confirmed RIS level
- Recommendations and corrective actions
- Validity period of assessment
- Appendices with supporting data
Reports SHALL be retained for the duration of the classification validity period.
10.8 Corrective Action Requirements
If deficiencies or violations are identified:
- corrective actions SHALL be documented
- the system owner SHALL remediate issues
- the system SHALL undergo partial or full reassessment
- major issues SHALL reset the RIS classification
Corrective actions MAY include:
- adjusting inference parameters
- redefining governance boundaries
- improving drift or variance controls
- updating tool or agent behavior
10.9 Auditor Qualifications
RIS audits MAY be conducted by:
- internal governance teams
- independent third-party evaluators
- qualified researchers
- automated auditing systems
Auditors SHOULD possess expertise in:
- LLM inference behavior
- statistical evaluation
- reasoning-chain analysis
- AI governance and risk frameworks
- tool, agent, or RAG system integrations (if applicable)
Organizations MAY impose additional requirements for auditor certification.
10.10 Continuous Audit Recommendations
Organizations SHOULD implement ongoing audit-aligned practices:
- periodic drift checks
- rolling variance sampling
- boundary adherence monitoring
- interference detection and logging
- longitudinal stability assessments
Continuous monitoring SHALL supplement, not replace, formal RIS audits.
10.11 Audit Validity
An audit is valid for:
- 12 months for general systems
- 6 months for agentic or tool-integrated systems
- 3 months for safety-critical systems
Audits SHALL be invalidated immediately if:
- the model or agent is updated
- new tools, retrieval sources, or agents are introduced
- significant drift is observed in production
- boundary violations occur repeatedly
10.12 Audit Transparency
Organizations MAY disclose:
- confirmed RIS level
- summary of findings
- audit date and validity period
Organizations SHOULD NOT publicly disclose:
- raw inference samples
- proprietary metrics
- internal configurations
- confidential logs
Unless required by contract, regulation, or security review.
Reference Implementation
11. Reference Implementation (LCAC)
11.1 Overview
This section describes the Least-Context Access Control (LCAC) framework as a reference implementation of the Reasoning Integrity Standard (RIS). LCAC demonstrates one practical method for enforcing RIS controls, measuring reasoning integrity, and maintaining predictable reasoning behavior in LLM-based systems.
LCAC is not required for RIS conformance. It is provided solely as an informative reference to illustrate feasibility, implementation patterns, and system design considerations aligned with RIS requirements.
11.2 Architectural Overview
LCAC is structured as a modular reasoning-governance framework consisting of:
- a centralized governor service
- a trust and variance model
- an immutable reasoning ledger
- governance boundary enforcement components
- monitoring and evaluation pipelines
- optional persona overlays
LCAC may operate as an independent service or be integrated directly into an LLM-driven application, multi-agent system, or inference pipeline.
11.3 Components and Functions
11.3.1 Governor
The LCAC governor performs the following functions:
- receives prompt and output pairs for evaluation
- computes trust, variance, and drift metrics
- determines reasoning verdicts (stable, watch, unstable)
- updates governance states (hold, elevate, lockdown)
- writes results to the ledger
The governor serves as the core enforcement point for RIS-aligned controls.
11.3.2 Trust Model
LCAC's trust model provides:
- a baseline trust score
- variance and drift calculations
- predictive stability measurements
- structural and semantic similarity analysis
The trust score is mapped to RIS metrics such as:
- chain stability
- semantic coherence
- drift sensitivity
Trust values range between 0.00 and 1.00, consistent with RIS scoring.
11.3.3 Variance and Drift Envelope
LCAC defines and enforces a variance envelope based on:
- repeated inference samples
- controlled perturbation testing
- deviation thresholds
- longitudinal observations
Metric outputs are compared against these envelopes to determine:
- compliance
- drift events
- stability degradation
11.3.4 Immutable Ledger
LCAC maintains an immutable reasoning ledger using:
- hash-linked entries
- timestamped records
- evaluation metrics
- baseline comparisons
- anomaly documentation
Each ledger entry includes:
- input prompt
- model output
- trust score
- variance metrics
- stability verdict
- previous entry hash
- current entry hash
This supports audit, reproducibility, and traceability.
11.4 Governance Modes
LCAC defines three governance modes that influence reasoning behavior:
11.4.1 Hold
- default system state
- stable or near-stable reasoning
- no significant anomalies detected
11.4.2 Elevate
- moderate drift or variance detected
- reasoning monitored more aggressively
- potential instability anticipated
11.4.3 Lockdown
- reasoning integrity breach detected
- drift or variance above thresholds
- boundary violation or interference event observed
- heightened restrictions applied
Governance modes demonstrate how RIS controls may be operationalized.
11.5 Boundary Enforcement
LCAC establishes governance boundaries using:
- contextual limits
- semantic domain constraints
- tool and agent access restrictions
- validation checks for external outputs
Boundary enforcement aligns with RIS boundary violation controls.
11.6 Persona Overlay (Informative Only)
LCAC includes an optional persona system that applies:
- stability bias adjustments
- risk posture tuning
- domain-specific behavioral constraints
Personas may influence:
- baseline trust
- drift sensitivity
- variance envelopes
- stability thresholds
Persona systems are informative and not required for RIS compliance.
11.7 Integration Patterns
LCAC supports the following integration models:
11.7.1 Request-Evaluation Pattern
The system sends:
- prompt
- model output
to LCAC for evaluation. LCAC returns:
- trust score
- variance data
- stability verdict
- governance mode
This pattern enables external inference pipelines to remain unchanged.
11.7.2 Inline Agent or LLM Integration
LCAC may be embedded into:
- agent toolchains
- orchestration frameworks
- multi-agent reasoning flows
- specialized LLM applications
Governance checks occur inline between inference steps.
11.7.3 Multi-Model or Router Integration
LCAC may act as:
- a routing constraint
- a stability filter
- a verification layer
for systems that dynamically select between models.
11.8 Operational Monitoring
LCAC supports RIS-aligned monitoring by:
- generating audit logs
- tracking drift events
- identifying variance breaches
- logging boundary violations
- maintaining historical baselines
Monitoring outputs MAY be used to trigger reassessment under RIS Section 9.
11.9 Example Workflow
A RIS-aligned workflow using LCAC includes:
- The LLM produces an output for a given prompt.
- The output and prompt are sent to LCAC.
- LCAC computes trust, variance, and drift metrics.
- LCAC assigns a verdict (stable, watch, unstable).
- LCAC writes an immutable ledger entry.
- Governance mode is updated based on results.
- Application logic uses the verdict to:
- accept the output
- reject the output
- request re-evaluation
- escalate to a fallback model
- enforce stricter controls
This workflow illustrates one method of operationalizing RIS requirements.
11.10 Compliance Relationship
LCAC demonstrates compliance with the following RIS controls:
- RS-1, RS-2, RS-3
- SC-1, SC-2, SC-3
- DR-1, DR-2, DR-3
- VE-1, VE-2, VE-3
- GB-1, GB-2, GB-3
- OP-1, OP-2, OP-3, OP-4
LCAC is a reference implementation and SHOULD NOT be considered the only method of achieving RIS compliance.
11.11 Implementation Notes
- LCAC uses Redis for metric storage and ledger hashing.
- All evaluation computations are deterministic given identical inputs.
- LCAC can operate in multi-node architectures with shared storage.
- The reference implementation is platform-agnostic.
- API endpoints and data schemas are documented in the LCAC repository.
These details are informative and may vary between deployments.
11.12 Limitations
LCAC, as a reference implementation:
- does not guarantee factual correctness
- does not address privacy or security
- does not enforce ethical or domain-specific constraints
- does not replace external governance frameworks
RIS conformance MUST be determined independently of LCAC adoption.
11.13 Informative Appendix Linkage
Sections of LCAC MAY be referenced in Appendix A for:
- example metrics
- sample ledger entries
- example drift baselines
- variance envelope templates
- integration code snippets
These references are informative only.
Standard Mapping
12. Mapping to Existing Standards
12.1 Overview
This section provides a crosswalk between the Reasoning Integrity Standard (RIS) and widely adopted security, governance, and AI risk management frameworks. The purpose of this mapping is to:
- support organizations aligning RIS with existing compliance programs
- demonstrate compatibility between RIS and established standards
- clarify which risks and controls RIS covers
- identify where RIS provides additional, reasoning-specific requirements
Mappings are informative and do not constitute equivalence or certification under referenced standards.
12.2 Relationship to NIST SP 800-53
RIS aligns with NIST SP 800-53 by extending its principles into the domain of reasoning integrity for AI systems.
12.2.1 Complementary Areas
- NIST RA (Risk Assessment): RIS provides a reasoning-specific risk model.
- NIST CA (Assessment, Authorization, Monitoring): RIS defines evaluation and audit procedures.
- NIST SI (System Integrity): RIS adds reasoning-chain integrity and drift detection.
- NIST PM (Program Management): RIS supports governance for LLM-based workflows.
12.2.2 Gaps Filled by RIS
NIST does not address:
- reasoning drift
- semantic coherence
- chain stability
- variance envelopes
- multi-agent interference
RIS provides controls specifically for these areas.
12.3 Relationship to ISO/IEC 27001
RIS complements ISO/IEC 27001 by introducing AI reasoning integrity controls that parallel information security requirements.
12.3.1 Complementary Areas
- ISO 27001 Annex A: controls for monitoring, logging, and operational integrity
- ISO 27001 Annex A: change management and configuration management align with RIS evaluation invariants
- ISO 27001 Annex A: audit logging aligns with RIS ledger requirements
12.3.2 Gaps Filled by RIS
ISO 27001 does not include:
- reasoning integrity assessment
- drift evaluation
- semantic stability requirements
- boundary adherence controls for LLMs
RIS extends governance into these areas.
12.4 Relationship to SOC 2 Trust Services Criteria
RIS aligns with SOC 2 across the following TSC categories:
- Security: LCAC-like governance boundaries enhance system integrity
- Availability: drift and variance controls improve predictability of behavior
- Processing Integrity: RIS focuses specifically on the integrity of reasoning processes
- Confidentiality: RIS does not address confidentiality explicitly but complements it
SOC 2 does not cover reasoning stability, which RIS supplies.
12.5 Relationship to OWASP Application Security Verification Standard (ASVS)
RIS extends ASVS principles into the domain of AI reasoning by providing:
- integrity validation of reasoning workflows
- boundary checks analogous to input validation
- monitoring comparable to application logic verification
Areas where RIS aligns with ASVS:
- session integrity (analogous to reasoning chain integrity)
- validation layers (similar to boundary enforcement)
- error and exception handling (mapped to drift detection)
ASVS does not include concepts such as:
- semantic drift
- reasoning instability
- variance envelopes
RIS adds these requirements.
12.6 Relationship to NIST AI RMF
RIS is closely aligned with the NIST AI Risk Management Framework.
12.6.1 Alignment Areas
- Govern Function: RIS defines governance boundaries and control families
- Map Function: RIS specifies reasoning-specific risks and threat models
- Measure Function: RIS provides standardized metrics and scoring
- Manage Function: RIS prescribes drift, variance, and stability controls
12.6.2 RIS Extensions Beyond NIST AI RMF
RIS introduces:
- explicit reasoning chain assessment
- structured variance and drift envelopes
- multi-level conformance classifications
- detailed evaluation methodology
- audit and verification standards
These elements provide more granular operational guidance.
12.7 Relationship to the EU AI Act
The EU AI Act categorizes AI systems according to risk. RIS provides a compatibility layer by offering objective measurements relevant to risk classification.
12.7.1 High-Risk Alignment
High-risk AI systems require:
- logging
- monitoring
- technical documentation
- risk mitigation
RIS supports these by:
- providing a reasoning ledger
- defining stability metrics
- enabling drift monitoring
- defining reproducible evaluation processes
12.7.2 RIS-4 and High-Risk Use Cases
RIS-4 provides the strictest reasoning integrity requirements and is appropriate for:
- safety-critical systems
- medical reasoning systems
- financial decision systems
- legal reasoning systems
- infrastructure and security systems
EU AI Act does not provide technical detail on reasoning integrity. RIS fills that gap.
12.8 Summary Table (RIS vs. External Standards)
| Framework | Overlap | RIS Contribution |
|---|---|---|
| NIST SP 800-53 | Monitoring, logging, system integrity | Adds reasoning-chain integrity and drift controls |
| ISO 27001 | Operational controls, audit logging | Adds reasoning stability requirements |
| SOC 2 | Processing integrity | Adds reasoning-specific integrity metrics |
| OWASP ASVS | Boundary and validation concepts | Adds semantic and structural reasoning requirements |
| NIST AI RMF | Risk, governance, measurement | Adds detailed evaluation and conformance system |
| EU AI Act | Documentation and oversight | Adds measurable technical integrity criteria |
12.9 General Observations
- RIS is compatible with existing frameworks but more granular in evaluating reasoning behavior.
- RIS introduces metrics and controls not present in any existing standard.
- Organizations adopting RIS can integrate it with existing compliance programs.
- RIS MAY serve as a technical layer beneath regulatory or governance frameworks.
12.10 Use of Mapping in Compliance Programs
Organizations MAY use this mapping to:
- integrate RIS evaluation with ISO or SOC 2 audits
- include RIS controls in security and governance documentation
- demonstrate operational integrity in high-risk AI deployments
- support regulatory submissions requiring technical evidence
- develop internal AI governance programs
Mapping does not imply certification under other standards.
Appendices
13. Appendices
13.1 Overview
This section contains informative appendices supporting the implementation, evaluation, and audit of the Reasoning Integrity Standard (RIS). Appendices are not normative unless explicitly referenced by a control or requirement.
The materials in this section include:
- definitions and terminology
- mathematical formulations
- example baseline structures
- sample drift and variance calculations
- reference ledger entries
- prompt set templates
- evaluation dataset formats
- implementation considerations
13.2 Key Definitions
13.2.1 Reasoning chain
A sequence of internal conceptual steps used by an LLM or agent to derive an output from an input.
13.2.2 Structural drift
A deviation in the ordering, length, or internal transitions of a reasoning chain beyond acceptable variance thresholds.
13.2.3 Semantic drift
A deviation in meaning, conceptual structure, or topic alignment across samples.
13.2.4 Variance envelope
The statistical boundary within which repeated inference behavior must remain to be considered stable.
13.2.5 Boundary violation
A reasoning event in which the model exceeds defined contextual, semantic, or operational constraints.
13.2.6 Interference event
An alteration of reasoning behavior caused by a tool, agent, retrieval source, or external system.
13.3 Mathematical Formulations
13.3.1 Chain stability score
Chain stability S is calculated as:
S = mean(similarity(Ri, Rj))
13.3.2 Semantic coherence score
Semantic coherence C is calculated as:
C = mean(cosine(Ei, Ej))
13.3.3 Drift sensitivity
Drift sensitivity D is calculated as:
D = variance(S) + variance(C)
13.3.4 Variance envelope compliance
Compliance V is:
V = (samples within envelope) / (total samples)
13.4 Example Variance Envelopes
13.4.1 General-purpose LLM envelope
- chain stability: 0.75 to 1.00
- semantic coherence: 0.80 to 1.00
- drift sensitivity: below 0.30
13.4.2 High-integrity envelope
- chain stability: 0.90 to 1.00
- semantic coherence: 0.90 to 1.00
- drift sensitivity: below 0.10
Evaluators SHALL adjust envelopes based on domain requirements.
13.5 Example Ledger Entry
Below is a sample hash-linked ledger entry.
{
"prompt": "Explain the impact of X on Y.",
"output": "Reasoning steps and conclusion...",
"trust": 0.92,
"variance": 0.05,
"verdict": "stable",
"timestamp": 1735861234,
"prev_hash": "0000c9beef78aa12fd39b4b2d4e9a18f",
"hash": "8ad1c4fd1e09aa7dfefc324b3987456b"
}
This format is informative and MAY be adapted.
13.6 Prompt Set Templates
The following templates MAY be used to construct RIS evaluation prompt sets.
13.6.1 Baseline prompts
- Explain a concept.
- Perform a step-by-step analysis.
- Compare two ideas.
- Solve a structured reasoning task.
- Describe implications of a policy or event.
13.6.2 Perturbation prompts
- synonym substitutions
- minor phrasal changes
- syntax adjustments
- clause reordering
All perturbations MUST preserve semantic meaning.
13.7 Evaluation Dataset Format
Evaluation datasets SHOULD be structured as follows:
{
"prompt_id": "P001",
"prompt": "Explain the role of Z.",
"baseline_samples": [],
"perturbation_prompts": [],
"perturbation_samples": [],
"metadata": { "domain": "general", "difficulty": "medium" }
}
This supports consistent analysis and cross-system comparison.
13.8 Example Drift Analysis Summary
A drift summary MAY include:
- baseline mean stability score
- new sample mean stability score
- deviation percentage
- variance across iterations
- threshold compliance
- drift triggers observed
Example:
{
"baseline_stability": 0.91,
"current_stability": 0.82,
"deviation": "9.8%",
"threshold": "10%",
"status": "borderline-compliant"
}
13.9 Boundary Violation Examples
Examples of boundary violations include:
- referencing information not present in prompt or context
- unexpected topic expansion outside allowed domain
- using tool outputs that exceed allowed semantic scope
- combining context from unrelated tasks
- reasoning dependent on long-term model memory
Boundary violations SHALL be documented and investigated.
13.10 Implementation Considerations
Organizations implementing RIS SHOULD consider:
- model-specific tuning of drift and variance thresholds
- architecture-specific behaviors (transformer vs. hybrid vs. agentic systems)
- differential stability between prompt types
- retrieval and tool-use artifacts influencing stability
- cross-agent reasoning contamination
- infrastructure effects on sampling consistency
These considerations inform evaluation and do not alter RIS conformance requirements.
13.11 Future Extensions (Informative)
Potential future RIS extensions MAY include:
- multi-agent interference scoring
- predictive trust-flow modeling
- coherence signature analysis
- temporal reasoning-chain compression metrics
- scenario-based benchmark suites
- extended metadata for multi-modal systems
These are informative only and not included in RIS v1.0.
13.12 Document Revision Log
Version: 1.0
Published by: Atom Labs
Status: Initial Release
Future revisions SHALL be recorded in this section when issued.