Claim evaluation is the process of deciding whether a submitted Claim is supported by the evidence and rules that govern it. Use this article when you need the architecture model behind agent-assisted verification on IXO. It explains the concepts, boundaries, and safety rules for evaluation workflows. Use the linked developer guides and reference pages for exact SDK methods, package identifiers, endpoint values, and protocol message shapes.Documentation Index
Fetch the complete documentation index at: https://docs.ixo.world/llms.txt
Use this file to discover all available pages before exploring further.
The core question
A claim evaluation workflow answers one practical question:- which Claim was evaluated
- which evidence was inspected
- which authority allowed the evaluator to act
- which rubric or protocol was applied
- which checks passed, failed, or require review
- what decision or recommendation was made
- what state transition, payment, credential, dispute, or review step is allowed next
Why this matters
Claims often represent real-world work, identity, compliance, impact, delivery, or eligibility. Without a governed evaluation model, verification can drift into screenshots, emails, spreadsheets, ad hoc chat messages, and opaque expert judgment. Agentic Oracles can help with evidence review and decision support, but automation introduces its own risks:- the agent may act outside delegated authority
- evidence may be incomplete, stale, or forged
- a model may summarize confidently without citing sources
- a rubric may be too vague to reproduce
- state changes or payments may happen before a valid determination exists
- reviewers may be unable to replay the decision
System model
The evaluation pattern connects IXO Protocol, IXO Graph, Qi Intelligent Cooperating System, and Agentic Oracles.A Claim enters a governed context
The workflow resolves authority and state
Evidence becomes typed facts
The rubric evaluates the facts
A UDID records the determination
Core concepts
Claim
Claim
Claim Collection
Claim Collection
Evidence
Evidence
Agentic Oracle
Agentic Oracle
Evaluation kit
Evaluation kit
Rubric
Rubric
Fact ledger
Fact ledger
UDID
UDID
Source-of-truth boundaries
Keep each layer responsible for one part of the evaluation system.IXO Protocol
IXO Protocol
- Owns: Claim lifecycle state, protocol messages, authorization, and on-chain records.
- Does not own: Private evidence payloads or model reasoning.
IXO Graph
IXO Graph
- Owns: Shared context for entities, Claims, evidence, authority, workflows, decisions, and outcomes.
- Does not own: Unstructured chat as canonical state.
Qi Intelligent Cooperating System
Qi Intelligent Cooperating System
- Owns: Human-agent workflow state, review routing, decision points, and next actions.
- Does not own: Raw protocol message definitions.
Agentic Oracles
Agentic Oracles
- Owns: Evidence review, fact production, rubric application, recommendations, and Evaluation Claims.
- Does not own: Unbounded authority to approve, pay, issue credentials, or update high-value state. The Agentic Oracle should be able to produce a deterministic, reproducible, and cryptographically signed UDID when a determination is made.
IXO Matrix
IXO Matrix
- Owns: Encrypted collaboration, human review rooms, alerts, and private evidence discussion.
- Does not own: The canonical rubric or final determination.
SDK and API references
SDK and API references
- Owns: Exact package identifiers, methods, endpoints, and request shapes for the evaluation kit.
- Does not own: Broad architecture ownership.
Evaluation kit structure
An evaluation kit should separate domain-specific evidence handling from shared evaluation mechanics.Shared runtime artifacts
Domain-specific kit artifacts
Fact ledger pattern
The fact ledger is the bridge between messy evidence and repeatable decisions. Raw evidence can include PDFs, images, sensor logs, API responses, credentials, signatures, spreadsheets, Matrix events, and external attestations. A rubric should not need to know how each source was parsed. It should receive stable facts with provenance. The fact ledger should be deterministic, reproducible, and cryptographically signed.- a stable identifier
- a typed value
- a producer and version
- source references
- confidence, when relevant
- failure behavior
- enough provenance for replay
Rubric pattern
A rubric should be explicit, ordered, and reason-coded. It should make escalation as concrete as approval or rejection. The source of truth for the rubric is typically a JSON file in the evaluation kit.- admissibility checks
- hard safety vetoes
- missing mandatory evidence
- invalid or conflicting evidence
- manual-review triggers
- partial-success logic
- approval logic
Outcome model
An evaluation profile should define outcomes in operational terms before mapping them to any exact protocol enum or service field.approved
approved
- Meaning: Evidence satisfies the governed rubric.
- Typical workflow behavior: Continue to the allowed state transition, settlement, credential step, or record update.
rejected
rejected
- Meaning: The Claim fails a hard rule or lacks mandatory support.
- Typical workflow behavior: Record the rejection and reason code; do not proceed to approval-only actions.
manual_review_required
manual_review_required
- Meaning: The evidence is ambiguous, conflicting, or outside automated authority.
- Typical workflow behavior: Pause automation and route to a human or governance review.
partial_success
partial_success
- Meaning: Part of the Claim is supported under a governed rule.
- Typical workflow behavior: Continue only if the rubric defines the allowed partial action.
disputed
disputed
- Meaning: A participant challenges the evaluation or determination.
- Typical workflow behavior: Route to the dispute workflow.
MsgEvaluateClaim fields, service API fields, or SDK helper methods, use the canonical developer guides and API references for the exact literals.
Human review
Human review is a controlled checkpoint, not an informal chat. A review request should include:- Claim ID and Claim Collection
- Claim subject and type
- evaluator DID or service identity (the Agentic Oracle DID)
- rubric ID and version (the rubric JSON file)
- reason code (the reason code for the outcome)
- evidence references or redacted evidence links (the evidence that was inspected)
- fact ledger summary (the typed facts that were evaluated)
- proposed outcome (the recommended or proposed next action)
- questions requiring human judgment (the questions that require human judgment)
- deadline or escalation policy (the deadline or escalation policy for the review)
- required response shape (the required response shape for the review)
Safety rules
Use these rules before allowing an Agentic Oracle to affect value, credentials, or state.Do not let an LLM directly approve a Claim
Do not let an LLM directly approve a Claim
Do not treat a CID as authenticity proof
Do not treat a CID as authenticity proof
Do not let the evaluator change its own rubric
Do not let the evaluator change its own rubric
Do not commit high-value actions on ambiguous evidence
Do not commit high-value actions on ambiguous evidence
Do not store sensitive full traces publicly
Do not store sensitive full traces publicly
First implementation move
Start with one narrow evaluation workflow that cannot directly approve, pay, issue credentials, or update high-value state. Define:- one Claim type
- one Claim Collection
- one Flow
- one rubric (the rubric JSON file)
- one evidence schema (the evidence schema JSON file)
- one fact ledger schema (the fact ledger schema JSON file)
- one Agentic Oracle or evaluator identity (the Agentic Oracle DID)
- one human review path (the human review JSON file)
- one dispute or correction path (the dispute or correction JSON file)
- one test suite with approval, rejection, ambiguity, and adversarial cases (the test suite JSON file)