All Docs
FeaturesDepositClearUpdated March 15, 2026

AI Assessment Feedback-to-Improvement Loop

AI Assessment Feedback-to-Improvement Loop

Available from: v0.1.350

Overview

The AI Assessment Feedback-to-Improvement Loop is an event-driven workflow that automatically captures the outcome of every adjudicated dispute and uses it to build a running picture of AI assessment accuracy — on a per-organisation basis.

Once a dispute is resolved through the post-adjudication process, the outcome's key signals (reasoningTags and awardRatio) are extracted and written into a performance snapshot. Over time, these snapshots reveal how well AI deduction assessments align with real adjudication decisions.


How It Works

Trigger

The workflow is activated by the ai.feedback.dispute_outcome event, which is already emitted automatically when dispute-outcome-post-adjudication completes. No configuration or manual triggering is needed.

Event: ai.feedback.dispute_outcome
Type:  Sync

Execution Flow

dispute-outcome-post-adjudication
        │
        └─► fires: ai.feedback.dispute_outcome
                        │
                        ▼
            Read disputeOutcomes
              - reasoningTags
              - awardRatio
                        │
                        ▼
            Update ai_performance_snapshot (per org)
              or fallback: analyticsSnapshots
                        │
                        ▼
            Persist updated metrics:
              - avg award ratio
              - tag frequency distribution
              - win/loss patterns by deduction category

Metrics Tracked

MetricDescription
Average Award RatioRunning average of how much of the claimed deduction was awarded, across all disputes for the organisation.
Tag Frequency DistributionHow often each reasoningTag appears in adjudicated outcomes — surfaces which reasoning patterns correlate with success or failure.
Win/Loss Patterns by Deduction CategoryDispute outcomes grouped by the type of deduction (e.g. cleaning, damage, rent arrears), showing where AI assessments are most and least aligned with adjudicator decisions.

Entities

EntityRole
disputeOutcomesSource of truth for adjudication results — provides reasoningTags and awardRatio.
aiDeductionAssessmentsThe original AI assessments being evaluated for accuracy.
analyticsSnapshotsFallback store for performance snapshots when ai_performance_snapshot is unavailable.
usageEventsUsed to log workflow execution for auditing and billing purposes.

Storage: ai_performance_snapshot

Each organisation has its own ai_performance_snapshot record that is updated (not replaced) on every event. This means metrics accumulate incrementally — each new dispute outcome shifts the running averages rather than overwriting previous data.

If no snapshot exists for the organisation yet, one is created automatically on first event receipt.


Relationship to Other Features

  • Dispute Outcome Post-Adjudication — The upstream process that emits the trigger event. This feature requires it to be active.
  • AI Deduction Assessments — The assessments being measured. Accuracy metrics produced here will inform future model tuning and threshold calibration.
  • Analytics Snapshots — Used as the fallback persistence layer; also queryable for reporting dashboards.

Notes

  • This workflow runs synchronously within the post-adjudication event pipeline.
  • No additional setup is required — the trigger event is already emitted by the existing adjudication flow.
  • Snapshots are scoped per organisation, so metrics do not bleed across tenancies or landlord accounts.