Workflows

Version 1.1 by Robert Schaub on 2025/12/18 12:03

Workflows

FactHarbor workflows are simple, automated, focused on continuous improvement.

1. Core Principles

  • Automated by default: AI processes everything
  • Publish immediately: No centralized approval (removed in V0.9.50)
  • Quality through monitoring: Not gatekeeping
  • Fix systems, not data: Errors trigger improvements
  • Human-in-loop: Only for edge cases and abuse

2. Claim Submission Workflow

```
User submits → Duplicate detection → Categorization → Processing queue → User receives ID
```
Timeline: Seconds
No approval needed

3. Automated Analysis Workflow

```
Claim from queue

Evidence gathering (AKEL)

Source evaluation (track record check)

Scenario generation

Verdict synthesis

Risk assessment

Quality gates (confidence > 40%? risk < 80%?)

Publish OR Flag for improvement
```
Timeline: 10-30 seconds
90%+ published automatically

3.5 Evidence and Verdict Workflow

Information

Current Implementation (v2.10.2) - Simplified model without versioning. Uses 7-point symmetric verdict scale.

Evidence and Verdict Data Model


erDiagram
    CLAIM ||--|| CLAIM_VERDICT : has
    CLAIM_VERDICT }o--o{ EVIDENCE_ITEM : supported_by
    EVIDENCE_ITEM }o--|| SOURCE : from

    CLAIM {
        string id_PK
        string text
        string type
        string claimRole
        boolean isCentral
        string_array dependsOn
    }

    CLAIM_VERDICT {
        string id_PK
        string claimId_FK
        string verdict
        int truthPercentage
        int confidence
        string explanation
        string_array supportingEvidenceIds
        string_array opposingEvidenceIds
        string contestationStatus
        float harmPotential
    }

    EVIDENCE_ITEM {
        string id_PK
        string sourceId_FK
        string statement
        string sourceExcerpt
        string category
        string claimDirection
        string contextId
    }

    SOURCE {
        string id_PK
        string title
        string domain
        string url
        float trackRecordScore
        string bias
        string factualReporting
    }

Verdict Generation Flow


flowchart TB
    subgraph Research[Research Phase]
        EVIDENCE[Collected Evidence]
        SOURCES[Source Metadata]
    end

    subgraph Analysis[Analysis]
        WEIGHT[Weight Evidence by source reliability]
        CONTEST[Check Contestation doubted vs contested]
        HARM[Assess Harm Potential]
    end

    subgraph Verdict[Verdict Generation]
        CALC[Calculate Truth Percentage]
        MAP[Map to 7-point Scale]
        CONF[Assign Confidence]
    end

    subgraph Output[Result]
        CLAIM_V[Claim Verdict]
        ARTICLE_V[Article Verdict]
    end

    EVIDENCE --> WEIGHT
    SOURCES --> WEIGHT
    WEIGHT --> CONTEST
    CONTEST --> HARM
    HARM --> CALC
    CALC --> MAP
    MAP --> CONF
    CONF --> CLAIM_V
    CLAIM_V --> ARTICLE_V

7-Point Verdict Scale

 Verdict  Truth % Range  Description
 TRUE  86-100%  Claim is well-supported by evidence
 MOSTLY-TRUE  72-85%  Largely accurate with minor caveats
 LEANING-TRUE  58-71%  More evidence supports than contradicts
 MIXED  43-57% (high conf)  Roughly equal evidence both ways
 UNVERIFIED  43-57% (low conf)  Insufficient evidence to determine
 LEANING-FALSE  29-42%  More evidence contradicts than supports
 MOSTLY-FALSE  15-28%  Largely inaccurate
 FALSE  0-14%  Claim is refuted by evidence

Contestation Status

  • Doubted: Evidence is weak, uncertain, or ambiguous
  • Contested: Strong evidence exists on both sides

Source Reliability

Source reliability scores use LLM + Cache architecture (v2.2):

  • LLM-based assessment with in-memory caching
  • Batch prefetch → in-memory map → sync lookup
  • Configurable via UCM SR config (source-reliability.ts)

4. Publication Workflow

Standard (90%+): Pass quality gates → Publish immediately with confidence scores
High Risk (<10%): Risk > 80% → Moderator review
Low Quality: Confidence < 40% → Improvement queue → Re-process

5. User Contribution Workflow

```
Contributor edits → System validates → Applied immediately → Logged → Reputation earned
```
No approval required (Wikipedia model)
New contributors (<50 reputation): Limited to minor edits

6. Flagging Workflow

```
User flags issue → Categorize (abuse/quality) → Automated or manual resolution
```
Quality issues: Add to improvement queue → System fix → Auto re-process
Abuse: Moderator review → Action taken

7. Moderation Workflow

Automated pre-moderation: 95% published automatically
Moderator queue: Only high-risk or flagged content
Appeal process: Different moderator → Governing Team if needed

8. System Improvement Workflow

Weekly cycle:
```
Monday: Review error patterns
Tuesday-Wednesday: Develop fixes
Thursday: Test improvements
Friday: Deploy & re-process
Weekend: Monitor metrics
```
Error capture:
```
Error detected → Categorize → Root cause → Improvement queue → Pattern analysis
```
A/B Testing:
```
New algorithm → Split traffic (90% control, 10% test) → Run 1 week → Compare metrics → Deploy if better
```

9. Quality Monitoring Workflow

Continuous: Every hour calculate metrics, detect anomalies
Daily: Update source track records, aggregate error patterns
Weekly: System improvement cycle, performance review

10. Source Track Record Workflow

Initial score: New source starts at 50 (neutral)
Daily updates: Calculate accuracy, correction frequency, update score
Continuous: All claims using source recalculated when score changes

11. Re-Processing Workflow

Triggers: System improvement deployed, source score updated, new evidence, error fixed
Process: Identify affected claims → Re-run AKEL → Compare → Update if better → Log change

12. Related Pages