Changes for page Automation
Last modified by Robert Schaub on 2026/02/08 08:22
Summary
-
Page properties (2 modified, 0 added, 0 removed)
Details
- Page properties
-
- Parent
-
... ... @@ -1,1 +1,1 @@ 1 -FactHarbor.Specification.WebHome 1 +Archive.FactHarbor.Specification.WebHome - Content
-
... ... @@ -1,21 +1,31 @@ 1 1 = Automation = 2 + 2 2 **How FactHarbor scales through automated claim evaluation.** 4 + 3 3 == 1. Automation Philosophy == 6 + 4 4 FactHarbor is **automation-first**: AKEL (AI Knowledge Extraction Layer) makes all content decisions. Humans monitor system performance and improve algorithms. 5 5 **Why automation:** 9 + 6 6 * **Scale**: Can process millions of claims 7 7 * **Consistency**: Same evaluation criteria applied uniformly 8 8 * **Transparency**: Algorithms are auditable 9 9 * **Speed**: Results in <20 seconds typically 10 10 See [[Automation Philosophy>>FactHarbor.Organisation.Automation-Philosophy]] for detailed principles. 15 + 11 11 == 2. Claim Processing Flow == 17 + 12 12 === 2.1 User Submits Claim === 19 + 13 13 * User provides claim text + source URLs 14 14 * System validates format 15 15 * Assigns processing ID 16 16 * Queues for AKEL processing 24 + 17 17 === 2.2 AKEL Processing === 26 + 18 18 **AKEL automatically:** 28 + 19 19 1. Parses claim into testable components 20 20 2. Extracts evidence from sources 21 21 3. Scores source credibility ... ... @@ -25,9 +25,12 @@ 25 25 7. Publishes result 26 26 **Processing time**: Typically <20 seconds 27 27 **No human approval required** - publication is automatic 38 + 28 28 === 2.3 Publication States === 40 + 29 29 **Processing**: AKEL working on claim (not visible to public) 30 30 **Published**: AKEL completed evaluation (public) 43 + 31 31 * Verdict displayed with confidence score 32 32 * Evidence and sources shown 33 33 * Risk tier indicated ... ... @@ -37,35 +37,128 @@ 37 37 * Detected manipulation attempt 38 38 * Unusual pattern 39 39 * Moderator reviews and may take action 53 + 54 +== 2.5 LLM-Based Processing Architecture == 55 + 56 +FactHarbor delegates complex reasoning and analysis tasks to Large Language Models (LLMs). The architecture evolves from POC to production: 57 + 58 +=== POC: Two-Phase Approach === 59 + 60 +**Phase 1: Claim Extraction** 61 + 62 +* Single LLM call to extract all claims from submitted content 63 +* Light structure, focused on identifying distinct verifiable claims 64 +* Output: List of claims with context 65 + 66 +**Phase 2: Claim Analysis (Parallel)** 67 + 68 +* Single LLM call per claim (parallelizable) 69 +* Full structured output: Evidence, Scenarios, Sources, Verdict, Risk 70 +* Each claim analyzed independently 71 + 72 +**Advantages:** 73 + 74 +* Fast to implement ( to working POC) 75 +* Only 2-3 API calls total (1 + N claims) 76 +* Simple to debug (claim-level isolation) 77 +* Proves concept viability 78 + 79 +=== Production: Three-Phase Approach === 80 + 81 +**Phase 1: Claim Extraction + Validation** 82 + 83 +* Extract distinct verifiable claims 84 +* Validate claim clarity and uniqueness 85 +* Remove duplicates and vague claims 86 + 87 +**Phase 2: Evidence Gathering (Parallel)** 88 + 89 +* For each claim independently: 90 +* Find supporting and contradicting evidence 91 +* Identify authoritative sources 92 +* Generate test scenarios 93 +* Validation: Check evidence quality and source validity 94 +* Error containment: Issues in one claim don't affect others 95 + 96 +**Phase 3: Verdict Generation (Parallel)** 97 + 98 +* For each claim: 99 +* Generate verdict based on validated evidence 100 +* Assess confidence and risk level 101 +* Flag low-confidence results for human review 102 +* Validation: Check verdict consistency with evidence 103 + 104 +**Advantages:** 105 + 106 +* Error containment between phases 107 +* Clear quality gates and validation 108 +* Observable metrics per phase 109 +* Scalable (parallel processing across claims) 110 +* Adaptable (can optimize each phase independently) 111 + 112 +=== LLM Task Delegation === 113 + 114 +All complex cognitive tasks are delegated to LLMs: 115 + 116 +* **Claim Extraction**: Understanding context, identifying distinct claims 117 +* **Evidence Finding**: Analyzing sources, assessing relevance 118 +* **Scenario Generation**: Creating testable hypotheses 119 +* **Source Evaluation**: Assessing reliability and authority 120 +* **Verdict Generation**: Synthesizing evidence into conclusions 121 +* **Risk Assessment**: Evaluating potential impact 122 + 123 +=== Error Mitigation === 124 + 125 +Research shows sequential LLM calls face compound error risks. FactHarbor mitigates this through: 126 + 127 +* **Validation gates** between phases 128 +* **Confidence thresholds** for quality control 129 +* **Parallel processing** to avoid error propagation across claims 130 +* **Human review queue** for low-confidence verdicts 131 +* **Independent claim processing** - errors in one claim don't cascade to others 132 + 40 40 == 3. Risk Tiers == 134 + 41 41 Risk tiers classify claims by potential impact and guide audit sampling rates. 136 + 42 42 === 3.1 Tier A (High Risk) === 138 + 43 43 **Domains**: Medical, legal, elections, safety, security 44 44 **Characteristics**: 141 + 45 45 * High potential for harm if incorrect 46 46 * Complex specialized knowledge required 47 47 * Often subject to regulation 48 48 **Publication**: AKEL publishes automatically with prominent risk warning 49 49 **Audit rate**: Higher sampling recommended 147 + 50 50 === 3.2 Tier B (Medium Risk) === 149 + 51 51 **Domains**: Complex policy, science, causality claims 52 52 **Characteristics**: 152 + 53 53 * Moderate potential impact 54 54 * Requires careful evidence evaluation 55 55 * Multiple valid interpretations possible 56 56 **Publication**: AKEL publishes automatically with standard risk label 57 57 **Audit rate**: Moderate sampling recommended 158 + 58 58 === 3.3 Tier C (Low Risk) === 160 + 59 59 **Domains**: Definitions, established facts, historical data 60 60 **Characteristics**: 163 + 61 61 * Low potential for harm 62 62 * Well-documented information 63 63 * Clear right/wrong answers typically 64 64 **Publication**: AKEL publishes by default 65 65 **Audit rate**: Lower sampling recommended 169 + 66 66 == 4. Quality Gates == 171 + 67 67 AKEL applies quality gates before publication. If any fail, claim is **flagged** (not blocked - still published). 68 68 **Quality gates**: 174 + 69 69 * Sufficient evidence extracted (≥2 sources) 70 70 * Sources meet minimum credibility threshold 71 71 * Confidence score calculable ... ... @@ -72,19 +72,22 @@ 72 72 * No detected manipulation patterns 73 73 * Claim parseable into testable form 74 74 **Failed gates**: Claim published with flag for moderator review 181 + 75 75 == 5. Automation Levels == 76 -{{include reference="FactHarbor.Specification.Diagrams.Automation Level.WebHome"/}} 183 + 184 +{{include reference="Archive.FactHarbor.Specification.Diagrams.Automation Level.WebHome"/}} 77 77 FactHarbor progresses through automation maturity levels: 78 78 **Release 0.5** (Proof-of-Concept): Tier C only, human review required 79 79 **Release 1.0** (Initial): Tier B/C auto-published, Tier A flagged for review 80 80 **Release 2.0** (Mature): All tiers auto-published with risk labels, sampling audits 81 -See [[Automation Roadmap>>FactHarbor.Specification.Diagrams.Automation Roadmap.WebHome]] for detailed progression. 189 +See [[Automation Roadmap>>Archive.FactHarbor.Specification.Diagrams.Automation Roadmap.WebHome]] for detailed progression. 82 82 83 83 == 5.5 Automation Roadmap == 84 84 85 -{{include reference="FactHarbor.Specification.Diagrams.Automation Roadmap.WebHome"/}} 193 +{{include reference="Archive.FactHarbor.Specification.Diagrams.Automation Roadmap.WebHome"/}} 86 86 87 87 == 6. Human Role == 196 + 88 88 Humans do NOT review content for approval. Instead: 89 89 **Monitoring**: Watch aggregate performance metrics 90 90 **Improvement**: Fix algorithms when patterns show issues ... ... @@ -97,6 +97,7 @@ 97 97 {{include reference="FactHarbor.Specification.Diagrams.Manual vs Automated matrix.WebHome"/}} 98 98 99 99 == 7. Moderation == 209 + 100 100 Moderators handle items AKEL flags: 101 101 **Abuse detection**: Spam, manipulation, harassment 102 102 **Safety issues**: Content that could cause immediate harm ... ... @@ -104,7 +104,9 @@ 104 104 **Action**: May temporarily hide content, ban users, or propose algorithm improvements 105 105 **Does NOT**: Routinely review claims or override verdicts 106 106 See [[Organisational Model>>FactHarbor.Organisation.Organisational-Model]] for moderator role details. 217 + 107 107 == 8. Continuous Improvement == 219 + 108 108 **Performance monitoring**: Track AKEL accuracy, speed, coverage 109 109 **Issue identification**: Find systematic errors from metrics 110 110 **Algorithm updates**: Deploy improvements to fix patterns ... ... @@ -111,15 +111,21 @@ 111 111 **A/B testing**: Validate changes before full rollout 112 112 **Retrospectives**: Learn from failures systematically 113 113 See [[Continuous Improvement>>FactHarbor.Organisation.How-We-Work-Together.Continuous-Improvement]] for improvement cycle. 226 + 114 114 == 9. Scalability == 228 + 115 115 Automation enables FactHarbor to scale: 230 + 116 116 * **Millions of claims** processable 117 117 * **Consistent quality** at any volume 118 118 * **Cost efficiency** through automation 119 119 * **Rapid iteration** on algorithms 120 120 Without automation: Human review doesn't scale, creates bottlenecks, introduces inconsistency. 236 + 121 121 == 10. Transparency == 238 + 122 122 All automation is transparent: 240 + 123 123 * **Algorithm parameters** documented 124 124 * **Evaluation criteria** public 125 125 * **Source scoring rules** explicit