Changes for page Automation
Last modified by Robert Schaub on 2025/12/24 20:34
Summary
-
Page properties (1 modified, 0 added, 0 removed)
Details
- Page properties
-
- Content
-
... ... @@ -1,296 +1,111 @@ 1 1 = Automation = 2 2 3 -Automation in FactHarbor amplifies human capability while implementing risk-based oversight. 3 +Automation in FactHarbor amplifies human capability but never replaces human oversight. 4 +All automated outputs require human review before publication. 4 4 5 5 This chapter defines: 6 - 7 -* Risk-based publication model 8 -* Quality gates for AI-generated content 9 9 * What must remain human-only 10 -* What AI (AKEL) can draft and publish8 +* What AI (AKEL) can draft 11 11 * What can be fully automated 12 12 * How automation evolves through POC → Beta 0 → Release 1.0 13 13 14 -== POC v1 (A I-GeneratedPublicationDemonstration) ==12 +== POC v1 (Fully Automated "Text to Truth Landscape") == 15 15 16 -The goal of POC v1 is to validate the automated reasoning capabilities and demonstrateAI-generated content publication.14 +The goal of POC v1 is to validate the automated reasoning capabilities of the data model without human intervention. 17 17 18 18 === Workflow === 19 19 20 20 1. **Input**: User pastes a block of raw text. 21 -1. **Deep Analysis (Background)**: The system autonomously performs the full pipeline **before** displaying the text: 19 +2. **Deep Analysis (Background)**: The system autonomously performs the full pipeline **before** displaying the text: 20 +** Extraction & Normalisation 21 +** Scenario & Sub-query generation 22 +** Evidence retrieval & Verdict computation 23 +3. **Visualisation (Extraction & Marking)**: The system displays the text with claims extracted and marked. 24 +** **Verdict-Based Coloring**: The extraction highlights (e.g. Orange/Green) are chosen **according to the computed verdict** for each claim. 25 +4. **Inspection**: User clicks a highlighted claim to see the **Reasoning Trail**, showing exactly which evidence and sub-queries led to that verdict. 22 22 23 -* Extraction & Normalisation 24 -* Scenario & Sub-query generation 25 -* Evidence retrieval with **contradiction search** 26 -* Quality gate validation 27 -* Verdict computation 28 - 29 -1. **Visualisation (Extraction & Marking)**: The system displays the text with claims extracted and marked. 30 - 31 -* **Verdict-Based Coloring**: The extraction highlights (e.g. Orange/Green) are chosen **according to the computed verdict** for each claim. 32 -* **AI-Generated Label**: Clear indication that content is AI-produced 33 - 34 -1. **Inspection**: User clicks a highlighted claim to see the **Reasoning Trail**, showing exactly which evidence and sub-queries led to that verdict. 35 - 36 36 === Technical Scope === 37 37 38 -* **AI-Generated Publication**: Content published as Mode 2 (AI-Generated, no prior human review) 39 -* **Quality Gates Active**: All automated quality checks enforced 40 -* **Contradiction Search Demonstrated**: Shows counter-evidence and reservation detection 41 -* **Risk Tier Classification**: POC shows tier assignment (demo purposes) 42 -* **No Human Approval Gate**: Demonstrates scalable AI publication 43 -* **Structured Sub-Queries**: Logic generated by decomposing claims into the FactHarbor data model 29 +* **Fully Automated**: No human-in-the-loop for this phase. 30 +* **Structured Sub-Queries**: Logic is generated by decomposing claims into the FactHarbor data model. 31 +* **Latency**: Focus on accuracy of reasoning over real-time speed for v1. 44 44 45 45 ---- 46 46 47 -== Publication Model ==35 +== Manual vs Automated Responsibilities == 48 48 49 - FactHarborimplementsarisk-basedpublication model with three modes:37 +=== Human-Only Tasks === 50 50 51 - ===Mode1:Draft-Only===39 +These require human judgment, ethics, or contextual interpretation: 52 52 53 -* Failed quality gates 54 -* High-risk content pending expert review 55 -* Internal review queue only 41 +* Definition of key terms in claims 42 +* Approval or rejection of scenarios 43 +* Interpretation of evidence in context 44 +* Final verdict approval 45 +* Governance decisions and dispute resolution 46 +* High-risk domain oversight 47 +* Ethical boundary decisions (especially medical, political, psychological) 56 56 57 -=== Mode2:AI-Generated(Public) ===49 +=== Semi-Automated (AI Draft → Human Review) === 58 58 59 -* Passed all quality gates 60 -* Risk tier B or C 61 -* Clear AI-generated labeling 62 -* Users can request human review 51 +AKEL can draft these, but humans must refine/approve: 63 63 64 -=== Mode 3: Human-Reviewed === 53 +* Scenario structures (definitions, assumptions, context) 54 +* Evaluation methods 55 +* Evidence relevance suggestions 56 +* Reliability hints 57 +* Verdict reasoning chains 58 +* Uncertainty and limitations 59 +* Scenario comparison explanations 60 +* Suggestions for merging or splitting scenarios 61 +* Draft public summaries 65 65 66 -* Validated by human reviewers/experts 67 -* "Human-Reviewed" status badge 68 -* Required for Tier A content publication 63 +=== Fully Automated Structural Tasks === 69 69 70 -See [[AKEL page>>FactHarbor.Specification.AI Knowledge Extraction Layer (AKEL).WebHome]] for detailed publication mode descriptions. 71 - 72 ----- 73 - 74 -== Risk Tiers and Automation Levels == 75 - 76 -=== Tier A (High Risk) === 77 - 78 -* **Domains**: Medical, legal, elections, safety, security 79 -* **Automation**: AI can draft, human review required for "Human-Reviewed" status 80 -* **AI publication**: Allowed with prominent disclaimers and warnings 81 -* **Audit rate**: Recommendation: 30-50% 82 - 83 -=== Tier B (Medium Risk) === 84 - 85 -* **Domains**: Complex policy, science, causality claims 86 -* **Automation**: AI can draft and publish (Mode 2) 87 -* **Human review**: Optional, audit-based 88 -* **Audit rate**: Recommendation: 10-20% 89 - 90 -=== Tier C (Low Risk) === 91 - 92 -* **Domains**: Definitions, established facts, historical data 93 -* **Automation**: AI publication default 94 -* **Human review**: On request or via sampling 95 -* **Audit rate**: Recommendation: 5-10% 96 - 97 ----- 98 - 99 -== Human-Only Tasks == 100 - 101 -These require human judgment and cannot be automated: 102 - 103 -* **Ethical boundary decisions** (especially medical, political, psychological harm assessment) 104 -* **Dispute resolution** between conflicting expert opinions 105 -* **Governance policy** setting and enforcement 106 -* **Final authority** on Tier A "Human-Reviewed" status 107 -* **Audit system oversight** and quality standard definition 108 -* **Risk tier policy** adjustments based on societal context 109 - 110 ----- 111 - 112 -== AI-Draft with Audit (Semi-Automated) == 113 - 114 -AKEL drafts these; humans validate via sampling audits: 115 - 116 -* **Scenario structures** (definitions, assumptions, context) 117 -* **Evaluation methods** and reasoning chains 118 -* **Evidence relevance** assessment and ranking 119 -* **Reliability scoring** and source evaluation 120 -* **Verdict reasoning** with uncertainty quantification 121 -* **Contradiction and reservation** identification 122 -* **Scenario comparison** explanations 123 -* **Public summaries** and accessibility text 124 - 125 -Most Tier B and C content remains in AI-draft status unless: 126 - 127 -* Users request human review 128 -* Audits identify errors 129 -* High engagement triggers review 130 -* Community flags issues 131 - 132 ----- 133 - 134 -== Fully Automated Structural Tasks == 135 - 136 136 These require no human interpretation: 137 137 138 -* **Claim normalization** (canonical form generation) 139 -* **Duplicate detection** (vector embeddings, clustering) 140 -* **Evidence metadata extraction** (dates, authors, publication info) 141 -* **Basic reliability heuristics** (source reputation scoring) 142 -* **Contradiction detection** (conflicting statements across sources) 143 -* **Re-evaluation triggers** (new evidence, source updates) 144 -* **Layout generation** (diagrams, summaries, UI presentation) 145 -* **Federation integrity checks** (cross-node data validation) 146 - 147 ----- 148 - 149 -== Quality Gates (Automated) == 150 - 151 -Before AI-draft publication (Mode 2), content must pass: 152 - 153 -1. **Source Quality Gate** 154 - 155 -* Primary sources verified 156 -* Citations complete and accessible 157 -* Source reliability scored 158 - 159 -2. **Contradiction Search Gate** (MANDATORY) 160 - 161 -* Counter-evidence actively sought 162 -* Reservations and limitations identified 163 -* Bubble detection (echo chambers, conspiracy theories) 164 -* Diverse perspective verification 165 - 166 -3. **Uncertainty Quantification Gate** 167 - 168 -* Confidence scores calculated 169 -* Limitations stated 170 -* Data gaps disclosed 171 - 172 -4. **Structural Integrity Gate** 173 - 174 -* No hallucinations detected 175 -* Logic chain valid 176 -* References verifiable 177 - 178 -See [[AKEL page>>FactHarbor.Specification.AI Knowledge Extraction Layer (AKEL).WebHome]] for detailed quality gate specifications. 179 - 180 ----- 181 - 182 -== Audit System == 183 - 184 -Instead of reviewing all AI output, systematic sampling audits ensure quality: 185 - 186 -=== Stratified Sampling === 187 - 188 -* Risk tier (A > B > C sampling rates) 189 -* Confidence scores (low confidence → more audits) 190 -* Traffic/engagement (popular content audited more) 191 -* Novelty (new topics/claim types prioritized) 192 -* User flags and disagreement signals 193 - 194 -=== Continuous Improvement Loop === 195 - 196 -Audit findings improve: 197 - 198 -* Query templates 199 -* Source reliability weights 200 -* Contradiction detection algorithms 201 -* Risk tier assignment rules 202 -* Bubble detection heuristics 203 - 204 -=== Transparency === 205 - 206 -* Audit statistics published 207 -* Accuracy rates by tier reported 208 -* System improvements documented 209 - 210 ----- 211 - 212 -== Automation Roadmap == 213 - 214 -Automation capabilities increase with system maturity while maintaining quality oversight. 215 - 216 -=== POC (Current Focus) === 217 - 218 -**Automated:** 219 - 220 220 * Claim normalization 221 -* Scenariotemplategeneration68 +* Duplicate & cluster detection (vector embeddings) 222 222 * Evidence metadata extraction 223 -* Simple verdict drafts 224 -* **AI-generated publication** (Mode 2, with quality gates) 225 -* **Contradiction search** 226 -* **Risk tier assignment** 70 +* Basic reliability heuristics 71 +* Contradiction detection 72 +* Re-evaluation triggers 73 +* Batch layout generation (diagrams, summaries) 74 +* Federation integrity checks 227 227 228 - **Human:**76 +== Automation Roadmap == 229 229 230 -* High-risk content validation (Tier A) 231 -* Sampling audits across all tiers 232 -* Quality standard refinement 233 -* Governance decisions 78 +Automation increases with maturity. 234 234 235 -=== Beta 0 (Enhanced Automation) === 80 +=== POC (Low Automation) === 81 +* **Automated**: Claim normalization, Light scenario templates, Metadata extraction, Internal drafts. 82 +* **Human**: All scenario definitions, Evidence interpretation, Verdict creation, Governance. 236 236 237 -**Automated:** 84 +=== Beta 0 (Medium Automation) === 85 +* **Automated**: Detailed scenario drafts, Evidence reliability scoring, Cross-scenario comparisons, Contradiction detection. 86 +* **Human**: Scenario approval, Final verdict validation. 238 238 239 -* Detailed scenario generation 240 -* Advanced evidence reliability scoring 241 -* Cross-scenario comparisons 242 -* Multi-source contradiction detection 243 -* Internal Truth Landscape generation 244 -* **Increased AI-draft coverage** (more Tier B content) 245 - 246 -**Human:** 247 - 248 -* Tier A final approval 249 -* Audit sampling (continued) 250 -* Expert validation of complex domains 251 -* Quality improvement oversight 252 - 253 253 === Release 1.0 (High Automation) === 89 +* **Automated**: Full scenario generation, Evidence relevance ranking, Bayesian verdict scoring, Anomaly detection, Federation sync. 90 +* **Human**: Final approval, Ethical decisions, Oversight. 254 254 255 - **Automated:**92 +== Automation Levels == 256 256 257 -* Full scenario generation (comprehensive) 258 -* Bayesian verdict scoring across scenarios 259 -* Multi-scenario summary generation 260 -* Anomaly detection across federated nodes 261 -* AKEL-assisted cross-node synchronization 262 -* **Most Tier B and all Tier C** auto-published 94 +* **Level 0 — Human-Centric (POC)**: AI is purely advisory, nothing auto-published. 95 +* **Level 1 — Assisted (Beta 0)**: AI drafts structures; humans approve each part. 96 +* **Level 2 — Structured (Release 1.0)**: AI produces near-complete drafts; humans refine. 97 +* **Level 3 — Distributed Intelligence (Future)**: Nodes exchange embeddings and alerts; humans still approve. 263 263 264 - **Human:**99 +== Automation Matrix == 265 265 266 -* Tier A oversight (still required) 267 -* Strategic audits (lower sampling rates, higher value) 268 -* Ethical decisions and policy 269 -* Conflict resolution 101 +* **Always Human**: Final verdict, Scenario validity, Ethics, Disputes. 102 +* **Mostly AI**: Normalization, Clustering, Metadata, Heuristics, Alerts. 103 +* **Mixed**: Definitions, Boundaries, Assumptions, Reasoning. 270 270 271 - ----105 +== Diagram References == 272 272 273 - == AutomationLevelsDiagram==107 +{{include reference="FactHarbor.Specification.Diagrams.Automation Roadmap.WebHome"/}} 274 274 275 -{{include reference="FactHarbor. Archive.FactHarbor V0\.9\.23 Lost Data.Specification.Diagrams.Automation Level.WebHome"/}}109 +{{include reference="FactHarbor.Specification.Diagrams.Automation Level.WebHome"/}} 276 276 277 ----- 278 - 279 -== Automation Roadmap Diagram == 280 - 281 -{{include reference="FactHarbor.Archive.FactHarbor V0\.9\.23 Lost Data.Specification.Diagrams.Automation Roadmap.WebHome"/}} 282 - 283 ----- 284 - 285 -== Manual vs Automated Matrix == 286 - 287 -{{include reference="Test.FactHarborV09.Specification.Diagrams.Manual vs Automated matrix.WebHome"/}} 288 - 289 ----- 290 - 291 -== Related Pages == 292 - 293 -* [[AKEL (AI Knowledge Extraction Layer)>>FactHarbor.Specification.AI Knowledge Extraction Layer (AKEL).WebHome]] 294 -* [[Requirements (Roles)>>FactHarbor.Specification.Requirements.WebHome]] 295 -* [[Workflows>>FactHarbor.Specification.Workflows.WebHome]] 296 -* [[Governance>>FactHarbor.Organisation.Governance]] 111 +{{include reference="FactHarbor.Specification.Diagrams.Manual vs Automated matrix.WebHome"/}}