Changes for page AI Knowledge Extraction Layer (AKEL)
Last modified by Robert Schaub on 2026/02/08 21:23
Summary
-
Page properties (1 modified, 0 added, 0 removed)
Details
- Page properties
-
- Content
-
... ... @@ -1,13 +1,18 @@ 1 1 = AKEL — AI Knowledge Extraction Layer = 2 + 2 2 AKEL is FactHarbor's automated intelligence subsystem. 3 3 Its purpose is to reduce human workload, enhance consistency, and enable scalable knowledge processing — **without ever replacing human judgment**. 4 4 AKEL outputs are marked with **AuthorType = AI** and published according to risk-based review policies (see Publication Modes below). 5 5 AKEL operates in two modes: 7 + 6 6 * **Single-node mode** (POC & Beta 0) 7 7 * **Federated multi-node mode** (Release 1.0+) 10 + 8 8 == 1. Purpose and Role == 12 + 9 9 AKEL transforms unstructured inputs into structured, publication-ready content. 10 10 Core responsibilities: 15 + 11 11 * Claim extraction from arbitrary text 12 12 * Claim classification (domain, type, evaluability, safety, **risk tier**) 13 13 * Scenario generation (definitions, boundaries, assumptions, methodology) ... ... @@ -17,7 +17,9 @@ 17 17 * **Bubble detection** (echo chambers, conspiracy theories, isolated sources) 18 18 * Re-evaluation proposal generation 19 19 * Cross-node embedding exchange (Release 1.0+) 25 + 20 20 == 2. Components == 27 + 21 21 * **AKEL Orchestrator** – central coordinator 22 22 * **Claim Extractor** 23 23 * **Claim Classifier** (with risk tier assignment) ... ... @@ -28,14 +28,19 @@ 28 28 * **Audit Sampling Scheduler** 29 29 * **Embedding Handler** (Release 1.0+) 30 30 * **Federation Sync Adapter** (Release 1.0+) 38 + 31 31 == 3. Inputs and Outputs == 40 + 32 32 === 3.1 Inputs === 42 + 33 33 * User-submitted claims or evidence 34 34 * Uploaded documents 35 35 * URLs or citations 36 36 * External LLM API (optional) 37 37 * Embeddings (from local or federated peers) 48 + 38 38 === 3.2 Outputs (publication mode varies by risk tier) === 50 + 39 39 * ClaimVersion (draft or AI-generated) 40 40 * ScenarioVersion (draft or AI-generated) 41 41 * EvidenceVersion (summary + metadata, draft or AI-generated) ... ... @@ -44,17 +44,25 @@ 44 44 * Reservation and limitation notices 45 45 * Re-evaluation proposals 46 46 * Updated embeddings 59 + 47 47 == 4. Publication Modes == 61 + 48 48 AKEL content is published according to three modes: 63 + 49 49 === 4.1 Mode 1: Draft-Only (Never Public) === 65 + 50 50 **Used for:** 67 + 51 51 * Failed quality gate checks 52 52 * Sensitive topics flagged for expert review 53 53 * Unclear scope or missing critical sources 54 54 * High reputational risk content 55 55 **Visibility:** Internal review queue only 73 + 56 56 === 4.2 Mode 2: Published as AI-Generated (No Prior Human Review) === 75 + 57 57 **Requirements:** 77 + 58 58 * All automated quality gates passed (see below) 59 59 * Risk tier permits AI-draft publication (Tier B or C) 60 60 * Contradiction search completed successfully ... ... @@ -69,44 +69,62 @@ 69 69 * Browse and read content 70 70 * Request human review (escalates to review queue) 71 71 * Flag for expert attention 92 + 72 72 == 5. Risk tiers == 94 + 73 73 AKEL assigns risk tiers to all content to determine appropriate review requirements: 96 + 74 74 === 5.1 Tier A — High Risk / High Impact === 98 + 75 75 **Domains:** Medical, legal, elections, safety/security, major reputational harm 76 76 **Publication policy:** 101 + 77 77 * Human review REQUIRED before "AKEL-Generated" status 78 78 * AI-generated content MAY be published but: 79 - ** Clearly flagged as AI-draft with prominent disclaimer80 - ** May have limited visibility81 - ** Auto-escalated to expert review queue82 - ** User warnings displayed104 +** Clearly flagged as AI-draft with prominent disclaimer 105 +** May have limited visibility 106 +** Auto-escalated to expert review queue 107 +** User warnings displayed 83 83 **Audit rate:** Recommendation: 30-50% of published AI-drafts sampled in first 6 months 109 + 84 84 === 5.2 Tier B — Medium Risk === 111 + 85 85 **Domains:** Contested public policy, complex science, causality claims, significant financial impact 86 86 **Publication policy:** 114 + 87 87 * AI-draft CAN publish immediately with clear labeling 88 88 * Sampling audits conducted (see Audit System below) 89 89 * High-engagement items auto-escalated to expert review 90 90 * Users can report issue for moderator review 91 91 **Audit rate:** Recommendation: 10-20% of published AI-drafts sampled 120 + 92 92 === 5.3 Tier C — Low Risk === 122 + 93 93 **Domains:** Definitions, simple factual lookups with strong primary sources, historical facts, established scientific consensus 94 94 **Publication policy:** 125 + 95 95 * AI-draft default publication mode 96 96 * Sampling audits sufficient 97 97 * Community flagging available 98 98 * Human review on request 99 99 **Audit rate:** Recommendation: 5-10% of published AI-drafts sampled 131 + 100 100 == 6. Quality Gates (Mandatory Before AI-Draft Publication) == 133 + 101 101 All AI-generated content must pass these automated checks before Mode 2 publication: 135 + 102 102 === 6.1 Gate 1: Source Quality === 137 + 103 103 * Primary sources identified and accessible 104 104 * Source reliability scored against whitelist 105 105 * Citation completeness verified 106 106 * Publication dates checked 107 107 * Author credentials validated (where applicable) 143 + 108 108 === 6.2 Gate 2: Contradiction Search (MANDATORY) === 145 + 109 109 **The system MUST actively search for:** 147 + 110 110 * **Counter-evidence** – Rebuttals, conflicting results, contradictory studies 111 111 * **Reservations** – Caveats, limitations, boundary conditions, applicability constraints 112 112 * **Alternative interpretations** – Different framings, definitions, contextual variations ... ... @@ -128,13 +128,17 @@ 128 128 * **Significant uncertainty detected** → Require uncertainty disclosure in verdict 129 129 * **Bubble indicators present** → Flag for expert review and human validation 130 130 * **Limited perspective diversity** → Expand search or flag for human review 169 + 131 131 === 6.3 Gate 3: Uncertainty Quantification === 171 + 132 132 * Confidence scores calculated for all claims and verdicts 133 133 * Limitations explicitly stated 134 134 * Data gaps identified and disclosed 135 135 * Strength of evidence assessed 136 136 * Alternative scenarios considered 177 + 137 137 === 6.4 Gate 4: Structural Integrity === 179 + 138 138 * No hallucinations detected (fact-checking against sources) 139 139 * Logic chain valid and traceable 140 140 * References accessible and verifiable ... ... @@ -145,29 +145,40 @@ 145 145 * Failure reason logged 146 146 * Human review required before publication 147 147 * Failure patterns analyzed for system improvement 190 + 148 148 == 7. Audit System (Sampling-Based Quality Assurance) == 192 + 149 149 Instead of reviewing ALL AI output, FactHarbor implements stratified sampling audits: 194 + 150 150 === 7.1 Sampling Strategy === 196 + 151 151 Audits prioritize: 198 + 152 152 * **Risk tier** (higher tiers get more frequent audits) 153 153 * **AI confidence score** (low confidence → higher sampling rate) 154 154 * **Traffic and engagement** (high-visibility content audited more) 155 155 * **Novelty** (new claim types, new domains, emerging topics) 156 156 * **Disagreement signals** (user flags, contradiction alerts, community reports) 204 + 157 157 === 7.2 Audit Process === 206 + 158 158 1. System selects content for audit based on sampling strategy 159 159 2. Human auditor reviews AI-generated content against quality standards 160 160 3. Moderator validates or corrects: 161 - * Claim extraction accuracy 162 - * Scenario appropriateness 163 - * Evidence relevance and interpretation 164 - * Verdict reasoning 165 - * Contradiction search completeness 210 + 211 +* Claim extraction accuracy 212 +* Scenario appropriateness 213 +* Evidence relevance and interpretation 214 +* Verdict reasoning 215 +* Contradiction search completeness 166 166 4. Audit outcome recorded (pass/fail + detailed feedback) 167 167 5. Failed audits trigger immediate content review 168 168 6. Audit results feed back into system improvement 219 + 169 169 === 7.3 Feedback Loop (Continuous Improvement) === 221 + 170 170 Audit outcomes systematically improve: 223 + 171 171 * **Query templates** – Refined based on missed evidence patterns 172 172 * **Retrieval source weights** – Adjusted for accuracy and reliability 173 173 * **Contradiction detection heuristics** – Enhanced to catch missed counter-evidence ... ... @@ -174,15 +174,22 @@ 174 174 * **Model prompts and extraction rules** – Tuned for better claim extraction 175 175 * **Risk tier assignments** – Recalibrated based on error patterns 176 176 * **Bubble detection algorithms** – Improved to identify echo chambers 230 + 177 177 === 7.4 Audit Transparency === 232 + 178 178 * Audit statistics published regularly 179 179 * Accuracy rates by risk tier tracked and reported 180 180 * System improvements documented 181 181 * Community can view aggregate audit performance 237 + 182 182 == 8. Architecture Overview == 239 + 183 183 {{include reference="FactHarbor.Specification.Diagrams.AKEL Architecture.WebHome"/}} 241 + 184 184 == 9. AKEL and Federation == 243 + 185 185 In Release 1.0+, AKEL participates in cross-node knowledge alignment: 245 + 186 186 * Shares embeddings 187 187 * Exchanges canonicalized claim forms 188 188 * Exchanges scenario templates ... ... @@ -194,8 +194,11 @@ 194 194 * Trusted nodes: auto-merge embeddings + templates 195 195 * Neutral nodes: require additional verification 196 196 * Untrusted nodes: fully manual import 257 + 197 197 == 10. Human Review Workflow (Mode 3 Publication) == 259 + 198 198 For content requiring human validation before "AKEL-Generated" status: 261 + 199 199 1. AKEL generates content and publishes as AI-draft (Mode 2) or keeps as draft (Mode 1) 200 200 2. Contributors inspect content in review queue 201 201 3. Contributors validate quality gates were correctly applied ... ... @@ -203,8 +203,11 @@ 203 203 5. Moderators finalize "AKEL-Generated" publication 204 204 6. Version numbers increment, full history preserved 205 205 **Note:** Most AI-generated content (Tier B and C) can remain in Mode 2 (AI-Generated) indefinitely. Human review is optional for these tiers unless users or audits flag issues. 269 + 206 206 == 11. POC v1 Behavior == 271 + 207 207 The POC explicitly demonstrates AI-generated content publication: 273 + 208 208 * Produces public AI-generated output (Mode 2) 209 209 * No human data sources required 210 210 * No human approval gate ... ... @@ -212,8 +212,10 @@ 212 212 * All quality gates active (including contradiction search) 213 213 * Users understand this demonstrates AI reasoning capabilities 214 214 * Risk tier classification shown (demo purposes) 281 + 215 215 == 12. Related Pages == 283 + 216 216 * [[Automation>>FactHarbor.Specification.Automation.WebHome]] 217 217 * [[Requirements (Roles)>>FactHarbor.Specification.Requirements.WebHome]] 218 218 * [[Workflows>>FactHarbor.Specification.Workflows.WebHome]] 219 -* [[Governance>>FactHarbor.Organisation.Governance.WebHome]] 287 +* [[Governance>>Archive.FactHarbor.Organisation.Governance.WebHome]]