Wiki source code of FAQ

Last modified by Robert Schaub on 2025/12/24 20:30

Show last authors
1 = Frequently Asked Questions (FAQ) =
2
3 Common questions about FactHarbor's design, functionality, and approach.
4
5
6 == 1. How do facts get input into the system? ==
7
8 FactHarbor uses a hybrid model combining three complementary approaches:
9
10 === 1.1 1. AI-Generated Content (Scalable) ===
11
12 **What**: System dynamically researches claims using AKEL (AI Knowledge Extraction Layer)
13
14 **Process**:
15 * Extracts claims from submitted text
16 * Generates structured sub-queries
17 * Performs **mandatory contradiction search** (actively seeks counter-evidence, not just confirmations)
18 * Runs automated quality gates
19 * Publishes with clear "AI-Generated" labels
20
21 **Publication**: Mode 2 (public, AI-labeled) when quality gates pass
22
23 **Purpose**: Handles scale — emerging claims get immediate responses with transparent reasoning
24
25 === 1.2 2. Expert-Authored Content (Authoritative) ===
26
27 **What**: Domain experts directly author, edit, and validate content
28
29 **Focus**: High-risk domains (medical, legal, safety-critical)
30
31 **Publication**: Mode 3 ("Human-Reviewed" status) with expert attribution
32
33 **Authority**: Tier A content requires expert approval
34
35 **Purpose**: Provides authoritative grounding for critical domains where errors have serious consequences
36
37 === 1.3 3. Audit-Improved Quality (Continuous) ===
38
39 **What**: Sampling audits where experts review AI-generated content
40
41 **Rates**:
42 * High-risk (Tier A): 30-50% sampling
43 * Medium-risk (Tier B): 10-20% sampling
44 * Low-risk (Tier C): 5-10% sampling
45
46 **Impact**: Expert feedback systematically improves AI research quality
47
48 **Purpose**: Ensures AI quality evolves based on expert validation patterns
49
50 === 1.4 Why All Three Matter ===
51
52 **Complementary Strengths**:
53 * **AI research**: Scale and speed for emerging claims
54 * **Expert authoring**: Authority and precision for critical domains
55 * **Audit feedback**: Continuous quality improvement
56
57 **Expert Time Optimization**:
58
59 Experts can choose where to focus their time:
60 * Author high-priority content directly
61 * Validate and edit AI-generated outputs
62 * Audit samples to improve system-wide AI performance
63
64 This focuses expert time where domain expertise matters most while leveraging AI for scale.
65
66 === 1.5 Current Status ===
67
68 **POC v1**: Demonstrates the AI research pipeline (fully automated with transparent reasoning and quality gates)
69
70 **Full System**: Will support all three pathways with integrated workflow
71
72
73 == 2. What prevents FactHarbor from becoming another echo chamber? ==
74
75 FactHarbor includes multiple safeguards against echo chambers and filter bubbles:
76
77 **Mandatory Contradiction Search**:
78 * AI must actively search for counter-evidence, not just confirmations
79 * System checks for echo chamber patterns in source clusters
80 * Flags tribal or ideological source clustering
81 * Requires diverse perspectives across political/ideological spectrum
82
83 **Multiple Scenarios**:
84 * Claims are evaluated under different interpretations
85 * Reveals how assumptions change conclusions
86 * Makes disagreements understandable, not divisive
87
88 **Transparent Reasoning**:
89 * All assumptions, definitions, and boundaries are explicit
90 * Evidence chains are traceable
91 * Uncertainty is quantified, not hidden
92
93 **Audit System**:
94 * Human auditors check for bubble patterns
95 * Feedback loop improves AI search diversity
96 * Community can flag missing perspectives
97
98 **Federation**:
99 * Multiple independent nodes with different perspectives
100 * No single entity controls "the truth"
101 * Cross-node contradiction detection
102
103
104 == 3. How does FactHarbor handle claims that are "true in one context but false in another"? ==
105
106 This is exactly what FactHarbor is designed for:
107
108 **Scenarios capture contexts**:
109 * Each scenario defines specific boundaries, definitions, and assumptions
110 * The same claim can have different verdicts in different scenarios
111 * Example: "Coffee is healthy" depends on:
112 ** Definition of "healthy" (reduces disease risk? improves mood? affects specific conditions?)
113 ** Population (adults? pregnant women? people with heart conditions?)
114 ** Consumption level (1 cup/day? 5 cups/day?)
115 ** Time horizon (short-term? long-term?)
116
117 **Truth Landscape**:
118 * Shows all scenarios and their verdicts side-by-side
119 * Users see *why* interpretations differ
120 * No forced consensus when legitimate disagreement exists
121
122 **Explicit Assumptions**:
123 * Every scenario states its assumptions clearly
124 * Users can compare how changing assumptions changes conclusions
125 * Makes context-dependence visible, not hidden
126
127
128 == 4. What makes FactHarbor different from traditional fact-checking sites? ==
129
130 **Traditional Fact-Checking**:
131 * Binary verdicts: True / Mostly True / False
132 * Single interpretation chosen by fact-checker
133 * Often hides legitimate contextual differences
134 * Limited ability to show *why* people disagree
135
136 **FactHarbor**:
137 * **Multi-scenario**: Shows multiple valid interpretations
138 * **Likelihood-based**: Ranges with uncertainty, not binary labels
139 * **Transparent assumptions**: Makes boundaries and definitions explicit
140 * **Version history**: Shows how understanding evolves
141 * **Contradiction search**: Actively seeks opposing evidence
142 * **Federated**: No single authority controls truth
143
144
145 == 5. How do you prevent manipulation or coordinated misinformation campaigns? ==
146
147 **Quality Gates**:
148 * Automated checks before AI-generated content publishes
149 * Source quality verification
150 * Mandatory contradiction search
151 * Bubble detection for coordinated campaigns
152
153 **Audit System**:
154 * Stratified sampling catches manipulation patterns
155 * Expert auditors validate AI research quality
156 * Failed audits trigger immediate review
157
158 **Transparency**:
159 * All reasoning chains are visible
160 * Evidence sources are traceable
161 * AKEL involvement clearly labeled
162 * Version history preserved
163
164 **Moderation**:
165 * Moderators handle abuse, spam, coordinated manipulation
166 * Content can be flagged by community
167 * Audit trail maintained even if content hidden
168
169 **Federation**:
170 * Multiple nodes with independent governance
171 * No single point of control
172 * Cross-node contradiction detection
173 * Trust model prevents malicious node influence
174
175
176 == 6. What happens when new evidence contradicts an existing verdict? ==
177
178 FactHarbor is designed for evolving knowledge:
179
180 **Automatic Re-evaluation**:
181 1. New evidence arrives
182 2. System detects affected scenarios and verdicts
183 3. AKEL proposes updated verdicts
184 4. Reviewers/experts validate
185 5. New verdict version published
186 6. Old versions remain accessible
187
188 **Version History**:
189 * Every verdict has complete history
190 * Users can see "as of date X, what did we know?"
191 * Timeline shows how understanding evolved
192
193 **Transparent Updates**:
194 * Reason for re-evaluation documented
195 * New evidence clearly linked
196 * Changes explained, not hidden
197
198 **User Notifications**:
199 * Users following claims are notified of updates
200 * Can compare old vs new verdicts
201 * Can see which evidence changed conclusions
202
203
204 == 7. Who can submit claims to FactHarbor? ==
205
206 **Anyone** - even without login:
207
208 **Readers** (no login required):
209 * Browse and search all published content
210 * Submit text for analysis
211 * New claims added automatically unless duplicates exist
212 * System deduplicates and normalizes
213
214 **Contributors** (logged in):
215 * Everything Readers can do
216 * Submit evidence sources
217 * Suggest scenarios
218 * Participate in discussions
219
220 **Workflow**:
221 1. User submits text (as Reader or Contributor)
222 2. AKEL extracts claims
223 3. Checks for existing duplicates
224 4. Normalizes claim text
225 5. Assigns risk tier
226 6. Generates scenarios (draft)
227 7. Runs quality gates
228 8. Publishes as AI-Generated (Mode 2) if passes
229
230
231 == 8. What are "risk tiers" and why do they matter? ==
232
233 Risk tiers determine review requirements and publication workflow:
234
235 **Tier A (High Risk)**:
236 * **Domains**: Medical, legal, elections, safety, security, major financial
237 * **Publication**: AI can publish with warnings, expert review required for "Human-Reviewed" status
238 * **Audit rate**: Recommendation 30-50%
239 * **Why**: Potential for significant harm if wrong
240
241 **Tier B (Medium Risk)**:
242 * **Domains**: Complex policy, science causality, contested issues
243 * **Publication**: AI can publish immediately with clear labeling
244 * **Audit rate**: Recommendation 10-20%
245 * **Why**: Nuanced but lower immediate harm risk
246
247 **Tier C (Low Risk)**:
248 * **Domains**: Definitions, established facts, historical data
249 * **Publication**: AI publication default
250 * **Audit rate**: Recommendation 5-10%
251 * **Why**: Well-established, low controversy
252
253 **Assignment**:
254 * AKEL suggests tier based on domain, keywords, impact
255 * Moderators and Experts can override
256 * Risk tiers reviewed based on audit outcomes
257
258
259 == 9. How does federation work and why is it important? ==
260
261 **Federation Model**:
262 * Multiple independent FactHarbor nodes
263 * Each node has own database, AKEL, governance
264 * Nodes exchange claims, scenarios, evidence, verdicts
265 * No central authority
266
267 **Why Federation Matters**:
268 * **Resilience**: No single point of failure or censorship
269 * **Autonomy**: Communities govern themselves
270 * **Scalability**: Add nodes to handle more users
271 * **Specialization**: Domain-focused nodes (health, energy, etc.)
272 * **Trust diversity**: Multiple perspectives, not single truth source
273
274 **How Nodes Exchange Data**:
275 1. Local node creates versions
276 2. Builds signed bundle
277 3. Pushes to trusted neighbor nodes
278 4. Remote nodes validate signatures and lineage
279 5. Accept or branch versions
280 6. Local re-evaluation if needed
281
282 **Trust Model**:
283 * Trusted nodes → auto-import
284 * Neutral nodes → import with review
285 * Untrusted nodes → manual only
286
287
288 == 10. Can experts disagree in FactHarbor? ==
289
290 **Yes - and that's a feature, not a bug**:
291
292 **Multiple Scenarios**:
293 * Experts can create different scenarios with different assumptions
294 * Each scenario gets its own verdict
295 * Users see *why* experts disagree (different definitions, boundaries, evidence weighting)
296
297 **Parallel Verdicts**:
298 * Same scenario, different expert interpretations
299 * Both verdicts visible with expert attribution
300 * No forced consensus
301
302 **Transparency**:
303 * Expert reasoning documented
304 * Assumptions stated explicitly
305 * Evidence chains traceable
306 * Users can evaluate competing expert opinions
307
308 **Federation**:
309 * Different nodes can have different expert conclusions
310 * Cross-node branching allowed
311 * Users can see how conclusions vary across nodes
312
313
314 == 11. What prevents AI from hallucinating or making up facts? ==
315
316 **Multiple Safeguards**:
317
318 **Quality Gate 4: Structural Integrity**:
319 * Fact-checking against sources
320 * No hallucinations allowed
321 * Logic chain must be valid and traceable
322 * References must be accessible and verifiable
323
324 **Evidence Requirements**:
325 * Primary sources required
326 * Citations must be complete
327 * Sources must be accessible
328 * Reliability scored
329
330 **Audit System**:
331 * Human auditors check AI-generated content
332 * Hallucinations caught and fed back into training
333 * Patterns of errors trigger system improvements
334
335 **Transparency**:
336 * All reasoning chains visible
337 * Sources linked
338 * Users can verify claims against sources
339 * AKEL outputs clearly labeled
340
341 **Human Oversight**:
342 * Tier A requires expert review for "Human-Reviewed" status
343 * Audit sampling catches errors
344 * Community can flag issues
345
346
347 == 12. How does FactHarbor make money / is it sustainable? ==
348
349 [ToDo: Business model and sustainability to be defined]
350
351 Potential models under consideration:
352 * Non-profit foundation with grants and donations
353 * Institutional subscriptions (universities, research organizations, media)
354 * API access for third-party integrations
355 * Premium features for power users
356 * Federated node hosting services
357
358 Core principle: **Public benefit** mission takes priority over profit.
359
360
361 == 13. Related Pages ==
362
363 * [[Requirements (Roles)>>FactHarbor.Specification.Requirements.WebHome]]
364 * [[AKEL (AI Knowledge Extraction Layer)>>FactHarbor.Specification.AI Knowledge Extraction Layer (AKEL).WebHome]]
365 * [[Automation>>FactHarbor.Specification.Automation.WebHome]]
366 * [[Federation & Decentralization>>FactHarbor.Specification.Federation & Decentralization.WebHome]]
367 * [[Mission & Purpose>>FactHarbor.Organisation.Core Problems FactHarbor Solves.WebHome]]
368
369
370 == 20. Glossary / Key Terms ==
371
372 === Phase 0 vs POC v1 ===
373
374 These terms refer to the same stage of FactHarbor's development:
375
376 * **Phase 0** - Organisational perspective: Pre-alpha stage with founder-led governance
377 * **POC v1** - Technical perspective: Proof of Concept demonstrating AI-generated publication
378
379 Both describe the current development stage where the platform is being built and initially validated.
380
381 === Beta 0 ===
382
383 The next development stage after POC, featuring:
384 * External testers
385 * Basic federation experiments
386 * Enhanced automation
387
388 === Release 1.0 ===
389
390 The first public release featuring:
391 * Full federation support
392 * 2000+ concurrent users
393 * Production-grade infrastructure