Wiki source code of FAQ

Last modified by Robert Schaub on 2025/12/24 20:34

Show last authors
1 = Frequently Asked Questions (FAQ) =
2
3 Common questions about FactHarbor's design, functionality, and approach.
4
5
6 == 1. How do facts get input into the system? ==
7
8 FactHarbor uses a hybrid model combining three complementary approaches:
9
10 === 1.1 1. AI-Generated Content (Scalable) ===
11
12 **What**: System dynamically researches claims using AKEL (AI Knowledge Extraction Layer)
13
14 **Process**:
15
16 * Extracts claims from submitted text
17 * Generates structured sub-queries
18 * Performs **mandatory contradiction search** (actively seeks counter-evidence, not just confirmations)
19 * Runs automated quality gates
20 * Publishes with clear "AI-Generated" labels
21
22 **Publication**: Mode 2 (public, AI-labeled) when quality gates pass
23
24 **Purpose**: Handles scale — emerging claims get immediate responses with transparent reasoning
25
26 === 1.2 2. Expert-Authored Content (Authoritative) ===
27
28 **What**: Domain experts directly author, edit, and validate content
29
30 **Focus**: High-risk domains (medical, legal, safety-critical)
31
32 **Publication**: Mode 3 ("Human-Reviewed" status) with expert attribution
33
34 **Authority**: Tier A content requires expert approval
35
36 **Purpose**: Provides authoritative grounding for critical domains where errors have serious consequences
37
38 === 1.3 3. Audit-Improved Quality (Continuous) ===
39
40 **What**: Sampling audits where experts review AI-generated content
41
42 **Rates**:
43
44 * High-risk (Tier A): 30-50% sampling
45 * Medium-risk (Tier B): 10-20% sampling
46 * Low-risk (Tier C): 5-10% sampling
47
48 **Impact**: Expert feedback systematically improves AI research quality
49
50 **Purpose**: Ensures AI quality evolves based on expert validation patterns
51
52 === 1.4 Why All Three Matter ===
53
54 **Complementary Strengths**:
55
56 * **AI research**: Scale and speed for emerging claims
57 * **Expert authoring**: Authority and precision for critical domains
58 * **Audit feedback**: Continuous quality improvement
59
60 **Expert Time Optimization**:
61
62 Experts can choose where to focus their time:
63
64 * Author high-priority content directly
65 * Validate and edit AI-generated outputs
66 * Audit samples to improve system-wide AI performance
67
68 This focuses expert time where domain expertise matters most while leveraging AI for scale.
69
70 === 1.5 Current Status ===
71
72 **POC v1**: Demonstrates the AI research pipeline (fully automated with transparent reasoning and quality gates)
73
74 **Full System**: Will support all three pathways with integrated workflow
75
76
77 == 2. What prevents FactHarbor from becoming another echo chamber? ==
78
79 FactHarbor includes multiple safeguards against echo chambers and filter bubbles:
80
81 **Mandatory Contradiction Search**:
82
83 * AI must actively search for counter-evidence, not just confirmations
84 * System checks for echo chamber patterns in source clusters
85 * Flags tribal or ideological source clustering
86 * Requires diverse perspectives across political/ideological spectrum
87
88 **Multiple Scenarios**:
89
90 * Claims are evaluated under different interpretations
91 * Reveals how assumptions change conclusions
92 * Makes disagreements understandable, not divisive
93
94 **Transparent Reasoning**:
95
96 * All assumptions, definitions, and boundaries are explicit
97 * Evidence chains are traceable
98 * Uncertainty is quantified, not hidden
99
100 **Audit System**:
101
102 * Human auditors check for bubble patterns
103 * Feedback loop improves AI search diversity
104 * Community can flag missing perspectives
105
106 **Federation**:
107
108 * Multiple independent nodes with different perspectives
109 * No single entity controls "the truth"
110 * Cross-node contradiction detection
111
112 == 3. How does FactHarbor handle claims that are "true in one context but false in another"? ==
113
114 This is exactly what FactHarbor is designed for:
115
116 **Scenarios capture contexts**:
117
118 * Each scenario defines specific boundaries, definitions, and assumptions
119 * The same claim can have different verdicts in different scenarios
120 * Example: "Coffee is healthy" depends on:
121 ** Definition of "healthy" (reduces disease risk? improves mood? affects specific conditions?)
122 ** Population (adults? pregnant women? people with heart conditions?)
123 ** Consumption level (1 cup/day? 5 cups/day?)
124 ** Time horizon (short-term? long-term?)
125
126 **Truth Landscape**:
127
128 * Shows all scenarios and their verdicts side-by-side
129 * Users see *why* interpretations differ
130 * No forced consensus when legitimate disagreement exists
131
132 **Explicit Assumptions**:
133
134 * Every scenario states its assumptions clearly
135 * Users can compare how changing assumptions changes conclusions
136 * Makes context-dependence visible, not hidden
137
138 == 4. What makes FactHarbor different from traditional fact-checking sites? ==
139
140 **Traditional Fact-Checking**:
141
142 * Binary verdicts: True / Mostly True / False
143 * Single interpretation chosen by fact-checker
144 * Often hides legitimate contextual differences
145 * Limited ability to show *why* people disagree
146
147 **FactHarbor**:
148
149 * **Multi-scenario**: Shows multiple valid interpretations
150 * **Likelihood-based**: Ranges with uncertainty, not binary labels
151 * **Transparent assumptions**: Makes boundaries and definitions explicit
152 * **Version history**: Shows how understanding evolves
153 * **Contradiction search**: Actively seeks opposing evidence
154 * **Federated**: No single authority controls truth
155
156 == 5. How do you prevent manipulation or coordinated misinformation campaigns? ==
157
158 **Quality Gates**:
159
160 * Automated checks before AI-generated content publishes
161 * Source quality verification
162 * Mandatory contradiction search
163 * Bubble detection for coordinated campaigns
164
165 **Audit System**:
166
167 * Stratified sampling catches manipulation patterns
168 * Expert auditors validate AI research quality
169 * Failed audits trigger immediate review
170
171 **Transparency**:
172
173 * All reasoning chains are visible
174 * Evidence sources are traceable
175 * AKEL involvement clearly labeled
176 * Version history preserved
177
178 **Moderation**:
179
180 * Moderators handle abuse, spam, coordinated manipulation
181 * Content can be flagged by community
182 * Audit trail maintained even if content hidden
183
184 **Federation**:
185
186 * Multiple nodes with independent governance
187 * No single point of control
188 * Cross-node contradiction detection
189 * Trust model prevents malicious node influence
190
191 == 6. What happens when new evidence contradicts an existing verdict? ==
192
193 FactHarbor is designed for evolving knowledge:
194
195 **Automatic Re-evaluation**:
196
197 1. New evidence arrives
198 2. System detects affected scenarios and verdicts
199 3. AKEL proposes updated verdicts
200 4. Reviewers/experts validate
201 5. New verdict version published
202 6. Old versions remain accessible
203
204 **Version History**:
205
206 * Every verdict has complete history
207 * Users can see "as of date X, what did we know?"
208 * Timeline shows how understanding evolved
209
210 **Transparent Updates**:
211
212 * Reason for re-evaluation documented
213 * New evidence clearly linked
214 * Changes explained, not hidden
215
216 **User Notifications**:
217
218 * Users following claims are notified of updates
219 * Can compare old vs new verdicts
220 * Can see which evidence changed conclusions
221
222 == 7. Who can submit claims to FactHarbor? ==
223
224 **Anyone** - even without login:
225
226 **Readers** (no login required):
227
228 * Browse and search all published content
229 * Submit text for analysis
230 * New claims added automatically unless duplicates exist
231 * System deduplicates and normalizes
232
233 **Contributors** (logged in):
234
235 * Everything Readers can do
236 * Submit evidence sources
237 * Suggest scenarios
238 * Participate in discussions
239
240 **Workflow**:
241
242 1. User submits text (as Reader or Contributor)
243 2. AKEL extracts claims
244 3. Checks for existing duplicates
245 4. Normalizes claim text
246 5. Assigns risk tier
247 6. Generates scenarios (draft)
248 7. Runs quality gates
249 8. Publishes as AI-Generated (Mode 2) if passes
250
251 == 8. What are "risk tiers" and why do they matter? ==
252
253 Risk tiers determine review requirements and publication workflow:
254
255 **Tier A (High Risk)**:
256
257 * **Domains**: Medical, legal, elections, safety, security, major financial
258 * **Publication**: AI can publish with warnings, expert review required for "Human-Reviewed" status
259 * **Audit rate**: Recommendation 30-50%
260 * **Why**: Potential for significant harm if wrong
261
262 **Tier B (Medium Risk)**:
263
264 * **Domains**: Complex policy, science causality, contested issues
265 * **Publication**: AI can publish immediately with clear labeling
266 * **Audit rate**: Recommendation 10-20%
267 * **Why**: Nuanced but lower immediate harm risk
268
269 **Tier C (Low Risk)**:
270
271 * **Domains**: Definitions, established facts, historical data
272 * **Publication**: AI publication default
273 * **Audit rate**: Recommendation 5-10%
274 * **Why**: Well-established, low controversy
275
276 **Assignment**:
277
278 * AKEL suggests tier based on domain, keywords, impact
279 * Moderators and Experts can override
280 * Risk tiers reviewed based on audit outcomes
281
282 == 9. How does federation work and why is it important? ==
283
284 **Federation Model**:
285
286 * Multiple independent FactHarbor nodes
287 * Each node has own database, AKEL, governance
288 * Nodes exchange claims, scenarios, evidence, verdicts
289 * No central authority
290
291 **Why Federation Matters**:
292
293 * **Resilience**: No single point of failure or censorship
294 * **Autonomy**: Communities govern themselves
295 * **Scalability**: Add nodes to handle more users
296 * **Specialization**: Domain-focused nodes (health, energy, etc.)
297 * **Trust diversity**: Multiple perspectives, not single truth source
298
299 **How Nodes Exchange Data**:
300
301 1. Local node creates versions
302 2. Builds signed bundle
303 3. Pushes to trusted neighbor nodes
304 4. Remote nodes validate signatures and lineage
305 5. Accept or branch versions
306 6. Local re-evaluation if needed
307
308 **Trust Model**:
309
310 * Trusted nodes → auto-import
311 * Neutral nodes → import with review
312 * Untrusted nodes → manual only
313
314 == 10. Can experts disagree in FactHarbor? ==
315
316 **Yes - and that's a feature, not a bug**:
317
318 **Multiple Scenarios**:
319
320 * Experts can create different scenarios with different assumptions
321 * Each scenario gets its own verdict
322 * Users see *why* experts disagree (different definitions, boundaries, evidence weighting)
323
324 **Parallel Verdicts**:
325
326 * Same scenario, different expert interpretations
327 * Both verdicts visible with expert attribution
328 * No forced consensus
329
330 **Transparency**:
331
332 * Expert reasoning documented
333 * Assumptions stated explicitly
334 * Evidence chains traceable
335 * Users can evaluate competing expert opinions
336
337 **Federation**:
338
339 * Different nodes can have different expert conclusions
340 * Cross-node branching allowed
341 * Users can see how conclusions vary across nodes
342
343 == 11. What prevents AI from hallucinating or making up facts? ==
344
345 **Multiple Safeguards**:
346
347 **Quality Gate 4: Structural Integrity**:
348
349 * Fact-checking against sources
350 * No hallucinations allowed
351 * Logic chain must be valid and traceable
352 * References must be accessible and verifiable
353
354 **Evidence Requirements**:
355
356 * Primary sources required
357 * Citations must be complete
358 * Sources must be accessible
359 * Reliability scored
360
361 **Audit System**:
362
363 * Human auditors check AI-generated content
364 * Hallucinations caught and fed back into training
365 * Patterns of errors trigger system improvements
366
367 **Transparency**:
368
369 * All reasoning chains visible
370 * Sources linked
371 * Users can verify claims against sources
372 * AKEL outputs clearly labeled
373
374 **Human Oversight**:
375
376 * Tier A requires expert review for "Human-Reviewed" status
377 * Audit sampling catches errors
378 * Community can flag issues
379
380 == 12. How does FactHarbor make money / is it sustainable? ==
381
382 [ToDo: Business model and sustainability to be defined]
383
384 Potential models under consideration:
385
386 * Non-profit foundation with grants and donations
387 * Institutional subscriptions (universities, research organizations, media)
388 * API access for third-party integrations
389 * Premium features for power users
390 * Federated node hosting services
391
392 Core principle: **Public benefit** mission takes priority over profit.
393
394
395 == 13. Related Pages ==
396
397 * [[Requirements (Roles)>>Archive.FactHarbor V0\.9\.23 Lost Data.Specification.Requirements.WebHome]]
398 * [[AKEL (AI Knowledge Extraction Layer)>>Archive.FactHarbor V0\.9\.23 Lost Data.Specification.AI Knowledge Extraction Layer (AKEL).WebHome]]
399 * [[Automation>>Archive.FactHarbor V0\.9\.23 Lost Data.Specification.Automation.WebHome]]
400 * [[Federation & Decentralization>>Archive.FactHarbor V0\.9\.23 Lost Data.Specification.Federation & Decentralization.WebHome]]
401 * [[Mission & Purpose>>Archive.FactHarbor V0\.9\.23 Lost Data.Organisation.Core Problems FactHarbor Solves.WebHome]]