Changes for page Ideal Customer Profile (ICP)
Last modified by Robert Schaub on 2026/02/08 08:32
Summary
-
Page properties (2 modified, 0 added, 0 removed)
Details
- Page properties
-
- Parent
-
... ... @@ -1,1 +1,1 @@ 1 - Archive.FactHarbor2026\.02\.08.Organisation.WebHome1 +FactHarbor.Organisation.WebHome - Content
-
... ... @@ -5,7 +5,6 @@ 5 5 == 1. Purpose == 6 6 7 7 Understanding who benefits most from FactHarbor helps us: 8 - 9 9 * **Product Development**: Prioritize features that serve core user needs 10 10 * **Marketing**: Communicate value effectively to target audiences 11 11 * **Partnerships**: Identify and cultivate strategic relationships ... ... @@ -18,13 +18,11 @@ 18 18 === 2.1 Journalists & Newsrooms === 19 19 20 20 **Profile**: 21 - 22 22 * Working journalists at news organizations (local to international) 23 23 * Fact-checkers and verification specialists 24 24 * Editorial teams producing investigative or political content 25 25 26 26 **Core Needs** (from User Needs documentation): 27 - 28 28 * **UN-4**: Fast social media fact-checking (≤15 seconds to initial verdict) 29 29 * **UN-14**: API integration into professional workflows 30 30 * **UN-5/UN-6**: Source provenance and publisher reliability tracking ... ... @@ -31,7 +31,6 @@ 31 31 * **UN-7**: Evidence transparency for editorial review 32 32 33 33 **Key Pain Points**: 34 - 35 35 * Time pressure with breaking news and viral content 36 36 * Need to verify claims quickly without sacrificing accuracy 37 37 * Difficulty tracing claims to original sources ... ... @@ -41,7 +41,6 @@ 41 41 FactHarbor provides structured, scenario-based analysis that reveals **how** conclusions are reached, saving time while providing the context needed for accurate reporting. 42 42 43 43 **Success Indicators**: 44 - 45 45 * Reduced time spent on claim verification 46 46 * Ability to cite FactHarbor analyses in published work 47 47 * Improved editorial confidence in complex stories ... ... @@ -49,7 +49,6 @@ 49 49 === 2.2 Researchers & Academics === 50 50 51 51 **Profile**: 52 - 53 53 * University researchers (political science, communications, media studies) 54 54 * Think tank analysts 55 55 * PhD students studying misinformation ... ... @@ -56,7 +56,6 @@ 56 56 * Data scientists working on verification systems 57 57 58 58 **Core Needs**: 59 - 60 60 * **UN-7**: Complete evidence transparency 61 61 * **UN-9**: Methodology transparency (auditable reasoning) 62 62 * **UN-13**: Ability to cite FactHarbor verdicts in academic work ... ... @@ -63,7 +63,6 @@ 63 63 * **UN-15**: Verdict evolution timeline (how assessments change with new evidence) 64 64 65 65 **Key Pain Points**: 66 - 67 67 * Existing fact-checks are methodologically opaque 68 68 * Need structured data for quantitative analysis 69 69 * Difficulty comparing how claims are assessed across sources ... ... @@ -73,7 +73,6 @@ 73 73 FactHarbor provides **transparent, structured methodology** that can be cited, analyzed, and built upon. The Evidence Model approach creates reusable data for academic research. 74 74 75 75 **Success Indicators**: 76 - 77 77 * Academic papers citing FactHarbor methodology 78 78 * Researchers using FactHarbor data in studies 79 79 * Methodology validation by academic institutions ... ... @@ -81,7 +81,6 @@ 81 81 === 2.3 Educators === 82 82 83 83 **Profile**: 84 - 85 85 * University professors (media literacy, critical thinking, journalism) 86 86 * High school teachers (civics, social studies, media studies) 87 87 * Librarians and information literacy specialists ... ... @@ -88,7 +88,6 @@ 88 88 * Corporate trainers (media literacy programs) 89 89 90 90 **Core Needs**: 91 - 92 92 * **UN-3**: Article summaries with FactHarbor analysis for teaching materials 93 93 * **UN-8**: Understanding disagreement and consensus (why experts differ) 94 94 * **UN-9**: Methodology transparency for pedagogical purposes ... ... @@ -95,7 +95,6 @@ 95 95 * **UN-7**: Evidence transparency to teach source evaluation 96 96 97 97 **Key Pain Points**: 98 - 99 99 * Fact-checks don't show reasoning process for teaching 100 100 * Hard to teach critical thinking with black-box verdicts 101 101 * Need tools that demonstrate **how** to evaluate claims ... ... @@ -105,7 +105,6 @@ 105 105 FactHarbor teaches the **process** of evidence evaluation, not just the answer. Students see explicit assumptions, multiple scenarios, and how confidence levels are determined. 106 106 107 107 **Success Indicators**: 108 - 109 109 * Educators integrating FactHarbor into curricula 110 110 * Student engagement with evidence exploration features 111 111 * Educational institution partnerships ... ... @@ -113,7 +113,6 @@ 113 113 === 2.4 Policy Analysts === 114 114 115 115 **Profile**: 116 - 117 117 * Government policy advisors 118 118 * NGO research staff 119 119 * Legislative aides ... ... @@ -120,7 +120,6 @@ 120 120 * Regulatory analysts 121 121 122 122 **Core Needs**: 123 - 124 124 * **UN-2/UN-3**: Context-dependent analysis (claims true under some conditions, false under others) 125 125 * **UN-8**: Understanding why reasonable people disagree 126 126 * **UN-1**: Trust assessment with explicit confidence ranges ... ... @@ -127,7 +127,6 @@ 127 127 * **UN-17**: In-article claim highlighting for briefing documents 128 128 129 129 **Key Pain Points**: 130 - 131 131 * Policy questions rarely have simple true/false answers 132 132 * Need to understand stakeholder perspectives and their evidence 133 133 * Difficulty synthesizing information from multiple sources ... ... @@ -137,7 +137,6 @@ 137 137 FactHarbor's **scenario-based analysis** explicitly maps how conclusions depend on assumptions, enabling policy analysts to present balanced, well-sourced briefings. 138 138 139 139 **Success Indicators**: 140 - 141 141 * Policy briefs citing FactHarbor analyses 142 142 * Repeat usage for complex policy questions 143 143 * Feedback on improved briefing quality ... ... @@ -145,7 +145,6 @@ 145 145 === 2.5 Content Consumers (General Public) === 146 146 147 147 **Profile**: 148 - 149 149 * Social media users seeking to verify viral claims 150 150 * Engaged citizens following news and politics 151 151 * People making decisions based on contested information ... ... @@ -152,7 +152,6 @@ 152 152 * Anyone who has been frustrated by oversimplified fact-checks 153 153 154 154 **Core Needs**: 155 - 156 156 * **UN-1**: Trust assessment at a glance (immediate visual understanding) 157 157 * **UN-4**: Fast social media fact-checking 158 158 * **UN-12**: Ability to submit unchecked claims ... ... @@ -159,7 +159,6 @@ 159 159 * **UN-17**: In-article claim highlighting when reading content 160 160 161 161 **Key Pain Points**: 162 - 163 163 * Don't trust fact-checkers' authority 164 164 * Want to understand reasoning, not just accept verdicts 165 165 * Time-constrained but want to make informed decisions ... ... @@ -169,7 +169,6 @@ 169 169 FactHarbor shows **reasoning you can inspect**. Trust comes from transparent methodology, not authority. You can form your own judgment based on visible evidence. 170 170 171 171 **Success Indicators**: 172 - 173 173 * User retention (return visits) 174 174 * Time spent exploring evidence details 175 175 * Claims submitted for verification ... ... @@ -182,7 +182,6 @@ 182 182 **Priority**: HIGH (Tier 1) 183 183 184 184 **Target Partners**: 185 - 186 186 * Swiss Broadcasting (SRG SSR, SRF, RTS, RSI) 187 187 * Major newspapers (Tamedia, NZZ) 188 188 * Regional news organizations ... ... @@ -189,12 +189,10 @@ 189 189 * Digital-first news outlets 190 190 191 191 **Partnership Value**: 192 - 193 193 * **For Partners**: Automated initial analysis saves journalist time; structured evidence for reader transparency 194 194 * **For FactHarbor**: Validation, use cases, credibility, potential funding 195 195 196 196 **Engagement Model**: 197 - 198 198 * API integration for newsroom tools 199 199 * Embedded analysis widgets 200 200 * Co-branded fact-checking initiatives ... ... @@ -205,7 +205,6 @@ 205 205 **Priority**: HIGH (Tier 1) 206 206 207 207 **Target Partners**: 208 - 209 209 * IFCN (International Fact-Checking Network) members 210 210 * EFCSN (European Fact-Checking Standards Network) members 211 211 * dpa Fact-Checking (DACH region) ... ... @@ -213,12 +213,10 @@ 213 213 * Full Fact (UK) 214 214 215 215 **Partnership Value**: 216 - 217 217 * **For Partners**: Technology platform, scalability, methodology alignment 218 218 * **For FactHarbor**: Credibility, network access, ecosystem integration 219 219 220 220 **Engagement Model**: 221 - 222 222 * Open-source technology sharing 223 223 * ClaimReview schema collaboration 224 224 * Joint methodology development ... ... @@ -229,7 +229,6 @@ 229 229 **Priority**: HIGH (Tier 1) 230 230 231 231 **Target Partners**: 232 - 233 233 * ETH Zurich / University of Zurich (Swiss, research collaboration) 234 234 * Duke Reporters' Lab (ClaimReview, Tech & Check) 235 235 * Harvard Shorenstein Center (network access) ... ... @@ -237,12 +237,10 @@ 237 237 * Oxford Reuters Institute (journalism research) 238 238 239 239 **Partnership Value**: 240 - 241 241 * **For Partners**: Research platform, real-world data, novel methodology to study 242 242 * **For FactHarbor**: Academic validation, grant access (Innosuisse), publications 243 243 244 244 **Engagement Model**: 245 - 246 246 * Research partnerships 247 247 * Student thesis projects 248 248 * Co-authored publications ... ... @@ -254,7 +254,6 @@ 254 254 **Priority**: MEDIUM (Tier 2) 255 255 256 256 **Target Partners**: 257 - 258 258 * Knight Foundation (journalism innovation) 259 259 * Google News Initiative (fact-checking fund) 260 260 * Swiss Innosuisse (research/innovation grants) ... ... @@ -262,12 +262,10 @@ 262 262 * Prototype Fund Switzerland 263 263 264 264 **Partnership Value**: 265 - 266 266 * **For Partners**: Support innovative, transparent approach to misinformation 267 267 * **For FactHarbor**: Operational funding, validation, network access 268 268 269 269 **Engagement Model**: 270 - 271 271 * Grant applications 272 272 * Progress reporting 273 273 * Impact documentation ... ... @@ -278,7 +278,6 @@ 278 278 === 4.1 Unifying Frustrations === 279 279 280 280 All ideal customers share frustration with: 281 - 282 282 * Binary "true/false" verdicts that hide complexity 283 283 * Opaque methodology ("trust us" authority model) 284 284 * Lack of explicit assumptions and confidence ranges ... ... @@ -288,7 +288,6 @@ 288 288 === 4.2 Unifying Values === 289 289 290 290 All ideal customers value: 291 - 292 292 * **Transparency**: Visible reasoning chains and methodology 293 293 * **Nuance**: Context-dependent truth (scenarios) 294 294 * **Independence**: Forming own judgment from evidence ... ... @@ -298,7 +298,6 @@ 298 298 === 4.3 Decision Criteria === 299 299 300 300 When evaluating fact-checking tools, ideal customers prioritize: 301 - 302 302 1. **Methodology Transparency**: Can I see how conclusions are reached? 303 303 2. **Evidence Quality**: Are sources traceable and credible? 304 304 3. **Nuance Handling**: Does it acknowledge complexity? ... ... @@ -310,7 +310,6 @@ 310 310 === 5.1 Awareness === 311 311 312 312 **How they find us**: 313 - 314 314 * Academic publications citing FactHarbor 315 315 * Referrals from fact-checking organizations 316 316 * Search engine results (ClaimReview schema visibility) ... ... @@ -320,7 +320,6 @@ 320 320 === 5.2 Evaluation === 321 321 322 322 **What they assess**: 323 - 324 324 * Methodology documentation (open and detailed?) 325 325 * Sample analyses (quality and transparency?) 326 326 * Open-source code (auditable?) ... ... @@ -330,7 +330,6 @@ 330 330 === 5.3 Adoption === 331 331 332 332 **How they start**: 333 - 334 334 * Submit a claim they're curious about 335 335 * Explore an existing analysis in depth 336 336 * Review methodology documentation ... ... @@ -340,7 +340,6 @@ 340 340 === 5.4 Retention === 341 341 342 342 **Why they return**: 343 - 344 344 * Consistent quality and transparency 345 345 * Time savings in verification workflow 346 346 * Unique value (scenario analysis not available elsewhere) ... ... @@ -354,13 +354,11 @@ 354 354 **Profile**: Users who want verdicts that confirm their existing beliefs 355 355 356 356 **Why Not Ideal**: 357 - 358 358 * Will be frustrated by nuanced, scenario-based analysis 359 359 * May reject conclusions that don't match expectations 360 360 * Not looking for transparent reasoning—looking for validation 361 361 362 362 **How to Handle**: 363 - 364 364 * Don't compromise methodology to satisfy them 365 365 * The transparency may eventually convert some 366 366 ... ... @@ -369,13 +369,11 @@ 369 369 **Profile**: Users who only want instant answers, no interest in evidence 370 370 371 371 **Why Not Ideal**: 372 - 373 373 * Don't value FactHarbor's core differentiator (transparency) 374 374 * Would be better served by simpler binary fact-checkers 375 375 * Won't engage with evidence or scenarios 376 376 377 377 **How to Handle**: 378 - 379 379 * Provide quick summary views (UN-1: trust at a glance) 380 380 * Make deeper exploration available but not required 381 381 ... ... @@ -384,13 +384,11 @@ 384 384 **Profile**: Users seeking to game or manipulate the system 385 385 386 386 **Why Not Ideal**: 387 - 388 388 * Waste resources 389 389 * Damage system integrity 390 390 * Not genuine users 391 391 392 392 **How to Handle**: 393 - 394 394 * AKEL detection of manipulation patterns 395 395 * Moderation for flagged escalations 396 396 * Transparent ban policies ... ... @@ -400,7 +400,6 @@ 400 400 === 7.1 Segment Metrics === 401 401 402 402 Track for each segment: 403 - 404 404 * **Acquisition**: How many from each segment? 405 405 * **Activation**: Do they complete first analysis? 406 406 * **Engagement**: Do they explore evidence? ... ... @@ -409,27 +409,28 @@ 409 409 410 410 === 7.2 Segment-Specific Success Indicators === 411 411 412 -| Segment | Key Success Metric | \\413 -|-|-| \\414 -| Journalists | API calls per newsroom; time saved per verification | \\415 -| Researchers | Papers citing FactHarbor; data downloads | \\416 -| Educators | Curricula integrations; student engagement | \\417 -| Policy Analysts | Briefings citing FactHarbor; repeat usage | \\365 +| Segment | Key Success Metric | 366 +|---------|-------------------| 367 +| Journalists | API calls per newsroom; time saved per verification | 368 +| Researchers | Papers citing FactHarbor; data downloads | 369 +| Educators | Curricula integrations; student engagement | 370 +| Policy Analysts | Briefings citing FactHarbor; repeat usage | 418 418 | Content Consumers | Retention rate; evidence exploration rate | 419 419 420 420 === 7.3 Partnership Metrics === 421 421 422 -| Partner Type | Success Metric | \\423 -|-|| \\424 -| Media | Integration count; co-published analyses | \\425 -| Fact-Checkers | Data sharing volume; methodology alignment | \\426 -| Academic | Papers published; grants received | \\375 +| Partner Type | Success Metric | 376 +|-------------|----------------| 377 +| Media | Integration count; co-published analyses | 378 +| Fact-Checkers | Data sharing volume; methodology alignment | 379 +| Academic | Papers published; grants received | 427 427 | Funders | Grants awarded; renewal rate | 428 428 429 429 == 8. Related Pages == 430 430 431 -* [[User Needs>> Archive.FactHarbor2026\.02\.08.Specification.Requirements.User Needs.WebHome]] - Detailed user need definitions432 -* [[Requirements>> Archive.FactHarbor2026\.02\.08.Specification.Requirements.WebHome]] - How user needs map to requirements384 +* [[User Needs>>FactHarbor.Specification.Requirements.User Needs.WebHome]] - Detailed user need definitions 385 +* [[Requirements>>FactHarbor.Specification.Requirements.WebHome]] - How user needs map to requirements 433 433 * [[Partnership Strategy>>FactHarbor.Organisation.Partnership-Strategy]] - Partnership opportunity details 434 434 * [[Funding & Partnerships>>FactHarbor.Organisation.Funding-Partnerships]] - Funding sources and contacts 435 435 * [[Organisational Model>>FactHarbor.Organisation.Organisational-Model]] - How FactHarbor is structured 389 +