Wiki source code of Continuous Improvement
Last modified by Robert Schaub on 2026/02/08 08:29
Hide last authors
| author | version | line-number | content |
|---|---|---|---|
| |
1.1 | 1 | = Continuous Improvement = |
| |
1.2 | 2 | |
| |
1.1 | 3 | **From Sociocracy 3.0**: Empirical approach to improving FactHarbor systems. |
| |
1.2 | 4 | |
| |
1.1 | 5 | == 1. Philosophy == |
| |
1.2 | 6 | |
| |
1.1 | 7 | **Continuous improvement** means: |
| |
1.2 | 8 | |
| |
1.1 | 9 | * We're never "done" - systems always improve |
| 10 | * Learn from data, not opinions | ||
| 11 | * Small experiments, frequent iteration | ||
| 12 | * Measure everything | ||
| 13 | * Build, measure, learn, repeat | ||
| 14 | **Inspired by**: | ||
| 15 | * Sociocracy 3.0 empiricism principle | ||
| 16 | * Agile/lean methodologies | ||
| 17 | * Scientific method | ||
| 18 | * DevOps continuous deployment | ||
| |
1.2 | 19 | |
| |
1.1 | 20 | == 2. What We Improve == |
| |
1.2 | 21 | |
| |
1.1 | 22 | === 2.1 AKEL Performance === |
| |
1.2 | 23 | |
| |
1.1 | 24 | **Processing speed**: |
| |
1.2 | 25 | |
| |
1.1 | 26 | * Faster claim parsing |
| 27 | * Optimized evidence extraction | ||
| 28 | * Efficient source lookups | ||
| 29 | * Reduced latency | ||
| 30 | **Quality**: | ||
| 31 | * Better evidence detection | ||
| 32 | * More accurate verdicts | ||
| 33 | * Improved source scoring | ||
| 34 | * Enhanced contradiction detection | ||
| 35 | **Reliability**: | ||
| 36 | * Fewer errors | ||
| 37 | * Better error handling | ||
| 38 | * Graceful degradation | ||
| 39 | * Faster recovery | ||
| |
1.2 | 40 | |
| |
1.1 | 41 | === 2.2 Policies === |
| |
1.2 | 42 | |
| |
1.1 | 43 | **Risk tier definitions**: |
| |
1.2 | 44 | |
| |
1.1 | 45 | * Clearer criteria |
| 46 | * Better domain coverage | ||
| 47 | * Edge case handling | ||
| 48 | **Evidence weighting**: | ||
| 49 | * More appropriate weights by domain | ||
| 50 | * Better peer-review recognition | ||
| 51 | * Improved recency handling | ||
| 52 | **Source scoring**: | ||
| 53 | * More nuanced credibility assessment | ||
| 54 | * Better handling of new sources | ||
| 55 | * Domain-specific adjustments | ||
| |
1.2 | 56 | |
| |
1.1 | 57 | === 2.3 Infrastructure === |
| |
1.2 | 58 | |
| |
1.1 | 59 | **Performance**: |
| |
1.2 | 60 | |
| |
1.1 | 61 | * Database optimization |
| 62 | * Caching strategies | ||
| 63 | * Network efficiency | ||
| 64 | * Resource utilization | ||
| 65 | **Scalability**: | ||
| 66 | * Handle more load | ||
| 67 | * Geographic distribution | ||
| 68 | * Cost efficiency | ||
| 69 | **Monitoring**: | ||
| 70 | * Better dashboards | ||
| 71 | * Faster alerts | ||
| 72 | * More actionable metrics | ||
| |
1.2 | 73 | |
| |
1.1 | 74 | === 2.4 Processes === |
| |
1.2 | 75 | |
| |
1.1 | 76 | **Contributor workflows**: |
| |
1.2 | 77 | |
| |
1.1 | 78 | * Easier onboarding |
| 79 | * Clearer documentation | ||
| 80 | * Better tools | ||
| 81 | **Decision-making**: | ||
| 82 | * Faster decisions | ||
| 83 | * Better documentation | ||
| 84 | * Clearer escalation | ||
| |
1.2 | 85 | |
| |
1.1 | 86 | == 3. Improvement Cycle == |
| |
1.2 | 87 | |
| |
1.1 | 88 | === 3.1 Observe === |
| |
1.2 | 89 | |
| |
1.1 | 90 | **Continuously monitor**: |
| |
1.2 | 91 | |
| |
1.1 | 92 | * Performance metrics dashboards |
| 93 | * User feedback patterns | ||
| 94 | * AKEL processing logs | ||
| 95 | * Error reports | ||
| 96 | * Community discussions | ||
| 97 | **Look for**: | ||
| 98 | * Metrics outside acceptable ranges | ||
| 99 | * Systematic patterns in errors | ||
| 100 | * User pain points | ||
| 101 | * Opportunities for optimization | ||
| |
1.2 | 102 | |
| |
1.1 | 103 | === 3.2 Analyze === |
| |
1.2 | 104 | |
| |
1.1 | 105 | **Dig deeper**: |
| |
1.2 | 106 | |
| |
1.1 | 107 | * Why is this metric problematic? |
| 108 | * Is this a systematic issue or one-off? | ||
| 109 | * What's the root cause? | ||
| 110 | * What patterns exist? | ||
| 111 | * How widespread is this? | ||
| 112 | **Tools**: | ||
| 113 | * Data analysis (SQL queries, dashboards) | ||
| 114 | * Code profiling | ||
| 115 | * A/B test results | ||
| 116 | * User interviews | ||
| 117 | * Historical comparison | ||
| |
1.2 | 118 | |
| |
1.1 | 119 | === 3.3 Hypothesize === |
| |
1.2 | 120 | |
| |
1.1 | 121 | **Propose explanation**: |
| |
1.2 | 122 | |
| |
1.1 | 123 | * "We believe X is happening because Y" |
| 124 | * "If we change Z, we expect W to improve" | ||
| 125 | * "The root cause is likely A, not B" | ||
| 126 | **Make testable**: | ||
| 127 | * What would prove this hypothesis? | ||
| 128 | * What would disprove it? | ||
| 129 | * What metrics would change? | ||
| |
1.2 | 130 | |
| |
1.1 | 131 | === 3.4 Design Solution === |
| |
1.2 | 132 | |
| |
1.1 | 133 | **Propose specific change**: |
| |
1.2 | 134 | |
| |
1.1 | 135 | * Algorithm adjustment |
| 136 | * Policy clarification | ||
| 137 | * Infrastructure upgrade | ||
| 138 | * Process refinement | ||
| 139 | **Consider**: | ||
| 140 | * Trade-offs | ||
| 141 | * Risks | ||
| 142 | * Rollback plan | ||
| 143 | * Success metrics | ||
| |
1.2 | 144 | |
| |
1.1 | 145 | === 3.5 Test === |
| |
1.2 | 146 | |
| |
1.1 | 147 | **Before full deployment**: |
| |
1.2 | 148 | |
| |
1.1 | 149 | * Test environment deployment |
| 150 | * Historical data validation | ||
| 151 | * A/B testing if feasible | ||
| 152 | * Load testing if infrastructure | ||
| 153 | **Measure**: | ||
| 154 | * Did metrics improve as expected? | ||
| 155 | * Any unexpected side effects? | ||
| 156 | * Is the improvement statistically significant? | ||
| |
1.2 | 157 | |
| |
1.1 | 158 | === 3.6 Deploy === |
| |
1.2 | 159 | |
| |
1.1 | 160 | **Gradual rollout**: |
| |
1.2 | 161 | |
| |
1.1 | 162 | * Deploy to small % of traffic first |
| 163 | * Monitor closely | ||
| 164 | * Increase gradually if successful | ||
| 165 | * Rollback if problems | ||
| 166 | **Deployment strategies**: | ||
| 167 | * Canary (1% → 5% → 25% → 100%) | ||
| 168 | * Blue-green (instant swap with rollback ready) | ||
| 169 | * Feature flags (enable for specific users first) | ||
| |
1.2 | 170 | |
| |
1.1 | 171 | === 3.7 Evaluate === |
| |
1.2 | 172 | |
| |
1.1 | 173 | **After deployment**: |
| |
1.2 | 174 | |
| |
1.1 | 175 | * Review metrics - did they improve? |
| 176 | * User feedback - positive or negative? | ||
| 177 | * Unexpected effects - any surprises? | ||
| 178 | * Lessons learned - what would we do differently? | ||
| |
1.2 | 179 | |
| |
1.1 | 180 | === 3.8 Iterate === |
| |
1.2 | 181 | |
| |
1.1 | 182 | **Based on results**: |
| |
1.2 | 183 | |
| |
1.1 | 184 | * If successful: Document, celebrate, move to next improvement |
| 185 | * If partially successful: Refine and iterate | ||
| 186 | * If unsuccessful: Rollback, analyze why, try different approach | ||
| 187 | **Document learnings**: Update RFC with actual outcomes. | ||
| |
1.2 | 188 | |
| |
1.1 | 189 | == 4. Improvement Cadence == |
| |
1.2 | 190 | |
| |
1.1 | 191 | === 4.1 Continuous (Ongoing) === |
| |
1.2 | 192 | |
| |
1.1 | 193 | **Daily/Weekly**: |
| |
1.2 | 194 | |
| |
1.1 | 195 | * Monitor dashboards |
| 196 | * Review user feedback | ||
| 197 | * Identify emerging issues | ||
| 198 | * Quick fixes and patches | ||
| 199 | **Who**: Technical Coordinator, Community Coordinator | ||
| |
1.2 | 200 | |
| |
1.1 | 201 | === 4.2 Sprint Cycles (2 weeks) === |
| |
1.2 | 202 | |
| |
1.1 | 203 | **Every 2 weeks**: |
| |
1.2 | 204 | |
| |
1.1 | 205 | * Sprint planning: Select improvements to tackle |
| 206 | * Implementation: Build and test | ||
| 207 | * Sprint review: Demo what was built | ||
| 208 | * Retrospective: How can we improve the improvement process? | ||
| 209 | **Who**: Core team + regular contributors | ||
| |
1.2 | 210 | |
| |
1.1 | 211 | === 4.3 Quarterly Reviews (3 months) === |
| |
1.2 | 212 | |
| |
1.1 | 213 | **Every quarter**: |
| |
1.2 | 214 | |
| |
1.1 | 215 | * Comprehensive performance review |
| 216 | * Policy effectiveness assessment | ||
| 217 | * Strategic improvement priorities | ||
| 218 | * Architectural decisions | ||
| 219 | **Who**: Governing Team + Technical Coordinator | ||
| 220 | **Output**: Quarterly report, next quarter priorities | ||
| |
1.2 | 221 | |
| |
1.1 | 222 | === 4.4 Annual Planning (Yearly) === |
| |
1.2 | 223 | |
| |
1.1 | 224 | **Annually**: |
| |
1.2 | 225 | |
| |
1.1 | 226 | * Major strategic direction |
| 227 | * Significant architectural changes | ||
| 228 | * Multi-quarter initiatives | ||
| 229 | * Budget allocation | ||
| 230 | **Who**: General Assembly | ||
| |
1.2 | 231 | |
| |
1.1 | 232 | == 5. Metrics-Driven Improvement == |
| |
1.2 | 233 | |
| |
1.1 | 234 | === 5.1 Key Performance Indicators (KPIs) === |
| |
1.2 | 235 | |
| |
1.1 | 236 | **AKEL Performance**: |
| |
1.2 | 237 | |
| |
1.1 | 238 | * Processing time (P50, P95, P99) |
| 239 | * Success rate | ||
| 240 | * Evidence completeness | ||
| 241 | * Confidence distribution | ||
| 242 | **Content Quality**: | ||
| 243 | * User feedback (helpful/unhelpful ratio) | ||
| 244 | * Contradiction rate | ||
| 245 | * Source diversity | ||
| 246 | * Scenario coverage | ||
| 247 | **System Health**: | ||
| 248 | * Uptime | ||
| 249 | * Error rate | ||
| 250 | * Response time | ||
| 251 | * Resource utilization | ||
| 252 | **See**: [[System Performance Metrics>>FactHarbor.Specification.System-Performance-Metrics]] | ||
| |
1.2 | 253 | |
| |
1.1 | 254 | === 5.2 Targets and Thresholds === |
| |
1.2 | 255 | |
| |
1.1 | 256 | **For each metric**: |
| |
1.2 | 257 | |
| |
1.1 | 258 | * Target: Where we want to be |
| 259 | * Acceptable range: What's OK | ||
| 260 | * Alert threshold: When to intervene | ||
| 261 | * Critical threshold: Emergency | ||
| 262 | **Example**: | ||
| 263 | * Processing time P95 | ||
| |
1.2 | 264 | * Target: 15 seconds |
| 265 | * Acceptable: 10-18 seconds | ||
| 266 | * Alert: >20 seconds | ||
| 267 | * Critical: >30 seconds | ||
| 268 | |||
| |
1.1 | 269 | === 5.3 Metric-Driven Decisions === |
| |
1.2 | 270 | |
| |
1.1 | 271 | **Improvements prioritized by**: |
| |
1.2 | 272 | |
| |
1.1 | 273 | * Impact on metrics |
| 274 | * Effort required | ||
| 275 | * Risk level | ||
| 276 | * Strategic importance | ||
| 277 | **Not by**: | ||
| 278 | * Personal preferences | ||
| 279 | * Loudest voice | ||
| 280 | * Political pressure | ||
| 281 | * Gut feeling | ||
| |
1.2 | 282 | |
| |
1.1 | 283 | == 6. Experimentation == |
| |
1.2 | 284 | |
| |
1.1 | 285 | === 6.1 A/B Testing === |
| |
1.2 | 286 | |
| |
1.1 | 287 | **When feasible**: |
| |
1.2 | 288 | |
| |
1.1 | 289 | * Run two versions simultaneously |
| 290 | * Randomly assign users/claims | ||
| 291 | * Measure comparative performance | ||
| 292 | * Choose winner based on data | ||
| 293 | **Good for**: | ||
| 294 | * Algorithm parameter tuning | ||
| 295 | * UI/UX changes | ||
| 296 | * Policy variations | ||
| |
1.2 | 297 | |
| |
1.1 | 298 | === 6.2 Canary Deployments === |
| |
1.2 | 299 | |
| |
1.1 | 300 | **Small-scale first**: |
| |
1.2 | 301 | |
| |
1.1 | 302 | * Deploy to 1% of traffic |
| 303 | * Monitor closely for issues | ||
| 304 | * Gradually increase if successful | ||
| 305 | * Full rollback if problems | ||
| 306 | **Benefits**: | ||
| 307 | * Limits blast radius of failures | ||
| 308 | * Real-world validation | ||
| 309 | * Quick feedback loop | ||
| |
1.2 | 310 | |
| |
1.1 | 311 | === 6.3 Feature Flags === |
| |
1.2 | 312 | |
| |
1.1 | 313 | **Controlled rollout**: |
| |
1.2 | 314 | |
| |
1.1 | 315 | * Deploy code but disable by default |
| 316 | * Enable for specific users/scenarios | ||
| 317 | * Gather feedback before full release | ||
| 318 | * Easy enable/disable without redeployment | ||
| |
1.2 | 319 | |
| |
1.1 | 320 | == 7. Retrospectives == |
| |
1.2 | 321 | |
| |
1.1 | 322 | === 7.1 Sprint Retrospectives (Every 2 weeks) === |
| |
1.2 | 323 | |
| |
1.1 | 324 | **Questions**: |
| |
1.2 | 325 | |
| |
1.1 | 326 | * What went well? |
| 327 | * What could be improved? | ||
| 328 | * What will we commit to improving? | ||
| 329 | **Format** (30 minutes): | ||
| 330 | * Gather data: Everyone writes thoughts (5 min) | ||
| 331 | * Generate insights: Discuss patterns (15 min) | ||
| 332 | * Decide actions: Pick 1-3 improvements (10 min) | ||
| 333 | **Output**: 1-3 concrete actions for next sprint | ||
| |
1.2 | 334 | |
| |
1.1 | 335 | === 7.2 Project Retrospectives (After major changes) === |
| |
1.2 | 336 | |
| |
1.1 | 337 | **After significant changes**: |
| |
1.2 | 338 | |
| |
1.1 | 339 | * What was the goal? |
| 340 | * What actually happened? | ||
| 341 | * What went well? | ||
| 342 | * What went poorly? | ||
| 343 | * What did we learn? | ||
| 344 | * What would we do differently? | ||
| 345 | **Document**: Update project documentation with learnings | ||
| |
1.2 | 346 | |
| |
1.1 | 347 | === 7.3 Incident Retrospectives (After failures) === |
| |
1.2 | 348 | |
| |
1.1 | 349 | **After incidents/failures**: |
| |
1.2 | 350 | |
| |
1.1 | 351 | * Timeline: What happened when? |
| 352 | * Root cause: Why did it happen? | ||
| 353 | * Impact: What was affected? | ||
| 354 | * Response: How did we handle it? | ||
| 355 | * Prevention: How do we prevent this? | ||
| 356 | **Blameless**: Focus on systems, not individuals. | ||
| 357 | **Output**: Action items to prevent recurrence | ||
| |
1.2 | 358 | |
| |
1.1 | 359 | == 8. Knowledge Management == |
| |
1.2 | 360 | |
| |
1.1 | 361 | === 8.1 Documentation === |
| |
1.2 | 362 | |
| |
1.1 | 363 | **Keep updated**: |
| |
1.2 | 364 | |
| |
1.1 | 365 | * Architecture docs |
| 366 | * API documentation | ||
| 367 | * Operational runbooks | ||
| 368 | * Decision records | ||
| 369 | * Retrospective notes | ||
| 370 | **Principle**: Future you/others need to understand why decisions were made. | ||
| |
1.2 | 371 | |
| |
1.1 | 372 | === 8.2 Decision Records === |
| |
1.2 | 373 | |
| |
1.1 | 374 | **For significant decisions, document**: |
| |
1.2 | 375 | |
| |
1.1 | 376 | * What was decided? |
| 377 | * What problem does this solve? | ||
| 378 | * What alternatives were considered? | ||
| 379 | * What are the trade-offs? | ||
| 380 | * What are the success metrics? | ||
| 381 | * Review date? | ||
| 382 | **See**: [[Decision Processes>>FactHarbor.Organisation.Decision-Processes]] | ||
| |
1.2 | 383 | |
| |
1.1 | 384 | === 8.3 Learning Library === |
| |
1.2 | 385 | |
| |
1.1 | 386 | **Collect**: |
| |
1.2 | 387 | |
| |
1.1 | 388 | * Failed experiments (what didn't work) |
| 389 | * Successful patterns (what worked well) | ||
| 390 | * External research relevant to FactHarbor | ||
| 391 | * Best practices from similar systems | ||
| 392 | **Share**: Make accessible to all contributors | ||
| |
1.2 | 393 | |
| |
1.1 | 394 | == 9. Continuous Improvement of Improvement == |
| |
1.2 | 395 | |
| |
1.1 | 396 | **Meta-improvement**: Improve how we improve. |
| 397 | **Questions to ask**: | ||
| |
1.2 | 398 | |
| |
1.1 | 399 | * Is our improvement cycle effective? |
| 400 | * Are we measuring the right things? | ||
| 401 | * Are decisions actually data-driven? | ||
| 402 | * Is knowledge being captured? | ||
| 403 | * Are retrospectives actionable? | ||
| 404 | * Are improvements sustained? | ||
| 405 | **Annually review**: How can our improvement process itself improve? | ||
| |
1.2 | 406 | |
| |
1.1 | 407 | == 10. Cultural Practices == |
| |
1.2 | 408 | |
| |
1.1 | 409 | === 10.1 Safe to Fail === |
| |
1.2 | 410 | |
| |
1.1 | 411 | **Encourage experimentation**: |
| |
1.2 | 412 | |
| |
1.1 | 413 | * ✅ Try new approaches |
| 414 | * ✅ Test hypotheses | ||
| 415 | * ✅ Learn from failures | ||
| 416 | * ✅ Share what didn't work | ||
| 417 | **Not blame**: | ||
| 418 | * ❌ "Who broke it?" | ||
| 419 | * ❌ "Why didn't you know?" | ||
| 420 | * ❌ "This was a stupid idea" | ||
| 421 | **Instead**: | ||
| 422 | * ✅ "What did we learn?" | ||
| 423 | * ✅ "How can we prevent this?" | ||
| 424 | * ✅ "What will we try next?" | ||
| |
1.2 | 425 | |
| |
1.1 | 426 | === 10.2 Data Over Opinions === |
| |
1.2 | 427 | |
| |
1.1 | 428 | **Settle debates with**: |
| |
1.2 | 429 | |
| |
1.1 | 430 | * ✅ Metrics and measurements |
| 431 | * ✅ A/B test results | ||
| 432 | * ✅ User feedback data | ||
| 433 | * ✅ Performance benchmarks | ||
| 434 | **Not with**: | ||
| 435 | * ❌ "I think..." | ||
| 436 | * ❌ "In my experience..." | ||
| 437 | * ❌ "I've seen this before..." | ||
| 438 | * ❌ "Trust me..." | ||
| |
1.2 | 439 | |
| |
1.1 | 440 | === 10.3 Bias Toward Action === |
| |
1.2 | 441 | |
| |
1.1 | 442 | **Good enough for now, safe enough to try**: |
| |
1.2 | 443 | |
| |
1.1 | 444 | * Don't wait for perfect solution |
| 445 | * Test and learn | ||
| 446 | * Iterate quickly | ||
| 447 | * Prefer reversible decisions | ||
| 448 | **But not reckless**: | ||
| 449 | * Do test before deploying | ||
| 450 | * Do monitor after deploying | ||
| 451 | * Do have rollback plan | ||
| 452 | * Do document decisions | ||
| |
1.2 | 453 | |
| |
1.1 | 454 | == 11. Tools and Infrastructure == |
| |
1.2 | 455 | |
| |
1.1 | 456 | **Support continuous improvement with**: |
| 457 | **Monitoring**: | ||
| |
1.2 | 458 | |
| |
1.1 | 459 | * Real-time dashboards |
| 460 | * Alerting systems | ||
| 461 | * Log aggregation | ||
| 462 | * Performance profiling | ||
| 463 | **Testing**: | ||
| 464 | * Automated testing (unit, integration, regression) | ||
| 465 | * Test environments | ||
| 466 | * A/B testing framework | ||
| 467 | * Load testing tools | ||
| 468 | **Deployment**: | ||
| 469 | * CI/CD pipelines | ||
| 470 | * Canary deployment support | ||
| 471 | * Feature flag system | ||
| 472 | * Quick rollback capability | ||
| 473 | **Collaboration**: | ||
| 474 | * RFC repository | ||
| 475 | * Decision log | ||
| 476 | * Knowledge base | ||
| 477 | * Retrospective notes | ||
| 478 | --- | ||
| |
1.2 | 479 | **Remember**: Continuous improvement means we're always learning, always testing, always getting better.-- |
| 480 | |||
| |
1.1 | 481 | == 12. Related Pages == |
| |
1.2 | 482 | |
| |
1.1 | 483 | * [[Automation Philosophy>>FactHarbor.Organisation.Automation-Philosophy]] - Why we automate |
| 484 | * [[System Performance Metrics>>FactHarbor.Specification.System-Performance-Metrics]] - What we measure | ||
| 485 | * [[Contributor Processes>>FactHarbor.Organisation.Contributor-Processes]] - How to propose improvements | ||
| |
1.2 | 486 | * [[Governance>>Archive.FactHarbor 2026\.02\.08.Organisation.Governance.WebHome]] - How improvements are approved |