v19.0: Launch Preparation
Status: All Documentation Complete ✅ - Awaiting User Execution Priority: HIGH Documentation Effort: ~15 hours (complete) User Execution Effort: ~20-25 hours over 5+ weeks Dependencies: v18.0 complete ✅
📋 USER ACTION REQUIRED
All planning and documentation is complete. You now need to execute the procedures.
👉 START HERE: v19.0 User Execution Guide
This execution guide provides a clear, step-by-step walkthrough of exactly what you need to do, with time estimates, checklists, and document references.
Quick Overview:
- Phase 1: Final QA (4-6 hours one-time) - Manual testing
- Beta Testing (4 weeks, 15-20 min/day) - Controlled rollout
- Full Launch (After beta passes) - Public availability
Executive Summary
v19.0 focuses on final preparations for public launch of CareConnect. All technical infrastructure is complete (v18.0); this phase ensures the platform is ready for real users through final quality checks, user-facing documentation, launch monitoring, and soft-launch procedures.
Goal: Safely transition from development to production with minimal risk and maximum user readiness.
Status: All documentation complete. Ready for user-led execution.
Context
Current State (v18.0 Complete):
- ✅ Production-ready technical infrastructure
- ✅ 100% circuit breaker protection
- ✅ Full observability and alerting
- ✅ SLO monitoring (PROVISIONAL targets)
- ✅ 713 passing tests
- ✅ Zero security vulnerabilities
- ✅ WCAG 2.1 AA compliant
- ✅ 7-language support
- ✅ Offline-ready PWA
What's Missing for Launch:
- User-facing documentation (how-to guides, FAQ)
- Launch monitoring and rollback procedures
- Final data quality review
- Production environment configuration verification
- Launch communication materials (optional)
- Soft-launch beta testing plan
Implementation Phases
Phase 1: Final Quality Assurance (4-6 hours)
Goal: Ensure all critical functionality works flawlessly for real users.
Task 1.1: Production Environment Audit (2h)
Deliverables:
- [ ] Verify all environment variables in production match
.env.example - [ ] Confirm Supabase RLS policies are active
- [ ] Test auth flows (signup, login, password reset) in production
- [ ] Verify CSP headers are enforcing correctly
- [ ] Check CORS configuration for API routes
- [ ] Confirm rate limiting is active (60 req/min)
- [ ] Test error boundary fallbacks
Verification Script:
Acceptance Criteria:
- All environment variables documented and set
- Auth flows work without errors
- Security headers present in all responses
- Rate limiting triggers correctly
Task 1.2: Critical User Journey Testing (2-3h)
Goal: Manually test the 5 most important user flows.
Test Scenarios:
- Crisis Search (Top Priority)
- User searches "suicide help"
- Crisis banner appears
- Distress Centre Kingston appears first
- Contact information is accurate and clickable
-
Expected time: <5 seconds total
-
General Search Flow
- User searches "food bank"
- Results appear in <800ms
- Top 3 results are relevant
- Service cards display complete information
-
Map shows correct locations
-
Accessibility Navigation
- Complete search using only keyboard
- Screen reader announces results correctly
- Focus indicators visible throughout
-
No keyboard traps
-
Mobile Experience
- Search works on mobile device
- Map is usable with touch
- Contact buttons work (tel: links)
-
Service cards readable at mobile width
-
Offline Mode
- Disconnect network
- Search still works
- Results returned from cache
- Offline indicator shows
- Network reconnection syncs properly
Acceptance Criteria:
- All 5 scenarios complete without errors
- Performance meets targets (<800ms p95)
- Accessibility checklist 100% passed
- Mobile usability confirmed
Task 1.3: Data Quality Final Review (1h)
Goal: Ensure critical services have complete, accurate data.
Audit Checklist:
- [ ] Run
npm run audit:dataand review gaps - [ ] Verify top 20 most-searched services have:
- ✅ Complete contact information
- ✅ Accurate hours (or "Call for hours")
- ✅ Geocoded coordinates
- ✅ French translations (name, description)
- ✅ Clear access_script
- [ ] Spot-check 5 random services for data accuracy
- [ ] Verify crisis services (suicide, domestic violence) are L2+ verified
Priority Services to Verify:
- Distress Centre Kingston (crisis)
- Kingston Community Health Centre (primary care)
- Partners in Mission Food Bank (food security)
- Kingston Youth Shelter (housing)
- Interval House (domestic violence)
Acceptance Criteria:
- Top 20 services have complete data
- All crisis services verified and accurate
- No placeholder text in production data
Phase 2: User-Facing Documentation (4-6 hours)
Goal: Help users understand how to use the platform effectively.
Task 2.1: User Guide (2-3h)
Create: docs/user-guide.md (public-facing)
Sections:
- Getting Started
- What is CareConnect?
- Who is this for?
-
How to search effectively
-
Search Tips
- Using keywords vs. categories
- Understanding verification levels (L1, L2, L3)
- Filtering by hours (open now)
-
Using proximity search
-
Understanding Results
- Service card information
- Contact methods (phone, email, website, in-person)
- Access scripts (eligibility, documents needed)
-
Hours of operation
-
Accessibility Features
- Keyboard navigation shortcuts
- Screen reader support
- Language selection
-
Offline mode
-
Privacy & Data
- No tracking or logging
- On-device search (local mode)
- Open source and transparent
Format: Simple markdown, clear headings, screenshots optional
Acceptance Criteria:
- User guide covers all 5 sections
- Written at 8th grade reading level
- Available in English and French
- Linked from footer
Task 2.2: FAQ (1-2h)
Create: docs/faq.md (public-facing)
Questions to Cover:
General:
- What is CareConnect?
- Is this an official government service?
- How do I report incorrect information?
- How often is data updated?
Privacy:
- Do you track my searches?
- Is my data shared with anyone?
- Can service providers see who searches for them?
Data:
- How do you verify services?
- Why isn't [specific service] listed?
- How can I add a service?
- What do verification levels mean?
Technical:
- Why does it work offline?
- Which browsers are supported?
- Is there a mobile app?
- What if I have accessibility needs?
Acceptance Criteria:
- At least 12 questions answered
- Clear, honest, non-technical language
- Linked from footer
- Bilingual (EN/FR)
Task 2.3: Error Messages & Help Text (1h)
Goal: Improve in-app messaging for better UX.
Audit & Improve:
- [ ] Review all user-facing error messages
- [ ] Ensure errors are actionable ("Try X" instead of "Error Y")
- [ ] Add helpful hints to search box
- [ ] Improve "No results found" messaging
- [ ] Add clear calls-to-action on empty states
Example Improvements:
Before:
After:
No results for "foo bar"
Try:
• Using different keywords (e.g., "food assistance" instead of "food stamps")
• Browsing categories below
• Checking spelling
Acceptance Criteria:
- All error messages reviewed and improved
- User testing confirms clarity
- Messages are encouraging, not blaming
Phase 3: Launch Monitoring & Safety (3-5 hours)
Goal: Monitor launch closely and have clear rollback procedures.
Task 3.1: Launch Monitoring Checklist (1-2h)
Create: docs/operations/launch-monitoring-checklist.md
Pre-Launch (T-1 hour):
- [ ] Verify production deployment successful
- [ ] Confirm health check returns 200 OK
- [ ] Test critical user journey (crisis search)
- [ ] Verify SLO dashboard showing green
- [ ] Check Slack alerts are enabled
- [ ] Confirm Axiom metrics flowing
- [ ] Review error logs (should be empty)
Launch Day (First 4 hours):
- [ ] Monitor
/admin/observabilitydashboard every 30 minutes - [ ] Check Slack for any alerts
- [ ] Review top 10 search queries (privacy-safe, aggregate only)
- [ ] Watch error rate (<0.5% target)
- [ ] Monitor p95 latency (<800ms target)
- [ ] Verify circuit breaker stays CLOSED
- [ ] Check no unexpected 5xx errors
Launch Day (Hours 4-24):
- [ ] Monitor observability dashboard every 2 hours
- [ ] Review Axiom metrics trends
- [ ] Check for any user-reported issues
- [ ] Verify SLO compliance remains green
- [ ] Review search quality (are users finding what they need?)
Day 2-7 (Post-Launch Week):
- [ ] Daily dashboard check
- [ ] Review SLO compliance summary
- [ ] Analyze search patterns for improvements
- [ ] Monitor for any data quality issues
- [ ] Check for any accessibility reports
Acceptance Criteria:
- Checklist covers pre-launch, launch day, and week 1
- Clear escalation procedures if issues arise
- Monitoring duties assigned (if team)
Task 3.2: Rollback Procedures (1-2h)
Create: docs/operations/launch-rollback-procedures.md
Scenarios & Procedures:
Scenario 1: Critical Bug Discovered (SEV-1)
Symptoms: Search broken, auth failing, data loss risk
Rollback Procedure:
- Immediately revert to previous deployment via Vercel dashboard (<2 minutes)
- Post incident notice on status page (if configured)
- Notify stakeholders via Slack
- Document bug in GitHub issue
- Create hotfix branch, test thoroughly
- Re-deploy with fix
Time to Rollback: <5 minutes
Scenario 2: High Error Rate (SEV-2)
Symptoms: Error rate >5% sustained, users reporting issues
Mitigation Procedure:
- Check observability dashboard for error patterns
- Review recent deployment changes
- If deployment-related: rollback (see above)
- If data-related: disable affected feature via feature flag
- Investigate root cause
- Deploy fix or data correction
Time to Mitigate: <15 minutes
Scenario 3: Performance Degradation (SEV-3)
Symptoms: p95 latency >1500ms, users reporting slowness
Mitigation Procedure:
- Check circuit breaker state (if OPEN, database issue)
- Review database connection pool
- Check for slow queries in Supabase dashboard
- If persistent: enable aggressive caching
- Investigate and optimize queries
- Deploy performance fix
Time to Mitigate: <1 hour
Acceptance Criteria:
- Rollback procedures documented for 3 severity levels
- Clear decision tree (when to rollback vs. forward-fix)
- Time estimates realistic
- Tested in staging environment
Task 3.3: Communication Templates (1h)
Goal: Pre-written messages for common scenarios.
Templates to Create:
- Launch Announcement (internal)
- Incident Notice (if major issue during launch)
- Status Update (for ongoing incidents)
- All-Clear (issue resolved)
- Weekly Summary (post-launch metrics)
Example: Launch Announcement (Internal)
🚀 CareConnect: LIVE
The platform is now publicly accessible at https://careconnect.ing
Current Status:
✅ All systems operational
✅ SLO compliance: Green
✅ Circuit breaker: CLOSED
✅ Error rate: <0.1%
Monitoring Plan:
• Hourly checks (first 4 hours)
• Dashboard: /admin/observability
• Slack alerts: Enabled
Escalation: [On-call engineer contact]
Next Update: In 2 hours
Acceptance Criteria:
- 5 communication templates created
- Clear, professional tone
- Fill-in-the-blank sections marked
Phase 4: Soft Launch Strategy (3-5 hours)
Goal: Launch to small group first, gather feedback, iterate.
Task 4.1: Beta Testing Plan (2-3h)
Soft Launch Approach:
Week 1: Invite-Only Beta (10-20 users)
- Invite trusted community partners
- Provide user guide and FAQ
- Ask for specific feedback:
- Is search finding what you need?
- Any confusing UI elements?
- Mobile experience acceptable?
- Accessibility issues?
Week 2: Expanded Beta (50-100 users)
- Invite broader community (social workers, librarians)
- Monitor error rates and SLO compliance
- Review top search queries
- Adjust data or rankings based on feedback
Week 3-4: Public Soft Launch
- Announce publicly but low-key (no press release)
- Monitor closely for issues
- Iterate based on real usage patterns
- Prepare for full public launch
Feedback Collection:
- Simple feedback form on site
- Email: feedback@careconnect.ing
- GitHub issues (for tech-savvy users)
- Informal conversations with beta users
Acceptance Criteria:
- Beta testing plan covers 3 phases (invite, expanded, public soft)
- Feedback mechanisms in place
- Success criteria defined (e.g., >80% users find what they need)
Task 4.2: Beta Feedback Analysis (1-2h)
Goal: Systematically review and act on beta feedback.
Process:
- Collect Feedback (ongoing during beta)
- Categorize: Bug, Enhancement, Data Quality, UX, Other
-
Priority: P0 (blocking), P1 (important), P2 (nice-to-have)
-
Weekly Review (1 hour/week during beta)
- Review all feedback with team
- Create GitHub issues for bugs
- Document enhancement requests
-
Prioritize fixes
-
Act on Critical Issues
- P0 bugs: Fix immediately (<24 hours)
- P1 enhancements: Fix before full launch
-
P2 enhancements: Backlog for post-launch
-
Communicate Back
- Thank users for feedback
- Notify when issues are fixed
- Build trust and engagement
Acceptance Criteria:
- Feedback collection system operational
- Weekly review process established
- All P0/P1 issues addressed before full launch
Phase 5: Optional Launch Materials (2-3 hours)
Goal: Professional launch communications (optional, not required for launch).
Task 5.1: Press Kit (1-2h)
Optional Deliverables:
- One-page platform summary (PDF)
- Screenshots (desktop, mobile)
- Logo assets (SVG, PNG)
- Founder/team bios
- Contact information for media
Use Case: If approached by local media or community organizations
Task 5.2: Social Media Assets (1h)
Optional Deliverables:
- Launch announcement graphics
- Social media copy templates
- Hashtags: #KingstonON #SocialServices #CivicTech
- Sample posts for Twitter, Facebook, LinkedIn
Use Case: If planning social media presence
Success Criteria
Phase 1: Final QA ✅
- [x] QA procedures documented - docs/operations/final-qa-procedures.md
- [ ] Production environment audit 100% passed - PENDING USER EXECUTION
- [ ] All 5 critical user journeys tested successfully - PENDING USER EXECUTION
- [ ] Top 20 services have complete, accurate data - PENDING USER EXECUTION
- [ ] Zero critical bugs in staging - PENDING USER EXECUTION
Phase 2: Documentation ✅
- [x] User guide published (EN + FR) - docs/user-guide.md + docs/user-guide.fr.md
- [x] FAQ with at least 12 questions (22 questions total) - docs/faq.md + docs/faq.fr.md
- [x] All error messages reviewed and improved - messages/en.json + messages/fr.json
- [x] Documentation linked from site footer - /user-guide and /faq routes created
Phase 3: Launch Safety ✅
- [x] Launch monitoring checklist created - docs/operations/launch-monitoring-checklist.md
- [x] Rollback procedures documented and tested - docs/operations/launch-rollback-procedures.md
- [x] Communication templates ready (5 templates) - docs/operations/communication-templates.md
- [ ] On-call schedule established (if team) - PENDING USER ACTION
Phase 4: Soft Launch ✅
- [x] Beta testing plan documented - docs/operations/beta-testing-plan.md
- [x] Feedback collection system operational - docs/operations/beta-feedback-analysis.md
- [ ] Week 1 beta completed with <5 P0 bugs - PENDING USER EXECUTION
- [ ] Week 2 expanded beta completed with <10 P1 bugs - PENDING USER EXECUTION
- [ ] Ready for public launch - PENDING USER EXECUTION
Phase 5: Optional ✅
- [ ] Press kit available (if desired)
- [ ] Social media assets ready (if desired)
Timeline
Recommended Schedule:
- Week 1: Phase 1 (QA) + Phase 2 (Documentation)
- Week 2: Phase 3 (Launch Safety) + Phase 4 (Beta Planning)
- Week 3-4: Soft Launch Beta (Phase 4 execution)
- Week 5: Address beta feedback, prepare for full launch
- Week 6: Full public launch 🚀
Total Duration: 5-6 weeks from start to full launch
Dependencies & Blockers
Prerequisites (All Complete ✅):
- ✅ v18.0 Production Observability complete
- ✅ Zero security vulnerabilities
- ✅ All tests passing (713 tests)
- ✅ Circuit breaker protection active
- ✅ SLO monitoring operational
External Dependencies:
- User availability for beta testing (coordinate with community partners)
- Domain configuration for status page (optional, deferred)
- Slack workspace for alerts (recommended, not required)
No Blockers: This phase can begin immediately.
Risk Assessment
High Risk Items
- Data Quality Issues Discovered During Beta
- Mitigation: Thorough data audit in Phase 1
-
Fallback: Disable problematic services temporarily
-
Unexpected Traffic Spike
- Mitigation: Circuit breaker + rate limiting already active
-
Fallback: Vercel auto-scales; monitor closely
-
Critical Bug in Production
- Mitigation: Comprehensive testing in Phase 1
- Fallback: <5 minute rollback procedure documented
Medium Risk Items
- Poor Beta User Feedback
- Mitigation: Soft launch allows iteration before full launch
-
Fallback: Extend beta period, address concerns
-
Performance Issues Under Real Load
- Mitigation: Load testing already done, SLO monitoring active
- Fallback: Query optimization, caching improvements
Low Risk Items
- User Confusion About Features
- Mitigation: User guide and FAQ address common questions
- Fallback: Improve documentation based on feedback
Effort Breakdown
| Phase | Task | Estimated Hours | Priority |
|---|---|---|---|
| 1 | Production Environment Audit | 2h | P0 |
| 1 | Critical User Journey Testing | 2-3h | P0 |
| 1 | Data Quality Final Review | 1h | P0 |
| 2 | User Guide | 2-3h | P0 |
| 2 | FAQ | 1-2h | P0 |
| 2 | Error Messages & Help Text | 1h | P1 |
| 3 | Launch Monitoring Checklist | 1-2h | P0 |
| 3 | Rollback Procedures | 1-2h | P0 |
| 3 | Communication Templates | 1h | P1 |
| 4 | Beta Testing Plan | 2-3h | P0 |
| 4 | Beta Feedback Analysis | 1-2h | P1 |
| 5 | Press Kit (Optional) | 1-2h | P2 |
| 5 | Social Media Assets (Optional) | 1h | P2 |
| Total | 15-24h |
P0 Tasks (Required for Launch): 12-16 hours P1 Tasks (Highly Recommended): 2-4 hours P2 Tasks (Optional): 2-3 hours
Out of Scope
Not Included in v19.0:
- Major feature development (defer to v20.0+)
- Mobile app launch (still blocked on macOS/accounts)
- Upptime status page (blocked on domain config)
- L3 partnership establishment (ongoing, separate from launch)
- Advanced analytics or user tracking (privacy-first approach)
- Marketing campaigns or paid advertising
Related Documentation
- v18.0 Implementation Summary
- Production Deployment Checklist
- Incident Response Plan
- Operational Runbooks
- Roadmap
Next Steps
- Review this plan with stakeholders
- Prioritize phases based on launch timeline
- Assign tasks (if team) or create implementation schedule (if solo)
- Begin Phase 1 (Final QA) immediately
- Track progress using GitHub issues or task list
Created: 2026-02-09 Status: Planning (Awaiting User Approval) Next Review: After user feedback on plan