← Back to Clients

Premera Blue Cross

Engagedhigh
Primary: Dave BrazaUpdated 2025-01-28

Premera Day 0–30 Execution Plan

Created: 2026-02-16 Status: DRAFT — needs Thomas answers on open questions before finalizing


Context

Premera engagement: FDE/consulting at $150/hr, ~40h/week, 3-6 months. Two work streams: Auto-Authorization (augmenting existing system for ~6→hundreds of conditions) and Appeals (going to production March 1 with Elementum workflow). Working inside Premera's environment (Azure AKS, their API keys, Phoenix tracing, data stays in VPC).


Week 1: Environment + Data + First Workflow Baseline (Days 1-5)

Outcomes

  1. Credentialed and connected — laptops/VPN/credentials provisioned, access to Facets, ADT feeds, medical policies, Elementum
  2. Colt/Jamie kickoff complete — daily standup cadence agreed, Slack/Teams channel live, POC confirmed
  3. Appeals workflow mapped end-to-end — current state documented (2 appeal types in production March 1), AI decision points identified within Elementum, data flow diagrammed
  4. Auto-auth current state understood — reviewed existing 6-condition auto-auth logic, identified how GenAI layers in, gap analysis for scaling to hundreds of policies
  5. First baseline measurements captured — current appeal processing time, current auto-auth coverage %, nurse review time per case

Daily Breakdown

DayFocusDeliverable
1Onboarding: credentials, environment setup, security orientationAccess confirmed, dev environment running
2Appeals deep-dive with Colt/Jamie: walk through Elementum workflow, 2 appeal types, AI touchpointsAppeals workflow diagram v1
3Auto-auth deep-dive: review existing system, 6 conditions, data sources, GenAI integration pointsAuto-auth architecture doc v1
4Data access: connect to ADT feeds, Facets, medical policies. Understand data schemas, quality, gapsData access inventory + gap list
5Baseline metrics: pull current processing times, volumes, error rates. Week 1 retro with ColtBaseline metrics doc + Week 2 plan

❓ Questions Thomas Needs to Answer

  • Q1: Provisioning timeline — Nathan warned SRP tickets take 4 weeks. Has Colt pre-staged any access? Do we need to submit requests NOW before contract signs?
  • Q2: Which appeals work stream first? — They go live March 1 with 2 types. Do we embed in the March 1 launch or start on the next appeal types?
  • Q3: Who is our daily POC? — Colt, Jamie, or someone on Nathan's team? This determines how fast we move.
  • Q4: Can we get a sandbox/staging environment? — Or are we working directly in their production pipeline from day 1?
  • Q5: Security review status — They haven't received our formal security deck yet. Is that a blocker for access?

Week 2: Pilot Workflow with Synthetic Data (Days 6-10)

Outcomes

  1. Appeals pilot running on synthetic data — AI parsing + reasoning pipeline processing sample appeal documents through Elementum decision points
  2. Auto-auth expansion prototype — selected 2-3 new conditions beyond the existing 6, built draft policy-to-criteria mapping
  3. LLM pipeline integrated with their security gateway — all calls routed through Premera's AI gateway, Phoenix tracing active
  4. First nurse feedback captured — showed appeal AI output to 1-2 nurses, documented what works / what's wrong
  5. Technical architecture documented — how our code fits into their stack, deployment approach, testing strategy

Key Activities

ActivityOwnerDependency
Build appeal document parser (extract key claims, supporting evidence, provider arguments)ThomasAccess to sample appeal docs
Implement criteria-matching logic against InterQual for appeal reviewThomasInterQual access/API or documentation
Create synthetic appeal dataset (10-20 cases covering 2 appeal types)BothUnderstanding of appeal types from Week 1
Auto-auth: map 2-3 new medical policies to automation logicMichaelMedical policy documents
Integrate LLM calls through Premera's AI security gatewayThomasGateway credentials + docs
Set up Phoenix tracing for all AI inference callsThomasPhoenix collector access
Demo to Colt: "here's what the AI sees when processing an appeal"BothWorking prototype

❓ Questions Thomas Needs to Answer

  • Q6: InterQual access model — Can we call InterQual programmatically? Or is it a UI-only tool nurses use manually? This changes the architecture significantly.
  • Q7: What does "synthetic data" mean here? — Can Premera provide de-identified real cases? Or do we generate from scratch? De-identified real data is 10x more useful.
  • Q8: LLM model choice — Premera has Anthropic + OpenAI relationships. Which models are approved? Any restrictions on Claude vs. GPT for PHI?
  • Q9: How do we handle the "no automated denials" constraint technically? — Every AI output that leans toward denial must route to human. Need to design the confidence threshold / escalation logic early.

Weeks 3-4: Quality Loop, Decision-Support Outputs, Internal Demos (Days 11-20)

Outcomes

  1. Appeals quality loop operational — AI outputs reviewed by nurses, feedback captured, prompts/logic tuned, measurable improvement across iterations
  2. Auto-auth expansion validated — 2-3 new conditions tested against real (de-identified) data, accuracy measured, ready for production review
  3. Decision-support outputs formatted for nurse workflow — outputs match Premera's templates, integrate into Elementum, require minimal manual editing
  4. Internal demo to Chad Murphy (CCO) — first time the business leader sees AI-augmented appeal review + auto-auth expansion in action
  5. Week 4 executive summary delivered — quantified results vs. baseline, roadmap for months 2-3, recommendation for expanded scope

Key Activities

ActivityOwnerDependency
Run 50+ appeal cases through pipeline, measure accuracy vs. nurse decisionsBothWorking pipeline + test cases
Build feedback UI/form for nurses to rate AI outputsThomasNurse availability
Tune prompts based on feedback (3+ iteration cycles)ThomasFeedback data
Format outputs to match Premera's existing summary templatesMichaelTemplate access (from Week 1)
Auto-auth: run new conditions against historical decisions, measure match rateBothHistorical decision data
Prep Chad Murphy demo: curated examples, before/after, metricsBothWorking prototypes
Draft Month 2-3 roadmap: what's next, what's needed, scaling planBothWeek 1-3 learnings
Deliver Week 4 executive summaryMichaelAll above

❓ Questions Thomas Needs to Answer

  • Q10: Have you met Chad Murphy yet? — He's the CCO and key business decision-maker. Colt's team is technical. Chad controls whether this scales. When do we get in front of him?
  • Q11: What does "success" look like to Chad vs. Colt? — Colt cares about technical capability. Chad cares about nurse productivity, compliance, cost. Need to present metrics that matter to both.
  • Q12: Vendor insourcing target — Colt mentioned "Caroline" for advanced imaging reviews as a disruption candidate. Is that a Week 3-4 deliverable or Month 2+?
  • Q13: What's the billing structure during ramp? — Full $150/hr from day 1? Or reduced rate during discovery? This affects how aggressive Week 1 can be.

Success Metrics

MetricBaseline (capture Week 1)Week 2 TargetWeek 4 TargetHow Measured
Appeal processing timeTBD (current manual)First AI-assisted time30%+ reductionTime tracking per case
Auto-auth condition coverage6 conditions6 (understanding)8-9 conditionsCount of automated policies
AI output accuracy (appeals)N/AFirst measurements>85% nurse agreementNurse review sample
AI output accuracy (auto-auth)Existing system baselineN/AMatch or exceed existingComparison to historical decisions
Review quality consistencyTBD (variance across nurses)First measurementsMeasurable reduction in varianceInter-reviewer agreement
Stakeholder confidenceLow (haven't seen it work)Colt team bought inChad demo positiveQualitative feedback
Cycle time: idea → deployedTBD (Nathan says weeks)First deployRepeatable deploy processCalendar tracking

Risk-Adjusted Timeline

RiskImpactMitigationPlan B
Provisioning takes 4+ weeks (Nathan's warning)Week 1 is blockedPre-submit access requests NOW, before contract signs. Colt to champion internally.Work on architecture/design docs + synthetic data while waiting
Data access is harder than expectedCan't baseline or buildStart with whatever Colt's team already has extracted. Use their existing data pipelines.Build against synthetic data, validate later with real
InterQual is UI-only (no API)Can't automate criteria matchingBuild our own criteria extraction from policy docsManual criteria encoding for pilot conditions
Appeals March 1 launch is chaoticTeam too busy for usFocus on auto-auth first, pick up appeals after launch stabilizesObserve/document March 1 launch, design improvements for post-launch
Chad Murphy doesn't engageBusiness side doesn't champion usUse Colt as bridge. Deliver results that force the conversation.Build relationship through Romilla (clinician team lead)

Pre-Contract Actions (Do NOW)

These don't require a signed contract:

  1. Submit security deck to Premera — they haven't seen it yet, and it may be required for access
  2. Ask Colt about pre-staging access requests — if SRP tickets take 4 weeks, starting now saves the entire first month
  3. Request sample data — de-identified appeal docs, medical policies, auto-auth decision logs
  4. Confirm appeals March 1 timeline — are we joining that launch or starting after?
  5. Schedule Chad Murphy intro — even a 15-min hello before contract signs builds trust
  6. Decide individual contractor vs. company contract — Nathan recommends company. How long does that add?

Daisy

v1

What do you need?

I can pull up the fundraise pipeline, CRM accounts, hot board, meeting notes — anything in the OS.

Sonnet · read-only