How I led the 0→1 product build of an AI-powered tool for oncologists, from discovery through alpha launch.
Oncologists managing complex cancer cases must simultaneously synthesize genomic sequencing reports, treatment history, published research, and clinical guidelines often across disconnected, legacy systems. The cognitive overhead is significant. Errors in this context aren't just expensive, they're life-threatening.
The core discovery insight: oncologists didn't need more information. They needed better synthesis, AI that could surface the right signal at the right moment in their workflow, without adding steps.
I owned the product function end-to-end for CancerAI from initial discovery through alpha launch. This spanned:
The single most important early call was establishing a formal Go/No-Go framework for alpha readiness, with explicit Hard Gates around clinical safety, hallucination rate thresholds, PHI compliance, and critical incident resolution. This was non-negotiable in a clinical setting and gave the entire team a shared, unambiguous definition of "ready."
Early prototypes were AI-first, impressive demos that didn't map to how oncologists actually work. I reoriented the product around the clinical decision moment: what question is the oncologist trying to answer right now, and how does this tool answer it without adding friction? This shifted us from a general assistant to a context-aware decision support tool.
With a small cohort of alpha users, the temptation was to ship broadly and iterate. Instead, I enforced strict scope, fewer features, higher confidence. The goal wasn't to impress. It was to generate usable signal about what was actually driving clinical value before scaling.
"Ship narrow, learn deep", the alpha wasn't about feature breadth. It was about getting enough real clinical interactions to validate the core value hypothesis before committing to a wider rollout.
User research with oncologists surfaced several non-obvious findings:
Clinical AI requires a different evaluation standard than consumer products. I built an evaluation framework that included:
The science team was the most critical cross-functional dependency, and the most complex to manage. I worked to align on shared definitions of model performance, evaluation criteria, and what "good enough for clinical use" actually meant. Getting this alignment early prevented significant rework downstream.
Engineering, data science, and clinical stakeholders each had different definitions of readiness. My job was to hold a single standard across all of them without letting any one function's definition dominate.
Building clinical AI is a different discipline from consumer product. The bar for safety, evidence, and trust is higher, and the timeline to earn adoption is longer. The PM's job is to hold that bar without letting it become an excuse for not shipping. Both things are true at once.