AI Engineering Leadership
Two decades building AI systems that inform consequential decisions.
Most AI and digital technologies never ship. They get presented in board decks, run in IDEs, make it through pilots, and then quietly disappear when the real constraints show up: the messy data, the interoperability challenges, the governance requirements, the stakeholders who all have to say yes, the infrastructure that was never built to run a model in production.
Over the past two decades, I have come to believe that the problem is rarely the data silos, the algorithms, or the infrastructure. It is almost always the absence of one person who can sit in the board conversation about AI strategy and then go and design the architecture the next morning, without losing anything in the translation.
Someone who has held the P&L, navigated the institutional politics, put smiles on partners' faces, and still knows exactly why their team's retrieval pipeline is underperforming or why the loss function they landed on is working against them.
That person is harder to find than any technology. And every single morning, I ask myself how much of that person I still have left to become.
Six domains spanning model training through production deployment.
End-to-end LLM engineering: domain-adaptive pre-training, SFT on 500K+ instruction pairs, DPO alignment with hard negative mining, Mixture of LoRA Experts across 6 scientific domains.
Fan-out / fan-in agentic pipelines with 6+ specialised agents. MCP server architecture. Streaming SSE output. Real-time orchestration with graceful degradation.
Hybrid retrieval (dense + BM25) with Reciprocal Rank Fusion. Cross-encoder re-ranking. 250M+ evidence spans in OpenSearch. GraphRAG for structured knowledge.
50TB+ S3 data lakes with Apache Iceberg, 700M+ row analytical stores, Athena serverless SQL, 24 EventBridge ingestion pipelines. Query time from 8s to <400ms.
Real-time ML inference at 30+ geography scale. P95 latency <150ms. Champion-challenger deployment. Canary rollouts with automated quality gates.
AI governance frameworks for regulated environments across 17 countries. Responsible AI embedded in pipelines. National expert committee leadership. $100M+ portfolio accountability.
Systems operating at scale across multiple countries under governance constraints.
A 397B-parameter domain-adapted foundation model trained through a 4-stage pipeline: domain-adaptive pre-training, supervised fine-tuning on 596K instruction pairs, DPO alignment on 192K preference pairs with 3-tier hard negative mining, Mixture of LoRA Experts across 6 scientific domains.
AI intelligence platform spanning 190+ countries for health, climate, and development finance. Funded by McGovern Foundation and Amazon.
Enterprise-grade data lake unifying 100M+ records across 17 national health systems. Iceberg on S3. 24 EventBridge pipelines. P95 from 8s to <400ms.
Production systems engineering: foundation model training, data platforms, and multi-agent orchestration.
Architecture selection, loss functions, model families, and design patterns from hands-on R&D.
| Loss | Formula | When |
|---|---|---|
| Cross-Entropy | -sum y.log(p) | Classification, LM |
| NLL | -log P(x|theta) | Sequence modeling |
| KL Divergence | sum p.log(p/q) | VAE, distillation |
| Wasserstein | inf E[|x-y|] | WGAN, Earth mover |
| Spectral | ||sigma(W)|| | GAN stabilization |
| Contrastive | -log(sim+/sim-) | Embeddings, CLIP |
| Triplet | max(d+-d-+m,0) | Metric learning |
| Focal | -a(1-p)^g.log p | Imbalanced data |
| DPO | -log sigma(b.dr) | Preference tuning |
| Hinge | max(0, 1-y.f) | SVM, GAN variants |
| Family | Sizes | Strength | Use Case |
|---|---|---|---|
| Llama 3 | 8-405B | General, code | Broad reasoning |
| Qwen 2.5 | 0.5-72B | Multilingual | Domain adaptation |
| Mistral | 7-8x22B | MoE efficiency | Cost-sensitive |
| DeepSeek | 7-236B | Math, code | Technical tasks |
| Gemma 2 | 2-27B | Compact, fast | Edge, mobile |
| Phi 3/4 | 3-14B | Small but capable | On-device |
| Command R+ | 35-104B | RAG, grounding | Enterprise RAG |
| Yi | 6-34B | Long context | Document analysis |
56 public repositories across personal and organization accounts. Production systems, research platforms, and foundational tools.
17 years of continuous open source development across AI, epidemiology, and global health platforms.
Technical roadmaps that became funded, multi-year programs.
Each highlighted country represents a program I designed, built, or delivered. Technical and strategic leadership on the ground.
Production-grade proficiency. Primary tools highlighted.
Open to conversations about AI engineering, foundation models, health systems, and climate intelligence.