Checkpoint and task definitions are downloaded automatically from HuggingFace on first use.
Computational pathology needs whole-slide image (WSI) foundation models that transfer across diverse clinical tasks, yet current approaches remain largely slide-centric, often depend on private data and expensive paired-report supervision, and do not explicitly model relationships among multiple slides from the same patient. We present MOOZY, a patient-first pathology foundation model in which the patient case, not the individual slide, is the core unit of representation. MOOZY explicitly models dependencies across all slides from the same patient via a case transformer during pretraining, combining multi-stage open self-supervision with scaled low-cost task supervision. In Stage 1, we pretrain a vision-only slide encoder on 77,134 public slide feature grids using masked self-distillation. In Stage 2, we align these representations with clinical semantics using a case transformer and multi-task supervision over 333 tasks from 56 public datasets, including 205 classification and 128 survival tasks across four endpoints. Across eight held-out tasks with five-fold frozen-feature probe evaluation, MOOZY achieves best or tied-best performance on most metrics and improves macro averages over TITAN by +7.37%, +5.50%, and +7.83% and over PRISM by +8.83%, +10.70%, and +9.78% for weighted F1, weighted ROC-AUC, and balanced accuracy, respectively.
Stage 1 (top): A frozen patch encoder extracts per-patch features arranged into a spatial grid. Multi-scale crops are sampled with spatial augmentations and block-based masking. A student slide encoder and EMA teacher are jointly trained via CLS-level self-distillation and masked patch prediction. Stage 2 (bottom): The pretrained slide encoder produces per-slide embeddings; a case transformer aggregates them into a unified case embedding, routed to task-specific classification and survival heads.
MOOZY is trained entirely on public data. Stage 1 uses 77,134 slide feature grids (53,286 at 20× and 23,848 at 40×) extracted from ~1.67 billion patches across ~31.8 TB of raw WSI data. Stage 2 uses 41,089 supervised cases across 333 tasks from 56 datasets — all 32 TCGA cohorts, all 10 CPTAC cohorts, REG, BC-Therapy, BRACS, CAMELYON17, DHMC Kidney, DHMC LUAD, EBRAINS, IMP Colorectum, IMP Cervix, MBC, MUT-HET-RCC, NADT Prostate, NAT-BRCA, and PANDA. Supervision covers 205 classification and 128 survival tasks across four endpoints (OS, DSS, DFI, PFI) and 23 anatomical sites.
Macro-average radar comparison across slide encoders on eight held-out tasks.
Parameter count vs. macro F1. Bubble size indicates total parameters.
Frozen-feature MLP probe on eight held-out tasks. Bold = best, underline = second best. Mean ± std over 5 folds.
| Task | Metric | CHIEF | GigaPath | PRISM | Madeleine | TITAN | MOOZY |
|---|---|---|---|---|---|---|---|
| Residual Cancer Burden | F1 | 0.46 | 0.45 | 0.46 | 0.51 | 0.43 | 0.56 |
| AUC | 0.60 | 0.55 | 0.58 | 0.63 | 0.58 | 0.74 | |
| Bal. Acc | 0.44 | 0.40 | 0.43 | 0.48 | 0.38 | 0.51 | |
| TP53 Mutation | F1 | 0.82 | 0.76 | 0.85 | 0.84 | 0.87 | 0.87 |
| AUC | 0.81 | 0.76 | 0.85 | 0.85 | 0.91 | 0.86 | |
| Bal. Acc | 0.83 | 0.76 | 0.84 | 0.84 | 0.88 | 0.86 | |
| BAP1 Mutation | F1 | 0.86 | 0.84 | 0.80 | 0.85 | 0.84 | 0.89 |
| AUC | 0.75 | 0.63 | 0.71 | 0.78 | 0.82 | 0.79 | |
| Bal. Acc | 0.75 | 0.66 | 0.66 | 0.75 | 0.75 | 0.78 | |
| ACVR2A Mutation | F1 | 0.89 | 0.80 | 0.85 | 0.89 | 0.87 | 0.91 |
| AUC | 0.80 | 0.74 | 0.83 | 0.76 | 0.79 | 0.91 | |
| Bal. Acc | 0.80 | 0.65 | 0.81 | 0.81 | 0.76 | 0.90 | |
| Histologic Grade | F1 | 0.71 | 0.77 | 0.73 | 0.75 | 0.73 | 0.78 |
| AUC | 0.71 | 0.77 | 0.67 | 0.74 | 0.71 | 0.75 | |
| Bal. Acc | 0.73 | 0.77 | 0.73 | 0.74 | 0.73 | 0.77 | |
| KRAS Mutation | F1 | 0.77 | 0.77 | 0.72 | 0.81 | 0.80 | 0.85 |
| AUC | 0.76 | 0.72 | 0.61 | 0.70 | 0.80 | 0.80 | |
| Bal. Acc | 0.74 | 0.76 | 0.63 | 0.77 | 0.81 | 0.79 | |
| IDH Status | F1 | 0.92 | 0.94 | 0.91 | 0.92 | 0.94 | 0.97 |
| AUC | 0.96 | 0.97 | 0.95 | 0.96 | 0.97 | 0.99 | |
| Bal. Acc | 0.92 | 0.94 | 0.91 | 0.91 | 0.94 | 0.97 | |
| Treatment Response | F1 | 0.53 | 0.51 | 0.57 | 0.49 | 0.49 | 0.58 |
| AUC | 0.70 | 0.68 | 0.69 | 0.59 | 0.60 | 0.68 | |
| Bal. Acc | 0.48 | 0.40 | 0.51 | 0.35 | 0.37 | 0.48 |
Macro-average across eight held-out tasks. Each entry averages over five MIL architectures (MeanMIL, ABMIL, CLAM, DSMIL, TransMIL).
| Metric | Backbone | UNI v2 | Phikon v2 | CONCH v1.5 | MUSK | MOOZY |
|---|---|---|---|---|---|---|
| F1 (weighted) | 0.733 | 0.716 | 0.715 | 0.746 | 0.729 | 0.801 |
| ROC-AUC (weighted) | 0.735 | 0.719 | 0.724 | 0.751 | 0.725 | 0.815 |
| Balanced Acc | 0.686 | 0.660 | 0.654 | 0.696 | 0.679 | 0.758 |
Macro-average across eight held-out tasks. Each stage and the case aggregator are toggled independently.
| Setting | Stage 1 | Stage 2 | Case Agg. | F1 | AUC | Bal. Acc |
|---|---|---|---|---|---|---|
| Stage 1 only | ✓ | ✗ | ✗ | 0.760 | 0.753 | 0.701 |
| Stage 2 only | ✗ | ✓ | ✓ | 0.748 | 0.725 | 0.701 |
| MOOZY w/o case agg. | ✓ | ✓ | ✗ | 0.771 | 0.789 | 0.729 |
| MOOZY | ✓ | ✓ | ✓ | 0.801 | 0.815 | 0.758 |
A board-certified pathologist reviewed attention maps across 20 representative WSIs and five encoders. MOOZY achieved the lowest mean semantic gap score (1.00) and near-balanced tumor vs. non-tumor attention (2.63), suggesting broad, diagnostically relevant coverage.
Dimensionality reduction of slide embeddings from four encoders. MOOZY shows the clearest class separation on cancer-type tasks.
@article{kotp2026moozy,
title = {MOOZY: A Patient-First Foundation Model for Computational Pathology},
author = {Kotp, Yousef and Trinh, Vincent Quoc-Huy and Pal, Christopher and Hosseini, Mahdi S.},
journal = {arXiv preprint arXiv:XXXX.XXXXX},
year = {2026}
}