CROSS-REFERENCES
CROSS-REFERENCE TO RELATED APPLICATIONS
This application is related to commonly owned U.S. Provisional Patent Application No. 63/967,576, filed January 25, 2026, entitled “SYSTEM AND METHOD FOR PHYSICS-CONSTRAINED SIM-TO-REAL TRANSFER LEARNING IN COMPUTATIONAL ONCOLOGY,” and to commonly owned disclosures directed to complementary aspects of the same platform architecture, including (i) ontology-guided gradient modulation for stable cross-species domain adaptation, and (ii) runtime enforcement of physical and statistical reliability envelopes in differentiable clinical digital twins. The present application addresses a distinct technical problem—distributionally robust training across heterogeneous data acquisition environments—and does not claim priority to the aforementioned applications.
I.FIELD OF THE INVENTION
The present invention relates generally to artificial intelligence, computational biology, and high-performance computing. More specifically, the invention relates to systems and methods for training neural networks on multi-center medical data using configurable risk set scopes—ranging from strictly isolated GPU memory structures to pooled cross-site ranking constraints—to optimize generalization in physics-constrained tumor dynamics models.
II.BACKGROUND OF THE INVENTION
1.Shortcut Learning and Batch Effects
Precision oncology increasingly relies on machine learning models to integrate high-dimensional patient data—such as genomics, transcriptomics, and histopathology—to predict clinical outcomes. A critical challenge in deploying such models is “shortcut learning,” where the model learns to identify the hospital or sequencing center from which the data originated rather than the underlying tumor biology. This “batch effect” occurs because different hospitals (“environments”) utilize different tissue staining protocols, slide scanners, and DNA sequencing platforms. A model trained via standard Empirical Risk Minimization (ERM) often minimizes global loss by overfitting to these institution-specific artifacts, causing performance degradation during external validation.
2.Limitations of Existing Harmonization
Existing solutions have significant limitations. Harmonization techniques like “ComBat” often remove biological signals correlated with the batch. Furthermore, the mathematical formulation of the training objective—specifically the Cox Proportional Hazards loss—presents a dilemma regarding “partial likelihood contamination.”
3.Pooled vs. Stratified Risk Sets
In standard deep learning frameworks, survival loss is computed by sorting a batch globally. This creates a “pooled” risk set where the denominator for a patient in Hospital A includes patients from Hospital B. While this can technically allow gradients to bleed across environmental boundaries (potentially encouraging shortcut learning), it also enforces a “global ranking” constraint, compelling the model to learn features that rank patients consistently across different institutions.
Conversely, strictly isolating risk sets by hospital (stratified Cox) prevents contamination but removes this global ranking pressure, potentially leading to models that rank well within a hospital but fail to generalize across hospitals.
4.Need for the Invention
There is an unmet need for a training framework that can flexibly switch between these regimes—supporting both strict memory-isolated computation for statistical independence and pooled computation for cross-site ranking invariance—while utilizing Distributionally Robust Optimization (DRO) to upweight the worst-performing environments. Furthermore, naive implementations of these losses on GPUs often scale poorly ( memory), creating a need for optimized, memory-efficient kernels for both modes. Finally, translating model predictions into clinical decisions requires rigorous safety bounds and standardized output formats for clinical integration.
III.SUMMARY OF THE INVENTION
The present invention provides a computer-implemented system and method for training cancer digital twins using Hospital-Environment Grouping and a survival-specific adaptation of Group Distributionally Robust Optimization (Group DRO), implemented via a dual-mode risk set construction engine.
- Dual-Mode Risk Set Construction
- A configurable training pipeline that extracts Tissue Source Site (TSS) codes to define environments and allows selection of two distinct risk set construction modes: Mode 1 (Pooled Cross-Site) computes a global cumulative log-sum-exp across all environments, enforcing cross-site ranking invariance. Mode 2 (Strictly Isolated) uses a Segmented LogCumSumExp Kernel operating on contiguous per-environment GPU buffers, physically preventing cross-environment memory reads.
- Group DRO with Exponentiated Gradient
- Regardless of risk set mode, the system employs an exponentiated gradient ascent mechanism to update environment weights based on computed losses . This focuses training on the “worst-case” environments, ensuring the model does not overfit the largest or easiest hospitals.
- Physics-Constrained Regimen Selection
- At inference, the system utilizes the robustly trained model to parameterize a Neural ODE. Parameters are strictly constrained to biological ranges (e.g., per day). The system numerically integrates the ODE for candidate dosage schedules and generates a treatment recommendation data structure stored in a patient-associated database record.
IV.BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram of the system architecture, illustrating the extraction of Tissue Source Site (TSS) codes and the parallel processing of omics and Whole Slide Imaging (WSI) data.
FIG. 2 illustrates the Dual-Mode Risk Set Construction logic, contrasting the Global LogCumSumExp (Mode 1) with the Segmented LogCumSumExp (Mode 2).
FIG. 3 details the inference-time pipeline, from Information Sufficiency Score (ISS) routing through mandatory ODE simulation to treatment recommendation output.
FIG. 1: System Architecture with TSS Environment Extraction
FIG. 2: Dual-Mode Risk Set Construction
FIG. 3: Inference Pipeline with ISS Routing and ODE Simulation
V.DETAILED DESCRIPTION OF THE INVENTION
[0001] System Architecture
[0001]The system comprises a computing platform (e.g., a GPU-accelerated server) configured to process multi-modal patient data. The data includes:
- Multi-omics data
- RNA-seq (e.g., 2,579 genes mapped to TME pathways), DNA mutation (500 genes), Copy Number Variation (CNV, 1,886 features), and Methylation (1,000 features).
- Histopathology data
- Whole Slide Images (WSI) processed into tile embeddings (e.g., 1536-dimensional vectors from UNI2-h).
- Provenance Metadata
- Patient barcodes containing Tissue Source Site (TSS) identifiers, extracted from fixed positions within TCGA-format barcode strings.
[0002] Foundation Model (VAE)
[0002]The system utilizes a hierarchical, biologically-disentangled Variational Autoencoder (VAE). The VAE encodes raw omics data into a latent space (, 328 dimensions). The latent space is structured into biologically meaningful slices, including (proliferation) and (pathway activity, 50 MSigDB Hallmark pathways). The VAE is preferably frozen during downstream training to ensure latent stability.
[0003] Hospital-Environment Grouping
[0003]To enable robust training, the system defines “environments” () representing distinct data acquisition distributions.
Environment Extraction Logic:
- The system parses a patient identifier string (e.g., TCGA-XX-YYYY) to extract the Tissue Source Site (TSS) code (e.g., XX).
- The TSS code maps to a specific hospital. In experimental validation, 132 distinct TSS environments were identified.
- Sample-Based Merging: Environments with fewer than 20 samples are merged into a composite “OTHER” environment to ensure statistical stability.
[0004] Mode 1: Pooled Cross-Site Construction (Production Default)
[0004]In this mode, the system enforces Cross-Site Ranking Invariance. The entire batch of size (containing patients from all environments) is sorted by observed time in descending order, and a global cumulative log-sum-exp term is computed:
Crucially, the summation includes patients from all environments. This means the risk of a patient in Hospital A is normalized against patients in Hospital B. To compute the per-environment loss required for Group DRO, the system applies a mask which is 1 if patient belongs to environment and has an observed event, and 0 otherwise:
This configuration forces the model to learn features that are robust enough to rank patients consistently across different hospitals, yielding superior external generalization.
[0005] Mode 2: Strictly Isolated Construction (Segmented Kernel)
[0005]In this mode, the system enforces Per-Environment Statistical Independence.
- Contiguous Buffer Allocation
- The system allocates a contiguous hazard buffer in GPU memory, partitioned into segments. Each segment spans indices .
- Segment Descriptor Array (S)
- A descriptor array stores tuples .
- Segmented LogCumSumExp
- The system executes a custom GPU kernel that resets the cumulative sum at segment boundaries defined by S. The kernel is structurally prevented from reading hazard values outside the current environment's segment.
The segmented kernel computes:
This mode is mathematically equivalent to stratified Cox loss and prevents any gradient interaction between hospitals. It is preferred when strict per-environment statistical independence is required.
[0006] Survival-Specific Distributionally Robust Optimization
[0006]The training objective is to minimize the maximum loss across all environments: . The system maintains environment weights and performs an exponentiated gradient update:
Exponentiated Gradient Update:
- Forward Pass: Compute using either Mode 1 or Mode 2.
- Weight Update:
where is a step size and is a normalization factor ensuring .
- Backward Pass: Update network parameters to minimize .
This mechanism focuses training on the “worst-case” environments, ensuring the model does not simply overfit the largest or easiest hospitals.
[0007] Physics-Constrained Parameter Generation
[0007]The Hypernetwork predicts parameters for a tumor dynamics equation. The system enforces strict constraints via bounded activation functions:
- Tumor Growth Rate (ρ)
- . Constrained to per day.
- Drug Sensitivity (β)
- . Constrained to .
- Immune Kill Rate (ω)
- . Constrained to .
These constraints ensure generated trajectories are physically plausible by design rather than by post-hoc rejection. Non-negativity is guaranteed at the architecture level, preventing numerical instability during ODE integration.
[0008] Inference and Treatment Recommendation
[0008]At inference time, the system performs the following steps:
A. Adaptive Routing (ISS):
The system calculates an Information Sufficiency Score (ISS) based on the Mahalanobis distance to the training centroid. If ISS < Threshold, the patient is routed to a fallback protocol.
B. Regimen Optimization Loop:
For each candidate regimen, the system numerically integrates the ODE:
where is tumor volume, is carrying capacity, is drug concentration, and is an immune state function.
The system performs Monte Carlo (MC) dropout passes (e.g., 50) to generate trajectory distributions and identifies the regimen minimizing tumor burden at a clinical horizon (e.g., 365 days).
C. System Output:
The system generates a treatment recommendation data structure comprising:
- selected_regimen_id: Unique identifier for the recommended regimen.
- predicted_trajectory_array: Tensor .
- uncertainty_bounds: Tensor (2.5th/97.5th percentiles from MC dropout).
- physics_compliance_flags: Boolean validation of parameter bounds.
- information_sufficiency_score: The computed ISS value.
The system stores this data structure in a database record associated with the patient identifier or transmits it to a clinical decision support interface.
[0009] Experimental Results
[0009]The system was validated on a multi-center dataset comprising 33 cancer types and 132 Tissue Source Site (TSS) environments.
Mode 1 (Pooled Cross-Site) Performance:
CPTAC External C-index
0.718
[0.684, 0.750]
TSS Held-Out C-index
0.866 ± 0.014
Physics Compliance
100%
TSS Environments
132
Mode 2 (Strictly Isolated) Performance:
CPTAC External C-index
0.623
[0.591, 0.654]
Internal Global C-index
0.733
Physics Compliance
100%
Delta vs Mode 1 (CPTAC)
-9.5pp
Mode Comparison:
| Metric | Mode 1 (Pooled) | Mode 2 (Isolated) | Delta |
|---|---|---|---|
| CPTAC External C-index | 0.718 | 0.623 | -9.5pp |
| Internal Global C-index | 0.727 | 0.733 | +0.6pp |
| TSS Held-Out C-index | 0.866 | — | — |
| Physics Compliance | 100% | 100% | 0 |
Mode 1 achieves superior external generalization; Mode 2 achieves slightly higher internal metrics.
These results demonstrate that cross-site ranking pressure (Mode 1) significantly improves generalization to unseen institutions, while strict isolation (Mode 2) optimizes for the training distribution but may sacrifice external transferability. Both modes maintain 100% physics compliance.
[0010] Specific Embodiment: Training Hyperparameters
[0010]In one specific embodiment, training hyperparameters comprise:
Batch Size
128
Max Epochs
200
Learning Rate
3e-4
Optimizer
AdamW
DRO Step Size (η)
0.01
Min Env Samples
20
VI.CLAIMS
What is claimed is:
Claim 1. (Training Method)
A computer-implemented method for training a distributionally robust cancer survival prediction model, comprising:
- receiving, by a processor, a training dataset comprising multi-modal patient data and associated patient identifier strings;
- parsing, by the processor, the patient identifier strings to extract a Tissue Source Site (TSS) code for each patient, wherein the TSS code indicates a data acquisition environment;
- assigning each patient to one of a plurality of environment groups () based on the extracted TSS code;
- executing, by a GPU, a forward pass of a neural network to generate risk scores and physics-constrained tumor dynamics parameters for the patients;
- computing, for each environment group , a partial log-likelihood loss () by applying a configured risk set construction protocol, wherein said protocol defines which patients in a batch constitute a risk set denominator for a given patient;
- updating a set of environment weights using an exponentiated gradient ascent rule defined as , wherein is a step size and is a normalization factor; and
- updating model parameters of the neural network to minimize a weighted sum of the per-environment partial log-likelihood losses weighted by the updated environment weights .
Claim 2.
The method of claim 1, wherein the configured risk set construction protocol is a pooled cross-site protocol comprising: computing, from risk scores of patients from all environment groups in the batch, a global cumulative log-sum-exp term defining a global risk denominator for each event time; and computing the partial log-likelihood loss by summing log-likelihood terms strictly for event observations assigned to environment group , such that the risk denominator for a patient assigned to environment group includes contributions from patients assigned to at least one other environment group, thereby enforcing cross-site ranking invariance.
Claim 3.
The method of claim 1, wherein the configured risk set construction protocol is a strictly isolated protocol comprising:
- allocating, in a GPU memory, a contiguous hazard buffer partitioned into segments, wherein each segment corresponds to an environment group and spans indices ;
- constructing a segment descriptor array containing tuples of ; and
- executing a segmented logcumsumexp reduction kernel parameterized by the segment descriptor array , wherein said kernel resets accumulation at segment boundaries defined by , such that the risk denominator for a patient in environment group is computed exclusively from patients within the same environment group .
Claim 4.
The method of claim 3, wherein the segmented logcumsumexp reduction kernel structurally prevents cross-environment memory access during denominator accumulation, thereby ensuring per-environment statistical independence.
Claim 5.
The method of claim 1, wherein the physics-constrained tumor dynamics parameters include a tumor growth rate () constrained to the range per day via a sigmoid scaling function, and a drug sensitivity coefficient () constrained to the range .
Claim 6.
The method of claim 1, further comprising merging any environment group containing fewer than a predetermined threshold of samples into a composite environment group prior to computing the partial log-likelihood loss.
Claim 7. (Inference & Therapy System)
A computer-implemented system for personalized oncology therapy selection, comprising: a memory storing a trained multi-modal neural network model; and a processor coupled to the memory and configured to:
- receive multi-modal input data for a target patient, including omics data and optional whole slide imaging (WSI) data;
- generate, using the trained multi-modal neural network model, a set of physics-constrained tumor dynamics parameters comprising a growth rate (), a drug sensitivity (), and an immune kill rate ();
- enforce biological constraints on the parameters such that , , and ;
- receive a definition of a plurality of candidate therapeutic regimens, wherein each regimen specifies a drug concentration function over a time horizon;
- for each candidate therapeutic regimen, simulate a tumor volume trajectory by numerically integrating a differential equation parameterized by the generated physics-constrained parameters and the drug concentration function ;
- select a recommended therapeutic regimen from the plurality of candidate therapeutic regimens that has a minimum predicted tumor burden at a specified clinical time horizon;
- generate a treatment recommendation data structure comprising at least: (i) a selected_regimen_id identifying the recommended therapeutic regimen; (ii) a predicted_trajectory_array of dimensions ; (iii) an uncertainty_bounds array of dimensions derived from Monte Carlo dropout forward passes; and (iv) physics_compliance_flags indicating adherence to the biological constraints; and
- store the treatment recommendation data structure in a database record associated with a patient identifier for the target patient, or transmit the treatment recommendation data structure to a clinical decision support interface configured to display the predicted trajectory array.
Claim 8.
The system of claim 7, wherein the trained multi-modal neural network model is trained using Distributionally Robust Optimization (DRO) over hospital-specific environment groups to ensure the generated parameters are robust to data acquisition artifacts.
Claim 9.
The system of claim 7, wherein the processor is further configured to:
- calculate an Information Sufficiency Score (ISS) for the target patient based on a distance between the target patient's latent representation and a training distribution centroid; and
- route the multi-modal input data to a robust generalist checkpoint of the neural network model if the ISS exceeds a sufficiency threshold, or to a fallback checkpoint if the ISS is below the sufficiency threshold.
Claim 10.
The system of claim 7, wherein numerically integrating the differential equation comprises using a fixed-step Euler solver or an adaptive Runge-Kutta solver.
Claim 11.
The system of claim 7, wherein the differential equation is a Lotka-Volterra equation of the form:
wherein is tumor volume, is carrying capacity, and is an immune state function.
Claim 12.
The system of claim 7, wherein the uncertainty_bounds are computed by executing a plurality of forward passes of the neural network with dropout enabled, and calculating the 2.5th and 97.5th percentiles of the resulting trajectory distribution at each time step.
Claim 13. (Computer-Readable Medium)
A non-transitory computer-readable medium storing instructions that, when executed by a processor, cause the processor to perform operations comprising:
- extracting environment labels from patient barcodes in a multi-center cancer dataset, wherein the environment labels correspond to Tissue Source Sites (TSS);
- assigning patients to environment groups based on the extracted environment labels;
- executing a forward pass of a neural network to predict tumor dynamics parameters;
- computing a per-environment partial log-likelihood loss () for each environment group using a pooled risk set construction, wherein said pooled risk set construction comprises: calculating a global cumulative log-sum-exp term across all environment groups in a batch; and masking event contributions such that aggregates loss only for patients in environment group while utilizing the global cumulative log-sum-exp term as a denominator, thereby enforcing cross-site ranking invariance;
- updating a set of environment weights using an exponentiated gradient ascent rule: ; and
- updating model parameters of the neural network to minimize a weighted sum of the per-environment partial log-likelihood losses weighted by .
Claim 14.
The non-transitory computer-readable medium of claim 13, wherein the operations further comprise projecting the predicted tumor dynamics parameters into a bounded feasible set prior to computing the loss, ensuring that the predicted parameters remain physically valid regardless of the environment weights.
Claim 15.
The non-transitory computer-readable medium of claim 13, wherein the neural network comprises a Variational Autoencoder (VAE) for genomic encoding and a separate hypernetwork for parameter prediction, and wherein the VAE is frozen during the updating of the neural network.
Claim 16.
The non-transitory computer-readable medium of claim 13, wherein the pooled risk set construction enables the neural network to learn biological features that rank patients consistently across different Tissue Source Sites.
Claim 17.
The non-transitory computer-readable medium of claim 13, wherein the patient barcodes are formatted according to The Cancer Genome Atlas (TCGA) standards, and the TSS is extracted from a fixed position within the barcode string.
VII.ABSTRACT OF THE DISCLOSURE
A system and method for training distributionally robust cancer digital twins and selecting therapeutic regimens. The system extracts Tissue Source Site (TSS) identifiers from patient barcodes to define acquisition environments. A dual-mode risk set construction engine allows configuration between (1) a pooled cross-site mode that computes global log-likelihood denominators to enforce cross-site ranking invariance, and (2) a strictly isolated mode using segmented GPU kernels to enforce statistical independence. An exponentiated gradient ascent loop updates environment weights to upweight the most difficult environments based on per-environment losses. At inference, the system generates physics-constrained parameters (e.g., growth rate ) and simulates therapeutic regimens via a differential equation. A treatment recommendation data structure comprising the selected regimen and Monte Carlo uncertainty bounds is stored in a patient-associated database record.
[End of Application]
Related Documentation
Patent 1: Sim-to-Real Transfer
Physics-constrained sim-to-real transfer learning in computational oncology.
Patent 2: Collapse Prevention
Preventing metabolic scaling-induced representational collapse in cross-species models.
Patent 3: Uncertainty & Modalities
Uncertainty-calibrated missing modality handling and multi-layer UQ.
Patent 4: Autogradient Modulation
Ontology-guided autogradient modulation for domain adaptation.
Patent 5: Adjoint Sensitivity
Control of adjoint sensitivity computation graphs in differentiable digital twins.
Foundation Whitepaper
Complete technical documentation of the DNAI neuro-symbolic architecture.
Interested in licensing this technology?
Contact us to discuss partnership and licensing opportunities for the Distributionally Robust Training System.