Multi-task learning model for gastrointestinal cancer histopathology demonstrates high classification and segmentation performance in preclinical validation.
This methodological paradigm development study utilized a preclinical validation system to assess a multi-task learning model for gastrointestinal cancer. The population consisted of histopathological patches from GasHisSDB and GCHTID datasets, totaling approximately 99,000 H&E-stained images. Evaluation occurred under controlled benchmark settings without a clinical follow-up period reported in the study.
The intervention involved a multi-task learning model integrating Swin Transformer, DeepLabV3+, and R2U-Net for joint classification and segmentation. Comparators included ResNet-50 and Swin-T. The model achieved a classification F1-score of 0.938 ± 0.007 and a segmentation Dice coefficient of 0.839 ± 0.009. Against ResNet-50, Swin-T showed a classification F1-score of 0.945 vs. 0.917 and an AUC of 0.965 vs. 0.907. Additional metrics included a LYM Dice of 0.781 for small-volume tissues and a cross-domain transfer F1-score of 0.902. A Dice decrease under staining perturbations was 2.4%.
Safety data regarding adverse events, serious adverse events, and discontinuations were not reported, consistent with the preclinical nature of the work. Key limitations include validation conducted under controlled benchmark conditions using strongly annotated patch-level data. The framework is regarded as a clinically relevant, preclinical validation system. Feasibility for routine clinical implementation requires further verification through large-scale whole-slide image cohorts, prospective multicenter studies, and workflow integration assessments and future research.