Mode
Text Size
Log in / Sign up

Narrative review reports Onca model performance on pancreatic cancer tasks versus other LLMs

Narrative review reports Onca model performance on pancreatic cancer tasks versus other LLMs
Photo by Dmytro Vynohradov / Unsplash
Key Takeaway
Note that open-source LLMs show strong text metrics but do not prove clinical efficacy for pancreatic cancer care.

This source is a narrative review and model development report focusing on the development of an open 9B dense language model, Onca, fine-tuned from Qwopus3.5-9B-v3. The model was compared against Woollie-7B, CancerLLM-7B, OpenBioLLM-8B, and the unmodified Qwopus base. The scope includes performance on secondary outcomes such as trial eligibility screening, case-specific clinical reasoning, structured pathology report extraction, and molecular variant evidence reasoning. The training dataset consisted of 37,364 rows.

Key synthesized findings indicate that Onca demonstrated the strongest overall results across all evaluated tasks. For trial screening, the model achieved an 81.6 F1 score. In clinical reasoning, it scored 14.1 composite. For structured pathology report extraction, the field exact-match score was 30.5. Additionally, the model achieved a 68.3 macro-F1 on PubMedQA Cancer and a 66.5 macro-F1 on PubMedQA.

The authors acknowledge that many existing oncology-focused language models depend on private institutional corpora, which limits reproducibility and practical reuse across centers. The report concludes that clinically targeted pancreatic-cancer language models can be built from open data with competitive performance while remaining practical to train on a single workstation-scale GPU setup. Practice relevance is tempered by the note that clinicians should not infer clinical efficacy or patient outcomes from model performance metrics on unstructured text tasks.

Study Details

EvidenceLevel 5
PublishedApr 2026
View Original Abstract ↓
Pancreatic ductal adenocarcinoma (PDAC) remains one of the deadliest solid tumors and continues to face low treatment-trial participation, fragmented evidence workflows, and labor-intensive ab- straction of unstructured clinical text. Existing oncology-focused language models show promise, but many depend on private institutional corpora, limiting reproducibility and practical reuse across centers. We present Onca, an open 9B dense model designed for four PDAC-relevant tasks: trial eligibility screening, case-specific clinical reasoning, structured pathology report extraction, and molecular variant evidence reasoning. Onca is fine-tuned from Qwopus3.5-9B-v3 with a single Un- sloth BF16 LoRA adapter on 37,364 training rows drawn from openly available sources. The evalu- ation spans 11 panels and compares Onca against Woollie-7B, CancerLLM-7B, OpenBioLLM-8B, and the unmodified Qwopus base. Onca achieves the strongest overall results on Trial Screening (81.6 F1), Clinical Reasoning (14.1 composite), Pathology Extraction (30.5 field exact-match), Pub- MedQA Cancer (68.3 macro-F1), and PubMedQA (66.5 macro-F1). The strongest gains appear in tasks closest to routine oncology workflow, especially trial review and pathology structuring. These findings suggest that clinically targeted pancreatic-cancer language models can be built from open data with competitive performance while remaining practical to train on a single workstation-scale GPU setup.
Free Newsletter

Clinical research that matters. Delivered to your inbox.

Join thousands of clinicians and researchers. No spam, unsubscribe anytime.