Early cancer diagnosis is critical for improving patient outcomes but is challenged by the disease?s profound heterogeneity. This paper introduces a unified, AI-powered framework that synergistically integrates histopathology, genomics, and proteomics data to enhance early cancer detection. Our architecture features a novel multi-transformer model with dedicated Vision and Genomic Transformers to encode modality-specific features, which are then fused by a cross-modal attention transformer. This intermediate fusion strategy enables the model to learn intricate genotype-phenotype correlations often missed by traditional methods. Validated on cohorts from The Cancer Genome Atlas (TCGA), our framework demonstrates a significant improvement in diagnostic performance over single-modality baselines. We also incorporate Explainable AI (XAI) techniques to ensure model transparency, a crucial step for clinical adoption. The framework serves as both a powerful diagnostic tool and a hypothesis- generation engine, uncovering novel biomarkers from complex multi-modal data and advancing computational pathology and personalized medicine.
Multi-Modal Learning, Transformers, Histopathology, Genomics, Proteomics, Explainable AI, Early Cancer Diagnosis
IRE Journals:
Ayush Mishra, Anadi Mishra, Adarsh Tiwari, Uttam Sharma, Nikhil Raj "A Unified Multi-Modal Transformer Framework for Synergistic Cancer Diagnosis" Iconic Research And Engineering Journals Volume 9 Issue 5 2025 Page 197-205
IEEE:
Ayush Mishra, Anadi Mishra, Adarsh Tiwari, Uttam Sharma, Nikhil Raj
"A Unified Multi-Modal Transformer Framework for Synergistic Cancer Diagnosis" Iconic Research And Engineering Journals, 9(5)