TY - GEN
T1 - Domain-specific pre-training improves confidence in whole slide image classification
AU - Rohit Chitnis, Soham
AU - Liu, Sidong
AU - Dash, Tirtharaj
AU - Verlekar, Tanmay
AU - Di Ieva, Antonio
AU - Berkovsky, Shlomo
AU - Vig, Lovekesh
AU - Srinivasan, Ashwin
PY - 2023/7/26
Y1 - 2023/7/26
N2 - Whole Slide Images (WSIs) or histopathology images are used in digital pathology. WSIs pose great challenges to deep learning models for clinical diagnosis, owing to their size and lack of pixel-level annotations. With the recent advancements in computational pathology, newer multiple-instance learning-based models have been proposed. Multiple-instance learning for WSIs necessitates creating patches and uses the encoding of these patches for diagnosis. These models use generic pre-trained models (ResNet-50 pre-trained on ImageNet) for patch encoding. The recently proposed KimiaNet, a DenseNet121 model pre-trained on TCGA slides, is a domain-specific pre-trained model. This paper shows the effect of domain-specific pre-training on WSI classification. To investigate the effect of domain-specific pre-training, we considered the current state-of-the-art multiple-instance learning models, 1) CLAM, an attention-based model, and 2) TransMIL, a self-attention-based model, and evaluated the models' confidence and predictive performance in detecting primary brain tumors - gliomas. Domain-specific pre-training improves the confidence of the models and also achieves a new state-of-the-art performance of WSI-based glioma subtype classification, showing a high clinical applicability in assisting glioma diagnosis. We will publicly share our code and experimental results at https://github.com/soham-chitnis10/WSI-domain-specific.
AB - Whole Slide Images (WSIs) or histopathology images are used in digital pathology. WSIs pose great challenges to deep learning models for clinical diagnosis, owing to their size and lack of pixel-level annotations. With the recent advancements in computational pathology, newer multiple-instance learning-based models have been proposed. Multiple-instance learning for WSIs necessitates creating patches and uses the encoding of these patches for diagnosis. These models use generic pre-trained models (ResNet-50 pre-trained on ImageNet) for patch encoding. The recently proposed KimiaNet, a DenseNet121 model pre-trained on TCGA slides, is a domain-specific pre-trained model. This paper shows the effect of domain-specific pre-training on WSI classification. To investigate the effect of domain-specific pre-training, we considered the current state-of-the-art multiple-instance learning models, 1) CLAM, an attention-based model, and 2) TransMIL, a self-attention-based model, and evaluated the models' confidence and predictive performance in detecting primary brain tumors - gliomas. Domain-specific pre-training improves the confidence of the models and also achieves a new state-of-the-art performance of WSI-based glioma subtype classification, showing a high clinical applicability in assisting glioma diagnosis. We will publicly share our code and experimental results at https://github.com/soham-chitnis10/WSI-domain-specific.
KW - Domain-specific Pre-training,
KW - Whole Slide Image Classification
KW - Multiple Instance Learning
KW - Brain Tumor
KW - Domain-specific Pre-training
UR - https://www.scopus.com/pages/publications/85179638339
U2 - 10.1109/EMBC40787.2023.10340659
DO - 10.1109/EMBC40787.2023.10340659
M3 - Conference proceeding contribution
C2 - 38083343
AN - SCOPUS:85179638339
VL - 2023
T3 - Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference
SP - 1
EP - 4
BT - IEEE EMBC 2023
PB - Institute of Electrical and Electronics Engineers (IEEE)
CY - Sydney
T2 - Annual International Conference of the IEEE Engineering in Medicine and Biology Conference (45th : 2023)
Y2 - 24 July 2023 through 27 July 2023
ER -