Skip to content
/ MoMA Public

Medical Image Analysis (MEDIA_2024) paper: MoMA: Momentum Contrastive Learning with Multi-head Attention-based Knowledge Distillation for Histopathology Image Analysis

Notifications You must be signed in to change notification settings

trinhvg/MoMA

Repository files navigation

MoMA: Momentum Contrastive Learning with Multi-head Attention-based Knowledge Distillation for Histopathology Image Analysis

Trinh Thi Le Vuong and Jin Tae Kwak. Medical Image Analysis (MEDIA) 2024.

Implementation of paper [arXiv]:

Release note: The CNN version has been released. We will release the ViT and SwinViT soon.

Overview of distillation flow across different tasks and datasets. 1) Supervised task is always conducted, 2) Feature distillation is applied if a well-trained teacher model is available, and 3) Vanilla ${L}_{KD}$ is employed if teacher and student models conduct the same task.

Overview of distillation flow across different tasks and datasets. 1) Supervised task is always conducted, 2) Feature distillation is applied if a well-trained teacher model is available, and 3) Vanilla ${L}_{KD}$ is employed if teacher and student models conduct the same task. SSL stands for self-supervised learning.

Train the teacher network (optional) or vanilla students

./scripts/run_vanilla.sh

Train the moma student network

If the student and teacher dataset vary in number of categories, you may need to use "--std_strict, --tec_strict".

./scripts/run_moma.sh

Train the student network using other KD methods

./scripts/run_comparison.sh

About

Medical Image Analysis (MEDIA_2024) paper: MoMA: Momentum Contrastive Learning with Multi-head Attention-based Knowledge Distillation for Histopathology Image Analysis

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published