Repository logo
 
Publication

Benchmarking Contrastive Learning for Multimodal Medical Imaging

datacite.subject.fosDepartamento de Informáticapt_PT
dc.contributor.advisorGarcia, Nuno Cruz
dc.contributor.authorSilva, Martim Dourado da
dc.date.accessioned2025-07-24T08:51:38Z
dc.date.available2025-07-24T08:51:38Z
dc.date.issued2025
dc.date.submitted2025
dc.descriptionTese de mestrado, Ciência de Dados, 2025, Universidade de Lisboa, Faculdade de Ciênciaspt_PT
dc.description.abstractDeep learning has achieved remarkable success in complex tasks, but its reliance on large, annotated datasets limits scalability in medical imaging, where expert labeling is costly and scarce. Contrastive learning, a self-supervised approach, offers a way to learn useful visual representations from unlabeled data by training models to distinguish between different images while aligning augmented views of the same instance. This thesis investigates the effectiveness of three state-of-the-art contrastive learning frameworks - SimCLR, MoCo, and BYOL - in generating transferable representations from medical images for two downstream tasks: multiclass classification and binary segmentation of breast tissue. It also examines whether combining ultrasound and mammography images during pretraining supports or hinders model generalization, reflecting real-world multimodal diagnostic workflows. Using seven public datasets, three modality-specific pretraining sets were constructed (ultrasound, mammography, and a balanced multimodal mix), each used to train the three frameworks. The resulting models were then fine-tuned for each downstream task. All networks shared a ResNet-18 backbone, and segmentation models used U-Net architectures with pretrained encoders. Results show that contrastive pretraining improves classification performance, particularly with ultrasound data and using BYOL or MoCo. These models outperformed randomly initialized baselines. For segmentation, however, random initialization yielded superior results, suggesting that standard contrastive objectives and augmentations do not capture the spatial precision needed for pixel-wise tasks. Mammography images posed further challenges due to small lesion size and detail loss from uniform resizing. This work underscores both the promise and the limitations of contrastive learning in clinical imaging. While effective for classification with limited labels, adapting contrastive methods for segmentation requires the design of specialized modality-aware pipelines.pt_PT
dc.identifier.urihttp://hdl.handle.net/10400.5/102399
dc.language.isoengpt_PT
dc.subjectVisão Computacionalpt_PT
dc.subjectAprendizagem Auto-Supervisionada Contrastivapt_PT
dc.subjectAnálise de Imagens Médicaspt_PT
dc.subjectDeteção/Diagnóstico do Cancro da Mamapt_PT
dc.subjectTeses de mestrado - 2025pt_PT
dc.titleBenchmarking Contrastive Learning for Multimodal Medical Imagingpt_PT
dc.typemaster thesis
dspace.entity.typePublication
rcaap.rightsopenAccesspt_PT
rcaap.typemasterThesispt_PT
thesis.degree.nameTese de mestrado em Ciência de Dadospt_PT

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
TM_Martim_Silva.pdf
Size:
32.12 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.2 KB
Format:
Item-specific license agreed upon to submission
Description: