arXiv Analytics

Sign in

arXiv:2003.00105 [cs.CV]AbstractReferencesReviewsResources

Self-supervised Representation Learning for Ultrasound Video

Jianbo Jiao, Richard Droste, Lior Drukker, Aris T. Papageorghiou, J. Alison Noble

Published 2020-02-28Version 1

Recent advances in deep learning have achieved promising performance for medical image analysis, while in most cases ground-truth annotations from human experts are necessary to train the deep model. In practice, such annotations are expensive to collect and can be scarce for medical imaging applications. Therefore, there is significant interest in learning representations from unlabelled raw data. In this paper, we propose a self-supervised learning approach to learn meaningful and transferable representations from medical imaging video without any type of human annotation. We assume that in order to learn such a representation, the model should identify anatomical structures from the unlabelled data. Therefore we force the model to address anatomy-aware tasks with free supervision from the data itself. Specifically, the model is designed to correct the order of a reshuffled video clip and at the same time predict the geometric transformation applied to the video clip. Experiments on fetal ultrasound video show that the proposed approach can effectively learn meaningful and strong representations, which transfer well to downstream tasks like standard plane detection and saliency prediction.

Related articles: Most relevant | Search more
arXiv:2004.10605 [cs.CV] (Published 2020-04-18)
Self-Supervised Representation Learning on Document Images
arXiv:2009.07994 [cs.CV] (Published 2020-09-17)
AAG: Self-Supervised Representation Learning by Auxiliary Augmentation with GNT-Xent Loss
arXiv:2311.03629 [cs.CV] (Published 2023-11-07)
Random Field Augmentations for Self-Supervised Representation Learning