arXiv Analytics

Sign in

arXiv:2306.06370 [cs.CV]AbstractReferencesReviewsResources

AutoSAM: Adapting SAM to Medical Images by Overloading the Prompt Encoder

Tal Shaharabany, Aviad Dahan, Raja Giryes, Lior Wolf

Published 2023-06-10Version 1

The recently introduced Segment Anything Model (SAM) combines a clever architecture and large quantities of training data to obtain remarkable image segmentation capabilities. However, it fails to reproduce such results for Out-Of-Distribution (OOD) domains such as medical images. Moreover, while SAM is conditioned on either a mask or a set of points, it may be desirable to have a fully automatic solution. In this work, we replace SAM's conditioning with an encoder that operates on the same input image. By adding this encoder and without further fine-tuning SAM, we obtain state-of-the-art results on multiple medical images and video benchmarks. This new encoder is trained via gradients provided by a frozen SAM. For inspecting the knowledge within it, and providing a lightweight segmentation solution, we also learn to decode it into a mask by a shallow deconvolution network.

Related articles: Most relevant | Search more
arXiv:2105.11333 [cs.CV] (Published 2021-05-24)
Multi-modal Understanding and Generation for Medical Images and Text via Vision-Language Pre-Training
arXiv:1506.00097 [cs.CV] (Published 2015-05-30)
A Review of Feature and Data Fusion with Medical Images
arXiv:1507.01251 [cs.CV] (Published 2015-07-05)
Autoencoding the Retrieval Relevance of Medical Images