Date of Award

5-28-2024

Publication Type

Thesis

Degree Name

M.Sc.

Department

Computer Science

Supervisor

Ziad Kobti

Abstract

The Segment Anything Model (SAM) by Meta AI Research, trained on an extensive collection of over 1 billion masks, has gained significant attention for its exceptional ability to segment ”anything” in ”any scene”. SAM integrates a sophisticated image encoder, prompt encoder, and lightweight mask decoder, enabling flexible prompting and rapid mask generation in segmentation tasks. This segmentation model excels in granular, component-level segmentation, enriching our understanding of pixel semantics, critical for local feature learning. On a different note, the challenge of classifying small-scale objects persists, especially in sectors like medical imaging and remote sensing where objects of interest typically represent a small fraction of the entire image. In this study, we investigate the potential applications of SAM in the classification of small objects despite its primary design as a segmentation model. We introduce an ensemble deep learning methodology that leverages SAM within our custom dataset, specifically targeting the classification of tiny objects. Through comparative analysis between segmented data (processed by SAM) and non-segmented data (original data), our findings indicate a performance improvement in favor of the segmented data, underscoring the efficacy of our proposed approach.

Share

COinS