Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/135915
Title: Efficient automatic annotation of binary masks for enhanced training of computer vision models
Authors: Seychell, Dylan
Kenely, Matthew
Bartolo, Matthias
Debono, Carl James
Bugeja, Mark
Sacco, Matthew
Keywords: Computer vision
Image processing -- Methodology
Reinforcement learning
Deep learning (Machine learning)
Data sets
Issue Date: 2023
Publisher: Institute of Electrical and Electronics Engineers Inc.
Citation: Seychell, D., Kenely, M., Bartolo, M., Debono, C. J., Bugeja, M., & Sacco, M. (2023, December). Efficient Automatic Annotation of Binary Masks for Enhanced Training of Computer Vision Models. In 2023 IEEE International Symposium on Multimedia (ISM) (pp. 256-259). IEEE.
Abstract: In modern computer vision models, the quality and quantity of training data have become crucial. Datasets deemed sufficient a few years ago now require data augmentation to increase their size. This presents a challenge, especially when these supplementary datasets lack annotations in standard formats like COCO, VGG, or YOLO. One solution to this problem is to learn semantic boundaries from binary images of unannotated datasets, thereby increasing the data available for training and evaluating models. However, choosing an efficient annotation method can be both time-consuming and effort-intensive. This research paper explores three approaches, ranging from traditional image processing algorithms to the recently introduced Segment Anything Model (SAM). The study demonstrates how these different algorithms perform on various datasets and concludes that the proposed image processing method strikes the best balance between performance and efficiency.
URI: https://www.um.edu.mt/library/oar/handle/123456789/135915
Appears in Collections:Scholarly Works - FacICTAI

Files in This Item:
File Description SizeFormat 
Efficient automatic annotation of binary masks for enhanced training of computer vision models.pdf
  Restricted Access
507.33 kBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.