Information Technology for Accelerated Annotation of Medical Images in Segmentation Tasks Based on Deep Learning Models
DOI:
https://doi.org/10.31649/1997-9266-2024-175-4-95-103Keywords:
information technology, artificial intelligence, deep learning, image segmentation, data annotation, pseudomasks, automatic validationAbstract
The paper analyzed tools for creating annotations of medical images in image segmentation tasks. The performance of the well-known tools Supervisely, CVAT, and Segments.ai is compared with the information technology proposed in the work, which uses the Language Segment-Anything model with relevant text prompts and an automatic validation mechanism, based on the EfficientNet-B2 classification model.
The main objective of the study was to determine the optimal approach to the automation of the image annotation process to ensure maximum speed, maintaining expert accuracy. The results showed that usage of the Supervisely tool reduced the initial annotation time to 39.7 seconds, but required additional 59.5 seconds to adjust the masks. CVAT, with its semi-automated tools, produced masks in 64.8 seconds, but required another 85.1 seconds for adjustments. In comparison, Segments.ai required a full manual annotation, which took 130.2 seconds. At the same time, the developed information technology, which uses the Language Segment-Anything model with task-specific text prompts and an additional automatic validation mechanism, significantly reduced the time for creating annotations to about 29.6 seconds per image, and also reduced the time for manual correction to 45.4 seconds.
The developed information technology demonstrated high speed and accuracy in creating pseudo-masks, confirmed by experimental results. The main advantages of this approach are the decrease of time, needed for manual correction and increase the efficiency of the medical image annotation process.
This work points out to the significant potential of using automated methods to accelerate annotation in the field of computer vision, improving the speed of performing medical data analysis tasks while maintaining the desired quality.
References
M. Aljabri, M. AlAmir, and M. AlGhamdi, “Towards a better understanding of annotation tools for medical imaging: a survey,” Multimed Tools Appl 81, pp. 25877-25911, 2022. https://doi.org/10.1007/s11042-022-12100-1 .
“Language Segment-Anything,” GitHub. [Electronic resource]. Available: https://github.com/luca-medeiros/lang-segment-anything .
“Supervisely,” GitHub, [Electronic resource]. Available: https://github.com/supervisely/supervisely .
“CVAT,” GitHub. [Electronic resource]. Available: https://github.com/cvat-ai/cvat .
“Segments.ai,” GitHub. [Electronic resource]. Available: https://github.com/segments-ai/segments-ai .
“Segment Anything Model (SAM),” GitHub. [Electronic resource]. Available: https://github.com/facebookresearch/segment-anything .
“OpenCV: Open Source Computer Vision Library,” GitHub. [Electronic resource]. Available: https://github.com/opencv/opencv
M. Tan, and Q.V. Le, “EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks,” Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, 2019, pp. 6105-6114. [Electronic resource]. Available: https://arxiv.org/pdf/1905.11946 .
Malhotra Priyanka, Gupta Sheifali, Koundal Deepika, Zaguia Atef, and Enbeyle Wegayehu, “Deep Neural Networks for Medical Image Segmentation,” Journal of Healthcare Engineering, 2022, 9580991, 15 p., 2022. https://doi.org/10.1155/2022/9580991 .
“Pulmonary Chest X-Ray Defect Detection,” Kaggle, [Electronic resource]. Available: https://www.kaggle.com/datasets/nikhilpandey360/chest-xray-masks-and-labels/data .
F. van Beers, A. Lindström, E. Okafor and M. Wiering, “Deep Neural Networks with Intersection over Union Loss for Binary Image Segmentation,” in Proceedings of the 8th International Conference on Pattern Recognition Applications and Methods, vol. 1 ICPRAM, 2019, pp. 438-445. SciTePress. [Electronic resource]. Available: https://pure.rug.nl/ws/portalfiles/portal/87088047/ICPRAM_2019_35.pdf .
Feng Li, et. al, “Visual In-Context Prompting,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2024, pp. 12861-12871. [Electronic resource]. Available: https://openaccess.thecvf.com/content/CVPR2024/papers/Li_Visual_In-Context_Prompting_CVPR_2024_paper.pdf .
О. В. Коменчук, і О. Б. Мокін, «Аналіз методів передоброблення панорамних стоматологічних рентгенівських знімків для задач сегментації зображень,» Вісник Вінницького політехнічного інституту, № 5, с. 41-49, 2023. https://doi.org/10.31649/1997-9266-2023-170-5-41-49 .
Downloads
-
pdf (Українська)
Downloads: 23
Published
How to Cite
Issue
Section
License
This work is licensed under a Creative Commons Attribution 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain copyright and grant the journal right of first publication.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See The Effect of Open Access).