Automatic Annotation of Training Datasets in Computer Vision using Machine Learning Methods
Main Article Content
Abstract
This paper addresses the issue of automatic annotation of training datasets in the field of computer vision using machine learning methods. Data annotation is a key stage in the development and training of deep learning models, yet the process of creating labeled data often requires significant time and labor. This paper proposes a mechanism for automatic annotation based on the use of convolutional neural networks (CNN) and active learning methods.
The proposed methodology includes the analysis and evaluation of existing approaches to automatic annotation. The effectiveness of the proposed solutions is assessed on publicly available datasets. The results demonstrate that the proposed method significantly reduces the time required for data annotation, although operator intervention is still necessary.
The literature review includes an analysis of modern annotation methods and existing automatic systems, providing a better understanding of the context and advantages of the proposed approach. The conclusion discusses achievements, limitations, and possible directions for future research in this field.
Article Details
References
2. LabelImg for Image Annotation. URL: https://viso.ai/computer-vision/labelimg-for-image-annotation/.
3. VGG Image Annotator. URL: https://www.robots.ox.ac.uk/~vgg/software/via/via_demo.html.
4. Everingham M. et al. The pascal visual object classes challenge: A retrospective // International Journal of Computer Vision. 2015. Vol. 111. P. 98–136.
5. Berg A. et al. Semi-automatic annotation of objects in visual-thermal video // Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, 2019. https://doi.org/10.1109/ICCVW.2019.00277
6. Sager C., Janiesch C., Zschech P. A survey of image labelling for computer vision applications // Journal of Business Analytics. 2021. Vol. 4, No. 2. P. 91–110.
7. Cao J., Zhao A., Zhang Z. Automatic image annotation method based on a convolutional neural network with threshold optimization // Plos one. 2020. V. 15, No. 9. e0238956. https://doi.org/10.1371/journal.pone.0238956
8. Vatani A., Ahvanooey M.T., Rahimi M. An effective automatic image annotation model via attention model and data equilibrium // arXiv preprint arXiv:2001.10590. 2020.
9. Gu Y. et al. Automatic lung nodule detection using a 3D deep convolutional neural network combined with a multi-scale prediction strategy in chest CTs // Computers in Biology and Medicine. 2018. Vol. 103. P. 220–231.
10. Levine S. et al. Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection // The International Journal of Robotics Research. 2018. Vpl. 37, No. 4-5. P. 421–436.
11. Kirillov A. et al. Segment anything // Proceedings of the IEEE/CVF International Conference on Computer Vision. 2023. P. 4015–4026.
12. Zou X. et al. Segment everything everywhere all at once // Advances in Neural Information Processing Systems. NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing Systems. Article No. 868. P.19769–19782. https://dl.acm.org/doi/10.5555/3666122.3666990
13. Ultralytics YOLOv8 Docs. URL: https://docs.ultralytics.com/ru.
14. COCO Dataset. URL: https://cocodataset.org/#home.
15. Cityscapes Dataset. URL: https://www.cityscapes-dataset.com/.
16. Auto-Label. URL: https://roboflow.com/auto-label.
This work is licensed under a Creative Commons Attribution 4.0 International License.
Presenting an article for publication in the Russian Digital Libraries Journal (RDLJ), the authors automatically give consent to grant a limited license to use the materials of the Kazan (Volga) Federal University (KFU) (of course, only if the article is accepted for publication). This means that KFU has the right to publish an article in the next issue of the journal (on the website or in printed form), as well as to reprint this article in the archives of RDLJ CDs or to include in a particular information system or database, produced by KFU.
All copyrighted materials are placed in RDLJ with the consent of the authors. In the event that any of the authors have objected to its publication of materials on this site, the material can be removed, subject to notification to the Editor in writing.
Documents published in RDLJ are protected by copyright and all rights are reserved by the authors. Authors independently monitor compliance with their rights to reproduce or translate their papers published in the journal. If the material is published in RDLJ, reprinted with permission by another publisher or translated into another language, a reference to the original publication.
By submitting an article for publication in RDLJ, authors should take into account that the publication on the Internet, on the one hand, provide unique opportunities for access to their content, but on the other hand, are a new form of information exchange in the global information society where authors and publishers is not always provided with protection against unauthorized copying or other use of materials protected by copyright.
RDLJ is copyrighted. When using materials from the log must indicate the URL: index.phtml page = elbib / rus / journal?. Any change, addition or editing of the author's text are not allowed. Copying individual fragments of articles from the journal is allowed for distribute, remix, adapt, and build upon article, even commercially, as long as they credit that article for the original creation.
Request for the right to reproduce or use any of the materials published in RDLJ should be addressed to the Editor-in-Chief A.M. Elizarov at the following address: amelizarov@gmail.com.
The publishers of RDLJ is not responsible for the view, set out in the published opinion articles.
We suggest the authors of articles downloaded from this page, sign it and send it to the journal publisher's address by e-mail scan copyright agreements on the transfer of non-exclusive rights to use the work.