kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
YOLOv4-object: an Efficient Model and Method for Object Discovery
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0001-6037-1661
Univ Bristol, Visual Informat Lab, Bristol, Avon, England..
KTH, School of Electrical Engineering and Computer Science (EECS), Electrical Engineering.
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Software and Computer systems, SCS.ORCID iD: 0000-0002-4722-0823
2021 (English)In: 2021 IEEE 45th Annual Computers, Software, and Applications Conference (COMPSAC) / [ed] Chan, WK Claycomb, B Takakura, H Yang, JJ Teranishi, Y Towey, D Segura, S Shahriar, H Reisman, S Ahamed, SI, Institute of Electrical and Electronics Engineers (IEEE) , 2021, p. 31-36Conference paper, Published paper (Refereed)
Abstract [en]

Object discovery refers to recognising all unknown objects in images, which is crucial for robotic systems to explore the unseen environment. Recently, object detection models based on deep learning have shown remarkable achievements in object classification and localisation. However, these models have difficulties handling the unseen environment because it is infeasible to exhaustively predefine all types of objects. In this paper, we propose the model YOLOv4-object to recognise all objects in images by modifying the output space of YOLOv4 and related image labels. Experiments on COCO dataset demonstrate the effectiveness of our method by achieving 67.97% recall (6.49% higher than vanilla YOLOv4). We point out that the incomplete labels (COCO only labels for 80 categories) hurt the learning process of object discovery and a higher recall can be achieved by our method if the dataset is fully labelled. Moreover, our approach is transferable, extensible, and compressible, showing broad application scenarios. Finally, we conduct extensive experiments to illustrate the factors that affect the object discovery performance of our model and some suggestions on practical implementations are elaborated.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2021. p. 31-36
Series
Proceedings International Computer Software and Applications Conference, ISSN 0730-3157
Keywords [en]
deep learning, object detection, object discovery, robotics system
National Category
Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:kth:diva-304774DOI: 10.1109/COMPSAC51774.2021.00016ISI: 000706529000005Scopus ID: 2-s2.0-85115823461OAI: oai:DiVA.org:kth-304774DiVA, id: diva2:1612600
Conference
45th Annual International IEEE-Computer-Society Computers, Software, and Applications Conference (COMPSAC), JUL 12-16, 2021, Electrical Network
Note

Part of proceedings:ISBN 978-1-6654-2463-9

QC 20211118

Available from: 2021-11-18 Created: 2021-11-18 Last updated: 2025-02-07Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Ning, MangHou, WenyuanMatskin, Mihhail

Search in DiVA

By author/editor
Ning, MangHou, WenyuanMatskin, Mihhail
By organisation
Robotics, Perception and Learning, RPLElectrical EngineeringSoftware and Computer systems, SCS
Computer graphics and computer vision

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 113 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf