kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
A Design of Smart Unmanned Vending Machine for New Retail Based on Binocular Camera and Machine Vision
KTH. Fudan University, China.ORCID iD: 0000-0002-6554-2041
Fudan University, China.
Fudan University, China.
Fudan University, China.
Show others and affiliations
2022 (English)In: IEEE Consumer Electronics Magazine, ISSN 2162-2248, E-ISSN 2162-2256, Vol. 11, no 4, p. 21-31Article in journal (Refereed) Published
Abstract [en]

The smart unmanned vending machine using machine vision technology suffers from the sharp decrease of detection accuracy due to the incomplete image collection of items by monocular camera in complex environment, and the lack of obvious features in dense stacking of items. In this paper, a binocular camera system is designed to effectively solve the problems of distortion and coverage caused by monocular camera. Besides, an image-stitching algorithm is developed to splice the images captured by the camera, which reliefs the burden of computation for back-end recognition processing brought by the binocular camera. A new model YOLOv3-TinyE is proposed based on YOLOv3-tiny model. Based on the data set of 21, 000 images captured in real scenarios that containing 20 different type of beverages, the comparison experimental results show that YOLOv3-TinyE model achieves the mean average precision of 99.15%, and the inference speed is 2.91 times faster than that of YOLOv3 model, and the detection accuracy of YOLOv3-TinyE model based on binocular vision is higher than that based on monocular vision. The results suggest that the designed method achieves the goal in terms of inference speed and average precision, that is, it is able to satisfy the requirements for real-world applications.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2022. Vol. 11, no 4, p. 21-31
Keywords [en]
Cameras, Containers, Face recognition, Feature extraction, Image sensors, Image stitching, Servers
National Category
Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:kth:diva-306091DOI: 10.1109/MCE.2021.3060722ISI: 000809395000011Scopus ID: 2-s2.0-85101767958OAI: oai:DiVA.org:kth-306091DiVA, id: diva2:1622105
Note

QC 20250508

Available from: 2021-12-21 Created: 2021-12-21 Last updated: 2025-05-08Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Hu, Xiaoming

Search in DiVA

By author/editor
Liu, LizhengHu, Xiaoming
By organisation
KTHOptimization and Systems Theory
In the same journal
IEEE Consumer Electronics Magazine
Computer graphics and computer vision

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 88 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf