Human-Readable Fiducial Marker Classification using Convolutional Neural Networks

被引:0
|
作者
Liu, Yanfeng [1 ]
Psota, Eric T. [1 ]
Perez, Lance C. [1 ]
机构
[1] Univ Nebraska, Dept Elect & Comp Engn, Lincoln, NE 68588 USA
关键词
computer vision; convolutional neural network; machine learning; fiducial marker;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Many applications require both the location and identity of objects in images and video. Most existing solutions, like QR codes, AprilTags, and ARTags use complex machine-readable fiducial markers with heuristically derived methods for detection and classification. However, in applications where humans are integral to the system and need to be capable of locating objects in the environment, fiducial markers must be human readable. An obvious and convenient choice for human readable fiducial markers are alphanumeric characters (Arabic numbers and English letters). Here, a method for classifying characters using a convolutional neural network (CNN) is presented. The network is trained with a large set of computer generated images of characters where each is subjected to a carefully designed set of augmentations designed to simulate the conditions inherent in video capture. These augmentations include rotation, scaling, shearing, and blur. Results demonstrate that training on large numbers of synthetic images produces a system that works on real images captured by a video camera. The result also reveal that certain characters are generally more reliable and easier to recognize than others, thus the results can be used to intelligently design a human-readable fiducial markers system that avoids confusing characters.
引用
收藏
页码:606 / 610
页数:5
相关论文
共 50 条
  • [1] Explainable classification by learning human-readable sentences in feature subsets
    Krishnamurthy, Prashanth
    Sarmadi, Alireza
    Khorrami, Farshad
    INFORMATION SCIENCES, 2021, 564 : 202 - 219
  • [2] Learning to Encode Text as Human-Readable Summaries using Generative Adversarial Networks
    Wang, Yau-Shian
    Lee, Hung-Yi
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4187 - 4195
  • [3] Synthetically Trained Neural Networks for Learning Human-Readable Plans from Real-World Demonstrations
    Tremblay, Jonathan
    To, Thang
    Molchanov, Artem
    Tyree, Stephen
    Kautz, Jan
    Birchfield, Stan
    2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 5659 - 5666
  • [4] Human-Readable and Machine-Readable Knowledge Bases Using Specialized Word Processors
    Molina, Martin
    Blasco, Gemma
    20TH IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, VOL 1, PROCEEDINGS, 2008, : 103 - +
  • [5] Human Classification in Aerial Images Using Convolutional Neural Networks
    Akshatha, K. R.
    Karunakar, A. K.
    Shenoy, B. Satish
    MACHINE LEARNING AND AUTONOMOUS SYSTEMS, 2022, 269 : 537 - 549
  • [6] Classification of Human Metaspread Images Using Convolutional Neural Networks
    Arora, Tanvi
    INTERNATIONAL JOURNAL OF IMAGE AND GRAPHICS, 2021, 21 (03)
  • [7] Generating Semantically Similar and Human-Readable Summaries With Generative Adversarial Networks
    Zhuang, Haojie
    Zhang, Weibin
    IEEE ACCESS, 2019, 7 : 169426 - 169433
  • [8] Detection and Classification of Human Stool Using Deep Convolutional Neural Networks
    Choy, Yin Pui
    Hu, Guoqing
    Chen, Jia
    IEEE ACCESS, 2021, 9 : 160485 - 160496
  • [9] CLASSIFICATION OF HUMAN GAIT ACCELERATION DATA USING CONVOLUTIONAL NEURAL NETWORKS
    Kreuter, Daniel
    Takahashi, Hirotaka
    Omae, Yuto
    Akiduki, Takuma
    Zhang, Zhong
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2020, 16 (02): : 609 - 619
  • [10] Plant Classification using Convolutional Neural Networks
    Yalcin, Hulya
    Razavi, Salar
    2016 FIFTH INTERNATIONAL CONFERENCE ON AGRO-GEOINFORMATICS (AGRO-GEOINFORMATICS), 2016, : 233 - 237