Skip to content

ervinyo/Signboard-datasets

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

40 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Signboard Datasets

Domain Adaptation Deep Attention Network for Automatic Logo Detection and Recognition in Google Street View

Ervin Yohannes, Chih-Yang Lin, Timothy K. Shih, Chen-Ya Hong, Avirmed Enkhbat, Fitri Utaminingrum


Abstract

Signboards are important location landmarks that provide services to a local community. They are difficult to detect and recognize due to the myriad of designs that combine text and images. Many people only see signboards as sign of place and attract attention of them. However, the impaired visually people can’t see like the usual people about signboards. They need an assistance system to guided them go to destination. Currently, the assistance system still remaining issue since limited datasets to create the best one and also need long computation time for reaches the best results. In this paper, we propose a novel framework that can automatically detect and recognize signboard logos. In addition, we utilizing Google street view for collecting our datasets in around Taiwan’s street. This framework consists of a domain adaptation that not only reduces the loss function between source-target datasets but also represents important source features adopted by the target dataset and deep learning techniques for the detection and recognition system. In our model, we add nonlocal blocks and attention mechanisms called deep attention networks to achieve the best final result. We perform extensive experiments on both our and public datasets to show superior performance and effectiveness of our proposed method. The experimental results show our proposed method outperform the baseline in all evaluation metrics of state-of-the-art detection and recognition method.

Overview

our dataset is a signboard dataset containing 29,727 images with VOC annotation format and a resolution of 500 × 400 pixels. Our dataset consists of 14 classes of store logos, including Carrefour, Domino, Family Mart, Gas, Hi-Life, KFC, McDonald, Mos Burger, Ok Mart, Post, Pxmart, 7-ELEVEN, Starbucks, and Wellcome. We already divided our datasets into training and validation.

Requirements

  1. install anaconda and python 3.6.10 See Here
  2. setting environment variable which using CUDAv10.0 and cudnn64_7(put in bin(.dll), include(.h), and lib x64(.lib))Download Here
  3. install tensorflow-gpu=1.14.0
  4. install keras=2.2.4

Download

The link for download the datasets and annotations as follow :

Download Datasets

Results

The comparison results for SSD, YOLOv2, YOLOv3, YOLOv4, and our proposed method are shown in below

SSD YOLOv2 YOLOv3 YOLOv4 Our proposed method I Our proposed method II

Citation

If you use this dataset in your research, please include the following citation in any published results.

@ARTICLE{9491158,
author={Yohannes, Ervin and Lin, Chih-Yang and Shih, Timothy K. and Hong, Chen-Ya and Enkhbat, Avirmed and Utaminingrum, Fitri},  
journal={IEEE Access},  
title={Domain Adaptation Deep Attention Network for Automatic Logo Detection and Recognition in Google Street View},  
year={2021},  
volume={9},
number={}, 
pages={102623-102635}, 
doi={10.1109/ACCESS.2021.3098713}}