Training Cnns With Image Patches for Object Localisation
Loading...
Date
Authors
Orhan, Semih
Baştanlar, Yalın
Journal Title
Journal ISSN
Volume Title
Open Access Color
GOLD
Green Open Access
Yes
OpenAIRE Downloads
OpenAIRE Views
Publicly Funded
No
Abstract
Recently, convolutional neural networks (CNNs) have shown great performance in different problems of computer vision including object detection and localisation. A novel training approach is proposed for CNNs to localise some animal species whose bodies have distinctive patterns such as leopards and zebras. To learn characteristic patterns, small patches which are taken from different body parts of animals are used to train models. To find object location, in a test image, all locations are visited in a sliding window fashion. Crops are fed into trained CNN and their classification scores are combined into a heat map. Later on, heat maps are converted to bounding box estimates for varying confidence scores. The localisation performance of the patch-based training approach is compared with Faster R-CNN – a state-of-the-art CNN-based object detection and localisation method. Experimental results reveal that the patch-based training outperforms Faster R-CNN, especially for classes with distinctive patterns.
Description
Keywords
Object detection, Computer vision, Object recognition, Convolutional Neural Networks, Animal species, Object detection, Convolutional Neural Networks, Computer vision, Object recognition, Animal species
Fields of Science
03 medical and health sciences, 0302 clinical medicine, 0202 electrical engineering, electronic engineering, information engineering, 02 engineering and technology
Citation
Orhan, S., and Baştanlar, Y. (2018). Training CNNs with image patches for object localisation. Electronics Letters, 54(7), 424-426. doi:10.1049/el.2017.4725
WoS Q
Scopus Q

OpenCitations Citation Count
8
Source
Volume
54
Issue
7
Start Page
424
End Page
426
PlumX Metrics
Citations
CrossRef : 9
Scopus : 13
Captures
Mendeley Readers : 14
SCOPUS™ Citations
13
checked on May 06, 2026
Web of Science™ Citations
9
checked on May 06, 2026
Page Views
33054
checked on May 06, 2026
Downloads
472
checked on May 06, 2026
Google Scholar™


