Searching for just a few words should be enough to get started. If you need to make more complex queries, use the tips below to guide you.
Article type: Research Article
Authors: Mehmood, Zahida; * | Rashid, Muhammadb | Rehman, Amjadc | Saba, Tanzilad | Dawood, Hassana | Dawood, Hussaine
Affiliations: [a] Department of Software Engineering, University of Engineering and Technology, Taxila, Pakistan | [b] Department of Computer Engineering, Umm Al-Qura University, Makkah, Saudi Arabia | [c] College of Computer and Information Systems, Al-Yamamah University, Riyadh, Saudi Arabia | [d] College of Computer and Information Sciences, Prince Sultan University, Riyadh, Saudi Arabia | [e] Faculty of Computing and Information Technology, University of Jeddah, Jeddah, Saudi Arabia
Correspondence: [*] Corresponding author. Zahid Mehmood, Department of Software Engineering, University of Engineering and Technology, Taxila 47050, Pakistan. E-mail: [email protected].
Abstract: Due to the fast growth of multimedia archives, the semantic gap is becoming a vital problem between machine learning based semantic concepts and local features of the image to retrieve images accurately. To address this issue, the proposed method of this article introduces two novel methods for effective image retrieval known as visual words integration after clustering (VWIaC) and feature integration before clustering (FIbC). These methods use complementary features of histograms of oriented gradients (HOG) and oriented FAST and rotated BRIEF (ORB) descriptors founded on the bag-of-words (BoW) model for salient objects within the images to build smaller and larger sizes of codebooks. To achieve higher efficiency in terms of specificity of the image retrieval system, the codebook of larger sizes are preferred, while larger sizes codebook produces low sensitivity and vice versa. The proposed method of VWIaC produces two smaller sizes codebooks to achieve higher sensitivity. After that visual words of both smaller size codebooks are integrated to produce larger size codebook, which improves the specificity of the proposed method. The performance of the proposed method is tested on three standard image benchmarks, which verifies its vigorous performance as compared to an FIbC method and recent CBIR methods.
Keywords: Image visuals search, complementary image visuals, object retrieval, clustering, complementary features
DOI: 10.3233/JIFS-171137
Journal: Journal of Intelligent & Fuzzy Systems, vol. 35, no. 5, pp. 5421-5434, 2018
IOS Press, Inc.
6751 Tepper Drive
Clifton, VA 20124
USA
Tel: +1 703 830 6300
Fax: +1 703 830 2300
[email protected]
For editorial issues, like the status of your submitted paper or proposals, write to [email protected]
IOS Press
Nieuwe Hemweg 6B
1013 BG Amsterdam
The Netherlands
Tel: +31 20 688 3355
Fax: +31 20 687 0091
[email protected]
For editorial issues, permissions, book requests, submissions and proceedings, contact the Amsterdam office [email protected]
Inspirees International (China Office)
Ciyunsi Beili 207(CapitaLand), Bld 1, 7-901
100025, Beijing
China
Free service line: 400 661 8717
Fax: +86 10 8446 7947
[email protected]
For editorial issues, like the status of your submitted paper or proposals, write to [email protected]
如果您在出版方面需要帮助或有任何建, 件至: [email protected]