IEEE Access (Jan 2020)

Deep Image Sensing and Retrieval Using Suppression, Scale Spacing and Division, Interpolation and Spatial Color Coordinates With Bag of Words for Large and Complex Datasets

  • Khawaja Tehseen Ahmed,
  • Humaira Afzal,
  • Muhammad Rafiq Mufti,
  • Arif Mehmood,
  • Gyu Sang Choi

DOI
https://doi.org/10.1109/ACCESS.2020.2993721
Journal volume & issue
Vol. 8
pp. 90351 – 90379

Abstract

Read online

Intelligent and efficient image retrieval from versatile image datasets is an inevitable requirement of the current era. Primitive image signatures are vital to reflect the visual attributes for content based image retrieval (CBIR). Algorithmically descriptive and well identified visual contents form the image signatures to correctly index and retrieve similar results. Hence feature vectors should contain ample image information with color, shape, objects, spatial information perspectives to distinguish image category as a qualifying candidate. This contribution presents a novel features detector by locating the interest points by applying non-maximum suppression to productive sum of derivative of pixels computed from differential of corner scores. The interest points are described by applying scale space interpolation to scale space division produced from Hessian blob detector resulted after Gaussian smoothing. The computed shape and object information is fused with color features extracted from the spatially arranged L2 normalized coefficients. High variance coefficients are selected for object based feature vectors to reduce the massive data which in fuse form transformed to bag-of-words (BoW) for efficient retrieval and ranking. To check the competitiveness of the presented approach it is experimented on nine well-known image datasets Caltech-101, ImageNet, Corel-10000, 17-Flowers, Columbia object image library (COIL), Corel-1000, Caltech-256, tropical fruits and Amsterdam library of textures (ALOT) belong to shape, color, texture, and spatial & complex objects categories. Extensive experimentation is conducted for seven benchmark descriptors including maximally stable extremal region (MSER), speeded up robust features (SURF), difference of Gaussian (DoG), red green blue local binary pattern (RGBLBP), histogram of oriented gradients (HOG), scale invariant feature transform (SIFT), and local binary pattern (LBP). Remarkable outcomes reported that the presented technique has significant precision rates, recall rates, average retrieval precision & recall, mean average precision & recall rates for many image semantic groups of the challenging datasets. Results comparison is presented with research techniques and reported improved results.

Keywords