Study of Benchmark Datasets for Performance Evaluation of Content Based Image Retrieval System for Remote Sensing Images
Keywords:
examination, determine, literature,Abstract
The development of new feature extraction techniques and the testing of fresh datasets have greatly improved image retrieval systems for remote sensing. When using Remote Sensing Image Retrieval (RSIR) techniques and assessing their effectiveness, benchmark datasets are essential. Notable advancements in creating benchmark datasets for RSIR systems are highlighted in the literature. Unisource retrieval are the two categories into which these datasets fall. While the query and the images that are retrieved come from the same source in Unisource retrieval, they come from distinct sources in cross-source retrieval. In order to determine which datasets are best suited for the use of deep learning Convolutional Neural Networks and contemporary transfer learning techniques, this research offers a thorough examination of the salient features of both types of remote sensing datasets.
Downloads
References
A. Latif, A. Rasheed, U. Sajid, J. Ahmed, N. Ali, N. I. Ratyal, B. Zafar, S. H. Dar, M. Sajid, T. Khalil, et al., “Content-based image retrieval and feature extraction: a comprehensive review,” Mathematical problems in engineering, vol. 2019, 2019.
Y. Li, J. Ma, and Y. Zhang, “Image retrieval from remote sensing big data: A survey,” Information Fusion, vol. 67, pp. 94–115, 2021.
Z. Shao, W. Zhou, X. Deng, M. Zhang, and Q. Cheng, “Multilabel remote sensing image retrieval based on fully convolutional network,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 13, pp. 318–328, 2020.
Y. Yang and S. Newsam, “Bag-of-visual-words and spatial extensions for land-use classification,” in Proceedings of the 18th SIGSPATIAL international conference on advances in geographic information systems, pp. 270–279, 2010.
G.-S. Xia, W. Yang, J. Delon, Y. Gousseau, H. Sun, and H. Maˆıtre, “Structural high-resolution satellite image indexing,” in ISPRS TC VII Symposium-100 Years ISPRS, vol. 38, pp. 298–303, 2010.
Q. Zou, L. Ni, T. Zhang, and Q. Wang, “Deep learning based feature selection for remote sensing scene classifi- cation,” IEEE Geoscience and remote sensing letters, vol. 12, no. 11, pp. 2321–2325, 2015.
G.-S. Xia, J. Hu, F. Hu, B. Shi, X. Bai, Y. Zhong, L. Zhang, and X. Lu, “Aid: A benchmark data set for performance evaluation of aerial scene classification,” IEEE Transactions on Geoscience and Remote Sensing, vol. 55, no. 7, pp. 3965–3981, 2017.
W. Zhou, S. Newsam, C. Li, and Z. Shao, “Patternnet: A benchmark dataset for performance evaluation of remote sensing image retrieval,” ISPRS journal of photogrammetry and remote sensing, vol. 145, pp. 197–209, 2018.
H. Li, X. Dou, C. Tao, Z. Hou, J. Chen, J. Peng, M. Deng, and L. Zhao, “Rsi-cb: A large scale remote sensing image classification benchmark via crowdsource data,” arXiv preprint arXiv:1705.10450, 2017.
B. Zhao, Y. Zhong, G.-S. Xia, and L. Zhang, “Dirichlet-derived multiple topic scene classification model for high spatial resolution remote sensing imagery,” IEEE Transactions on Geoscience and Remote Sensing, vol. 54, no. 4,
pp. 2108–2123, 2015.
G. Cheng, J. Han, and X. Lu, “Remote sensing image scene classification: Benchmark and state of the art,”
Proceedings of the IEEE, vol. 105, no. 10, pp. 1865–1883, 2017.
Z. Shao, K. Yang, and W. Zhou, “Performance evaluation of single-label and multi-label remote sensing image retrieval using a dense labeling dataset,” Remote Sensing, vol. 10, no. 6, p. 964, 2018.
G. Sumbul, M. Charfuelan, B. Demir, and V. Markl, “Bigearthnet: A large-scale benchmark archive for re- mote sensing image understanding,” in IGARSS 2019-2019 IEEE International Geoscience and Remote Sensing Symposium, pp. 5901–5904, IEEE, 2019.
Z. Shao, K. Yang, and W. Zhou, “A benchmark dataset for performance evaluation of multi-label remote sensing image retrieval.,” Remote Sensing, vol. 10, no. 6, 2018.
U. Chaudhuri, B. Banerjee, A. Bhattacharya, and M. Datcu, “Cmir-net: A deep learning based model for cross-modal retrieval in remote sensing,” Pattern recognition letters, vol. 131, pp. 456–462, 2020.
F. Xu, W. Yang, T. Jiang, S. Lin, H. Luo, and G.-S. Xia, “Mental retrieval of remote sensing images via adversarial sketch-image feature learning,” IEEE Transactions on Geoscience and Remote Sensing, vol. 58, no. 11, pp. 7801–7814, 2020.
B. Qu, X. Li, D. Tao, and X. Lu, “Deep semantic understanding of high resolution remote sensing image,” in 2016 International conference on computer, information and telecommunication systems (Cits), pp. 1–5, IEEE, 2016.
X. Lu, B. Wang, X. Zheng, and X. Li, “Exploring models and data for remote sensing image caption generation,”
IEEE Transactions on Geoscience and Remote Sensing, vol. 56, no. 4, pp. 2183–2195, 2017.
G. Mao, Y. Yuan, and L. Xiaoqiang, “Deep cross-modal retrieval for remote sensing image and audio,” in 2018 10th IAPR workshop on pattern recognition in remote sensing (PRRS), pp. 1–7, IEEE, 2018.
T. Abdullah, Y. Bazi, M. M. Al Rahhal, M. L. Mekhalfi, L. Rangarajan, and M. Zuair, “Textrs: Deep bidi- rectional triplet network for matching text to remote sensing images,” Remote Sensing, vol. 12, no. 3, p. 405, 2020.
Y. Sun, S. Feng, Y. Ye, X. Li, J. Kang, Z. Huang, and C. Luo, “Multisensor fusion and explicit semantic preserving-based deep hashing for cross-modal remote sensing image retrieval,” IEEE Transactions on Geo- science and Remote Sensing, vol. 60, pp. 1–14, 2021.
Z. Zheng, Y. Wei, and Y. Yang, “University-1652: A multi-view multi-source benchmark for drone-based geo- localization,” in Proceedings of the 28th ACM international conference on Multimedia, pp. 1395–1403, 2020.
L. Liu and H. Li, “Lending orientation to neural networks for cross-view geo-localization,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 5624–5633, 2019.
G. Machado, E. Ferreira, K. Nogueira, H. Oliveira, M. Brito, P. H. T. Gama, and J. A.
Downloads
Published
How to Cite
Issue
Section
License

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.