Graph Relation Network: Modeling Relations Between Scenes for Multilabel Remote-Sensing Image Classification and Retrieval
Visualitza/
Impacte
Scholar |
Altres documents de l'autoria: kang, jian; Fernandez-Beltran, Ruben; Danfeng, Hong; Chanussot, Jocelyn; Plaza, Antonio
Metadades
Mostra el registre complet de l'elementcomunitat-uji-handle:10234/9
comunitat-uji-handle2:10234/43662
comunitat-uji-handle3:10234/43643
comunitat-uji-handle4:
INVESTIGACIONMetadades
Títol
Graph Relation Network: Modeling Relations Between Scenes for Multilabel Remote-Sensing Image Classification and RetrievalData de publicació
2020-08-21Editor
IEEEISSN
0196-2892Cita bibliogràfica
KANG, Jian, et al. Graph relation network: Modeling relations between scenes for multilabel remote-sensing image classification and retrieval. IEEE Transactions on Geoscience and Remote Sensing, 2020.Tipus de document
info:eu-repo/semantics/articleVersió de l'editorial
https://ieeexplore.ieee.org/abstract/document/9173783Versió
info:eu-repo/semantics/acceptedVersionParaules clau / Matèries
Resum
Due to the proliferation of large-scale remote-sensing (RS) archives with multiple annotations, multilabel RS scene classification and retrieval are becoming increasingly popular. Although some recent deep learning-based ... [+]
Due to the proliferation of large-scale remote-sensing (RS) archives with multiple annotations, multilabel RS scene classification and retrieval are becoming increasingly popular. Although some recent deep learning-based methods are able to achieve promising results in this context, the lack of research on how to learn embedding spaces under the multilabel assumption often makes these models unable to preserve complex semantic relations pervading aerial scenes, which is an important limitation in RS applications. To fill this gap, we propose a new graph relation network (GRN) for multilabel RS scene categorization. Our GRN is able to model the relations between samples (or scenes) by making use of a graph structure which is fed into network learning. For this purpose, we define a new loss function called scalable neighbor discriminative loss with binary cross entropy (SNDL-BCE) that is able to embed the graph structures through the networks more effectively. The proposed approach can guide deep learning techniques (such as convolutional neural networks) to a more discriminative metric space, where semantically similar RS scenes are closely embedded and dissimilar images are separated from a novel multilabel viewpoint. To achieve this goal, our GRN jointly maximizes a weighted leave-one-out K-nearest neighbors (KNN) score in the training set, where the weight matrix describes the contributions of the nearest neighbors associated with each RS image on its class decision, and the likelihood of the class discrimination in the multilabel scenario. An extensive experimental comparison, conducted on three multilabel RS scene data archives, validates the effectiveness of the proposed GRN in terms of KNN classification and image retrieval. The codes of this article will be made publicly available for reproducible research in the community. [-]
Proyecto de investigación
RTI2018-098651-B-C54; GR18060; H2020 EOXPOSURE project Grant 734541; AXA Research FundDrets d'accés
© 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.
http://rightsstatements.org/vocab/InC/1.0/
info:eu-repo/semantics/openAccess
http://rightsstatements.org/vocab/InC/1.0/
info:eu-repo/semantics/openAccess
Apareix a les col.leccions
- INIT_Articles [754]