Graph Relation Network: Modeling Relations Between Scenes for Multilabel Remote-Sensing Image Classification and Retrieval
Ver/ Abrir
Impacto
Scholar |
Otros documentos de la autoría: kang, jian; Fernandez-Beltran, Ruben; Danfeng, Hong; Chanussot, Jocelyn; Plaza, Antonio
Metadatos
Mostrar el registro completo del ítemcomunitat-uji-handle:10234/9
comunitat-uji-handle2:10234/43662
comunitat-uji-handle3:10234/43643
comunitat-uji-handle4:
INVESTIGACIONMetadatos
Título
Graph Relation Network: Modeling Relations Between Scenes for Multilabel Remote-Sensing Image Classification and RetrievalFecha de publicación
2020-08-21Editor
IEEEISSN
0196-2892Cita bibliográfica
KANG, Jian, et al. Graph relation network: Modeling relations between scenes for multilabel remote-sensing image classification and retrieval. IEEE Transactions on Geoscience and Remote Sensing, 2020.Tipo de documento
info:eu-repo/semantics/articleVersión de la editorial
https://ieeexplore.ieee.org/abstract/document/9173783Versión
info:eu-repo/semantics/acceptedVersionPalabras clave / Materias
Resumen
Due to the proliferation of large-scale remote-sensing (RS) archives with multiple annotations, multilabel RS scene classification and retrieval are becoming increasingly popular. Although some recent deep learning-based ... [+]
Due to the proliferation of large-scale remote-sensing (RS) archives with multiple annotations, multilabel RS scene classification and retrieval are becoming increasingly popular. Although some recent deep learning-based methods are able to achieve promising results in this context, the lack of research on how to learn embedding spaces under the multilabel assumption often makes these models unable to preserve complex semantic relations pervading aerial scenes, which is an important limitation in RS applications. To fill this gap, we propose a new graph relation network (GRN) for multilabel RS scene categorization. Our GRN is able to model the relations between samples (or scenes) by making use of a graph structure which is fed into network learning. For this purpose, we define a new loss function called scalable neighbor discriminative loss with binary cross entropy (SNDL-BCE) that is able to embed the graph structures through the networks more effectively. The proposed approach can guide deep learning techniques (such as convolutional neural networks) to a more discriminative metric space, where semantically similar RS scenes are closely embedded and dissimilar images are separated from a novel multilabel viewpoint. To achieve this goal, our GRN jointly maximizes a weighted leave-one-out K-nearest neighbors (KNN) score in the training set, where the weight matrix describes the contributions of the nearest neighbors associated with each RS image on its class decision, and the likelihood of the class discrimination in the multilabel scenario. An extensive experimental comparison, conducted on three multilabel RS scene data archives, validates the effectiveness of the proposed GRN in terms of KNN classification and image retrieval. The codes of this article will be made publicly available for reproducible research in the community. [-]
Proyecto de investigación
RTI2018-098651-B-C54; GR18060; H2020 EOXPOSURE project Grant 734541; AXA Research FundDerechos de acceso
© 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.
http://rightsstatements.org/vocab/InC/1.0/
info:eu-repo/semantics/openAccess
http://rightsstatements.org/vocab/InC/1.0/
info:eu-repo/semantics/openAccess
Aparece en las colecciones
- INIT_Articles [751]