MDAG-Net: Multi-Domain Association-Guided Network for Image-based Long-term Visual Localization

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)
89 Downloads (Pure)

Abstract

In the case of long-term changing environment, long-term visual localization is a challenging problem in autonomous driving and mobile robots. Due to the influence of season, illumination and other changing weather conditions, the traditional image retrieval methods are difficult to achieve ideal results in long-term visual localization. Therefore, inspired by the human brain associative recognition function, an image retrieval based on a multidomain association-guided network is proposed to solve the long-term visual localization problem. The key idea is to extract the discriminative domain-invariant features in different scenes through multidomain image transformation of the perceptual network and the conceptual network. In addition, in order to better associate image features of different scenes in the conceptual network and guide the perceptual network to obtain more robust domain invariant features, an association-guided module is designed without the need for external datasets. On this basis, the domain feature loss function and the guidance mechanism of the loss function are introduced to assist these two network models training to obtain better performance. Finally, experiments are carried out on the CMU-Seasons dataset and the RobotCar-Seasons dataset. Compared with some state-of-the-art methods, the proposed method improved the high-precision localization result of urban, suburban, and park scenes in the CMU-Seasons dataset by 1.5%, 0.5%, and 0.7%, respectively, which also can verify the effectiveness of the proposed method under various seasonal and illumination conditions.

Original languageEnglish
Pages (from-to)859-873
Number of pages15
JournalIEEE Transactions on Cognitive and Developmental Systems
Volume17
Issue number4
Early online date3 Jan 2025
DOIs
Publication statusPublished (in print/issue) - 31 Aug 2025

Bibliographical note

Publisher Copyright:
© 2016 IEEE.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 61973066, in part by the Major Science and Technology Projects of Liaoning Province under Grant 2021JH1/10400049. The work of Sonya Coleman was supported by the EPSRC, The Nuffield Foundation, The Leverhulme Trust, and the EU. She was involved in the EU FP7 funded projects RUBICON, VISUALISE, and SLANDAIL. The work of Dermot Kerr was supported by the EU FP7 funded projects VISUALISE and SLANDAIL.

FundersFunder number
Leverhulme Trust
Nuffield Foundation
European Commission
Engineering and Physical Sciences Research Council
National Natural Science Foundation of China61973066
2021JH1/10400049

    Keywords

    • Visual localization
    • image retrieval
    • association guided
    • changing environment
    • Location awareness
    • Visualization
    • Feature extraction
    • Translation
    • Image retrieval
    • Training
    • Transfer learning
    • Simultaneous localization and mapping
    • Image recognition
    • Mobile robots
    • associationguided
    • Association-guided
    • visual localization

    Fingerprint

    Dive into the research topics of 'MDAG-Net: Multi-Domain Association-Guided Network for Image-based Long-term Visual Localization'. Together they form a unique fingerprint.

    Cite this