Multimodal image fusion : a systematic review
Shrida Kalamkar, Geetha Mary A.
Multimodal image fusion combines information from multiple modalities to generate a composite image containing complementary information. Multimodal image fusion is challenging due to the heterogeneous nature of data, misalignment and nonlinear relationships between input data, or incomplete data during the fusion process. In recent years, several attention mechanisms have been introduced to enhance the performance of deep learning models. However, little literature is available on multimodal image fusion using attention mechanisms. This paper aims to study and analyze the latest deep-learning approaches, including attention mechanisms for multimodal image fusion. As a result of this study, the graphical taxonomy based on the different image modalities, various fusion strategies, fusion levels, and metrics for fusion tasks has been put forth. The focus has been on various Multimodal image fusion frameworks based on deep-learning techniques as their core methodology. This paper also sheds light on the challenges and future research directions in this field, application domains, and benchmark datasets used for multimodal fusion tasks. This paper contributes to the research on Multimodal image fusion and can help researchers select a suitable methodology for their applications.
Year of publication: |
2023
|
---|---|
Authors: | Kalamkar, Shrida ; Geetha Mary A. |
Published in: |
Decision analytics journal. - Amsterdam : Elsevier, ISSN 2772-6622, ZDB-ID 3106160-6. - Vol. 9.2023, Art.-No. 100327, p. 1-16
|
Subject: | Deep learning | Image fusion | Multimodal data | Transformers | Wavelet transforms |
Saved in:
freely available
Saved in favorites
Similar items by subject
-
Hedonic prices and quality adjusted price indices powered by AI
Bajari, Patrick L., (2023)
-
Ampel, Benjamin M., (2024)
-
A transformer-based model for default prediction in mid-cap corporate markets
Korangi, Kamesh, (2023)
- More ...