Volume 32, Issue 1 pp. 324-342
RESEARCH ARTICLE
Free to Read

A novel multimodality anatomical image fusion method based on contrast and structure extraction

Arbab Sufyan

Arbab Sufyan

Department of Computer Science, BUITEMS, Quetta, Pakistan

Search for more papers by this author
Muhammad Imran

Corresponding Author

Muhammad Imran

Department of Electrical Engineering, BUITEMS, Quetta, Pakistan

Control, Automotive, and Robotics Lab, National Center of Robotics and Automation, Rawalpindi, Pakistan

Correspondence

Muhammad Imran, Department of Electrical Engineering, BUITEMS, Quetta, 87300, Pakistan.

Email: [email protected]

Search for more papers by this author
Syed Attique Shah

Syed Attique Shah

Department of Computer Science, BUITEMS, Quetta, Pakistan

Search for more papers by this author
Hamayoun Shahwani

Hamayoun Shahwani

Department of Telecom Engineering, BUITEMS, Quetta, Pakistan

Search for more papers by this author
Arbab Abdul Wadood

Arbab Abdul Wadood

Quetta Institute of Medical Sciences, Quetta, Pakistan

Search for more papers by this author
First published: 27 August 2021
Citations: 14

Funding information: National Center of Robotics and Automation, Pakistan

Abstract

Image modalities, such as computed tomography (CT), magnetic resonance imaging (MRI), single-photon emission computed tomography (SPECT), and so on, reflect various levels of details about objects of interest that help medical practitioners to examine patients' diseases from different perspectives. A single medical image, at times, may not be sufficient for making a critical decision; therefore, providing detailed information from a different perspective may help in making a better decision. Image fusion techniques play a vital role in this regard by combining important details from different medical images into a single, information enhanced image. In this article, we present a novel weighted term multimodality anatomical medical image fusion method. The proposed method, as a first step, eliminates the distortions from the source images and afterward, extracts two pieces of crucial information: the local contrast and the salient structure. Both the local contrast and salient structure are later combined to obtain the final weight map. The obtained weights are then passed through a fast guided filter to remove the discontinuities and noise. Lastly, the refined weight map is fused with source images using pyramid decomposition to get the final fused image. The proposed method is accessed and compared both qualitatively and quantitatively with state-of-the-art techniques. The result illustrates the performance superiority and efficiency of the proposed method.

CONFLICT OF INTEREST

The authors declare no conflicts of interest.

DATA AVAILABILITY STATEMENT

Data will be provided if required.

The full text of this article hosted at iucr.org is unavailable due to technical difficulties.