Abstract
Image fusion aims to merge two or more images captured via various sensors of the same scene to construct a more informative image by integrating their details. Generally, such integration is achieved through the manipulation of the representations of the images concerned. Sparse representation plays an important role in the effective description of images, offering a great potential in a variety of image processing tasks, including image fusion. Supported by sparse representation, in this paper, an approach for image fusion by the use of a novel dictionary learning scheme is proposed. The nonlocal self-similarity property of the images is exploited, not only at the stage of learning the underlying description dictionary but during the process of image fusion. In particular, the property of nonlocal self-similarity is combined with the traditional sparse dictionary. This results in an improved learned dictionary, hereafter referred to as the nonlocal sparse K-SVD dictionary (where K-SVD stands for the K times singular value decomposition that is commonly used in the literature), and abbreviated to NL_SK_SVD. The performance of the NL_SK_SVD dictionary is applied for image fusion using simultaneous orthogonal matching pursuit. The proposed approach is evaluated with different types of images, and compared with a number of alternative image fusion techniques. The resultant superior fused images using the present approach demonstrates the efficacy of the NL_SK_SVD dictionary in sparse image representation.
© 2016 Optical Society of America
Full Article | PDF ArticleMore Like This
L. N. Smith, C. C. Olson, K. P. Judd, and J. M. Nichols
Appl. Opt. 51(17) 3941-3949 (2012)
Lihong Chang, Xiangchu Feng, Rui Zhang, Hua Huang, Weiwei Wang, and Chen Xu
Appl. Opt. 56(28) 7969-7977 (2017)
Xiang Yan, Hanlin Qin, and Jia Li
J. Opt. Soc. Am. A 34(3) 432-440 (2017)