Fusion methods based on multi-scale transforms have become the mainstream of the pixel-level image fusion. However,most of these methods cannot fully exploit spatial domain information of source images, which lead to ...Fusion methods based on multi-scale transforms have become the mainstream of the pixel-level image fusion. However,most of these methods cannot fully exploit spatial domain information of source images, which lead to the degradation of image.This paper presents a fusion framework based on block-matching and 3D(BM3D) multi-scale transform. The algorithm first divides the image into different blocks and groups these 2D image blocks into 3D arrays by their similarity. Then it uses a 3D transform which consists of a 2D multi-scale and a 1D transform to transfer the arrays into transform coefficients, and then the obtained low-and high-coefficients are fused by different fusion rules. The final fused image is obtained from a series of fused 3D image block groups after the inverse transform by using an aggregation process. In the experimental part, we comparatively analyze some existing algorithms and the using of different transforms, e.g. non-subsampled Contourlet transform(NSCT), non-subsampled Shearlet transform(NSST), in the 3D transform step. Experimental results show that the proposed fusion framework can not only improve subjective visual effect, but also obtain better objective evaluation criteria than state-of-the-art methods.展开更多
An improved block matching approach to fast disparity estimation in machine vision applications is proposed, where the matching criterion is the sum of the absolute difference(SAD).By evaluating the lower bounds, wh...An improved block matching approach to fast disparity estimation in machine vision applications is proposed, where the matching criterion is the sum of the absolute difference(SAD).By evaluating the lower bounds, which become increasingly tighter for the matching criteria, the method tries to successively terminate unnecessary computations of the matching criteria between the reference block in one image and the ineligible candidate blocks in another image.It also eliminates the ineligible blocks as early as possible, while ensuring the optimal disparity of each pixel.Also, the proposed method can further speed up the elimination of ineligible candidate blocks by efficiently using the continuous constraint of disparity to predict the initial disparity of each pixel.The performance of the new algorithm is evaluated by carrying out a theoretical analysis, and by comparing its performance with the disparity estimation method based on the standard block matching.Simulated results demonstrate that the proposed algorithm achieves a computational cost reduction of over 50.5% in comparision with the standard block matching method.展开更多
基金supported by the National Natural Science Foundation of China(6157206361401308)+6 种基金the Fundamental Research Funds for the Central Universities(2016YJS039)the Natural Science Foundation of Hebei Province(F2016201142F2016201187)the Natural Social Foundation of Hebei Province(HB15TQ015)the Science Research Project of Hebei Province(QN2016085ZC2016040)the Natural Science Foundation of Hebei University(2014-303)
文摘Fusion methods based on multi-scale transforms have become the mainstream of the pixel-level image fusion. However,most of these methods cannot fully exploit spatial domain information of source images, which lead to the degradation of image.This paper presents a fusion framework based on block-matching and 3D(BM3D) multi-scale transform. The algorithm first divides the image into different blocks and groups these 2D image blocks into 3D arrays by their similarity. Then it uses a 3D transform which consists of a 2D multi-scale and a 1D transform to transfer the arrays into transform coefficients, and then the obtained low-and high-coefficients are fused by different fusion rules. The final fused image is obtained from a series of fused 3D image block groups after the inverse transform by using an aggregation process. In the experimental part, we comparatively analyze some existing algorithms and the using of different transforms, e.g. non-subsampled Contourlet transform(NSCT), non-subsampled Shearlet transform(NSST), in the 3D transform step. Experimental results show that the proposed fusion framework can not only improve subjective visual effect, but also obtain better objective evaluation criteria than state-of-the-art methods.
基金supported by the Opening Project of State Key Laboratory for Manufacturing Systems EngineeringFoundation for Youth Teacher of School of Mechanical Engineering, Xi’an Jiaotong University Brain Korea 21(BK21) Program of Ministry of Education and Human Resources Development
文摘An improved block matching approach to fast disparity estimation in machine vision applications is proposed, where the matching criterion is the sum of the absolute difference(SAD).By evaluating the lower bounds, which become increasingly tighter for the matching criteria, the method tries to successively terminate unnecessary computations of the matching criteria between the reference block in one image and the ineligible candidate blocks in another image.It also eliminates the ineligible blocks as early as possible, while ensuring the optimal disparity of each pixel.Also, the proposed method can further speed up the elimination of ineligible candidate blocks by efficiently using the continuous constraint of disparity to predict the initial disparity of each pixel.The performance of the new algorithm is evaluated by carrying out a theoretical analysis, and by comparing its performance with the disparity estimation method based on the standard block matching.Simulated results demonstrate that the proposed algorithm achieves a computational cost reduction of over 50.5% in comparision with the standard block matching method.