In this paper, we propose a new shape-coding algorithm called wavelet-based shape coding (WBSC). Performing wavelet transform on the orientation of original planar curve gives the corners called corner-1 points and en...In this paper, we propose a new shape-coding algorithm called wavelet-based shape coding (WBSC). Performing wavelet transform on the orientation of original planar curve gives the corners called corner-1 points and end of arcs that belong to the original curve. Each arc is represented by a broken line and the corners called corner-2 points of the broken line are extracted. A polygonal approximation of a contour is an ordered list of corner-1 points, ends of arcs and corner-2 points which are extracted by using the above algorithm. All of the points are called polygonal vertices which will be compressed by our adaptive arithmetic encoding. Experimental results show that our method reduces code bits by about 26% compared with the context-based arithmetic encoding (CAE) of MPEG-4, and the subjective quality of the reconstructed shape is better than that of CAE at the same Dn.展开更多
In order to improve the acquisition probability of satellite navigation signals, this paper proposes a novel code acquisition method based on wavelet transform filtering. Firstly, the signal vector based on the signal...In order to improve the acquisition probability of satellite navigation signals, this paper proposes a novel code acquisition method based on wavelet transform filtering. Firstly, the signal vector based on the signal passing through a set of partial matched filters (PMFs) is built. Then, wavelet domain filtering is performed on the signal vector value. Since the correlation signal is low in frequency and narrow in bandwidth, the noise out-of-band can be filtered out and the most of the useful signal energy is retained. Thus this process greatly improves the signal to noise ratio (SNR). Finally, the detection variable when the filtered signal goes through the combination process is constructed and the detection based on signal energy is made. Moreover, for the better retaining useful signal energy, the rule of selection of wavelet function has been made. Simulation results show the proposed method has a better detection performance than the normal code acquisition methods under the same false alarm probability.展开更多
With the advances of display technology, three-dimensional(3-D) imaging systems are becoming increasingly popular. One way of stimulating 3-D perception is to use stereo pairs, a pair of images of the same scene acqui...With the advances of display technology, three-dimensional(3-D) imaging systems are becoming increasingly popular. One way of stimulating 3-D perception is to use stereo pairs, a pair of images of the same scene acquired from different perspectives. Since there is an inherent redundancy between the images of a stereo pairs, data compression algorithms should be employed to represent stereo pairs efficiently. The proposed techniques generally use block-based disparity compensation. In order to get the higher compression ratio, this paper employs the wavelet-based mixed-resolution coding technique to incorporate with SPT-based disparity-compensation to compress the stereo image data. The mixed-resolution coding is a perceptually justified technique that is achieved by presenting one eye with a low-resolution image and the other with a high-resolution image. Psychophysical experiments show that the stereo image pairs with one high-resolution image and one low-resolution image provide almost the same stereo depth to that of a stereo image with two high-resolution images. By combining the mixed-resolution coding and SPT-based disparity-compensation techniques, one reference (left) high-resolution image can be compressed by a hierarchical wavelet transform followed by vector quantization and Huffman encoder. After two level wavelet decompositions, for the low-resolution right image and low-resolution left image, subspace projection technique using the fixed block size disparity compensation estimation is used. At the decoder, the low-resolution right subimage is estimated using the disparity from the low-resolution left subimage. A full-size reconstruction is obtained by upsampling a factor of 4 and reconstructing with the synthesis low pass filter. Finally, experimental results are presented, which show that our scheme achieves a PSNR gain (about 0.92dB) as compared to the current block-based disparity compensation coding techniques.展开更多
In this paper a novel coding method based on fuzzy vector quantization for noised image with Gaussian white-noise pollution is presented. By restraining the high frequency subbands of wavelet image the noise is signif...In this paper a novel coding method based on fuzzy vector quantization for noised image with Gaussian white-noise pollution is presented. By restraining the high frequency subbands of wavelet image the noise is significantly removed and coded with fuzzy vector quantization. The experimental result shows that the method can not only achieve high compression ratio but also remove noise dramatically.展开更多
To compress hyperspectral images, a low complexity discrete cosine transform (DCT)-based distributed source coding (DSC) scheme with Gray code is proposed. Unlike most of the existing DSC schemes, which utilize tr...To compress hyperspectral images, a low complexity discrete cosine transform (DCT)-based distributed source coding (DSC) scheme with Gray code is proposed. Unlike most of the existing DSC schemes, which utilize transform in spatial domain, the proposed algorithm applies transform in spectral domain. Set-partitioning-based approach is applied to reorganize DCT coefficients into waveletlike tree structure and extract the sign, refinement, and significance bitplanes. The extracted refinement bits are Gray encoded. Because of the dependency along the line dimension of hyperspectral images, low density paritycheck-(LDPC)-based Slepian-Wolf coder is adopted to implement the DSC strategy. Experimental results on airborne visible/infrared imaging spectrometer (AVIRIS) dataset show that the proposed paradigm achieves up to 6 dB improvement over DSC-based coders which apply transform in spatial domain, with significantly reduced computational complexity and memory storage.展开更多
Through research for image compression based on wavelet analysis in recent years, we put forward an adaptive wavelet decomposition strategy. Whether sub-images are to be decomposed or not are decided by their energy d...Through research for image compression based on wavelet analysis in recent years, we put forward an adaptive wavelet decomposition strategy. Whether sub-images are to be decomposed or not are decided by their energy defined by certain criterion. Then we derive the adaptive wavelet decomposition tree (AWDT) and the way of adjustable compression ratio. According to the feature of AWDT, this paper also deals with the strategies which are used to handle different sub-images in the procedure of quantification and coding of the wavelet coefficients. Through experiments, not only the algorithm in the paper can adapt to various images, but also the quality of recovered image is improved though compression ratio is higher and adjustable. When their compression ratios are near, the quality of subjective vision and PSNR of the algorithm are better than those of JPEG algorithm.展开更多
Considering the relatively poor robustness of quality scores for different types of distortion and the lack of mechanism for determining distortion types, a no-reference image quality assessment(NR-IQA) method based o...Considering the relatively poor robustness of quality scores for different types of distortion and the lack of mechanism for determining distortion types, a no-reference image quality assessment(NR-IQA) method based on the Ada Boost BP neural network in the wavelet domain(WABNN) is proposed. A 36-dimensional image feature vector is constructed by extracting natural scene statistics(NSS) features and local information entropy features of the distorted image wavelet sub-band coefficients in three scales. The ABNN classifier is obtained by learning the relationship between image features and distortion types. The ABNN scorer is obtained by learning the relationship between image features and image quality scores. A series of contrast experiments are carried out in the laboratory of image and video engineering(LIVE) database and TID2013 database. Experimental results show the high accuracy of the distinguishing distortion type, the high consistency with subjective scores and the high robustness of the method for distorted images. Experiment results also show the independence of the database and the relatively high operation efficiency of this method.展开更多
为了解决Transformer编码器在行人重识别中因图像块信息丢失以及行人局部特征表达不充分导致模型识别准确率低的问题,本文提出改进型Transformer编码器和特征融合的行人重识别算法。针对Transformer在注意力运算时会丢失行人图像块相对...为了解决Transformer编码器在行人重识别中因图像块信息丢失以及行人局部特征表达不充分导致模型识别准确率低的问题,本文提出改进型Transformer编码器和特征融合的行人重识别算法。针对Transformer在注意力运算时会丢失行人图像块相对位置信息的问题,引入相对位置编码,促使网络关注行人图像块语义化的特征信息,以增强行人特征的提取能力。为了突出包含行人区域的显著特征,将局部patch注意力机制模块嵌入到Transformer网络中,对局部关键特征信息进行加权强化。最后,利用全局与局部信息特征融合实现特征间的优势互补,提高模型识别能力。训练阶段使用Softmax及三元组损失函数联合优化网络,本文算法在Market1501和DukeMTMC⁃reID两大主流数据集中评估测试,Rank⁃1指标分别达到97.5%和93.5%,平均精度均值(mean Average precision,mAP)分别达到92.3%和83.1%,实验结果表明改进型Transformer编码器和特征融合算法能够有效提高行人重识别的准确率。展开更多
文摘In this paper, we propose a new shape-coding algorithm called wavelet-based shape coding (WBSC). Performing wavelet transform on the orientation of original planar curve gives the corners called corner-1 points and end of arcs that belong to the original curve. Each arc is represented by a broken line and the corners called corner-2 points of the broken line are extracted. A polygonal approximation of a contour is an ordered list of corner-1 points, ends of arcs and corner-2 points which are extracted by using the above algorithm. All of the points are called polygonal vertices which will be compressed by our adaptive arithmetic encoding. Experimental results show that our method reduces code bits by about 26% compared with the context-based arithmetic encoding (CAE) of MPEG-4, and the subjective quality of the reconstructed shape is better than that of CAE at the same Dn.
基金supported by the National Natural Science Foundation of China(6117213861401340)the Fundamental Research Funds for the Central Universities(K5051302015)
文摘In order to improve the acquisition probability of satellite navigation signals, this paper proposes a novel code acquisition method based on wavelet transform filtering. Firstly, the signal vector based on the signal passing through a set of partial matched filters (PMFs) is built. Then, wavelet domain filtering is performed on the signal vector value. Since the correlation signal is low in frequency and narrow in bandwidth, the noise out-of-band can be filtered out and the most of the useful signal energy is retained. Thus this process greatly improves the signal to noise ratio (SNR). Finally, the detection variable when the filtered signal goes through the combination process is constructed and the detection based on signal energy is made. Moreover, for the better retaining useful signal energy, the rule of selection of wavelet function has been made. Simulation results show the proposed method has a better detection performance than the normal code acquisition methods under the same false alarm probability.
基金This project was supported by the National Natural Science Foundation (No. 69972027).
文摘With the advances of display technology, three-dimensional(3-D) imaging systems are becoming increasingly popular. One way of stimulating 3-D perception is to use stereo pairs, a pair of images of the same scene acquired from different perspectives. Since there is an inherent redundancy between the images of a stereo pairs, data compression algorithms should be employed to represent stereo pairs efficiently. The proposed techniques generally use block-based disparity compensation. In order to get the higher compression ratio, this paper employs the wavelet-based mixed-resolution coding technique to incorporate with SPT-based disparity-compensation to compress the stereo image data. The mixed-resolution coding is a perceptually justified technique that is achieved by presenting one eye with a low-resolution image and the other with a high-resolution image. Psychophysical experiments show that the stereo image pairs with one high-resolution image and one low-resolution image provide almost the same stereo depth to that of a stereo image with two high-resolution images. By combining the mixed-resolution coding and SPT-based disparity-compensation techniques, one reference (left) high-resolution image can be compressed by a hierarchical wavelet transform followed by vector quantization and Huffman encoder. After two level wavelet decompositions, for the low-resolution right image and low-resolution left image, subspace projection technique using the fixed block size disparity compensation estimation is used. At the decoder, the low-resolution right subimage is estimated using the disparity from the low-resolution left subimage. A full-size reconstruction is obtained by upsampling a factor of 4 and reconstructing with the synthesis low pass filter. Finally, experimental results are presented, which show that our scheme achieves a PSNR gain (about 0.92dB) as compared to the current block-based disparity compensation coding techniques.
文摘In this paper a novel coding method based on fuzzy vector quantization for noised image with Gaussian white-noise pollution is presented. By restraining the high frequency subbands of wavelet image the noise is significantly removed and coded with fuzzy vector quantization. The experimental result shows that the method can not only achieve high compression ratio but also remove noise dramatically.
基金supported by the National Natural Science Foundationof China (60702012)the Scientific Research Foundation for the Re-turned Overseas Chinese Scholars, State Education Ministry
文摘To compress hyperspectral images, a low complexity discrete cosine transform (DCT)-based distributed source coding (DSC) scheme with Gray code is proposed. Unlike most of the existing DSC schemes, which utilize transform in spatial domain, the proposed algorithm applies transform in spectral domain. Set-partitioning-based approach is applied to reorganize DCT coefficients into waveletlike tree structure and extract the sign, refinement, and significance bitplanes. The extracted refinement bits are Gray encoded. Because of the dependency along the line dimension of hyperspectral images, low density paritycheck-(LDPC)-based Slepian-Wolf coder is adopted to implement the DSC strategy. Experimental results on airborne visible/infrared imaging spectrometer (AVIRIS) dataset show that the proposed paradigm achieves up to 6 dB improvement over DSC-based coders which apply transform in spatial domain, with significantly reduced computational complexity and memory storage.
文摘Through research for image compression based on wavelet analysis in recent years, we put forward an adaptive wavelet decomposition strategy. Whether sub-images are to be decomposed or not are decided by their energy defined by certain criterion. Then we derive the adaptive wavelet decomposition tree (AWDT) and the way of adjustable compression ratio. According to the feature of AWDT, this paper also deals with the strategies which are used to handle different sub-images in the procedure of quantification and coding of the wavelet coefficients. Through experiments, not only the algorithm in the paper can adapt to various images, but also the quality of recovered image is improved though compression ratio is higher and adjustable. When their compression ratios are near, the quality of subjective vision and PSNR of the algorithm are better than those of JPEG algorithm.
基金supported by the National Natural Science Foundation of China(61471194 61705104)+1 种基金the Science and Technology on Avionics Integration Laboratory and Aeronautical Science Foundation of China(20155552050)the Natural Science Foundation of Jiangsu Province(BK20170804)
文摘Considering the relatively poor robustness of quality scores for different types of distortion and the lack of mechanism for determining distortion types, a no-reference image quality assessment(NR-IQA) method based on the Ada Boost BP neural network in the wavelet domain(WABNN) is proposed. A 36-dimensional image feature vector is constructed by extracting natural scene statistics(NSS) features and local information entropy features of the distorted image wavelet sub-band coefficients in three scales. The ABNN classifier is obtained by learning the relationship between image features and distortion types. The ABNN scorer is obtained by learning the relationship between image features and image quality scores. A series of contrast experiments are carried out in the laboratory of image and video engineering(LIVE) database and TID2013 database. Experimental results show the high accuracy of the distinguishing distortion type, the high consistency with subjective scores and the high robustness of the method for distorted images. Experiment results also show the independence of the database and the relatively high operation efficiency of this method.
文摘为了解决Transformer编码器在行人重识别中因图像块信息丢失以及行人局部特征表达不充分导致模型识别准确率低的问题,本文提出改进型Transformer编码器和特征融合的行人重识别算法。针对Transformer在注意力运算时会丢失行人图像块相对位置信息的问题,引入相对位置编码,促使网络关注行人图像块语义化的特征信息,以增强行人特征的提取能力。为了突出包含行人区域的显著特征,将局部patch注意力机制模块嵌入到Transformer网络中,对局部关键特征信息进行加权强化。最后,利用全局与局部信息特征融合实现特征间的优势互补,提高模型识别能力。训练阶段使用Softmax及三元组损失函数联合优化网络,本文算法在Market1501和DukeMTMC⁃reID两大主流数据集中评估测试,Rank⁃1指标分别达到97.5%和93.5%,平均精度均值(mean Average precision,mAP)分别达到92.3%和83.1%,实验结果表明改进型Transformer编码器和特征融合算法能够有效提高行人重识别的准确率。