With the advent of the information security era,it is necessary to guarantee the privacy,accuracy,and dependable transfer of pictures.This study presents a new approach to the encryption and compression of color image...With the advent of the information security era,it is necessary to guarantee the privacy,accuracy,and dependable transfer of pictures.This study presents a new approach to the encryption and compression of color images.It is predicated on 2D compressed sensing(CS)and the hyperchaotic system.First,an optimized Arnold scrambling algorithm is applied to the initial color images to ensure strong security.Then,the processed images are con-currently encrypted and compressed using 2D CS.Among them,chaotic sequences replace traditional random measurement matrices to increase the system’s security.Third,the processed images are re-encrypted using a combination of permutation and diffusion algorithms.In addition,the 2D projected gradient with an embedding decryption(2DPG-ED)algorithm is used to reconstruct images.Compared with the traditional reconstruction algorithm,the 2DPG-ED algorithm can improve security and reduce computational complexity.Furthermore,it has better robustness.The experimental outcome and the performance analysis indicate that this algorithm can withstand malicious attacks and prove the method is effective.展开更多
The act of transmitting photos via the Internet has become a routine and significant activity.Enhancing the security measures to safeguard these images from counterfeiting and modifications is a critical domain that c...The act of transmitting photos via the Internet has become a routine and significant activity.Enhancing the security measures to safeguard these images from counterfeiting and modifications is a critical domain that can still be further enhanced.This study presents a system that employs a range of approaches and algorithms to ensure the security of transmitted venous images.The main goal of this work is to create a very effective system for compressing individual biometrics in order to improve the overall accuracy and security of digital photographs by means of image compression.This paper introduces a content-based image authentication mechanism that is suitable for usage across an untrusted network and resistant to data loss during transmission.By employing scale attributes and a key-dependent parametric Long Short-Term Memory(LSTM),it is feasible to improve the resilience of digital signatures against image deterioration and strengthen their security against malicious actions.Furthermore,the successful implementation of transmitting biometric data in a compressed format over a wireless network has been accomplished.For applications involving the transmission and sharing of images across a network.The suggested technique utilizes the scalability of a structural digital signature to attain a satisfactory equilibrium between security and picture transfer.An effective adaptive compression strategy was created to lengthen the overall lifetime of the network by sharing the processing of responsibilities.This scheme ensures a large reduction in computational and energy requirements while minimizing image quality loss.This approach employs multi-scale characteristics to improve the resistance of signatures against image deterioration.The proposed system attained a Gaussian noise value of 98%and a rotation accuracy surpassing 99%.展开更多
Many classical encoding algorithms of vector quantization (VQ) of image compression that can obtain global optimal solution have computational complexity O(N). A pure quantum VQ encoding algorithm with probability...Many classical encoding algorithms of vector quantization (VQ) of image compression that can obtain global optimal solution have computational complexity O(N). A pure quantum VQ encoding algorithm with probability of success near 100% has been proposed, that performs operations 45√N times approximately. In this paper, a hybrid quantum VQ encoding algorithm between the classical method and the quantum algorithm is presented. The number of its operations is less than √N for most images, and it is more efficient than the pure quantum algorithm.展开更多
A nonlinear data analysis algorithm, namely empirical data decomposition (EDD) is proposed, which can perform adaptive analysis of observed data. Analysis filter, which is not a linear constant coefficient filter, i...A nonlinear data analysis algorithm, namely empirical data decomposition (EDD) is proposed, which can perform adaptive analysis of observed data. Analysis filter, which is not a linear constant coefficient filter, is automatically determined by observed data, and is able to implement multi-resolution analysis as wavelet transform. The algorithm is suitable for analyzing non-stationary data and can effectively wipe off the relevance of observed data. Then through discussing the applications of EDD in image compression, the paper presents a 2-dimension data decomposition framework and makes some modifications of contexts used by Embedded Block Coding with Optimized Truncation (EBCOT) . Simulation results show that EDD is more suitable for non-stationary image data compression.展开更多
This paper utilizes a spatial texture correlation and the intelligent classification algorithm (ICA) search strategy to speed up the encoding process and improve the bit rate for fractal image compression. Texture f...This paper utilizes a spatial texture correlation and the intelligent classification algorithm (ICA) search strategy to speed up the encoding process and improve the bit rate for fractal image compression. Texture features is one of the most important properties for the representation of an image. Entropy and maximum entry from co-occurrence matrices are used for representing texture features in an image. For a range block, concerned domain blocks of neighbouring range blocks with similar texture features can be searched. In addition, domain blocks with similar texture features are searched in the ICA search process. Experiments show that in comparison with some typical methods, the proposed algorithm significantly speeds up the encoding process and achieves a higher compression ratio, with a slight diminution in the quality of the reconstructed image; in comparison with a spatial correlation scheme, the proposed scheme spends much less encoding time while the compression ratio and the quality of the reconstructed image are almost the same.展开更多
The paper presents a class of nonlinear adaptive wavelet transforms for lossless image compression. In update step of the lifting the different operators are chosen by the local gradient of original image. A nonlinear...The paper presents a class of nonlinear adaptive wavelet transforms for lossless image compression. In update step of the lifting the different operators are chosen by the local gradient of original image. A nonlinear morphological predictor follows the update adaptive lifting to result in fewer large wavelet coefficients near edges for reducing coding. The nonlinear adaptive wavelet transforms can also allow perfect reconstruction without any overhead cost. Experiment results are given to show lower entropy of the adaptive transformed images than those of the non-adaptive case and great applicable potentiality in lossless image compresslon.展开更多
A new method using plane fitting to decide whether a domain block is similar enough to a given range block is proposed in this paper. First, three coefficients are computed for describing each range and domain block. ...A new method using plane fitting to decide whether a domain block is similar enough to a given range block is proposed in this paper. First, three coefficients are computed for describing each range and domain block. Then, the best-matched one for every range block is obtained by analysing the relation between their coefficients. Experimental results show that the proposed method can shorten encoding time markedly, while the retrieved image quality is still acceptable. In the decoding step, a kind of simple line fitting on block boundaries is used to reduce blocking effects. At the same time, the proposed method can also achieve a high compression ratio.展开更多
This paper presents a novel method utilizing wavelets with particle swarm optimization(PSO)for medical image compression.Our method utilizes PSO to overcome the wavelets discontinuity which occurs when compressing ima...This paper presents a novel method utilizing wavelets with particle swarm optimization(PSO)for medical image compression.Our method utilizes PSO to overcome the wavelets discontinuity which occurs when compressing images using thresholding.It transfers images into subband details and approximations using a modified Haar wavelet(MHW),and then applies a threshold.PSO is applied for selecting a particle assigned to the threshold values for the subbands.Nine positions assigned to particles values are used to represent population.Every particle updates its position depending on the global best position(gbest)(for all details subband)and local best position(pbest)(for a subband).The fitness value is developed to terminate PSO when the difference between two local best(pbest)successors is smaller than a prescribe value.The experiments are applied on five different medical image types,i.e.,MRI,CT,and X-ray.Results show that the proposed algorithm can be more preferably to compress medical images than other existing wavelets techniques from peak signal to noise ratio(PSNR)and compression ratio(CR)points of views.展开更多
An edge oriented image sequence coding scheme is presented. On the basis of edge detecting, an image could be divided into the sensitized region and the smooth region. In this scheme, the architecture of sensitized r...An edge oriented image sequence coding scheme is presented. On the basis of edge detecting, an image could be divided into the sensitized region and the smooth region. In this scheme, the architecture of sensitized region is approximated with linear type of segments. Then a rectangle belt is constructed for each segment. Finally, the gray value distribution in the region is fitted by normal forms polynomials. The model matching and motion analysis are also based on the architecture of sensitized region. For the smooth region we use the run length scanning and linear approximating. By means of normal forms polynomial fitting and motion prediction by matching, the images are compressed. It is shown through the simulations that the subjective quality of reconstructed picture is excellent at 0.0075 bit per pel.展开更多
Data compression is one of the core fields of study for applications of image and video processing.The raw data to be transmitted consumes large bandwidth and requires huge storage space as a result,it is desirable to...Data compression is one of the core fields of study for applications of image and video processing.The raw data to be transmitted consumes large bandwidth and requires huge storage space as a result,it is desirable to represent the information in the data with considerably fewer bits by the mean of data compression techniques,the data must be reconstituted very similarly to the initial form.In this paper,a hybrid compression based on Discrete Cosine Transform(DCT),DiscreteWavelet Transform(DWT)is used to enhance the quality of the reconstructed image.These techniques are followed by entropy encoding such as Huffman coding to give additional compression.Huffman coding is optimal prefix code because of its implementation is more simple,faster,and easier than other codes.It needs less execution time and it is the shortest average length and the measurements for analysis are based upon Compression Ratio,Mean Square Error(MSE),and Peak Signal to Noise Ratio(PSNR).We applied a hybrid algorithm on(DWT–DCT 2×2,4×4,8×8,16×16,32×32)blocks.Finally,we show that by using a hybrid(DWT–DCT)compression technique,the PSNR is reconstructed for the image by using the proposed hybrid algorithm(DWT–DCT 8×8 block)is quite high than DCT.展开更多
The better compression rate can be achieved by the traditional vector quantization (VQ) method, and the quality of the recovered image can also be accepted. But the decompressed image quality can not be promoted eff...The better compression rate can be achieved by the traditional vector quantization (VQ) method, and the quality of the recovered image can also be accepted. But the decompressed image quality can not be promoted efficiently, so how to balance the image compression rate and image recovering quality is an important issue, in this paper, an image is transformed by discrete wavelet transform (DWT) to generate its DWT transformed image which can be compressed by the VQ method further. Besides, we compute the values between the DWT transformed image and decompressed DWT transformed image as the difference matrix which is the adjustable basis of the decompressed image quality. By controlling the deviation of the difference matrix, there can be nearly Iossless compression for the VQ method. Experimental results show that when the number of compressed bits by our method is equal to the number of those bits compressed by the VQ method, the quality of our recovered image is better. Moreover, the proposed method has more compression capability comparing with the VQ scheme.展开更多
In this paper, image compression and decompression are realized on a personal computer based on fractal theory. The algorithm is effectiveas as the reconstructed image is similar to the original. In the algorithm, the...In this paper, image compression and decompression are realized on a personal computer based on fractal theory. The algorithm is effectiveas as the reconstructed image is similar to the original. In the algorithm, the formulas for contrast scaling and luminance shift are simplified,and the Hausdorff distance is replaced by the Euclidean distance. Thus, the calculation load is reduced. The formula for compression ratio is presented for an ideal situation, from which one can analyze how the different factors influence image compression ratio.展开更多
In this paper, the second generation wavelet transform is applied to image lossless coding, according to its characteristic of reversible integer wavelet transform. The second generation wavelet transform can provide ...In this paper, the second generation wavelet transform is applied to image lossless coding, according to its characteristic of reversible integer wavelet transform. The second generation wavelet transform can provide higher compression ratio than Huffman coding while it reconstructs image without loss compared with the first generation wavelet transform. The experimental results show that the se cond generation wavelet transform can obtain excellent performance in medical image compression coding.展开更多
Compressing encrypted images remains a challenge.As illustrated in our previous work on compression of encrypted binary images,it is preferable to exploit statistical characteristics at the receiver.Through this line,...Compressing encrypted images remains a challenge.As illustrated in our previous work on compression of encrypted binary images,it is preferable to exploit statistical characteristics at the receiver.Through this line,we characterize statistical correlations between adjacent bitplanes of a gray image with the Markov random field(MRF),represent it with a factor graph,and integrate the constructed MRF factor graph in that for binary image reconstruction,which gives rise to a joint factor graph for gray images reconstruction(JFGIR).By exploiting the JFGIR at the receiver to facilitate the reconstruction of the original bitplanes and deriving theoretically the sum-product algorithm(SPA)adapted to the JFGIR,a novel MRF-based encryption-then-compression(ETC)scheme is thus proposed.After preferable universal parameters of the MRF between adjacent bitplanes are sought via a numerical manner,extensive experimental simulations are then carried out to show that the proposed scheme successfully compresses the first 3 and 4 most significant bitplanes(MSBs)for most test gray images and the others with a large portion of smooth area,respectively.Thus,the proposed scheme achieves significant improvement against the state-of-the-art leveraging the 2-D Markov source model at the receiver and is comparable or somewhat inferior to that using the resolution-progressive strategy in recovery.展开更多
Synthetic aperture radar (SAR) images are corrupted by multiplicative speckle noise which limits the performance of the classical coder/decoder algorithm in spatial domain. The relatively new transform of multiwavel...Synthetic aperture radar (SAR) images are corrupted by multiplicative speckle noise which limits the performance of the classical coder/decoder algorithm in spatial domain. The relatively new transform of multiwavelets can possess desirable features simultaneously, such as orthogonality and symmetry, while scalar wavelets cannot. In this paper we propose a compression scheme combining with speckle noise reduction within the multiwavelet framework. Compared with classical set partitioning in hierarchical trees (SPIHT) algorithm, our method achieves favorable peak signal to noise ratio (PSNR) and superior speckle noise reduction performances.展开更多
A new image compression algorithm is proposed based on local visual activity classification and the investigation of the histograms of the small non-overlapping blocks of the differential and angle image. Histograms o...A new image compression algorithm is proposed based on local visual activity classification and the investigation of the histograms of the small non-overlapping blocks of the differential and angle image. Histograms of the differential blocks are classified according to their visual activities as Unimodal, Bimodal and Multimodal blocks. According to the histogram shape of the differential block and the mean angle of the same block an optimized quantization table with special coding is applied by taking the advantages of the local visual activities within the block. A considerable compression ratio and visual output improvement compared with the DCT compression algorithm are gained.展开更多
In this paper, three techniques, line run coding, quadtree DF (Depth-First) representation and H coding for compressing classified satellite cloud images with no distortion are presented. In these three codings, the f...In this paper, three techniques, line run coding, quadtree DF (Depth-First) representation and H coding for compressing classified satellite cloud images with no distortion are presented. In these three codings, the first two were invented by other persons and the third one, by ourselves. As a result, the comparison among their compression rates is. given at the end of this paper. Further application of these image compression technique to satellite data and other meteorological data looks promising.展开更多
The amount of image data generated in multimedia applications is ever increasing. The image compression plays vital role in multimedia applications. The ultimate aim of image compression is to reduce storage space wit...The amount of image data generated in multimedia applications is ever increasing. The image compression plays vital role in multimedia applications. The ultimate aim of image compression is to reduce storage space without degrading image quality. Compression is required whenever the data handled is huge they may be required to sent or transmitted and also stored. The New Edge Directed Interpolation (NEDI)-based lifting Discrete Wavelet Transfrom (DWT) scheme with modified Set Partitioning In Hierarchical Trees (MSPIHT) algorithm is proposed in this paper. The NEDI algorithm gives good visual quality image particularly at edges. The main objective of this paper is to be preserving the edges while performing image compression which is a challenging task. The NEDI with lifting DWT has achieved 99.18% energy level in the low frequency ranges which has 1.07% higher than 5/3 Wavelet decomposition and 0.94% higher than traditional DWT. To implement this NEDI with Lifting DWT along with MSPIHT algorithm which gives higher Peak Signal to Noise Ratio (PSNR) value and minimum Mean Square Error (MSE) and hence better image quality. The experimental results proved that the proposed method gives better PSNR value (39.40 dB for rate 0.9 bpp without arithmetic coding) and minimum MSE value is 7.4.展开更多
When an image, which is decomposed by bi-orthogonal wavelet bases, is reconstructed, some information will be lost at the four edges of the image. At the same time, artificial discontinuities will be introduced. We us...When an image, which is decomposed by bi-orthogonal wavelet bases, is reconstructed, some information will be lost at the four edges of the image. At the same time, artificial discontinuities will be introduced. We use a method called symmetric extension to solve the problem. We only consider the case of the two-band filter banks, and the results can be applied to M-band filter banks. There are only two types of symmetric extension in analysis phrase, namely the whole-sample symmetry (WS), the half-sample symmetry (HS), while there are four types of symmetric extension in synthesis phrase, namely the WS, HS, the whole-sample anti-symmetry (WA), and the half-sample anti-symmetry (HA) respectively. We can select the exact type according to the image length and the filter length, and we will show how to do these. The image can be perfectly reconstructed without any edge effects in this way. Finally, simulation results are reported. Key words edge effect - image compression - wavelet - biorthogonal bases - symmetric extension CLC number TP 37 Foundation item: Supported by the National 863 Project (20021111901010)Biography: Yu Sheng-sheng (1944-), male, Professor, research direction: multimedia information processing, SAN.展开更多
be stored or transmitted in an efficient form.In this work,a new idea is proposed,where we take advantage of the redundancy that appears in a group of images to be all compressed together,instead of compressing each i...be stored or transmitted in an efficient form.In this work,a new idea is proposed,where we take advantage of the redundancy that appears in a group of images to be all compressed together,instead of compressing each image by itself.In our proposed technique,a classification process is applied,where the set of the input images are classified into groups based on existing technique like L1 and L2 norms,color histograms.All images that belong to the same group are compressed based on dividing the images of the same group into sub-images of equal sizes and saving the references into a codebook.In the process of extracting the different sub-images,we used the mean squared error for comparison and three blurring methods(simple,middle and majority blurring)to increase the compression ratio.Experiments show that varying blurring values,as well as MSE thresholds,enhanced the compression results in a group of images compared to JPEG and PNG compressors.展开更多
基金This work was supported in part by the National Natural Science Foundation of China under Grants 71571091,71771112the State Key Laboratory of Synthetical Automation for Process Industries Fundamental Research Funds under Grant PAL-N201801the Excellent Talent Training Project of University of Science and Technology Liaoning under Grant 2019RC05.
文摘With the advent of the information security era,it is necessary to guarantee the privacy,accuracy,and dependable transfer of pictures.This study presents a new approach to the encryption and compression of color images.It is predicated on 2D compressed sensing(CS)and the hyperchaotic system.First,an optimized Arnold scrambling algorithm is applied to the initial color images to ensure strong security.Then,the processed images are con-currently encrypted and compressed using 2D CS.Among them,chaotic sequences replace traditional random measurement matrices to increase the system’s security.Third,the processed images are re-encrypted using a combination of permutation and diffusion algorithms.In addition,the 2D projected gradient with an embedding decryption(2DPG-ED)algorithm is used to reconstruct images.Compared with the traditional reconstruction algorithm,the 2DPG-ED algorithm can improve security and reduce computational complexity.Furthermore,it has better robustness.The experimental outcome and the performance analysis indicate that this algorithm can withstand malicious attacks and prove the method is effective.
文摘The act of transmitting photos via the Internet has become a routine and significant activity.Enhancing the security measures to safeguard these images from counterfeiting and modifications is a critical domain that can still be further enhanced.This study presents a system that employs a range of approaches and algorithms to ensure the security of transmitted venous images.The main goal of this work is to create a very effective system for compressing individual biometrics in order to improve the overall accuracy and security of digital photographs by means of image compression.This paper introduces a content-based image authentication mechanism that is suitable for usage across an untrusted network and resistant to data loss during transmission.By employing scale attributes and a key-dependent parametric Long Short-Term Memory(LSTM),it is feasible to improve the resilience of digital signatures against image deterioration and strengthen their security against malicious actions.Furthermore,the successful implementation of transmitting biometric data in a compressed format over a wireless network has been accomplished.For applications involving the transmission and sharing of images across a network.The suggested technique utilizes the scalability of a structural digital signature to attain a satisfactory equilibrium between security and picture transfer.An effective adaptive compression strategy was created to lengthen the overall lifetime of the network by sharing the processing of responsibilities.This scheme ensures a large reduction in computational and energy requirements while minimizing image quality loss.This approach employs multi-scale characteristics to improve the resistance of signatures against image deterioration.The proposed system attained a Gaussian noise value of 98%and a rotation accuracy surpassing 99%.
文摘Many classical encoding algorithms of vector quantization (VQ) of image compression that can obtain global optimal solution have computational complexity O(N). A pure quantum VQ encoding algorithm with probability of success near 100% has been proposed, that performs operations 45√N times approximately. In this paper, a hybrid quantum VQ encoding algorithm between the classical method and the quantum algorithm is presented. The number of its operations is less than √N for most images, and it is more efficient than the pure quantum algorithm.
基金This project was supported by the National Natural Science Foundation of China (60532060)Hainan Education Bureau Research Project (Hjkj200602)Hainan Natural Science Foundation (80551).
文摘A nonlinear data analysis algorithm, namely empirical data decomposition (EDD) is proposed, which can perform adaptive analysis of observed data. Analysis filter, which is not a linear constant coefficient filter, is automatically determined by observed data, and is able to implement multi-resolution analysis as wavelet transform. The algorithm is suitable for analyzing non-stationary data and can effectively wipe off the relevance of observed data. Then through discussing the applications of EDD in image compression, the paper presents a 2-dimension data decomposition framework and makes some modifications of contexts used by Embedded Block Coding with Optimized Truncation (EBCOT) . Simulation results show that EDD is more suitable for non-stationary image data compression.
基金supported by the National Natural Science Foundation of China (Grant Nos. 60573172 and 60973152)the Superior University Doctor Subject Special Scientific Research Foundation of China (Grant No. 20070141014)the Natural Science Foundation of Liaoning Province of China (Grant No. 20082165)
文摘This paper utilizes a spatial texture correlation and the intelligent classification algorithm (ICA) search strategy to speed up the encoding process and improve the bit rate for fractal image compression. Texture features is one of the most important properties for the representation of an image. Entropy and maximum entry from co-occurrence matrices are used for representing texture features in an image. For a range block, concerned domain blocks of neighbouring range blocks with similar texture features can be searched. In addition, domain blocks with similar texture features are searched in the ICA search process. Experiments show that in comparison with some typical methods, the proposed algorithm significantly speeds up the encoding process and achieves a higher compression ratio, with a slight diminution in the quality of the reconstructed image; in comparison with a spatial correlation scheme, the proposed scheme spends much less encoding time while the compression ratio and the quality of the reconstructed image are almost the same.
基金Supported by the National Natural Science Foundation of China (69983005)
文摘The paper presents a class of nonlinear adaptive wavelet transforms for lossless image compression. In update step of the lifting the different operators are chosen by the local gradient of original image. A nonlinear morphological predictor follows the update adaptive lifting to result in fewer large wavelet coefficients near edges for reducing coding. The nonlinear adaptive wavelet transforms can also allow perfect reconstruction without any overhead cost. Experiment results are given to show lower entropy of the adaptive transformed images than those of the non-adaptive case and great applicable potentiality in lossless image compresslon.
基金Project supported by the National Natural Science Foundation of China (Grant Nos. 61173183, 60973152, and 60573172)the Special Scientific Research Fund for the Doctoral Program of Higher Education of China (Grant No. 20070141014)the Natural Science Foundation of Liaoning Province, China (Grant No. 20082165)
文摘A new method using plane fitting to decide whether a domain block is similar enough to a given range block is proposed in this paper. First, three coefficients are computed for describing each range and domain block. Then, the best-matched one for every range block is obtained by analysing the relation between their coefficients. Experimental results show that the proposed method can shorten encoding time markedly, while the retrieved image quality is still acceptable. In the decoding step, a kind of simple line fitting on block boundaries is used to reduce blocking effects. At the same time, the proposed method can also achieve a high compression ratio.
基金funded by the University of Jeddah,Saudi Arabia,under Grant No.UJ-20-043-DR。
文摘This paper presents a novel method utilizing wavelets with particle swarm optimization(PSO)for medical image compression.Our method utilizes PSO to overcome the wavelets discontinuity which occurs when compressing images using thresholding.It transfers images into subband details and approximations using a modified Haar wavelet(MHW),and then applies a threshold.PSO is applied for selecting a particle assigned to the threshold values for the subbands.Nine positions assigned to particles values are used to represent population.Every particle updates its position depending on the global best position(gbest)(for all details subband)and local best position(pbest)(for a subband).The fitness value is developed to terminate PSO when the difference between two local best(pbest)successors is smaller than a prescribe value.The experiments are applied on five different medical image types,i.e.,MRI,CT,and X-ray.Results show that the proposed algorithm can be more preferably to compress medical images than other existing wavelets techniques from peak signal to noise ratio(PSNR)and compression ratio(CR)points of views.
文摘An edge oriented image sequence coding scheme is presented. On the basis of edge detecting, an image could be divided into the sensitized region and the smooth region. In this scheme, the architecture of sensitized region is approximated with linear type of segments. Then a rectangle belt is constructed for each segment. Finally, the gray value distribution in the region is fitted by normal forms polynomials. The model matching and motion analysis are also based on the architecture of sensitized region. For the smooth region we use the run length scanning and linear approximating. By means of normal forms polynomial fitting and motion prediction by matching, the images are compressed. It is shown through the simulations that the subjective quality of reconstructed picture is excellent at 0.0075 bit per pel.
文摘Data compression is one of the core fields of study for applications of image and video processing.The raw data to be transmitted consumes large bandwidth and requires huge storage space as a result,it is desirable to represent the information in the data with considerably fewer bits by the mean of data compression techniques,the data must be reconstituted very similarly to the initial form.In this paper,a hybrid compression based on Discrete Cosine Transform(DCT),DiscreteWavelet Transform(DWT)is used to enhance the quality of the reconstructed image.These techniques are followed by entropy encoding such as Huffman coding to give additional compression.Huffman coding is optimal prefix code because of its implementation is more simple,faster,and easier than other codes.It needs less execution time and it is the shortest average length and the measurements for analysis are based upon Compression Ratio,Mean Square Error(MSE),and Peak Signal to Noise Ratio(PSNR).We applied a hybrid algorithm on(DWT–DCT 2×2,4×4,8×8,16×16,32×32)blocks.Finally,we show that by using a hybrid(DWT–DCT)compression technique,the PSNR is reconstructed for the image by using the proposed hybrid algorithm(DWT–DCT 8×8 block)is quite high than DCT.
文摘The better compression rate can be achieved by the traditional vector quantization (VQ) method, and the quality of the recovered image can also be accepted. But the decompressed image quality can not be promoted efficiently, so how to balance the image compression rate and image recovering quality is an important issue, in this paper, an image is transformed by discrete wavelet transform (DWT) to generate its DWT transformed image which can be compressed by the VQ method further. Besides, we compute the values between the DWT transformed image and decompressed DWT transformed image as the difference matrix which is the adjustable basis of the decompressed image quality. By controlling the deviation of the difference matrix, there can be nearly Iossless compression for the VQ method. Experimental results show that when the number of compressed bits by our method is equal to the number of those bits compressed by the VQ method, the quality of our recovered image is better. Moreover, the proposed method has more compression capability comparing with the VQ scheme.
文摘In this paper, image compression and decompression are realized on a personal computer based on fractal theory. The algorithm is effectiveas as the reconstructed image is similar to the original. In the algorithm, the formulas for contrast scaling and luminance shift are simplified,and the Hausdorff distance is replaced by the Euclidean distance. Thus, the calculation load is reduced. The formula for compression ratio is presented for an ideal situation, from which one can analyze how the different factors influence image compression ratio.
基金Supported by the National Natural Science Foundation of China!( 6 9875 0 0 9)
文摘In this paper, the second generation wavelet transform is applied to image lossless coding, according to its characteristic of reversible integer wavelet transform. The second generation wavelet transform can provide higher compression ratio than Huffman coding while it reconstructs image without loss compared with the first generation wavelet transform. The experimental results show that the se cond generation wavelet transform can obtain excellent performance in medical image compression coding.
基金This work is supported in part by the National Natural Science Foundation of China under contracts 61672242 and 61702199in part by China Spark Program under Grant 2015GA780002+1 种基金in part by The National Key Research and Development Program of China under Grant 2017YFD0701601in part by Natural Science Foundation of Guangdong Province under Grant 2015A030313413.
文摘Compressing encrypted images remains a challenge.As illustrated in our previous work on compression of encrypted binary images,it is preferable to exploit statistical characteristics at the receiver.Through this line,we characterize statistical correlations between adjacent bitplanes of a gray image with the Markov random field(MRF),represent it with a factor graph,and integrate the constructed MRF factor graph in that for binary image reconstruction,which gives rise to a joint factor graph for gray images reconstruction(JFGIR).By exploiting the JFGIR at the receiver to facilitate the reconstruction of the original bitplanes and deriving theoretically the sum-product algorithm(SPA)adapted to the JFGIR,a novel MRF-based encryption-then-compression(ETC)scheme is thus proposed.After preferable universal parameters of the MRF between adjacent bitplanes are sought via a numerical manner,extensive experimental simulations are then carried out to show that the proposed scheme successfully compresses the first 3 and 4 most significant bitplanes(MSBs)for most test gray images and the others with a large portion of smooth area,respectively.Thus,the proposed scheme achieves significant improvement against the state-of-the-art leveraging the 2-D Markov source model at the receiver and is comparable or somewhat inferior to that using the resolution-progressive strategy in recovery.
基金This work was supported by the National Natural Science Foundation of China under Grant No. 60472048.
文摘Synthetic aperture radar (SAR) images are corrupted by multiplicative speckle noise which limits the performance of the classical coder/decoder algorithm in spatial domain. The relatively new transform of multiwavelets can possess desirable features simultaneously, such as orthogonality and symmetry, while scalar wavelets cannot. In this paper we propose a compression scheme combining with speckle noise reduction within the multiwavelet framework. Compared with classical set partitioning in hierarchical trees (SPIHT) algorithm, our method achieves favorable peak signal to noise ratio (PSNR) and superior speckle noise reduction performances.
文摘A new image compression algorithm is proposed based on local visual activity classification and the investigation of the histograms of the small non-overlapping blocks of the differential and angle image. Histograms of the differential blocks are classified according to their visual activities as Unimodal, Bimodal and Multimodal blocks. According to the histogram shape of the differential block and the mean angle of the same block an optimized quantization table with special coding is applied by taking the advantages of the local visual activities within the block. A considerable compression ratio and visual output improvement compared with the DCT compression algorithm are gained.
文摘In this paper, three techniques, line run coding, quadtree DF (Depth-First) representation and H coding for compressing classified satellite cloud images with no distortion are presented. In these three codings, the first two were invented by other persons and the third one, by ourselves. As a result, the comparison among their compression rates is. given at the end of this paper. Further application of these image compression technique to satellite data and other meteorological data looks promising.
文摘The amount of image data generated in multimedia applications is ever increasing. The image compression plays vital role in multimedia applications. The ultimate aim of image compression is to reduce storage space without degrading image quality. Compression is required whenever the data handled is huge they may be required to sent or transmitted and also stored. The New Edge Directed Interpolation (NEDI)-based lifting Discrete Wavelet Transfrom (DWT) scheme with modified Set Partitioning In Hierarchical Trees (MSPIHT) algorithm is proposed in this paper. The NEDI algorithm gives good visual quality image particularly at edges. The main objective of this paper is to be preserving the edges while performing image compression which is a challenging task. The NEDI with lifting DWT has achieved 99.18% energy level in the low frequency ranges which has 1.07% higher than 5/3 Wavelet decomposition and 0.94% higher than traditional DWT. To implement this NEDI with Lifting DWT along with MSPIHT algorithm which gives higher Peak Signal to Noise Ratio (PSNR) value and minimum Mean Square Error (MSE) and hence better image quality. The experimental results proved that the proposed method gives better PSNR value (39.40 dB for rate 0.9 bpp without arithmetic coding) and minimum MSE value is 7.4.
文摘When an image, which is decomposed by bi-orthogonal wavelet bases, is reconstructed, some information will be lost at the four edges of the image. At the same time, artificial discontinuities will be introduced. We use a method called symmetric extension to solve the problem. We only consider the case of the two-band filter banks, and the results can be applied to M-band filter banks. There are only two types of symmetric extension in analysis phrase, namely the whole-sample symmetry (WS), the half-sample symmetry (HS), while there are four types of symmetric extension in synthesis phrase, namely the WS, HS, the whole-sample anti-symmetry (WA), and the half-sample anti-symmetry (HA) respectively. We can select the exact type according to the image length and the filter length, and we will show how to do these. The image can be perfectly reconstructed without any edge effects in this way. Finally, simulation results are reported. Key words edge effect - image compression - wavelet - biorthogonal bases - symmetric extension CLC number TP 37 Foundation item: Supported by the National 863 Project (20021111901010)Biography: Yu Sheng-sheng (1944-), male, Professor, research direction: multimedia information processing, SAN.
文摘be stored or transmitted in an efficient form.In this work,a new idea is proposed,where we take advantage of the redundancy that appears in a group of images to be all compressed together,instead of compressing each image by itself.In our proposed technique,a classification process is applied,where the set of the input images are classified into groups based on existing technique like L1 and L2 norms,color histograms.All images that belong to the same group are compressed based on dividing the images of the same group into sub-images of equal sizes and saving the references into a codebook.In the process of extracting the different sub-images,we used the mean squared error for comparison and three blurring methods(simple,middle and majority blurring)to increase the compression ratio.Experiments show that varying blurring values,as well as MSE thresholds,enhanced the compression results in a group of images compared to JPEG and PNG compressors.