This paper presents an investigation on the effect of JPEG compression on the similarity between the target image and the background,where the similarity is further used to determine the degree of clutter in the image...This paper presents an investigation on the effect of JPEG compression on the similarity between the target image and the background,where the similarity is further used to determine the degree of clutter in the image.Four new clutter metrics based on image quality assessment are introduced,among which the Haar wavelet-based perceptual similarity index,known as HaarPSI,provides the best target acquisition prediction results.It is shown that the similarity between the target and the background at the boundary between visually lossless and visually lossy compression does not change significantly compared to the case when an uncompressed image is used.In future work,through subjective tests,it is necessary to check whether this presence of compression at the threshold of just noticeable differences will affect the human target acquisition performance.Similarity values are compared with the results of subjective tests of the well-known target Search_2 database,where the degree of agreement between objective and subjective scores,measured through linear correlation,reached a value of 90%.展开更多
Blind image quality assessment(BIQA)is of fundamental importance in low-level computer vision community.Increasing interest has been drawn in exploiting deep neural networks for BIQA.Despite of the notable success ach...Blind image quality assessment(BIQA)is of fundamental importance in low-level computer vision community.Increasing interest has been drawn in exploiting deep neural networks for BIQA.Despite of the notable success achieved,there is a broad consensus that training deep convolutional neural networks(DCNN)heavily relies on massive annotated data.Unfortunately,BIQA is typically a small sample problem,resulting the generalization ability of BIQA severely restricted.In order to improve the accuracy and generalization ability of BIQA metrics,this work proposed a totally opinion-unaware BIQA in which no subjective annotations are involved in the training stage.Multiple full-reference image quality assessment(FR-IQA)metrics are employed to label the distorted image as a substitution of subjective quality annotation.A deep neural network(DNN)is trained to blindly predict the multiple FR-IQA score in absence of corresponding pristine image.In the end,a selfsupervised FR-IQA score aggregator implemented by adversarial auto-encoder pools the predictions of multiple FR-IQA scores into the final quality predicting score.Even though none of subjective scores are involved in the training stage,experimental results indicate that our proposed full reference induced BIQA framework is as competitive as state-of-the-art BIQA metrics.展开更多
Image quality assessment(IQA)is constantly innovating,but there are still three types of stickers that have not been resolved:the“content sticker”-limitation of training set,the“annotation sticker”-subjective inst...Image quality assessment(IQA)is constantly innovating,but there are still three types of stickers that have not been resolved:the“content sticker”-limitation of training set,the“annotation sticker”-subjective instability in opinion scores and the“distortion sticker”-disordered distortion settings.In this paper,a No-Reference Image Quality Assessment(NR IQA)approach is proposed to deal with the problems.For“content sticker”,we introduce the idea of pairwise comparison and generate a largescale ranking set to pre-train the network;For“annotation sticker”,the absolute noise-containing subjective scores are transformed into ranking comparison results,and we design an indirect unsupervised regression based on EigenValue Decomposition(EVD);For“distortion sticker”,we propose a perception-based distortion classification method,which makes the distortion types clear and refined.Experiments have proved that our NR IQA approach Experiments show that the algorithm performs well and has good generalization ability.Furthermore,the proposed perception based distortion classification method would be able to provide insights on how the visual related studies may be developed and to broaden our understanding of human visual system.展开更多
Multi-modality medical image fusion has more and more important applications in medical image analysis and understanding. In this paper, we develop and apply a multi-resolution method based on wavelet pyramid to fuse ...Multi-modality medical image fusion has more and more important applications in medical image analysis and understanding. In this paper, we develop and apply a multi-resolution method based on wavelet pyramid to fuse medical images from different modalities such as PET-MRI and CT-MRI. In particular, we evaluate the different fusion results when applying different selection rules and obtain optimum combination of fusion parameters.展开更多
The traditional Range Doppler(RD)algorithm is unable to meet practical needs owing to the limit of resolution.The order of fractional Fourier Transform(FrFT)and the length of sampling signals affect SAR imaging perfor...The traditional Range Doppler(RD)algorithm is unable to meet practical needs owing to the limit of resolution.The order of fractional Fourier Transform(FrFT)and the length of sampling signals affect SAR imaging performance when FrFT is applied to RD algorithm.To overcome the above shortcomings,the purpose of this paper is to propose a high-resolution SAR image algorithm by using the optimal order of FrFT and the sample length constraints for the range direction.The expression of the optimal order of SAR range signals via FrFT is deduced in detail.The initial sample length and its constraints are proposed to obtain the best sample length of SAR range signals.Experimental results demonstrate that,when the range sampling-length changes in a certain interval,the best sampling-length will be obtained,which the best values of the range resolution,PSLR and ISLR,will be derived respectively.Compared with traditional RD algorithm,the main-lobe width of the peak-point target of the proposed algorithm is narrow in the range direction.While the peak amplitude of the first side-lobe is reduced significantly,those of other side-lobes also drop in various degrees.展开更多
In this paper,a new approach is proposed to determine whether the content of an image is authentic or modified with a focus on detecting complex image tampering.Detecting image tampering without any prior information ...In this paper,a new approach is proposed to determine whether the content of an image is authentic or modified with a focus on detecting complex image tampering.Detecting image tampering without any prior information of the original image is a challenging problem since unknown diverse manipulations may have different characteristics and so do various formats of images.Our principle is that image processing,no matter how complex,may affect image quality,so image quality metrics can be used to distinguish tampered images.In particular,based on the alteration of image quality in modified blocks,the proposed method can locate the tampered areas.Referring to four types of effective no-reference image quality metrics,we obtain 13 features to present an image.The experimental results show that the proposed method is very promising on detecting image tampering and locating the locally tampered areas especially in realistic scenarios.展开更多
Most of Image Quality Assessment (IQA) metrics consist of two processes. In the first process, quality map of image is measured locally. In the second process, the last quality score is converted from the quality map ...Most of Image Quality Assessment (IQA) metrics consist of two processes. In the first process, quality map of image is measured locally. In the second process, the last quality score is converted from the quality map by using the pooling strategy. The first process had been made effective and significant progresses, while the second process was always done in simple ways. In the second process of the pooling strategy, the optimal perceptual pooling weights should be determined and computed according to Human Visual System (HVS). Thus, a reliable spatial pooling mathematical model based on HVS is an important issue worthy of study. In this paper, a new Visual Perceptual Pooling Strategy (VPPS) for IQA is presented based on contrast sensitivity and luminance sensitivity of HVS. Experimental results with the LIVE database show that the visual perceptual weights, obtained by the proposed pooling strategy, can effectively and significantly improve the performances of the IQA metrics with Mean Structural SIMilarity (MSSIM) or Phase Quantization Code (PQC). It is confirmed that the proposed VPPS demonstrates promising results for improving the performances of existing IQA metrics.展开更多
Based on compressive sampling transmission model, we demonstrate here a method of quality evaluation for the reconstruction images, which is promising for the transmission of unstructured signal with reduced dimension...Based on compressive sampling transmission model, we demonstrate here a method of quality evaluation for the reconstruction images, which is promising for the transmission of unstructured signal with reduced dimension. By this method, the auxiliary information of the recovery image quality is obtained as a feedback to control number of measurements from compressive sampling video stream. Therefore, the number of measurements can be easily derived at the condition of the absence of information sparsity, and the recovery image quality is effectively improved. Theoretical and experimental results show that this algorithm can estimate the quality of images effectively and is in well consistency with the traditional objective evaluation algorithm.展开更多
In this paper, we propose a novel image recompression frame- work and image quality assessment (IQA) method to efficiently recompress Internet images. With this framework image size is significantly reduced without ...In this paper, we propose a novel image recompression frame- work and image quality assessment (IQA) method to efficiently recompress Internet images. With this framework image size is significantly reduced without affecting spatial resolution or perceptible quality of the image. With the help of IQA, the relationship between image quality and image evaluation scores can be quickly established, and the optimal quality factor can be obtained quickly and accurately within a pre - determined perceptual quality range. This process ensures the image's perceptual quality, which is applied to each input image. The test results show that, using the proposed method, the file size of images can be reduced by about 45%-60% without affecting their visual quality. Moreover, our new image -reeompression framework can be used in to many different application scenarios.展开更多
Traditional image quality assessment methods use the hand-crafted features to predict the image quality score,which cannot perform well in many scenes.Since deep learning promotes the development of many computer visi...Traditional image quality assessment methods use the hand-crafted features to predict the image quality score,which cannot perform well in many scenes.Since deep learning promotes the development of many computer vision tasks,many IQA methods start to utilize the deep convolutional neural networks(CNN)for IQA task.In this paper,a CNN-based multi-scale blind image quality predictor is proposed to extract more effectivity multi-scale distortion features through the pyramidal convolution,which consists of two tasks:A distortion recognition task and a quality regression task.For the first task,image distortion type is obtained by the fully connected layer.For the second task,the image quality score is predicted during the distortion recognition progress.Experimental results on three famous IQA datasets show that the proposed method has better performance than the previous traditional algorithms for quality prediction and distortion recognition.展开更多
Satellite vibrations during exposure will lead to pixel aliasing of remote sensors, resulting in the deterioration of image quality. In this paper, we expose the problem and discuss the characteristics of satellite vi...Satellite vibrations during exposure will lead to pixel aliasing of remote sensors, resulting in the deterioration of image quality. In this paper, we expose the problem and discuss the characteristics of satellite vibrations, and then present a pixel mixing model. The idea of mean mixing ratio (MMR) is proposed. MMR computations for different frequencies are implemented. In the mixing model, a coefficient matrix is introduced to estimate each mixed pixel. Thus, the simulation of degraded image can be performed when the vibration attitudes are known. The computation of MMR takes into considera- tion the influences of various frequencies and amplitudes. Therefore, the roles of these parameters played in the degradation progress are identified. Computations show that under the same vibration amplitude, the influence of vibrations fluctuates with the variation of frequency. The fluctuation becomes smaller as the frequency rises. Two kinds of vibration imaging experiments are performed: different amplitudes with the same frequency and different frequencies with the same amplitude. Results are found to be in very good agreement with the theoretical results. MMR has a better description of image quality than modulation transfer function (MTF). The influence of vibrations is determined mainly by the amplitude rather than the frequency. The influence of vibrations on image quality becomes gradually stable with the increase of frequency.展开更多
Low contrast of Magnetic Resonance(MR)images limits the visibility of subtle structures and adversely affects the outcome of both subjective and automated diagnosis.State-of-the-art contrast boosting techniques intole...Low contrast of Magnetic Resonance(MR)images limits the visibility of subtle structures and adversely affects the outcome of both subjective and automated diagnosis.State-of-the-art contrast boosting techniques intolerably alter inherent features of MR images.Drastic changes in brightness features,induced by post-processing are not appreciated in medical imaging as the grey level values have certain diagnostic meanings.To overcome these issues this paper proposes an algorithm that enhance the contrast of MR images while preserving the underlying features as well.This method termed as Power-law and Logarithmic Modification-based Histogram Equalization(PLMHE)partitions the histogram of the image into two sub histograms after a power-law transformation and a log compression.After a modification intended for improving the dispersion of the sub-histograms and subsequent normalization,cumulative histograms are computed.Enhanced grey level values are computed from the resultant cumulative histograms.The performance of the PLMHE algorithm is comparedwith traditional histogram equalization based algorithms and it has been observed from the results that PLMHE can boost the image contrast without causing dynamic range compression,a significant change in mean brightness,and contrast-overshoot.展开更多
The bandwidth of internet connections is still a bottleneck when transmitting large amounts of images,making the image quality assessment essential.Neurophysiological assessment of image quality has highlight advantag...The bandwidth of internet connections is still a bottleneck when transmitting large amounts of images,making the image quality assessment essential.Neurophysiological assessment of image quality has highlight advantages for it does not interfere with natural viewing behavior.However,in JPEG compression,the previous study is hard to tell the difference between the electroencephalogram(EEG)evoked by different quality images.In this paper,we propose an EEG analysis approach based on algebraic topology analysis,and the result shows that the difference between Euler characteristics of EEG evoked by different distortion images is striking both in the alpha and beta band.Moreover,we further discuss the relationship between the images and the EEG signals,and the results implied that the algebraic topological properties of images are consistent with that of brain perception,which is possible to give birth to braininspired image compression based on algebraic topological features.In general,an algebraic topologybased approach was proposed in this paper to analyze the perceptual characteristics of image quality,which will be beneficial to provide a reliable score for data compression in the network and improve the network transmission capacity.展开更多
Through research for image compression based on wavelet analysis in recent years, we put forward an adaptive wavelet decomposition strategy. Whether sub-images are to be decomposed or not are decided by their energy d...Through research for image compression based on wavelet analysis in recent years, we put forward an adaptive wavelet decomposition strategy. Whether sub-images are to be decomposed or not are decided by their energy defined by certain criterion. Then we derive the adaptive wavelet decomposition tree (AWDT) and the way of adjustable compression ratio. According to the feature of AWDT, this paper also deals with the strategies which are used to handle different sub-images in the procedure of quantification and coding of the wavelet coefficients. Through experiments, not only the algorithm in the paper can adapt to various images, but also the quality of recovered image is improved though compression ratio is higher and adjustable. When their compression ratios are near, the quality of subjective vision and PSNR of the algorithm are better than those of JPEG algorithm.展开更多
To minimize radiation risk,dose reduction is important in the diagnostic and therapeutic applications of computed tomography(CT).However,image noise degrades image quality owing to the reduced X-ray dose and a possibl...To minimize radiation risk,dose reduction is important in the diagnostic and therapeutic applications of computed tomography(CT).However,image noise degrades image quality owing to the reduced X-ray dose and a possible unacceptably reduced diagnostic performance.Deep learning approaches with convolutional neural networks(CNNs)have been proposed for natural image denoising;however,these approaches might introduce image blurring or loss of original gradients.The aim of this study was to compare the dose-dependent properties of a CNN-based denoising method for low-dose CT with those of other noise-reduction methods on unique CT noise-simulation images.To simulate a low-dose CT image,a Poisson noise distribution was introduced to normal-dose images while convoluting the CT unit-specific modulation transfer function.An abdominal CT of 100 images obtained from a public database was adopted,and simulated dose-reduction images were created from the original dose at equal 10-step dose-reduction intervals with a final dose of 1/100.These images were denoised using the denoising network structure of CNN(DnCNN)as the general CNN model and for transfer learning.To evaluate the image quality,image similarities determined by the structural similarity index(SSIM)and peak signal-to-noise ratio(PSNR)were calculated for the denoised images.Significantly better denoising,in terms of SSIM and PSNR,was achieved by the DnCNN than by other image denoising methods,especially at the ultra-low-dose levels used to generate the 10%and 5%dose-equivalent images.Moreover,the developed CNN model can eliminate noise and maintain image sharpness at these dose levels and improve SSIM by approximately 10%from that of the original method.In contrast,under small dose-reduction conditions,this model also led to excessive smoothing of the images.In quantitative evaluations,the CNN denoising method improved the low-dose CT and prevented over-smoothing by tailoring the CNN model.展开更多
By embedding secret data into cover images,image steganography can produce non-discriminable stego-images.The turtle shell model for data hiding is an excellent method that uses a reference matrix to make a good balan...By embedding secret data into cover images,image steganography can produce non-discriminable stego-images.The turtle shell model for data hiding is an excellent method that uses a reference matrix to make a good balance between image quality and embedding capacity.However,increasing the embedding capacity by extending the area of basic structures of the turtle shell model usually leads to severe degradation of image quality.In this research,we innovatively extend the basic structure of the turtle shell model into a three-dimensional(3D)space.Some intrinsic properties of the original turtle shell model are well preserved in the 3D version.Theoretic analysis shows that the new proposed models have good performance both in the image quality and in the complexity of the reference matrix.Our experimental results justify the theoretic conclusions.展开更多
The evaluation of geometric calibration accuracy of high resolution satellite images has been increasingly recognized in recent years.In order to evaluate geometric accuracy for dual-camera satellite images based on t...The evaluation of geometric calibration accuracy of high resolution satellite images has been increasingly recognized in recent years.In order to evaluate geometric accuracy for dual-camera satellite images based on the ground control points(GCP),a rigorous geometric imaging model,which was based on the collinear equation of the probe directional angle and the optimized tri-axial attitude determination(TRIAD)algorithm,is presented.Two reliable test fields in Tianjin and Jinan(China)were utilized for geometric accuracy validation of Pakistan Remote Sensing Satellite-1.The experimental results demonstrate a certain deviation of the on-orbit calibration result from the initial design values of the calibration parameters.Therefore,on-orbit geometric calibration is necessary for optical satellite imagery.Within this research,the geometrical performances including positioning accuracy without/with GCP and band registration of the dual-camera satellite were analyzed in detail,and the results of geometric image quality are assessed and discussed.As a result,it is feasible and necessary to establish such a geometric calibration model to evaluate the geometric quality of dual-camera satellite.展开更多
A new method for no-reference image quality assessment based on hybrid fuzzy-genetic technique is proposed. Noise variance and edge sharpness level of the restored image are two basic metrics for assessing the perform...A new method for no-reference image quality assessment based on hybrid fuzzy-genetic technique is proposed. Noise variance and edge sharpness level of the restored image are two basic metrics for assessing the performance of the restoration algorithm, then a fuzzy if-then inference system is developed to combine the two metrics to get a final quality score, and the parameters of the fuzzy membership function are trained with genetic algorithms. Experiments results show that the image quality score correlates well with mean opinion score and the proposed approach is robust and effective.展开更多
Invoice document digitization is crucial for efficient management in industries.The scanned invoice image is often noisy due to various reasons.This affects the OCR(optical character recognition)detection accuracy.In ...Invoice document digitization is crucial for efficient management in industries.The scanned invoice image is often noisy due to various reasons.This affects the OCR(optical character recognition)detection accuracy.In this paper,letter data obtained from images of invoices are denoised using a modified autoencoder based deep learning method.A stacked denoising autoencoder(SDAE)is implemented with two hidden layers each in encoder network and decoder network.In order to capture the most salient features of training samples,a undercomplete autoencoder is designed with non-linear encoder and decoder function.This autoencoder is regularized for denoising application using a combined loss function which considers both mean square error and binary cross entropy.A dataset consisting of 59,119 letter images,which contains both English alphabets(upper and lower case)and numbers(0 to 9)is prepared from many scanned invoices images and windows true type(.ttf)files,are used for training the neural network.Performance is analyzed in terms of Signal to Noise Ratio(SNR),Peak Signal to Noise Ratio(PSNR),Structural Similarity Index(SSIM)and Universal Image Quality Index(UQI)and compared with other filtering techniques like Nonlocal Means filter,Anisotropic diffusion filter,Gaussian filters and Mean filters.Denoising performance of proposed SDAE is compared with existing SDAE with single loss function in terms of SNR and PSNR values.Results show the superior performance of proposed SDAE method.展开更多
In order to establish a stereoscopic image quality assessment method which is consistent with human visual perception,we propose an objective stereoscopic image quality assessment method.It takes into account the stro...In order to establish a stereoscopic image quality assessment method which is consistent with human visual perception,we propose an objective stereoscopic image quality assessment method.It takes into account the strong correlation and high degree of structural between pixels of image.This method contains two models.One is the quality synthetic assessment of left-right view images,which is based on human visual characteristics,we use the Singular Value Decomposition(SVD)that can represent the degree of the distortion,and combine the qualities of left and right images by the characteristics of binocular superposition.The other model is stereoscopic perception quality assessment,due to strong stability of image’s singular value characteristics,we calculate the distance of the singular values and structural characteristic similarity of the absolute difference maps,and utilize the statistical value of the global error to evaluate stereoscopic perception.Finally,we combine two models to describe the stereoscopic image quality.Experimental results show that the correlation coefficients of the proposed assessment method and the human subjective perception are above 0.93,and the mean square errors are all less than 6.2,under JPEG,JP2K compression,Gaussian blurring,Gaussian white noise,H.264 coding distortion,and hybrid cross distortion.It indicates that the proposed stereoscopic objective method is consistent with human visual properties and also of availability.展开更多
文摘This paper presents an investigation on the effect of JPEG compression on the similarity between the target image and the background,where the similarity is further used to determine the degree of clutter in the image.Four new clutter metrics based on image quality assessment are introduced,among which the Haar wavelet-based perceptual similarity index,known as HaarPSI,provides the best target acquisition prediction results.It is shown that the similarity between the target and the background at the boundary between visually lossless and visually lossy compression does not change significantly compared to the case when an uncompressed image is used.In future work,through subjective tests,it is necessary to check whether this presence of compression at the threshold of just noticeable differences will affect the human target acquisition performance.Similarity values are compared with the results of subjective tests of the well-known target Search_2 database,where the degree of agreement between objective and subjective scores,measured through linear correlation,reached a value of 90%.
基金supported by the Public Welfare Technology Application Research Project of Zhejiang Province,China(No.LGF21F010001)the Key Research and Development Program of Zhejiang Province,China(Grant No.2019C01002)the Key Research and Development Program of Zhejiang Province,China(Grant No.2021C03138)。
文摘Blind image quality assessment(BIQA)is of fundamental importance in low-level computer vision community.Increasing interest has been drawn in exploiting deep neural networks for BIQA.Despite of the notable success achieved,there is a broad consensus that training deep convolutional neural networks(DCNN)heavily relies on massive annotated data.Unfortunately,BIQA is typically a small sample problem,resulting the generalization ability of BIQA severely restricted.In order to improve the accuracy and generalization ability of BIQA metrics,this work proposed a totally opinion-unaware BIQA in which no subjective annotations are involved in the training stage.Multiple full-reference image quality assessment(FR-IQA)metrics are employed to label the distorted image as a substitution of subjective quality annotation.A deep neural network(DNN)is trained to blindly predict the multiple FR-IQA score in absence of corresponding pristine image.In the end,a selfsupervised FR-IQA score aggregator implemented by adversarial auto-encoder pools the predictions of multiple FR-IQA scores into the final quality predicting score.Even though none of subjective scores are involved in the training stage,experimental results indicate that our proposed full reference induced BIQA framework is as competitive as state-of-the-art BIQA metrics.
基金supported by the Specialized Research Fund for the Doctoral Program of Higher Education of China, "Research of Visual Perception for Impairments of Color Information in High-Definition Images" (No.20110018110001)
文摘Image quality assessment(IQA)is constantly innovating,but there are still three types of stickers that have not been resolved:the“content sticker”-limitation of training set,the“annotation sticker”-subjective instability in opinion scores and the“distortion sticker”-disordered distortion settings.In this paper,a No-Reference Image Quality Assessment(NR IQA)approach is proposed to deal with the problems.For“content sticker”,we introduce the idea of pairwise comparison and generate a largescale ranking set to pre-train the network;For“annotation sticker”,the absolute noise-containing subjective scores are transformed into ranking comparison results,and we design an indirect unsupervised regression based on EigenValue Decomposition(EVD);For“distortion sticker”,we propose a perception-based distortion classification method,which makes the distortion types clear and refined.Experiments have proved that our NR IQA approach Experiments show that the algorithm performs well and has good generalization ability.Furthermore,the proposed perception based distortion classification method would be able to provide insights on how the visual related studies may be developed and to broaden our understanding of human visual system.
基金the National Natural Science Foundation of China (No. 19675005).
文摘Multi-modality medical image fusion has more and more important applications in medical image analysis and understanding. In this paper, we develop and apply a multi-resolution method based on wavelet pyramid to fuse medical images from different modalities such as PET-MRI and CT-MRI. In particular, we evaluate the different fusion results when applying different selection rules and obtain optimum combination of fusion parameters.
基金This work is supported by the 13th Five-Year Plan for Jiangsu Education Science(D/2020/01/22)JSPIGKZ and Natural Science Research Projects of Colleges and Universities in Jiangsu Province(19KJB510022)。
文摘The traditional Range Doppler(RD)algorithm is unable to meet practical needs owing to the limit of resolution.The order of fractional Fourier Transform(FrFT)and the length of sampling signals affect SAR imaging performance when FrFT is applied to RD algorithm.To overcome the above shortcomings,the purpose of this paper is to propose a high-resolution SAR image algorithm by using the optimal order of FrFT and the sample length constraints for the range direction.The expression of the optimal order of SAR range signals via FrFT is deduced in detail.The initial sample length and its constraints are proposed to obtain the best sample length of SAR range signals.Experimental results demonstrate that,when the range sampling-length changes in a certain interval,the best sampling-length will be obtained,which the best values of the range resolution,PSLR and ISLR,will be derived respectively.Compared with traditional RD algorithm,the main-lobe width of the peak-point target of the proposed algorithm is narrow in the range direction.While the peak amplitude of the first side-lobe is reduced significantly,those of other side-lobes also drop in various degrees.
基金Sponsored by the National Natural Science Foundation of China(Grant No.60971095 and No.61172109)Artificial Intelligence Key Laboratory of Sichuan Province(Grant No.2012RZJ01)the Fundamental Research Funds for the Central Universities(Grant No.DUT13RC201)
文摘In this paper,a new approach is proposed to determine whether the content of an image is authentic or modified with a focus on detecting complex image tampering.Detecting image tampering without any prior information of the original image is a challenging problem since unknown diverse manipulations may have different characteristics and so do various formats of images.Our principle is that image processing,no matter how complex,may affect image quality,so image quality metrics can be used to distinguish tampered images.In particular,based on the alteration of image quality in modified blocks,the proposed method can locate the tampered areas.Referring to four types of effective no-reference image quality metrics,we obtain 13 features to present an image.The experimental results show that the proposed method is very promising on detecting image tampering and locating the locally tampered areas especially in realistic scenarios.
基金Supported by the National Natural Science Foundation of China (No. 60832003, 60902096, 61171163, 61071120)the Scientific Research Foundation of Graduate School of Ningbo University
文摘Most of Image Quality Assessment (IQA) metrics consist of two processes. In the first process, quality map of image is measured locally. In the second process, the last quality score is converted from the quality map by using the pooling strategy. The first process had been made effective and significant progresses, while the second process was always done in simple ways. In the second process of the pooling strategy, the optimal perceptual pooling weights should be determined and computed according to Human Visual System (HVS). Thus, a reliable spatial pooling mathematical model based on HVS is an important issue worthy of study. In this paper, a new Visual Perceptual Pooling Strategy (VPPS) for IQA is presented based on contrast sensitivity and luminance sensitivity of HVS. Experimental results with the LIVE database show that the visual perceptual weights, obtained by the proposed pooling strategy, can effectively and significantly improve the performances of the IQA metrics with Mean Structural SIMilarity (MSSIM) or Phase Quantization Code (PQC). It is confirmed that the proposed VPPS demonstrates promising results for improving the performances of existing IQA metrics.
基金Supported by the National Natural Science Foundation of China (No. 60972039)Jiangsu Province Natural Science Fund Project (BK2010077)Innovation Project of SCI & Tech for College Graduates of Jiangsu Province(CXLX12 _0475)
文摘Based on compressive sampling transmission model, we demonstrate here a method of quality evaluation for the reconstruction images, which is promising for the transmission of unstructured signal with reduced dimension. By this method, the auxiliary information of the recovery image quality is obtained as a feedback to control number of measurements from compressive sampling video stream. Therefore, the number of measurements can be easily derived at the condition of the absence of information sparsity, and the recovery image quality is effectively improved. Theoretical and experimental results show that this algorithm can estimate the quality of images effectively and is in well consistency with the traditional objective evaluation algorithm.
基金supported in part by China"973"Program under Grant No.2014CB340303
文摘In this paper, we propose a novel image recompression frame- work and image quality assessment (IQA) method to efficiently recompress Internet images. With this framework image size is significantly reduced without affecting spatial resolution or perceptible quality of the image. With the help of IQA, the relationship between image quality and image evaluation scores can be quickly established, and the optimal quality factor can be obtained quickly and accurately within a pre - determined perceptual quality range. This process ensures the image's perceptual quality, which is applied to each input image. The test results show that, using the proposed method, the file size of images can be reduced by about 45%-60% without affecting their visual quality. Moreover, our new image -reeompression framework can be used in to many different application scenarios.
文摘Traditional image quality assessment methods use the hand-crafted features to predict the image quality score,which cannot perform well in many scenes.Since deep learning promotes the development of many computer vision tasks,many IQA methods start to utilize the deep convolutional neural networks(CNN)for IQA task.In this paper,a CNN-based multi-scale blind image quality predictor is proposed to extract more effectivity multi-scale distortion features through the pyramidal convolution,which consists of two tasks:A distortion recognition task and a quality regression task.For the first task,image distortion type is obtained by the fully connected layer.For the second task,the image quality score is predicted during the distortion recognition progress.Experimental results on three famous IQA datasets show that the proposed method has better performance than the previous traditional algorithms for quality prediction and distortion recognition.
基金supported by the National Basic Research Program of China(Grant No.2013CB329202)the Basic Industrial Technology Project of China(Grant No.J312012B002)
文摘Satellite vibrations during exposure will lead to pixel aliasing of remote sensors, resulting in the deterioration of image quality. In this paper, we expose the problem and discuss the characteristics of satellite vibrations, and then present a pixel mixing model. The idea of mean mixing ratio (MMR) is proposed. MMR computations for different frequencies are implemented. In the mixing model, a coefficient matrix is introduced to estimate each mixed pixel. Thus, the simulation of degraded image can be performed when the vibration attitudes are known. The computation of MMR takes into considera- tion the influences of various frequencies and amplitudes. Therefore, the roles of these parameters played in the degradation progress are identified. Computations show that under the same vibration amplitude, the influence of vibrations fluctuates with the variation of frequency. The fluctuation becomes smaller as the frequency rises. Two kinds of vibration imaging experiments are performed: different amplitudes with the same frequency and different frequencies with the same amplitude. Results are found to be in very good agreement with the theoretical results. MMR has a better description of image quality than modulation transfer function (MTF). The influence of vibrations is determined mainly by the amplitude rather than the frequency. The influence of vibrations on image quality becomes gradually stable with the increase of frequency.
基金This work was supported by Taif university Researchers Supporting Project Number(TURSP-2020/114),Taif University,Taif,Saudi Arabia.
文摘Low contrast of Magnetic Resonance(MR)images limits the visibility of subtle structures and adversely affects the outcome of both subjective and automated diagnosis.State-of-the-art contrast boosting techniques intolerably alter inherent features of MR images.Drastic changes in brightness features,induced by post-processing are not appreciated in medical imaging as the grey level values have certain diagnostic meanings.To overcome these issues this paper proposes an algorithm that enhance the contrast of MR images while preserving the underlying features as well.This method termed as Power-law and Logarithmic Modification-based Histogram Equalization(PLMHE)partitions the histogram of the image into two sub histograms after a power-law transformation and a log compression.After a modification intended for improving the dispersion of the sub-histograms and subsequent normalization,cumulative histograms are computed.Enhanced grey level values are computed from the resultant cumulative histograms.The performance of the PLMHE algorithm is comparedwith traditional histogram equalization based algorithms and it has been observed from the results that PLMHE can boost the image contrast without causing dynamic range compression,a significant change in mean brightness,and contrast-overshoot.
基金supported by the Key Research and Development Program of Zhejiang Province(Grant No.2019C03138 and No.2019C01002)。
文摘The bandwidth of internet connections is still a bottleneck when transmitting large amounts of images,making the image quality assessment essential.Neurophysiological assessment of image quality has highlight advantages for it does not interfere with natural viewing behavior.However,in JPEG compression,the previous study is hard to tell the difference between the electroencephalogram(EEG)evoked by different quality images.In this paper,we propose an EEG analysis approach based on algebraic topology analysis,and the result shows that the difference between Euler characteristics of EEG evoked by different distortion images is striking both in the alpha and beta band.Moreover,we further discuss the relationship between the images and the EEG signals,and the results implied that the algebraic topological properties of images are consistent with that of brain perception,which is possible to give birth to braininspired image compression based on algebraic topological features.In general,an algebraic topologybased approach was proposed in this paper to analyze the perceptual characteristics of image quality,which will be beneficial to provide a reliable score for data compression in the network and improve the network transmission capacity.
文摘Through research for image compression based on wavelet analysis in recent years, we put forward an adaptive wavelet decomposition strategy. Whether sub-images are to be decomposed or not are decided by their energy defined by certain criterion. Then we derive the adaptive wavelet decomposition tree (AWDT) and the way of adjustable compression ratio. According to the feature of AWDT, this paper also deals with the strategies which are used to handle different sub-images in the procedure of quantification and coding of the wavelet coefficients. Through experiments, not only the algorithm in the paper can adapt to various images, but also the quality of recovered image is improved though compression ratio is higher and adjustable. When their compression ratios are near, the quality of subjective vision and PSNR of the algorithm are better than those of JPEG algorithm.
基金This work was supported by JSPS KAKENHI,No.18 K15563.
文摘To minimize radiation risk,dose reduction is important in the diagnostic and therapeutic applications of computed tomography(CT).However,image noise degrades image quality owing to the reduced X-ray dose and a possible unacceptably reduced diagnostic performance.Deep learning approaches with convolutional neural networks(CNNs)have been proposed for natural image denoising;however,these approaches might introduce image blurring or loss of original gradients.The aim of this study was to compare the dose-dependent properties of a CNN-based denoising method for low-dose CT with those of other noise-reduction methods on unique CT noise-simulation images.To simulate a low-dose CT image,a Poisson noise distribution was introduced to normal-dose images while convoluting the CT unit-specific modulation transfer function.An abdominal CT of 100 images obtained from a public database was adopted,and simulated dose-reduction images were created from the original dose at equal 10-step dose-reduction intervals with a final dose of 1/100.These images were denoised using the denoising network structure of CNN(DnCNN)as the general CNN model and for transfer learning.To evaluate the image quality,image similarities determined by the structural similarity index(SSIM)and peak signal-to-noise ratio(PSNR)were calculated for the denoised images.Significantly better denoising,in terms of SSIM and PSNR,was achieved by the DnCNN than by other image denoising methods,especially at the ultra-low-dose levels used to generate the 10%and 5%dose-equivalent images.Moreover,the developed CNN model can eliminate noise and maintain image sharpness at these dose levels and improve SSIM by approximately 10%from that of the original method.In contrast,under small dose-reduction conditions,this model also led to excessive smoothing of the images.In quantitative evaluations,the CNN denoising method improved the low-dose CT and prevented over-smoothing by tailoring the CNN model.
文摘By embedding secret data into cover images,image steganography can produce non-discriminable stego-images.The turtle shell model for data hiding is an excellent method that uses a reference matrix to make a good balance between image quality and embedding capacity.However,increasing the embedding capacity by extending the area of basic structures of the turtle shell model usually leads to severe degradation of image quality.In this research,we innovatively extend the basic structure of the turtle shell model into a three-dimensional(3D)space.Some intrinsic properties of the original turtle shell model are well preserved in the 3D version.Theoretic analysis shows that the new proposed models have good performance both in the image quality and in the complexity of the reference matrix.Our experimental results justify the theoretic conclusions.
基金supported by the National Natural Science Foundation of China(No.41801291)。
文摘The evaluation of geometric calibration accuracy of high resolution satellite images has been increasingly recognized in recent years.In order to evaluate geometric accuracy for dual-camera satellite images based on the ground control points(GCP),a rigorous geometric imaging model,which was based on the collinear equation of the probe directional angle and the optimized tri-axial attitude determination(TRIAD)algorithm,is presented.Two reliable test fields in Tianjin and Jinan(China)were utilized for geometric accuracy validation of Pakistan Remote Sensing Satellite-1.The experimental results demonstrate a certain deviation of the on-orbit calibration result from the initial design values of the calibration parameters.Therefore,on-orbit geometric calibration is necessary for optical satellite imagery.Within this research,the geometrical performances including positioning accuracy without/with GCP and band registration of the dual-camera satellite were analyzed in detail,and the results of geometric image quality are assessed and discussed.As a result,it is feasible and necessary to establish such a geometric calibration model to evaluate the geometric quality of dual-camera satellite.
文摘A new method for no-reference image quality assessment based on hybrid fuzzy-genetic technique is proposed. Noise variance and edge sharpness level of the restored image are two basic metrics for assessing the performance of the restoration algorithm, then a fuzzy if-then inference system is developed to combine the two metrics to get a final quality score, and the parameters of the fuzzy membership function are trained with genetic algorithms. Experiments results show that the image quality score correlates well with mean opinion score and the proposed approach is robust and effective.
文摘Invoice document digitization is crucial for efficient management in industries.The scanned invoice image is often noisy due to various reasons.This affects the OCR(optical character recognition)detection accuracy.In this paper,letter data obtained from images of invoices are denoised using a modified autoencoder based deep learning method.A stacked denoising autoencoder(SDAE)is implemented with two hidden layers each in encoder network and decoder network.In order to capture the most salient features of training samples,a undercomplete autoencoder is designed with non-linear encoder and decoder function.This autoencoder is regularized for denoising application using a combined loss function which considers both mean square error and binary cross entropy.A dataset consisting of 59,119 letter images,which contains both English alphabets(upper and lower case)and numbers(0 to 9)is prepared from many scanned invoices images and windows true type(.ttf)files,are used for training the neural network.Performance is analyzed in terms of Signal to Noise Ratio(SNR),Peak Signal to Noise Ratio(PSNR),Structural Similarity Index(SSIM)and Universal Image Quality Index(UQI)and compared with other filtering techniques like Nonlocal Means filter,Anisotropic diffusion filter,Gaussian filters and Mean filters.Denoising performance of proposed SDAE is compared with existing SDAE with single loss function in terms of SNR and PSNR values.Results show the superior performance of proposed SDAE method.
基金Supported by the National Natural Science Foundation of China(Nos.6117116361271270+2 种基金6127102161111140392)National Science and Technology Support Program(2012BAH67F01)
文摘In order to establish a stereoscopic image quality assessment method which is consistent with human visual perception,we propose an objective stereoscopic image quality assessment method.It takes into account the strong correlation and high degree of structural between pixels of image.This method contains two models.One is the quality synthetic assessment of left-right view images,which is based on human visual characteristics,we use the Singular Value Decomposition(SVD)that can represent the degree of the distortion,and combine the qualities of left and right images by the characteristics of binocular superposition.The other model is stereoscopic perception quality assessment,due to strong stability of image’s singular value characteristics,we calculate the distance of the singular values and structural characteristic similarity of the absolute difference maps,and utilize the statistical value of the global error to evaluate stereoscopic perception.Finally,we combine two models to describe the stereoscopic image quality.Experimental results show that the correlation coefficients of the proposed assessment method and the human subjective perception are above 0.93,and the mean square errors are all less than 6.2,under JPEG,JP2K compression,Gaussian blurring,Gaussian white noise,H.264 coding distortion,and hybrid cross distortion.It indicates that the proposed stereoscopic objective method is consistent with human visual properties and also of availability.