Single Image Super-Resolution(SISR)technology aims to reconstruct a clear,high-resolution image with more information from an input low-resolution image that is blurry and contains less information.This technology has...Single Image Super-Resolution(SISR)technology aims to reconstruct a clear,high-resolution image with more information from an input low-resolution image that is blurry and contains less information.This technology has significant research value and is widely used in fields such as medical imaging,satellite image processing,and security surveillance.Despite significant progress in existing research,challenges remain in reconstructing clear and complex texture details,with issues such as edge blurring and artifacts still present.The visual perception effect still needs further enhancement.Therefore,this study proposes a Pyramid Separable Channel Attention Network(PSCAN)for the SISR task.Thismethod designs a convolutional backbone network composed of Pyramid Separable Channel Attention blocks to effectively extract and fuse multi-scale features.This expands the model’s receptive field,reduces resolution loss,and enhances the model’s ability to reconstruct texture details.Additionally,an innovative artifact loss function is designed to better distinguish between artifacts and real edge details,reducing artifacts in the reconstructed images.We conducted comprehensive ablation and comparative experiments on the Arabidopsis root image dataset and several public datasets.The experimental results show that the proposed PSCAN method achieves the best-known performance in both subjective visual effects and objective evaluation metrics,with improvements of 0.84 in Peak Signal-to-Noise Ratio(PSNR)and 0.017 in Structural Similarity Index(SSIM).This demonstrates that the method can effectively preserve high-frequency texture details,reduce artifacts,and have good generalization performance.展开更多
The numerous photos captured by low-price Internet of Things(IoT)sensors are frequently affected by meteorological factors,especially rainfall.It causes varying sizes of white streaks on the image,destroying the image...The numerous photos captured by low-price Internet of Things(IoT)sensors are frequently affected by meteorological factors,especially rainfall.It causes varying sizes of white streaks on the image,destroying the image texture and ruining the performance of the outdoor computer vision system.Existing methods utilise training with pairs of images,which is difficult to cover all scenes and leads to domain gaps.In addition,the network structures adopt deep learning to map rain images to rain-free images,failing to use prior knowledge effectively.To solve these problems,we introduce a single image derain model in edge computing that combines prior knowledge of rain patterns with the learning capability of the neural network.Specifically,the algorithm first uses Residue Channel Prior to filter out the rainfall textural features then it uses the Feature Fusion Module to fuse the original image with the background feature information.This results in a pre-processed image which is fed into Half Instance Net(HINet)to recover a high-quality rain-free image with a clear and accurate structure,and the model does not rely on any rainfall assumptions.Experimental results on synthetic and real-world datasets show that the average peak signal-to-noise ratio of the model decreases by 0.37 dB on the synthetic dataset and increases by 0.43 dB on the real-world dataset,demonstrating that a combined model reduces the gap between synthetic data and natural rain scenes,improves the generalization ability of the derain network,and alleviates the overfitting problem.展开更多
Visual degradation of captured images caused by rainy streaks under rainy weather can adversely affect the performance of many open-air vision systems.Hence,it is necessary to address the problem of eliminating rain s...Visual degradation of captured images caused by rainy streaks under rainy weather can adversely affect the performance of many open-air vision systems.Hence,it is necessary to address the problem of eliminating rain streaks from the individual rainy image.In this work,a deep convolution neural network(CNN)based method is introduced,called Rain-Removal Net(R2N),to solve the single image de-raining issue.Firstly,we decomposed the rainy image into its high-frequency detail layer and lowfrequency base layer.Then,we used the high-frequency detail layer to input the carefully designed CNN architecture to learn the mapping between it and its corresponding derained high-frequency detail layer.The CNN architecture consists of four convolution layers and four deconvolution layers,as well as three skip connections.The experiments on synthetic and real-world rainy images show that the performance of our architecture outperforms the compared state-of-the-art de-raining models with respects to the quality of de-rained images and computing efficiency.展开更多
Outdoor haze has adverse impact on outdoor image quality,including contrast loss and poor visibility.In this paper,a novel dehazing algorithm based on the decomposition strategy is proposed.It combines the advantages ...Outdoor haze has adverse impact on outdoor image quality,including contrast loss and poor visibility.In this paper,a novel dehazing algorithm based on the decomposition strategy is proposed.It combines the advantages of the two-dimensional variational mode decomposition(2DVMD)algorithm and dark channel prior.The original hazy image is adaptively decom-posed into low-frequency and high-frequency images according to the image frequency band by using the 2DVMD algorithm.The low-frequency image is dehazed by using the improved dark channel prior,and then fused with the high-frequency image.Furthermore,we optimize the atmospheric light and transmit-tance estimation method to obtain a defogging effect with richer details and stronger contrast.The proposed algorithm is com-pared with the existing advanced algorithms.Experiment results show that the proposed algorithm has better performance in comparison with the state-of-the-art algorithms.展开更多
Single image super-resolution(SISR)is a fundamentally challenging problem because a low-resolution(LR)image can correspond to a set of high-resolution(HR)images,while most are not expected.Recently,SISR can be achieve...Single image super-resolution(SISR)is a fundamentally challenging problem because a low-resolution(LR)image can correspond to a set of high-resolution(HR)images,while most are not expected.Recently,SISR can be achieved by a deep learning-based method.By constructing a very deep super-resolution convolutional neural network(VDSRCNN),the LR images can be improved to HR images.This study mainly achieves two objectives:image super-resolution(ISR)and deblurring the image from VDSRCNN.Firstly,by analyzing ISR,we modify different training parameters to test the performance of VDSRCNN.Secondly,we add the motion blurred images to the training set to optimize the performance of VDSRCNN.Finally,we use image quality indexes to evaluate the difference between the images from classical methods and VDSRCNN.The results indicate that the VDSRCNN performs better in generating HR images from LR images using the optimized VDSRCNN in a proper method.展开更多
The employment of deep convolutional neural networks has recently contributed to significant progress in single image super-resolution(SISR)research.However,the high computational demands of most SR techniques hinder ...The employment of deep convolutional neural networks has recently contributed to significant progress in single image super-resolution(SISR)research.However,the high computational demands of most SR techniques hinder their applicability to edge devices,despite their satisfactory reconstruction performance.These methods commonly use standard convolutions,which increase the convolutional operation cost of the model.In this paper,a lightweight Partial Separation and Multiscale Fusion Network(PSMFNet)is proposed to alleviate this problem.Specifically,this paper introduces partial convolution(PConv),which reduces the redundant convolution operations throughout the model by separating some of the features of an image while retaining features useful for image reconstruction.Additionally,it is worth noting that the existing methods have not fully utilized the rich feature information,leading to information loss,which reduces the ability to learn feature representations.Inspired by self-attention,this paper develops a multiscale feature fusion block(MFFB),which can better utilize the non-local features of an image.MFFB can learn long-range dependencies from the spatial dimension and extract features from the channel dimension,thereby obtaining more comprehensive and rich feature information.As the role of the MFFB is to capture rich global features,this paper further introduces an efficient inverted residual block(EIRB)to supplement the local feature extraction ability of PSMFNet.A comprehensive analysis of the experimental results shows that PSMFNet maintains a better performance with fewer parameters than the state-of-the-art models.展开更多
Due to the presence of turbid media, such as microdust and water vapor in the environment, outdoor pictures taken under hazy weather circumstances are typically degraded. To enhance the quality of such images, this wo...Due to the presence of turbid media, such as microdust and water vapor in the environment, outdoor pictures taken under hazy weather circumstances are typically degraded. To enhance the quality of such images, this work proposes a new hybrid λ2-λ0 penalty model for image dehazing. This model performs a weighted fusion of two distinct transmission maps, generated by imposing λ2 and λ0 norm penalties on the approximate regression coefficients of the transmission map. This approach effectively balances the sparsity and smoothness associated with the λ0 and λ2 norms, thereby optimizing the transmittance map. Specifically, when the λ2 norm is penalized in the model, an updated guided image is obtained after implementing λ0 penalty. The resulting optimization problem is effectively solved using the least square method and the alternating direction algorithm. The dehazing framework combines the advantages of λ2 and λ0 norms, enhancing sparse and smoothness, resulting in higher quality images with clearer details and preserved edges.展开更多
In this letter, we present a novel integrated feature that incorporates traditional parameters, and adopt a parallel cascading fashion network Haze Net for enhancing image quality. Our unified feature is a complete in...In this letter, we present a novel integrated feature that incorporates traditional parameters, and adopt a parallel cascading fashion network Haze Net for enhancing image quality. Our unified feature is a complete integration, and its role is to directly describe the effects of haze. In Haze Net, we design two separate structures including backbone and auxiliary networks to extract feature map. Backbone network is responsible for extracting high-level feature map, and low-level feature learned by the auxiliary network can be interpreted as fine-grained feature. After cascading two features with different accuracy, final performance can be effectively improved. Extensive experimental results on both synthetic datasets and real-world images prove the superiority of the proposed method, and demonstrate more favorable performance compared with the existing state-of-art methods.展开更多
We synthesize animations from a single image by transferring fluid motion of a video example globally.Given a target image of a fluid scene,an alpha matte is required to extract the fluid region.Our method needs to ad...We synthesize animations from a single image by transferring fluid motion of a video example globally.Given a target image of a fluid scene,an alpha matte is required to extract the fluid region.Our method needs to adjust a user-specified video example for producing the fluid motion suitable for the extracted fluid region.Employing the fluid video database,the flow field of the target image is obtained by warping the optical flow of a video frame that has a visually similar scene to the target image according to their scene correspondences,which assigns fluid orientation and speed automatically.Results show that our method is successful in preserving large fluid features in the synthesized animations.In comparison to existing approaches,it is both possible and useful to utilize our method to create flow animations with higher quality.展开更多
There is a steadily growing range of applications that can benefit from facial reconstruction techniques,leading to an increasing demand for reconstruction of high-quality 3D face models.While it is an important expre...There is a steadily growing range of applications that can benefit from facial reconstruction techniques,leading to an increasing demand for reconstruction of high-quality 3D face models.While it is an important expressive part of the human face,the nose has received less attention than other expressive regions in the face reconstruction literature.When applying existing reconstruction methods to facial images,the reconstructed nose models are often inconsistent with the desired shape and expression.In this paper,we propose a coarse-to-fine 3D nose reconstruction and correction pipeline to build a nose model from a single image,where 3D and 2D nose curve correspondences are adaptively updated and refined.We first correct the reconstruction result coarsely using constraints of 3D-2D sparse landmark correspondences,and then heuristically update a dense 3D-2D curve correspondence based on the coarsely corrected result.A final refinement step is performed to correct the shape based on the updated 3D-2D dense curve constraints.Experimental results show the advantages of our method for 3D nose reconstruction over existing methods.展开更多
The quality of photos is highly susceptible to severe weather such as heavy rain;it can also degrade the performance of various visual tasks like object detection.Rain removal is a challenging problem because rain str...The quality of photos is highly susceptible to severe weather such as heavy rain;it can also degrade the performance of various visual tasks like object detection.Rain removal is a challenging problem because rain streaks have different appearances even in one image.Regions where rain accumulates appear foggy or misty,while rain streaks can be clearly seen in areas where rain is less heavy.We propose removing various rain effects in pictures using a hybrid multiscale loss guided multiple feature fusion de-raining network(MSGMFFNet).Specially,to deal with rain streaks,our method generates a rain streak attention map,while preprocessing uses gamma correction and contrast enhancement to enhanced images to address the problem of rain accumulation.Using these tools,the model can restore a result with abundant details.Furthermore,a hybrid multiscale loss combining L1 loss and edge loss is used to guide the training process to pay attention to edge and content information.Comprehensive experiments conducted on both synthetic and real-world datasets demonstrate the effectiveness of our method.展开更多
An improved single image dehazing method based on dark channel prior and wavelet transform is proposed. This proposed method employs wavelet transform and guided filter instead of the soft matting procedure to estimat...An improved single image dehazing method based on dark channel prior and wavelet transform is proposed. This proposed method employs wavelet transform and guided filter instead of the soft matting procedure to estimate and refine the depth map of haze images. Moreover, a contrast enhancement method based on just noticeable difference(JND) and quadratic function is adopted to enhance the contrast for the dehazed image, since the scene radiance is usually not as bright as the atmospheric light,and the dehazed image looks dim. The experimental results show that the proposed approach can effectively enhance the haze image and is well suitable for implementing on the surveillance and obstacle detection systems.展开更多
Cities are in constant change and city managers aim to keep an updated digital model of the city for city governance. There are a lot of images uploaded daily on image sharing platforms (as “Flickr”, “Twitter”, et...Cities are in constant change and city managers aim to keep an updated digital model of the city for city governance. There are a lot of images uploaded daily on image sharing platforms (as “Flickr”, “Twitter”, etc.). These images feature a rough localization and no orientation information. Nevertheless, they can help to populate an active collaborative database of street images usable to maintain a city 3D model, but their localization and orientation need to be known. Based on these images, we propose the Data Gathering system for image Pose Estimation (DGPE) that helps to find the pose (position and orientation) of the camera used to shoot them with better accuracy than the sole GPS localization that may be embedded in the image header. DGPE uses both visual and semantic information, existing in a single image processed by a fully automatic chain composed of three main layers: Data retrieval and preprocessing layer, Features extraction layer, Decision Making layer. In this article, we present the whole system details and compare its detection results with a state of the art method. Finally, we show the obtained localization, and often orientation results, combining both semantic and visual information processing on 47 images. Our multilayer system succeeds in 26% of our test cases in finding a better localization and orientation of the original photo. This is achieved by using only the image content and associated metadata. The use of semantic information found on social media such as comments, hash tags, etc. has doubled the success rate to 59%. It has reduced the search area and thus made the visual search more accurate.展开更多
Dissolution kinetics of sodium carbonate is investigated with the image analysis method at the approach of single particle.The dissolution experiments are carried out in an aqueous solution under a series of controlle...Dissolution kinetics of sodium carbonate is investigated with the image analysis method at the approach of single particle.The dissolution experiments are carried out in an aqueous solution under a series of controlled temperature and p H.The selected sodium carbonate particles are all spherical with the same mass and diameter.The dissolution process is quantified with the measurement of particle diameter from dissolution images.The concentration of dissolved sodium carbonate in solvent is calculated with the measured diameter of particle.Both surface reaction model and mass transport model are implemented to determine the dissolution mechanism and quantify the dissolution rate constant at each experimental condition.According to the fitting results with both two models,it is clarified that the dissolution process at the increasing temperature is controlled by the mass transport of dissolved sodium carbonate travelling from particle surface into solvent.The dissolution process at the increasing pH is controlled by the chemical reaction on particle surface.Furthermore,the dissolution rate constant for each single spherical sodium carbonate particle is quantified and the results show that the dissolution rate constant of single spherical sodium carbonate increases significantly with the rising of temperature,but decreases with the increasing of pH conversely.展开更多
The thermal front in the oceanic system is believed to have a significant effect on biological activity.During an era of climate change,changes in heat regulation between the atmosphere and oceanic interior can alter ...The thermal front in the oceanic system is believed to have a significant effect on biological activity.During an era of climate change,changes in heat regulation between the atmosphere and oceanic interior can alter the characteristics of this important feature.Using the simulation results of the 3D Regional Ocean Modelling System(ROMS),we identified the location of thermal fronts and determined their dynamic variability in the area between the southern Andaman Sea and northern Malacca Strait.The Single Image Edge Detection(SIED)algorithm was used to detect the thermal front from model-derived temperature.Results show that a thermal front occurred every year from 2002 to 2012 with the temperature gradient at the location of the front was 0.3°C/km.Compared to the years affected by El Ni?o and negative Indian Ocean Dipole(IOD),the normal years(e.g.,May 2003)show the presence of the thermal front at every selected depth(10,25,50,and 75 m),whereas El Ni?o and negative IOD during 2010 show the presence of the thermal front only at depth of 75 m due to greater warming,leading to the thermocline deepening and enhanced stratification.During May 2003,the thermal front was separated by cooler SST in the southern Andaman Sea and warmer SST in the northern Malacca Strait.The higher SST in the northern Malacca Strait was believed due to the besieged Malacca Strait,which trapped the heat and make it difficult to release while higher chlorophyll a in Malacca Strait is due to the freshwater conduit from nearby rivers(Klang,Langat,Perak,and Selangor).Furthermore,compared to the southern Andaman Sea,the chlorophyll a in the northern Malacca Strait is easier to reach the surface area due to the shallower thermocline,which allows nutrients in the area to reach the surface faster.展开更多
Single image super resolution(SISR)is an important research content in the field of computer vision and image processing.With the rapid development of deep neural networks,different image super-resolution models have ...Single image super resolution(SISR)is an important research content in the field of computer vision and image processing.With the rapid development of deep neural networks,different image super-resolution models have emerged.Compared to some traditional SISR methods,deep learning-based methods can complete the super-resolution tasks through a single image.In addition,compared with the SISR methods using traditional convolutional neural networks,SISR based on generative adversarial networks(GAN)has achieved the most advanced visual performance.In this review,we first explore the challenges faced by SISR and introduce some common datasets and evaluation metrics.Then,we review the improved network structures and loss functions of GAN-based perceptual SISR.Subsequently,the advantages and disadvantages of different networks are analyzed by multiple comparative experiments.Finally,we summarize the paper and look forward to the future development trends of GAN-based perceptual SISR.展开更多
Sparse representation has attracted extensive attention and performed well on image super-resolution(SR) in the last decade. However, many current image SR methods face the contradiction of detail recovery and artif...Sparse representation has attracted extensive attention and performed well on image super-resolution(SR) in the last decade. However, many current image SR methods face the contradiction of detail recovery and artifact suppression. We propose a multi-resolution dictionary learning(MRDL) model to solve this contradiction, and give a fast single image SR method based on the MRDL model. To obtain the MRDL model, we first extract multi-scale patches by using our proposed adaptive patch partition method(APPM). The APPM divides images into patches of different sizes according to their detail richness. Then, the multiresolution dictionary pairs, which contain structural primitives of various resolutions, can be trained from these multi-scale patches.Owing to the MRDL strategy, our SR algorithm not only recovers details well, with less jag and noise, but also significantly improves the computational efficiency. Experimental results validate that our algorithm performs better than other SR methods in evaluation metrics and visual perception.展开更多
Although there has been a great breakthrough in the accuracy and speed of super-resolution(SR)reconstruction of a single image by using a convolutional neural network,an important problem remains unresolved:how to res...Although there has been a great breakthrough in the accuracy and speed of super-resolution(SR)reconstruction of a single image by using a convolutional neural network,an important problem remains unresolved:how to restore finer texture details during image super-resolution reconstruction?This paper proposes an Enhanced Laplacian Pyramid Generative Adversarial Network(ELSRGAN),based on the Laplacian pyramid to capture the high-frequency details of the image.By combining Laplacian pyramids and generative adversarial networks,progressive reconstruction of super-resolution images can be made,making model applications more flexible.In order to solve the problem of gradient disappearance,we introduce the Residual-in-Residual Dense Block(RRDB)as the basic network unit.Network capacity benefits more from dense connections,is able to capture more visual features with better reconstruction effects,and removes BN layers to increase calculation speed and reduce calculation complexity.In addition,a loss of content driven by perceived similarity is used instead of content loss driven by spatial similarity,thereby enhancing the visual effect of the super-resolution image,making it more consistent with human visual perception.Extensive qualitative and quantitative evaluation of the baseline datasets shows that the proposed algorithm has higher mean-sort-score(MSS)than any state-of-the-art method and has better visual perception.展开更多
Background This study presents a neural hand reconstruction method for monocular 3D hand pose and shape estimation.Methods Alternate to directly representing hand with 3D data,a novel UV position map is used to repres...Background This study presents a neural hand reconstruction method for monocular 3D hand pose and shape estimation.Methods Alternate to directly representing hand with 3D data,a novel UV position map is used to represent a hand pose and shape with 2D data that maps 3D hand surface points to 2D image space.Furthermore,an encoder-decoder neural network is proposed to infer such UV position map from a single image.To train this network with inadequate ground truth training pairs,we propose a novel MANOReg module that employs MANO model as a prior shape to constrain high dimensional space of the UV position map.Results The quantitative and qualitative experiments demonstrate the effectiveness of our UV position map representation and MANOReg module.展开更多
Obtaining clear images of underwater scenes with descriptive details is an arduous task.Conventional imaging techniques fail to provide clear cut features and attributes that ultimately result in object recognition er...Obtaining clear images of underwater scenes with descriptive details is an arduous task.Conventional imaging techniques fail to provide clear cut features and attributes that ultimately result in object recognition errors.Consequently,a need for a system that produces clear images for underwater image study has been necessitated.To overcome problems in resolution and to make better use of the Super-Resolution(SR)method,this paper introduces a novel method that has been derived from the Alpha Generative Adversarial Network(AlphaGAN)model,named Alpha Super Resolution Generative Adversarial Network(AlphaSRGAN).The model put forth in this paper helps in enhancing the quality of underwater imagery and yields images with greater resolution and more concise details.Images undergo pre-processing before they are fed into a generator network that optimizes and reforms the structure of the network while enhancing the stability of the network that acts as the generator.After the images are processed by the generator network,they are passed through an adversarial method for training models.The dataset used in this paper to learn Single Image Super Resolution(SISR)is the USR 248 dataset.Training supervision is performed by an unprejudiced function that simultaneously scrutinizes and improves the image quality.Appraisal of images is done with reference to factors like local style information,global content and color.The dataset USR 248 which has a huge collection of images has been used for the study is composed of three collections of images—high(640×480)and low(80×60,160×120,and 320×240).Paired instances of different sizes—2×,4×and 8×—are also present in the dataset.Parameters like Mean Opinion Score(MOS),Peak Signal-to-Noise Ratio(PSNR),Structural Similarity(SSIM)and Underwater Image Quality Measure(UIQM)scores have been compared to validate the improved efficiency of our model when compared to existing works.展开更多
基金supported by Beijing Municipal Science and Technology Project(No.Z221100007122003).
文摘Single Image Super-Resolution(SISR)technology aims to reconstruct a clear,high-resolution image with more information from an input low-resolution image that is blurry and contains less information.This technology has significant research value and is widely used in fields such as medical imaging,satellite image processing,and security surveillance.Despite significant progress in existing research,challenges remain in reconstructing clear and complex texture details,with issues such as edge blurring and artifacts still present.The visual perception effect still needs further enhancement.Therefore,this study proposes a Pyramid Separable Channel Attention Network(PSCAN)for the SISR task.Thismethod designs a convolutional backbone network composed of Pyramid Separable Channel Attention blocks to effectively extract and fuse multi-scale features.This expands the model’s receptive field,reduces resolution loss,and enhances the model’s ability to reconstruct texture details.Additionally,an innovative artifact loss function is designed to better distinguish between artifacts and real edge details,reducing artifacts in the reconstructed images.We conducted comprehensive ablation and comparative experiments on the Arabidopsis root image dataset and several public datasets.The experimental results show that the proposed PSCAN method achieves the best-known performance in both subjective visual effects and objective evaluation metrics,with improvements of 0.84 in Peak Signal-to-Noise Ratio(PSNR)and 0.017 in Structural Similarity Index(SSIM).This demonstrates that the method can effectively preserve high-frequency texture details,reduce artifacts,and have good generalization performance.
基金supported by the National Natural Science Foundation of China under Grant no.41975183,and Grant no.41875184 and Supported by a grant from State Key Laboratory of Resources and Environmental Information System.
文摘The numerous photos captured by low-price Internet of Things(IoT)sensors are frequently affected by meteorological factors,especially rainfall.It causes varying sizes of white streaks on the image,destroying the image texture and ruining the performance of the outdoor computer vision system.Existing methods utilise training with pairs of images,which is difficult to cover all scenes and leads to domain gaps.In addition,the network structures adopt deep learning to map rain images to rain-free images,failing to use prior knowledge effectively.To solve these problems,we introduce a single image derain model in edge computing that combines prior knowledge of rain patterns with the learning capability of the neural network.Specifically,the algorithm first uses Residue Channel Prior to filter out the rainfall textural features then it uses the Feature Fusion Module to fuse the original image with the background feature information.This results in a pre-processed image which is fed into Half Instance Net(HINet)to recover a high-quality rain-free image with a clear and accurate structure,and the model does not rely on any rainfall assumptions.Experimental results on synthetic and real-world datasets show that the average peak signal-to-noise ratio of the model decreases by 0.37 dB on the synthetic dataset and increases by 0.43 dB on the real-world dataset,demonstrating that a combined model reduces the gap between synthetic data and natural rain scenes,improves the generalization ability of the derain network,and alleviates the overfitting problem.
基金This work was supported by the National Natural Science Foundation of China(Grant No.61673222)Jiangsu Universities Natural Science Research Project(Grant No.13KJA510001)Major Program of the National Social Science Fund of China(Grant No.17ZDA092).
文摘Visual degradation of captured images caused by rainy streaks under rainy weather can adversely affect the performance of many open-air vision systems.Hence,it is necessary to address the problem of eliminating rain streaks from the individual rainy image.In this work,a deep convolution neural network(CNN)based method is introduced,called Rain-Removal Net(R2N),to solve the single image de-raining issue.Firstly,we decomposed the rainy image into its high-frequency detail layer and lowfrequency base layer.Then,we used the high-frequency detail layer to input the carefully designed CNN architecture to learn the mapping between it and its corresponding derained high-frequency detail layer.The CNN architecture consists of four convolution layers and four deconvolution layers,as well as three skip connections.The experiments on synthetic and real-world rainy images show that the performance of our architecture outperforms the compared state-of-the-art de-raining models with respects to the quality of de-rained images and computing efficiency.
基金supported by the National Defense Technology Advance Research Project of China(004040204).
文摘Outdoor haze has adverse impact on outdoor image quality,including contrast loss and poor visibility.In this paper,a novel dehazing algorithm based on the decomposition strategy is proposed.It combines the advantages of the two-dimensional variational mode decomposition(2DVMD)algorithm and dark channel prior.The original hazy image is adaptively decom-posed into low-frequency and high-frequency images according to the image frequency band by using the 2DVMD algorithm.The low-frequency image is dehazed by using the improved dark channel prior,and then fused with the high-frequency image.Furthermore,we optimize the atmospheric light and transmit-tance estimation method to obtain a defogging effect with richer details and stronger contrast.The proposed algorithm is com-pared with the existing advanced algorithms.Experiment results show that the proposed algorithm has better performance in comparison with the state-of-the-art algorithms.
文摘Single image super-resolution(SISR)is a fundamentally challenging problem because a low-resolution(LR)image can correspond to a set of high-resolution(HR)images,while most are not expected.Recently,SISR can be achieved by a deep learning-based method.By constructing a very deep super-resolution convolutional neural network(VDSRCNN),the LR images can be improved to HR images.This study mainly achieves two objectives:image super-resolution(ISR)and deblurring the image from VDSRCNN.Firstly,by analyzing ISR,we modify different training parameters to test the performance of VDSRCNN.Secondly,we add the motion blurred images to the training set to optimize the performance of VDSRCNN.Finally,we use image quality indexes to evaluate the difference between the images from classical methods and VDSRCNN.The results indicate that the VDSRCNN performs better in generating HR images from LR images using the optimized VDSRCNN in a proper method.
基金Guangdong Science and Technology Program under Grant No.202206010052Foshan Province R&D Key Project under Grant No.2020001006827Guangdong Academy of Sciences Integrated Industry Technology Innovation Center Action Special Project under Grant No.2022GDASZH-2022010108.
文摘The employment of deep convolutional neural networks has recently contributed to significant progress in single image super-resolution(SISR)research.However,the high computational demands of most SR techniques hinder their applicability to edge devices,despite their satisfactory reconstruction performance.These methods commonly use standard convolutions,which increase the convolutional operation cost of the model.In this paper,a lightweight Partial Separation and Multiscale Fusion Network(PSMFNet)is proposed to alleviate this problem.Specifically,this paper introduces partial convolution(PConv),which reduces the redundant convolution operations throughout the model by separating some of the features of an image while retaining features useful for image reconstruction.Additionally,it is worth noting that the existing methods have not fully utilized the rich feature information,leading to information loss,which reduces the ability to learn feature representations.Inspired by self-attention,this paper develops a multiscale feature fusion block(MFFB),which can better utilize the non-local features of an image.MFFB can learn long-range dependencies from the spatial dimension and extract features from the channel dimension,thereby obtaining more comprehensive and rich feature information.As the role of the MFFB is to capture rich global features,this paper further introduces an efficient inverted residual block(EIRB)to supplement the local feature extraction ability of PSMFNet.A comprehensive analysis of the experimental results shows that PSMFNet maintains a better performance with fewer parameters than the state-of-the-art models.
文摘Due to the presence of turbid media, such as microdust and water vapor in the environment, outdoor pictures taken under hazy weather circumstances are typically degraded. To enhance the quality of such images, this work proposes a new hybrid λ2-λ0 penalty model for image dehazing. This model performs a weighted fusion of two distinct transmission maps, generated by imposing λ2 and λ0 norm penalties on the approximate regression coefficients of the transmission map. This approach effectively balances the sparsity and smoothness associated with the λ0 and λ2 norms, thereby optimizing the transmittance map. Specifically, when the λ2 norm is penalized in the model, an updated guided image is obtained after implementing λ0 penalty. The resulting optimization problem is effectively solved using the least square method and the alternating direction algorithm. The dehazing framework combines the advantages of λ2 and λ0 norms, enhancing sparse and smoothness, resulting in higher quality images with clearer details and preserved edges.
基金supported by the National Natural Science Foundation of China (No.61561030)the Gansu Provincial F inance Department (No.214138)。
文摘In this letter, we present a novel integrated feature that incorporates traditional parameters, and adopt a parallel cascading fashion network Haze Net for enhancing image quality. Our unified feature is a complete integration, and its role is to directly describe the effects of haze. In Haze Net, we design two separate structures including backbone and auxiliary networks to extract feature map. Backbone network is responsible for extracting high-level feature map, and low-level feature learned by the auxiliary network can be interpreted as fine-grained feature. After cascading two features with different accuracy, final performance can be effectively improved. Extensive experimental results on both synthetic datasets and real-world images prove the superiority of the proposed method, and demonstrate more favorable performance compared with the existing state-of-art methods.
基金Project supported by the National Basic Research Program (973) of China (No.2011CB302203)the Innovation Program of the Science and Technology Commission of Shanghai Municipality,China (No.10511501200)
文摘We synthesize animations from a single image by transferring fluid motion of a video example globally.Given a target image of a fluid scene,an alpha matte is required to extract the fluid region.Our method needs to adjust a user-specified video example for producing the fluid motion suitable for the extracted fluid region.Employing the fluid video database,the flow field of the target image is obtained by warping the optical flow of a video frame that has a visually similar scene to the target image according to their scene correspondences,which assigns fluid orientation and speed automatically.Results show that our method is successful in preserving large fluid features in the synthesized animations.In comparison to existing approaches,it is both possible and useful to utilize our method to create flow animations with higher quality.
基金supported by the National Natural Science Foundation of China(Grant Nos.61972342,61602402,and 61902334)Zhejiang Provincial Basic Public Welfare Research(Grant No.LGG19F020001)+1 种基金Shenzhen Fundamental Research(General Project)(Grant No.JCYJ20190814112007258)the Royal Society(Grant No.IES\R1\180126).
文摘There is a steadily growing range of applications that can benefit from facial reconstruction techniques,leading to an increasing demand for reconstruction of high-quality 3D face models.While it is an important expressive part of the human face,the nose has received less attention than other expressive regions in the face reconstruction literature.When applying existing reconstruction methods to facial images,the reconstructed nose models are often inconsistent with the desired shape and expression.In this paper,we propose a coarse-to-fine 3D nose reconstruction and correction pipeline to build a nose model from a single image,where 3D and 2D nose curve correspondences are adaptively updated and refined.We first correct the reconstruction result coarsely using constraints of 3D-2D sparse landmark correspondences,and then heuristically update a dense 3D-2D curve correspondence based on the coarsely corrected result.A final refinement step is performed to correct the shape based on the updated 3D-2D dense curve constraints.Experimental results show the advantages of our method for 3D nose reconstruction over existing methods.
基金This work was supported in part by the National Key R&D Program of China under No.2017YFB1003000the National Natural Science Foundation of China under No.61872047 and No.61720106007+2 种基金the Beijing Nova Program under No.Z201100006820124the Beijing Natural Science Foundation(L191004)the 111 Project(B18008).
文摘The quality of photos is highly susceptible to severe weather such as heavy rain;it can also degrade the performance of various visual tasks like object detection.Rain removal is a challenging problem because rain streaks have different appearances even in one image.Regions where rain accumulates appear foggy or misty,while rain streaks can be clearly seen in areas where rain is less heavy.We propose removing various rain effects in pictures using a hybrid multiscale loss guided multiple feature fusion de-raining network(MSGMFFNet).Specially,to deal with rain streaks,our method generates a rain streak attention map,while preprocessing uses gamma correction and contrast enhancement to enhanced images to address the problem of rain accumulation.Using these tools,the model can restore a result with abundant details.Furthermore,a hybrid multiscale loss combining L1 loss and edge loss is used to guide the training process to pay attention to edge and content information.Comprehensive experiments conducted on both synthetic and real-world datasets demonstrate the effectiveness of our method.
基金supported by the National Natural Science Foundation of China(61075013)the Joint Funds of the Civil Aviation(61139003)
文摘An improved single image dehazing method based on dark channel prior and wavelet transform is proposed. This proposed method employs wavelet transform and guided filter instead of the soft matting procedure to estimate and refine the depth map of haze images. Moreover, a contrast enhancement method based on just noticeable difference(JND) and quadratic function is adopted to enhance the contrast for the dehazed image, since the scene radiance is usually not as bright as the atmospheric light,and the dehazed image looks dim. The experimental results show that the proposed approach can effectively enhance the haze image and is well suitable for implementing on the surveillance and obstacle detection systems.
文摘Cities are in constant change and city managers aim to keep an updated digital model of the city for city governance. There are a lot of images uploaded daily on image sharing platforms (as “Flickr”, “Twitter”, etc.). These images feature a rough localization and no orientation information. Nevertheless, they can help to populate an active collaborative database of street images usable to maintain a city 3D model, but their localization and orientation need to be known. Based on these images, we propose the Data Gathering system for image Pose Estimation (DGPE) that helps to find the pose (position and orientation) of the camera used to shoot them with better accuracy than the sole GPS localization that may be embedded in the image header. DGPE uses both visual and semantic information, existing in a single image processed by a fully automatic chain composed of three main layers: Data retrieval and preprocessing layer, Features extraction layer, Decision Making layer. In this article, we present the whole system details and compare its detection results with a state of the art method. Finally, we show the obtained localization, and often orientation results, combining both semantic and visual information processing on 47 images. Our multilayer system succeeds in 26% of our test cases in finding a better localization and orientation of the original photo. This is achieved by using only the image content and associated metadata. The use of semantic information found on social media such as comments, hash tags, etc. has doubled the success rate to 59%. It has reduced the search area and thus made the visual search more accurate.
基金the Institute of Particle and Science Engineering,University of Leeds and Procter&Gamble Newcastle Innovation Centre(UK)for partially funding the project
文摘Dissolution kinetics of sodium carbonate is investigated with the image analysis method at the approach of single particle.The dissolution experiments are carried out in an aqueous solution under a series of controlled temperature and p H.The selected sodium carbonate particles are all spherical with the same mass and diameter.The dissolution process is quantified with the measurement of particle diameter from dissolution images.The concentration of dissolved sodium carbonate in solvent is calculated with the measured diameter of particle.Both surface reaction model and mass transport model are implemented to determine the dissolution mechanism and quantify the dissolution rate constant at each experimental condition.According to the fitting results with both two models,it is clarified that the dissolution process at the increasing temperature is controlled by the mass transport of dissolved sodium carbonate travelling from particle surface into solvent.The dissolution process at the increasing pH is controlled by the chemical reaction on particle surface.Furthermore,the dissolution rate constant for each single spherical sodium carbonate particle is quantified and the results show that the dissolution rate constant of single spherical sodium carbonate increases significantly with the rising of temperature,but decreases with the increasing of pH conversely.
基金the Higher Education Ministry research grant,under the Long-Term Research Grant Scheme(No.LRGS/1/2020/UMT/01/1/2)the Universiti Malaysia Terengganu Scholarship(BUMT)。
文摘The thermal front in the oceanic system is believed to have a significant effect on biological activity.During an era of climate change,changes in heat regulation between the atmosphere and oceanic interior can alter the characteristics of this important feature.Using the simulation results of the 3D Regional Ocean Modelling System(ROMS),we identified the location of thermal fronts and determined their dynamic variability in the area between the southern Andaman Sea and northern Malacca Strait.The Single Image Edge Detection(SIED)algorithm was used to detect the thermal front from model-derived temperature.Results show that a thermal front occurred every year from 2002 to 2012 with the temperature gradient at the location of the front was 0.3°C/km.Compared to the years affected by El Ni?o and negative Indian Ocean Dipole(IOD),the normal years(e.g.,May 2003)show the presence of the thermal front at every selected depth(10,25,50,and 75 m),whereas El Ni?o and negative IOD during 2010 show the presence of the thermal front only at depth of 75 m due to greater warming,leading to the thermocline deepening and enhanced stratification.During May 2003,the thermal front was separated by cooler SST in the southern Andaman Sea and warmer SST in the northern Malacca Strait.The higher SST in the northern Malacca Strait was believed due to the besieged Malacca Strait,which trapped the heat and make it difficult to release while higher chlorophyll a in Malacca Strait is due to the freshwater conduit from nearby rivers(Klang,Langat,Perak,and Selangor).Furthermore,compared to the southern Andaman Sea,the chlorophyll a in the northern Malacca Strait is easier to reach the surface area due to the shallower thermocline,which allows nutrients in the area to reach the surface faster.
基金The authors are highly thankful to the Development Research Center of Guangxi Relatively Sparse-populated Minorities(ID:GXRKJSZ201901)to the Natural Science Foundation of Guangxi Province(No.2018GXNSFAA281164)This research was financially supported by the project of outstanding thousand young teachers’training in higher education institutions of Guangxi,Guangxi Colleges and Universities Key Laboratory Breeding Base of System Control and Information Processing.
文摘Single image super resolution(SISR)is an important research content in the field of computer vision and image processing.With the rapid development of deep neural networks,different image super-resolution models have emerged.Compared to some traditional SISR methods,deep learning-based methods can complete the super-resolution tasks through a single image.In addition,compared with the SISR methods using traditional convolutional neural networks,SISR based on generative adversarial networks(GAN)has achieved the most advanced visual performance.In this review,we first explore the challenges faced by SISR and introduce some common datasets and evaluation metrics.Then,we review the improved network structures and loss functions of GAN-based perceptual SISR.Subsequently,the advantages and disadvantages of different networks are analyzed by multiple comparative experiments.Finally,we summarize the paper and look forward to the future development trends of GAN-based perceptual SISR.
文摘Sparse representation has attracted extensive attention and performed well on image super-resolution(SR) in the last decade. However, many current image SR methods face the contradiction of detail recovery and artifact suppression. We propose a multi-resolution dictionary learning(MRDL) model to solve this contradiction, and give a fast single image SR method based on the MRDL model. To obtain the MRDL model, we first extract multi-scale patches by using our proposed adaptive patch partition method(APPM). The APPM divides images into patches of different sizes according to their detail richness. Then, the multiresolution dictionary pairs, which contain structural primitives of various resolutions, can be trained from these multi-scale patches.Owing to the MRDL strategy, our SR algorithm not only recovers details well, with less jag and noise, but also significantly improves the computational efficiency. Experimental results validate that our algorithm performs better than other SR methods in evaluation metrics and visual perception.
基金This work was supported in part by the National Science Foundation of China under Grant 61572526.
文摘Although there has been a great breakthrough in the accuracy and speed of super-resolution(SR)reconstruction of a single image by using a convolutional neural network,an important problem remains unresolved:how to restore finer texture details during image super-resolution reconstruction?This paper proposes an Enhanced Laplacian Pyramid Generative Adversarial Network(ELSRGAN),based on the Laplacian pyramid to capture the high-frequency details of the image.By combining Laplacian pyramids and generative adversarial networks,progressive reconstruction of super-resolution images can be made,making model applications more flexible.In order to solve the problem of gradient disappearance,we introduce the Residual-in-Residual Dense Block(RRDB)as the basic network unit.Network capacity benefits more from dense connections,is able to capture more visual features with better reconstruction effects,and removes BN layers to increase calculation speed and reduce calculation complexity.In addition,a loss of content driven by perceived similarity is used instead of content loss driven by spatial similarity,thereby enhancing the visual effect of the super-resolution image,making it more consistent with human visual perception.Extensive qualitative and quantitative evaluation of the baseline datasets shows that the proposed algorithm has higher mean-sort-score(MSS)than any state-of-the-art method and has better visual perception.
文摘Background This study presents a neural hand reconstruction method for monocular 3D hand pose and shape estimation.Methods Alternate to directly representing hand with 3D data,a novel UV position map is used to represent a hand pose and shape with 2D data that maps 3D hand surface points to 2D image space.Furthermore,an encoder-decoder neural network is proposed to infer such UV position map from a single image.To train this network with inadequate ground truth training pairs,we propose a novel MANOReg module that employs MANO model as a prior shape to constrain high dimensional space of the UV position map.Results The quantitative and qualitative experiments demonstrate the effectiveness of our UV position map representation and MANOReg module.
文摘Obtaining clear images of underwater scenes with descriptive details is an arduous task.Conventional imaging techniques fail to provide clear cut features and attributes that ultimately result in object recognition errors.Consequently,a need for a system that produces clear images for underwater image study has been necessitated.To overcome problems in resolution and to make better use of the Super-Resolution(SR)method,this paper introduces a novel method that has been derived from the Alpha Generative Adversarial Network(AlphaGAN)model,named Alpha Super Resolution Generative Adversarial Network(AlphaSRGAN).The model put forth in this paper helps in enhancing the quality of underwater imagery and yields images with greater resolution and more concise details.Images undergo pre-processing before they are fed into a generator network that optimizes and reforms the structure of the network while enhancing the stability of the network that acts as the generator.After the images are processed by the generator network,they are passed through an adversarial method for training models.The dataset used in this paper to learn Single Image Super Resolution(SISR)is the USR 248 dataset.Training supervision is performed by an unprejudiced function that simultaneously scrutinizes and improves the image quality.Appraisal of images is done with reference to factors like local style information,global content and color.The dataset USR 248 which has a huge collection of images has been used for the study is composed of three collections of images—high(640×480)and low(80×60,160×120,and 320×240).Paired instances of different sizes—2×,4×and 8×—are also present in the dataset.Parameters like Mean Opinion Score(MOS),Peak Signal-to-Noise Ratio(PSNR),Structural Similarity(SSIM)and Underwater Image Quality Measure(UIQM)scores have been compared to validate the improved efficiency of our model when compared to existing works.