Due to improper acquisition settings and other noise artifacts,the image degraded to yield poor mean preservation in brightness.The simplest way to improve the preservation is the implementation of histogram equalizat...Due to improper acquisition settings and other noise artifacts,the image degraded to yield poor mean preservation in brightness.The simplest way to improve the preservation is the implementation of histogram equalization.Because of over-enhancement,it failed to preserve the mean brightness and produce the poor quality of the image.This paper proposes a multi-scale decomposi-tion for brightness preservation using gamma correction.After transformation to hue,saturation and intensity(HSI)channel,the 2D-discrete wavelet transform decomposed the intensity component into low and high-pass coefficients.At the next phase,gamma correction is used by auto-tuning the scale value.The scale is the modified constant value used in the logarithmic function.Further,the scale value is optimized to obtain better visual quality in the image.The optimized value is the weighted distribution of standard deviation-mean of low pass coefficients.Finally,the experimental result is estimated in terms of quality assessment measures used as absolute mean brightness error,the measure of information detail,signal to noise ratio and patch-based contrast quality in the image.By comparison,the proposed method proved to be suitably remarkable in retaining the mean brightness and better visual quality of the image.展开更多
With the development of computer graphics,realistic computer graphics(CG)have become more and more common in our field of vision.This rendered image is invisible to the naked eye.How to effectively identify CG and nat...With the development of computer graphics,realistic computer graphics(CG)have become more and more common in our field of vision.This rendered image is invisible to the naked eye.How to effectively identify CG and natural images(NI)has been become a new issue in the field of digital forensics.In recent years,a series of deep learning network frameworks have shown great advantages in the field of images,which provides a good choice for us to solve this problem.This paper aims to track the latest developments and applications of deep learning in the field of CG and NI forensics in a timely manner.Firstly,it introduces the background of deep learning and the knowledge of convolutional neural networks.The purpose is to understand the basic model structure of deep learning applications in the image field,and then outlines the mainstream framework;secondly,it briefly introduces the application of deep learning in CG and NI forensics,and finally points out the problems of deep learning in this field and the prospects for the future.展开更多
With the recent tremendous advances of computer graphics rendering and image editing technologies,computergenerated fake images,which in general do not reflect what happens in the reality,can now easily deceive the in...With the recent tremendous advances of computer graphics rendering and image editing technologies,computergenerated fake images,which in general do not reflect what happens in the reality,can now easily deceive the inspection of human visual system.In this work,we propose a convolutional neural network(CNN)-based model to distinguish computergenerated(CG)images from natural images(NIs)with channel and pixel correlation.The key component of the proposed CNN architecture is a self-coding module that takes the color images as input to extract the correlation between color channels explicitly.Unlike previous approaches that directly apply CNN to solve this problem,we consider the generality of the network(or subnetwork),i.e.,the newly introduced hybrid correlation module can be directly combined with existing CNN models for enhancing the discrimination capacity of original networks.Experimental results demonstrate that the proposed network outperforms state-of-the-art methods in terms of classification performance.We also show that the newly introduced hybrid correlation module can improve the classification accuracy of different CNN architectures.展开更多
In today’s real world, an important research part in image processing isscene text detection and recognition. Scene text can be in different languages,fonts, sizes, colours, orientations and structures. Moreover, the...In today’s real world, an important research part in image processing isscene text detection and recognition. Scene text can be in different languages,fonts, sizes, colours, orientations and structures. Moreover, the aspect ratios andlayouts of a scene text may differ significantly. All these variations appear assignificant challenges for the detection and recognition algorithms that are consideredfor the text in natural scenes. In this paper, a new intelligent text detection andrecognition method for detectingthe text from natural scenes and forrecognizingthe text by applying the newly proposed Conditional Random Field-based fuzzyrules incorporated Convolutional Neural Network (CR-CNN) has been proposed.Moreover, we have recommended a new text detection method for detecting theexact text from the input natural scene images. For enhancing the presentation ofthe edge detection process, image pre-processing activities such as edge detectionand color modeling have beenapplied in this work. In addition, we have generatednew fuzzy rules for making effective decisions on the processes of text detectionand recognition. The experiments have been directedusing the standard benchmark datasets such as the ICDAR 2003, the ICDAR 2011, the ICDAR2005 and the SVT and have achieved better detection accuracy intext detectionand recognition. By using these three datasets, five different experiments havebeen conducted for evaluating the proposed model. And also, we have comparedthe proposed system with the other classifiers such as the SVM, the MLP and theCNN. In these comparisons, the proposed model has achieved better classificationaccuracywhen compared with the other existing works.展开更多
Detecting and recognizing text from natural scene images presents a challenge because the image quality depends on the conditions in which the image is captured,such as viewing angles,blurring,sensor noise,etc.However...Detecting and recognizing text from natural scene images presents a challenge because the image quality depends on the conditions in which the image is captured,such as viewing angles,blurring,sensor noise,etc.However,in this paper,a prototype for text detection and recognition from natural scene images is proposed.This prototype is based on the Raspberry Pi 4 and the Universal Serial Bus(USB)camera and embedded our text detection and recognition model,which was developed using the Python language.Our model is based on the deep learning text detector model through the Efficient and Accurate Scene Text Detec-tor(EAST)model for text localization and detection and the Tesseract-OCR,which is used as an Optical Character Recognition(OCR)engine for text recog-nition.Our prototype is controlled by the Virtual Network Computing(VNC)tool through a computer via a wireless connection.The experiment results show that the recognition rate for the captured image through the camera by our prototype can reach 99.75%with low computational complexity.Furthermore,our proto-type is more performant than the Tesseract software in terms of the recognition rate.Besides,it provides the same performance in terms of the recognition rate with a huge decrease in the execution time by an average of 89%compared to the EasyOCR software on the Raspberry Pi 4 board.展开更多
Fuzzy C-means(FCM)is a clustering method that falls under unsupervised machine learning.The main issues plaguing this clustering algorithm are the number of the unknown clusters within a particular dataset and initial...Fuzzy C-means(FCM)is a clustering method that falls under unsupervised machine learning.The main issues plaguing this clustering algorithm are the number of the unknown clusters within a particular dataset and initialization sensitivity of cluster centres.Artificial Bee Colony(ABC)is a type of swarm algorithm that strives to improve the members’solution quality as an iterative process with the utilization of particular kinds of randomness.However,ABC has some weaknesses,such as balancing exploration and exploitation.To improve the exploration process within the ABC algorithm,the mean artificial bee colony(MeanABC)by its modified search equation that depends on solutions of mean previous and global best is used.Furthermore,to solve the main issues of FCM,Automatic clustering algorithm was proposed based on the mean artificial bee colony called(AC-MeanABC).It uses the MeanABC capability of balancing between exploration and exploitation and its capacity to explore the positive and negative directions in search space to find the best value of clusters number and centroids value.A few benchmark datasets and a set of natural images were used to evaluate the effectiveness of AC-MeanABC.The experimental findings are encouraging and indicate considerable improvements compared to other state-of-the-art approaches in the same domain.展开更多
文摘Due to improper acquisition settings and other noise artifacts,the image degraded to yield poor mean preservation in brightness.The simplest way to improve the preservation is the implementation of histogram equalization.Because of over-enhancement,it failed to preserve the mean brightness and produce the poor quality of the image.This paper proposes a multi-scale decomposi-tion for brightness preservation using gamma correction.After transformation to hue,saturation and intensity(HSI)channel,the 2D-discrete wavelet transform decomposed the intensity component into low and high-pass coefficients.At the next phase,gamma correction is used by auto-tuning the scale value.The scale is the modified constant value used in the logarithmic function.Further,the scale value is optimized to obtain better visual quality in the image.The optimized value is the weighted distribution of standard deviation-mean of low pass coefficients.Finally,the experimental result is estimated in terms of quality assessment measures used as absolute mean brightness error,the measure of information detail,signal to noise ratio and patch-based contrast quality in the image.By comparison,the proposed method proved to be suitably remarkable in retaining the mean brightness and better visual quality of the image.
基金supported by National Natural Science Foundation of China(62072250).
文摘With the development of computer graphics,realistic computer graphics(CG)have become more and more common in our field of vision.This rendered image is invisible to the naked eye.How to effectively identify CG and natural images(NI)has been become a new issue in the field of digital forensics.In recent years,a series of deep learning network frameworks have shown great advantages in the field of images,which provides a good choice for us to solve this problem.This paper aims to track the latest developments and applications of deep learning in the field of CG and NI forensics in a timely manner.Firstly,it introduces the background of deep learning and the knowledge of convolutional neural networks.The purpose is to understand the basic model structure of deep learning applications in the image field,and then outlines the mainstream framework;secondly,it briefly introduces the application of deep learning in CG and NI forensics,and finally points out the problems of deep learning in this field and the prospects for the future.
基金supported by the National Key Research and Development Program of China under Grant No.2019YFB2204104the Beijing Natural Science Foundation of China under Grant No.L182059+1 种基金the National Natural Science Foundation of China under Grant Nos.61772523,61620106003,and 61802406,Alibaba Group through Alibaba Innovative Research Programthe Joint Open Research Fund Program of State Key Laboratory of Hydroscience and Engineering and Tsinghua-Ningxia Yinchuan Joint Institute of Internet of Waters on Digital Water Governance.
文摘With the recent tremendous advances of computer graphics rendering and image editing technologies,computergenerated fake images,which in general do not reflect what happens in the reality,can now easily deceive the inspection of human visual system.In this work,we propose a convolutional neural network(CNN)-based model to distinguish computergenerated(CG)images from natural images(NIs)with channel and pixel correlation.The key component of the proposed CNN architecture is a self-coding module that takes the color images as input to extract the correlation between color channels explicitly.Unlike previous approaches that directly apply CNN to solve this problem,we consider the generality of the network(or subnetwork),i.e.,the newly introduced hybrid correlation module can be directly combined with existing CNN models for enhancing the discrimination capacity of original networks.Experimental results demonstrate that the proposed network outperforms state-of-the-art methods in terms of classification performance.We also show that the newly introduced hybrid correlation module can improve the classification accuracy of different CNN architectures.
文摘In today’s real world, an important research part in image processing isscene text detection and recognition. Scene text can be in different languages,fonts, sizes, colours, orientations and structures. Moreover, the aspect ratios andlayouts of a scene text may differ significantly. All these variations appear assignificant challenges for the detection and recognition algorithms that are consideredfor the text in natural scenes. In this paper, a new intelligent text detection andrecognition method for detectingthe text from natural scenes and forrecognizingthe text by applying the newly proposed Conditional Random Field-based fuzzyrules incorporated Convolutional Neural Network (CR-CNN) has been proposed.Moreover, we have recommended a new text detection method for detecting theexact text from the input natural scene images. For enhancing the presentation ofthe edge detection process, image pre-processing activities such as edge detectionand color modeling have beenapplied in this work. In addition, we have generatednew fuzzy rules for making effective decisions on the processes of text detectionand recognition. The experiments have been directedusing the standard benchmark datasets such as the ICDAR 2003, the ICDAR 2011, the ICDAR2005 and the SVT and have achieved better detection accuracy intext detectionand recognition. By using these three datasets, five different experiments havebeen conducted for evaluating the proposed model. And also, we have comparedthe proposed system with the other classifiers such as the SVM, the MLP and theCNN. In these comparisons, the proposed model has achieved better classificationaccuracywhen compared with the other existing works.
基金This work was funded by the Deanship of Scientific Research at Jouf University(Kingdom of Saudi Arabia)under Grant No.DSR-2021-02-0392.
文摘Detecting and recognizing text from natural scene images presents a challenge because the image quality depends on the conditions in which the image is captured,such as viewing angles,blurring,sensor noise,etc.However,in this paper,a prototype for text detection and recognition from natural scene images is proposed.This prototype is based on the Raspberry Pi 4 and the Universal Serial Bus(USB)camera and embedded our text detection and recognition model,which was developed using the Python language.Our model is based on the deep learning text detector model through the Efficient and Accurate Scene Text Detec-tor(EAST)model for text localization and detection and the Tesseract-OCR,which is used as an Optical Character Recognition(OCR)engine for text recog-nition.Our prototype is controlled by the Virtual Network Computing(VNC)tool through a computer via a wireless connection.The experiment results show that the recognition rate for the captured image through the camera by our prototype can reach 99.75%with low computational complexity.Furthermore,our proto-type is more performant than the Tesseract software in terms of the recognition rate.Besides,it provides the same performance in terms of the recognition rate with a huge decrease in the execution time by an average of 89%compared to the EasyOCR software on the Raspberry Pi 4 board.
基金supported by the Research Management Center,Xiamen University Malaysia under XMUM Research Program Cycle 4(Grant No:XMUMRF/2019-C4/IECE/0012).
文摘Fuzzy C-means(FCM)is a clustering method that falls under unsupervised machine learning.The main issues plaguing this clustering algorithm are the number of the unknown clusters within a particular dataset and initialization sensitivity of cluster centres.Artificial Bee Colony(ABC)is a type of swarm algorithm that strives to improve the members’solution quality as an iterative process with the utilization of particular kinds of randomness.However,ABC has some weaknesses,such as balancing exploration and exploitation.To improve the exploration process within the ABC algorithm,the mean artificial bee colony(MeanABC)by its modified search equation that depends on solutions of mean previous and global best is used.Furthermore,to solve the main issues of FCM,Automatic clustering algorithm was proposed based on the mean artificial bee colony called(AC-MeanABC).It uses the MeanABC capability of balancing between exploration and exploitation and its capacity to explore the positive and negative directions in search space to find the best value of clusters number and centroids value.A few benchmark datasets and a set of natural images were used to evaluate the effectiveness of AC-MeanABC.The experimental findings are encouraging and indicate considerable improvements compared to other state-of-the-art approaches in the same domain.