期刊文献+
共找到15,506篇文章
< 1 2 250 >
每页显示 20 50 100
Background removal from global auroral images:Data-driven dayglow modeling 被引量:1
1
作者 A.Ohma M.Madelaire +4 位作者 K.M.Laundal J.P.Reistad S.M.Hatch S.Gasparini S.J.Walker 《Earth and Planetary Physics》 EI CSCD 2024年第1期247-257,共11页
Global images of auroras obtained by cameras on spacecraft are a key tool for studying the near-Earth environment.However,the cameras are sensitive not only to auroral emissions produced by precipitating particles,but... Global images of auroras obtained by cameras on spacecraft are a key tool for studying the near-Earth environment.However,the cameras are sensitive not only to auroral emissions produced by precipitating particles,but also to dayglow emissions produced by photoelectrons induced by sunlight.Nightglow emissions and scattered sunlight can contribute to the background signal.To fully utilize such images in space science,background contamination must be removed to isolate the auroral signal.Here we outline a data-driven approach to modeling the background intensity in multiple images by formulating linear inverse problems based on B-splines and spherical harmonics.The approach is robust,flexible,and iteratively deselects outliers,such as auroral emissions.The final model is smooth across the terminator and accounts for slow temporal variations and large-scale asymmetries in the dayglow.We demonstrate the model by using the three far ultraviolet cameras on the Imager for Magnetopause-to-Aurora Global Exploration(IMAGE)mission.The method can be applied to historical missions and is relevant for upcoming missions,such as the Solar wind Magnetosphere Ionosphere Link Explorer(SMILE)mission. 展开更多
关键词 AURORA dayglow modeling global auroral images far ultraviolet images dayglow removal
下载PDF
Road Traffic Monitoring from Aerial Images Using Template Matching and Invariant Features 被引量:1
2
作者 Asifa Mehmood Qureshi Naif Al Mudawi +2 位作者 Mohammed Alonazi Samia Allaoua Chelloug Jeongmin Park 《Computers, Materials & Continua》 SCIE EI 2024年第3期3683-3701,共19页
Road traffic monitoring is an imperative topic widely discussed among researchers.Systems used to monitor traffic frequently rely on cameras mounted on bridges or roadsides.However,aerial images provide the flexibilit... Road traffic monitoring is an imperative topic widely discussed among researchers.Systems used to monitor traffic frequently rely on cameras mounted on bridges or roadsides.However,aerial images provide the flexibility to use mobile platforms to detect the location and motion of the vehicle over a larger area.To this end,different models have shown the ability to recognize and track vehicles.However,these methods are not mature enough to produce accurate results in complex road scenes.Therefore,this paper presents an algorithm that combines state-of-the-art techniques for identifying and tracking vehicles in conjunction with image bursts.The extracted frames were converted to grayscale,followed by the application of a georeferencing algorithm to embed coordinate information into the images.The masking technique eliminated irrelevant data and reduced the computational cost of the overall monitoring system.Next,Sobel edge detection combined with Canny edge detection and Hough line transform has been applied for noise reduction.After preprocessing,the blob detection algorithm helped detect the vehicles.Vehicles of varying sizes have been detected by implementing a dynamic thresholding scheme.Detection was done on the first image of every burst.Then,to track vehicles,the model of each vehicle was made to find its matches in the succeeding images using the template matching algorithm.To further improve the tracking accuracy by incorporating motion information,Scale Invariant Feature Transform(SIFT)features have been used to find the best possible match among multiple matches.An accuracy rate of 87%for detection and 80%accuracy for tracking in the A1 Motorway Netherland dataset has been achieved.For the Vehicle Aerial Imaging from Drone(VAID)dataset,an accuracy rate of 86%for detection and 78%accuracy for tracking has been achieved. 展开更多
关键词 Unmanned Aerial Vehicles(UAV) aerial images DATASET object detection object tracking data elimination template matching blob detection SIFT VAID
下载PDF
Enhancing Dense Small Object Detection in UAV Images Based on Hybrid Transformer 被引量:1
3
作者 Changfeng Feng Chunping Wang +2 位作者 Dongdong Zhang Renke Kou Qiang Fu 《Computers, Materials & Continua》 SCIE EI 2024年第3期3993-4013,共21页
Transformer-based models have facilitated significant advances in object detection.However,their extensive computational consumption and suboptimal detection of dense small objects curtail their applicability in unman... Transformer-based models have facilitated significant advances in object detection.However,their extensive computational consumption and suboptimal detection of dense small objects curtail their applicability in unmanned aerial vehicle(UAV)imagery.Addressing these limitations,we propose a hybrid transformer-based detector,H-DETR,and enhance it for dense small objects,leading to an accurate and efficient model.Firstly,we introduce a hybrid transformer encoder,which integrates a convolutional neural network-based cross-scale fusion module with the original encoder to handle multi-scale feature sequences more efficiently.Furthermore,we propose two novel strategies to enhance detection performance without incurring additional inference computation.Query filter is designed to cope with the dense clustering inherent in drone-captured images by counteracting similar queries with a training-aware non-maximum suppression.Adversarial denoising learning is a novel enhancement method inspired by adversarial learning,which improves the detection of numerous small targets by counteracting the effects of artificial spatial and semantic noise.Extensive experiments on the VisDrone and UAVDT datasets substantiate the effectiveness of our approach,achieving a significant improvement in accuracy with a reduction in computational complexity.Our method achieves 31.9%and 21.1%AP on the VisDrone and UAVDT datasets,respectively,and has a faster inference speed,making it a competitive model in UAV image object detection. 展开更多
关键词 UAV images TRANSFORMER dense small object detection
下载PDF
DeepSVDNet:A Deep Learning-Based Approach for Detecting and Classifying Vision-Threatening Diabetic Retinopathy in Retinal Fundus Images 被引量:1
4
作者 Anas Bilal Azhar Imran +4 位作者 Talha Imtiaz Baig Xiaowen Liu Haixia Long Abdulkareem Alzahrani Muhammad Shafiq 《Computer Systems Science & Engineering》 2024年第2期511-528,共18页
Artificial Intelligence(AI)is being increasingly used for diagnosing Vision-Threatening Diabetic Retinopathy(VTDR),which is a leading cause of visual impairment and blindness worldwide.However,previous automated VTDR ... Artificial Intelligence(AI)is being increasingly used for diagnosing Vision-Threatening Diabetic Retinopathy(VTDR),which is a leading cause of visual impairment and blindness worldwide.However,previous automated VTDR detection methods have mainly relied on manual feature extraction and classification,leading to errors.This paper proposes a novel VTDR detection and classification model that combines different models through majority voting.Our proposed methodology involves preprocessing,data augmentation,feature extraction,and classification stages.We use a hybrid convolutional neural network-singular value decomposition(CNN-SVD)model for feature extraction and selection and an improved SVM-RBF with a Decision Tree(DT)and K-Nearest Neighbor(KNN)for classification.We tested our model on the IDRiD dataset and achieved an accuracy of 98.06%,a sensitivity of 83.67%,and a specificity of 100%for DR detection and evaluation tests,respectively.Our proposed approach outperforms baseline techniques and provides a more robust and accurate method for VTDR detection. 展开更多
关键词 Diabetic retinopathy(DR) fundus images(FIs) support vector machine(SVM) medical image analysis convolutional neural networks(CNN) singular value decomposition(SVD) classification
下载PDF
Fuzzy Difference Equations in Diagnoses of Glaucoma from Retinal Images Using Deep Learning
5
作者 D.Dorathy Prema Kavitha L.Francis Raj +3 位作者 Sandeep Kautish Abdulaziz S.Almazyad Karam M.Sallam Ali Wagdy Mohamed 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第4期801-816,共16页
The intuitive fuzzy set has found important application in decision-making and machine learning.To enrich and utilize the intuitive fuzzy set,this study designed and developed a deep neural network-based glaucoma eye ... The intuitive fuzzy set has found important application in decision-making and machine learning.To enrich and utilize the intuitive fuzzy set,this study designed and developed a deep neural network-based glaucoma eye detection using fuzzy difference equations in the domain where the retinal images converge.Retinal image detections are categorized as normal eye recognition,suspected glaucomatous eye recognition,and glaucomatous eye recognition.Fuzzy degrees associated with weighted values are calculated to determine the level of concentration between the fuzzy partition and the retinal images.The proposed model was used to diagnose glaucoma using retinal images and involved utilizing the Convolutional Neural Network(CNN)and deep learning to identify the fuzzy weighted regularization between images.This methodology was used to clarify the input images and make them adequate for the process of glaucoma detection.The objective of this study was to propose a novel approach to the early diagnosis of glaucoma using the Fuzzy Expert System(FES)and Fuzzy differential equation(FDE).The intensities of the different regions in the images and their respective peak levels were determined.Once the peak regions were identified,the recurrence relationships among those peaks were then measured.Image partitioning was done due to varying degrees of similar and dissimilar concentrations in the image.Similar and dissimilar concentration levels and spatial frequency generated a threshold image from the combined fuzzy matrix and FDE.This distinguished between a normal and abnormal eye condition,thus detecting patients with glaucomatous eyes. 展开更多
关键词 Convolutional Neural Network(CNN) glaucomatous eyes fuzzy difference equation intuitive fuzzy sets image segmentation retinal images
下载PDF
CrossLinkNet: An Explainable and Trustworthy AI Framework for Whole-Slide Images Segmentation
6
作者 Peng Xiao Qi Zhong +3 位作者 Jingxue Chen Dongyuan Wu Zhen Qin Erqiang Zhou 《Computers, Materials & Continua》 SCIE EI 2024年第6期4703-4724,共22页
In the intelligent medical diagnosis area,Artificial Intelligence(AI)’s trustworthiness,reliability,and interpretability are critical,especially in cancer diagnosis.Traditional neural networks,while excellent at proc... In the intelligent medical diagnosis area,Artificial Intelligence(AI)’s trustworthiness,reliability,and interpretability are critical,especially in cancer diagnosis.Traditional neural networks,while excellent at processing natural images,often lack interpretability and adaptability when processing high-resolution digital pathological images.This limitation is particularly evident in pathological diagnosis,which is the gold standard of cancer diagnosis and relies on a pathologist’s careful examination and analysis of digital pathological slides to identify the features and progression of the disease.Therefore,the integration of interpretable AI into smart medical diagnosis is not only an inevitable technological trend but also a key to improving diagnostic accuracy and reliability.In this paper,we introduce an innovative Multi-Scale Multi-Branch Feature Encoder(MSBE)and present the design of the CrossLinkNet Framework.The MSBE enhances the network’s capability for feature extraction by allowing the adjustment of hyperparameters to configure the number of branches and modules.The CrossLinkNet Framework,serving as a versatile image segmentation network architecture,employs cross-layer encoder-decoder connections for multi-level feature fusion,thereby enhancing feature integration and segmentation accuracy.Comprehensive quantitative and qualitative experiments on two datasets demonstrate that CrossLinkNet,equipped with the MSBE encoder,not only achieves accurate segmentation results but is also adaptable to various tumor segmentation tasks and scenarios by replacing different feature encoders.Crucially,CrossLinkNet emphasizes the interpretability of the AI model,a crucial aspect for medical professionals,providing an in-depth understanding of the model’s decisions and thereby enhancing trust and reliability in AI-assisted diagnostics. 展开更多
关键词 Explainable AI security TRUSTWORTHY CrossLinkNet whole slide images
下载PDF
Double quantum images encryption scheme based on chaotic system
7
作者 蒋社想 李杨 +1 位作者 石锦 张茹 《Chinese Physics B》 SCIE EI CAS CSCD 2024年第4期305-320,共16页
This paper explores a double quantum images representation(DNEQR)model that allows for simultaneous storage of two digital images in a quantum superposition state.Additionally,a new type of two-dimensional hyperchaoti... This paper explores a double quantum images representation(DNEQR)model that allows for simultaneous storage of two digital images in a quantum superposition state.Additionally,a new type of two-dimensional hyperchaotic system based on sine and logistic maps is investigated,offering a wider parameter space and better chaotic behavior compared to the sine and logistic maps.Based on the DNEQR model and the hyperchaotic system,a double quantum images encryption algorithm is proposed.Firstly,two classical plaintext images are transformed into quantum states using the DNEQR model.Then,the proposed hyperchaotic system is employed to iteratively generate pseudo-random sequences.These chaotic sequences are utilized to perform pixel value and position operations on the quantum image,resulting in changes to both pixel values and positions.Finally,the ciphertext image can be obtained by qubit-level diffusion using two XOR operations between the position-permutated image and the pseudo-random sequences.The corresponding quantum circuits are also given.Experimental results demonstrate that the proposed scheme ensures the security of the images during transmission,improves the encryption efficiency,and enhances anti-interference and anti-attack capabilities. 展开更多
关键词 double quantum images encryption chaotic system pixel scrambling XOR operation
下载PDF
Mapping soil organic matter in cultivated land based on multi-year composite images on monthly time scales
8
作者 Jie Song Dongsheng Yu +4 位作者 Siwei Wang Yanhe Zhao Xin Wang Lixia Ma Jiangang Li 《Journal of Integrative Agriculture》 SCIE CAS CSCD 2024年第4期1393-1408,共16页
Rapid and accurate acquisition of soil organic matter(SOM)information in cultivated land is important for sustainable agricultural development and carbon balance management.This study proposed a novel approach to pred... Rapid and accurate acquisition of soil organic matter(SOM)information in cultivated land is important for sustainable agricultural development and carbon balance management.This study proposed a novel approach to predict SOM with high accuracy using multiyear synthetic remote sensing variables on a monthly scale.We obtained 12 monthly synthetic Sentinel-2 images covering the study area from 2016 to 2021 through the Google Earth Engine(GEE)platform,and reflectance bands and vegetation indices were extracted from these composite images.Then the random forest(RF),support vector machine(SVM)and gradient boosting regression tree(GBRT)models were tested to investigate the difference in SOM prediction accuracy under different combinations of monthly synthetic variables.Results showed that firstly,all monthly synthetic spectral bands of Sentinel-2 showed a significant correlation with SOM(P<0.05)for the months of January,March,April,October,and November.Secondly,in terms of single-monthly composite variables,the prediction accuracy was relatively poor,with the highest R^(2)value of 0.36 being observed in January.When monthly synthetic environmental variables were grouped in accordance with the four quarters of the year,the first quarter and the fourth quarter showed good performance,and any combination of three quarters was similar in estimation accuracy.The overall best performance was observed when all monthly synthetic variables were incorporated into the models.Thirdly,among the three models compared,the RF model was consistently more accurate than the SVM and GBRT models,achieving an R^(2)value of 0.56.Except for band 12 in December,the importance of the remaining bands did not exhibit significant differences.This research offers a new attempt to map SOM with high accuracy and fine spatial resolution based on monthly synthetic Sentinel-2 images. 展开更多
关键词 soil organic matter Sentinel-2 monthly synthetic images machine learning model spatial prediction
下载PDF
Robust and Discriminative Feature Learning via Mutual Information Maximization for Object Detection in Aerial Images
9
作者 Xu Sun Yinhui Yu Qing Cheng 《Computers, Materials & Continua》 SCIE EI 2024年第9期4149-4171,共23页
Object detection in unmanned aerial vehicle(UAV)aerial images has become increasingly important in military and civil applications.General object detection models are not robust enough against interclass similarity an... Object detection in unmanned aerial vehicle(UAV)aerial images has become increasingly important in military and civil applications.General object detection models are not robust enough against interclass similarity and intraclass variability of small objects,and UAV-specific nuisances such as uncontrolledweather conditions.Unlike previous approaches focusing on high-level semantic information,we report the importance of underlying features to improve detection accuracy and robustness fromthe information-theoretic perspective.Specifically,we propose a robust and discriminative feature learning approach through mutual information maximization(RD-MIM),which can be integrated into numerous object detection methods for aerial images.Firstly,we present the rank sample mining method to reduce underlying feature differences between the natural image domain and the aerial image domain.Then,we design a momentum contrast learning strategy to make object features similar to the same category and dissimilar to different categories.Finally,we construct a transformer-based global attention mechanism to boost object location semantics by leveraging the high interrelation of different receptive fields.We conduct extensive experiments on the VisDrone and Unmanned Aerial Vehicle Benchmark Object Detection and Tracking(UAVDT)datasets to prove the effectiveness of the proposed method.The experimental results show that our approach brings considerable robustness gains to basic detectors and advanced detection methods,achieving relative growth rates of 51.0%and 39.4%in corruption robustness,respectively.Our code is available at https://github.com/cq100/RD-MIM(accessed on 2 August 2024). 展开更多
关键词 Aerial images object detection mutual information contrast learning attention mechanism
下载PDF
U-Net Inspired Deep Neural Network-Based Smoke Plume Detection in Satellite Images
10
作者 Ananthakrishnan Balasundaram Ayesha Shaik +1 位作者 Japmann Kaur Banga Aman Kumar Singh 《Computers, Materials & Continua》 SCIE EI 2024年第4期779-799,共21页
Industrial activities, through the human-induced release of Green House Gas (GHG) emissions, have beenidentified as the primary cause of global warming. Accurate and quantitative monitoring of these emissions isessent... Industrial activities, through the human-induced release of Green House Gas (GHG) emissions, have beenidentified as the primary cause of global warming. Accurate and quantitative monitoring of these emissions isessential for a comprehensive understanding of their impact on the Earth’s climate and for effectively enforcingemission regulations at a large scale. This work examines the feasibility of detecting and quantifying industrialsmoke plumes using freely accessible geo-satellite imagery. The existing systemhas so many lagging factors such aslimitations in accuracy, robustness, and efficiency and these factors hinder the effectiveness in supporting timelyresponse to industrial fires. In this work, the utilization of grayscale images is done instead of traditional colorimages for smoke plume detection. The dataset was trained through a ResNet-50 model for classification and aU-Net model for segmentation. The dataset consists of images gathered by European Space Agency’s Sentinel-2 satellite constellation from a selection of industrial sites. The acquired images predominantly capture scenesof industrial locations, some of which exhibit active smoke plume emissions. The performance of the abovementionedtechniques and models is represented by their accuracy and IOU (Intersection-over-Union) metric.The images are first trained on the basic RGB images where their respective classification using the ResNet-50model results in an accuracy of 94.4% and segmentation using the U-Net Model with an IOU metric of 0.5 andaccuracy of 94% which leads to the detection of exact patches where the smoke plume has occurred. This work hastrained the classification model on grayscale images achieving a good increase in accuracy of 96.4%. 展开更多
关键词 Smoke plume ResNet-50 U-Net geo satellite images early warning global monitoring
下载PDF
CMMCAN:Lightweight Feature Extraction and Matching Network for Endoscopic Images Based on Adaptive Attention
11
作者 Nannan Chong Fan Yang 《Computers, Materials & Continua》 SCIE EI 2024年第8期2761-2783,共23页
In minimally invasive surgery,endoscopes or laparoscopes equipped with miniature cameras and tools are used to enter the human body for therapeutic purposes through small incisions or natural cavities.However,in clini... In minimally invasive surgery,endoscopes or laparoscopes equipped with miniature cameras and tools are used to enter the human body for therapeutic purposes through small incisions or natural cavities.However,in clinical operating environments,endoscopic images often suffer from challenges such as low texture,uneven illumination,and non-rigid structures,which affect feature observation and extraction.This can severely impact surgical navigation or clinical diagnosis due to missing feature points in endoscopic images,leading to treatment and postoperative recovery issues for patients.To address these challenges,this paper introduces,for the first time,a Cross-Channel Multi-Modal Adaptive Spatial Feature Fusion(ASFF)module based on the lightweight architecture of EfficientViT.Additionally,a novel lightweight feature extraction and matching network based on attention mechanism is proposed.This network dynamically adjusts attention weights for cross-modal information from grayscale images and optical flow images through a dual-branch Siamese network.It extracts static and dynamic information features ranging from low-level to high-level,and from local to global,ensuring robust feature extraction across different widths,noise levels,and blur scenarios.Global and local matching are performed through a multi-level cascaded attention mechanism,with cross-channel attention introduced to simultaneously extract low-level and high-level features.Extensive ablation experiments and comparative studies are conducted on the HyperKvasir,EAD,M2caiSeg,CVC-ClinicDB,and UCL synthetic datasets.Experimental results demonstrate that the proposed network improves upon the baseline EfficientViT-B3 model by 75.4%in accuracy(Acc),while also enhancing runtime performance and storage efficiency.When compared with the complex DenseDescriptor feature extraction network,the difference in Acc is less than 7.22%,and IoU calculation results on specific datasets outperform complex dense models.Furthermore,this method increases the F1 score by 33.2%and accelerates runtime by 70.2%.It is noteworthy that the speed of CMMCAN surpasses that of comparative lightweight models,with feature extraction and matching performance comparable to existing complex models but with faster speed and higher cost-effectiveness. 展开更多
关键词 Feature extraction and matching lightweighted network medical images ENDOSCOPIC ATTENTION
下载PDF
A new two-step variational model for multiplicative noise removal with applications to texture images
12
作者 ZHANG Long-hui YAO Wen-juan +2 位作者 SHI Sheng-zhu GUO Zhi-chang ZHANG Da-zhi 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 2024年第3期486-501,共16页
Multiplicative noise removal problems have attracted much attention in recent years.Unlike additive noise,multiplicative noise destroys almost all information of the original image,especially for texture images.Motiva... Multiplicative noise removal problems have attracted much attention in recent years.Unlike additive noise,multiplicative noise destroys almost all information of the original image,especially for texture images.Motivated by the TV-Stokes model,we propose a new two-step variational model to denoise the texture images corrupted by multiplicative noise with a good geometry explanation in this paper.In the first step,we convert the multiplicative denoising problem into an additive one by the logarithm transform and propagate the isophote directions in the tangential field smoothing.Once the isophote directions are constructed,an image is restored to fit the constructed directions in the second step.The existence and uniqueness of the solution to the variational problems are proved.In these two steps,we use the gradient descent method and construct finite difference schemes to solve the problems.Especially,the augmented Lagrangian method and the fast Fourier transform are adopted to accelerate the calculation.Experimental results show that the proposed model can remove the multiplicative noise efficiently and protect the texture well. 展开更多
关键词 multiplicative noise removal texture images total variation two-step variational method aug-mented Lagrangian method
下载PDF
EDU-GAN:Edge Enhancement Generative Adversarial Networks with Dual-Domain Discriminators for Inscription Images Denoising
13
作者 Yunjing Liu Erhu Zhang +2 位作者 Jingjing Wang Guangfeng Lin Jinghong Duan 《Computers, Materials & Continua》 SCIE EI 2024年第7期1633-1653,共21页
Recovering high-quality inscription images from unknown and complex inscription noisy images is a challenging research issue.Different fromnatural images,character images pay more attention to stroke information.Howev... Recovering high-quality inscription images from unknown and complex inscription noisy images is a challenging research issue.Different fromnatural images,character images pay more attention to stroke information.However,existingmodelsmainly consider pixel-level informationwhile ignoring structural information of the character,such as its edge and glyph,resulting in reconstructed images with mottled local structure and character damage.To solve these problems,we propose a novel generative adversarial network(GAN)framework based on an edge-guided generator and a discriminator constructed by a dual-domain U-Net framework,i.e.,EDU-GAN.Unlike existing frameworks,the generator introduces the edge extractionmodule,guiding it into the denoising process through the attention mechanism,which maintains the edge detail of the restored inscription image.Moreover,a dual-domain U-Net-based discriminator is proposed to learn the global and local discrepancy between the denoised and the label images in both image and morphological domains,which is helpful to blind denoising tasks.The proposed dual-domain discriminator and generator for adversarial training can reduce local artifacts and keep the denoised character structure intact.Due to the lack of a real-inscription image,we built the real-inscription dataset to provide an effective benchmark for studying inscription image denoising.The experimental results show the superiority of our method both in the synthetic and real-inscription datasets. 展开更多
关键词 Dual-domain discriminators inscription images DENOISING edge-guided generator
下载PDF
Deep Transfer Learning Models for Mobile-Based Ocular Disorder Identification on Retinal Images
14
作者 Roseline Oluwaseun Ogundokun Joseph Bamidele Awotunde +2 位作者 Hakeem Babalola Akande Cheng-Chi Lee Agbotiname Lucky Imoize 《Computers, Materials & Continua》 SCIE EI 2024年第7期139-161,共23页
Mobile technology is developing significantly.Mobile phone technologies have been integrated into the healthcare industry to help medical practitioners.Typically,computer vision models focus on image detection and cla... Mobile technology is developing significantly.Mobile phone technologies have been integrated into the healthcare industry to help medical practitioners.Typically,computer vision models focus on image detection and classification issues.MobileNetV2 is a computer vision model that performs well on mobile devices,but it requires cloud services to process biometric image information and provide predictions to users.This leads to increased latency.Processing biometrics image datasets on mobile devices will make the prediction faster,but mobiles are resource-restricted devices in terms of storage,power,and computational speed.Hence,a model that is small in size,efficient,and has good prediction quality for biometrics image classification problems is required.Quantizing pre-trained CNN(PCNN)MobileNetV2 architecture combined with a Support Vector Machine(SVM)compacts the model representation and reduces the computational cost and memory requirement.This proposed novel approach combines quantized pre-trained CNN(PCNN)MobileNetV2 architecture with a Support Vector Machine(SVM)to represent models efficiently with low computational cost and memory.Our contributions include evaluating three CNN models for ocular disease identification in transfer learning and deep feature plus SVM approaches,showing the superiority of deep features from MobileNetV2 and SVM classification models,comparing traditional methods,exploring six ocular diseases and normal classification with 20,111 images postdata augmentation,and reducing the number of trainable models.The model is trained on ocular disorder retinal fundus image datasets according to the severity of six age-related macular degeneration(AMD),one of the most common eye illnesses,Cataract,Diabetes,Glaucoma,Hypertension,andMyopia with one class Normal.From the experiment outcomes,it is observed that the suggested MobileNetV2-SVM model size is compressed.The testing accuracy for MobileNetV2-SVM,InceptionV3,and MobileNetV2 is 90.11%,86.88%,and 89.76%respectively while MobileNetV2-SVM,InceptionV3,and MobileNetV2 accuracy are observed to be 92.59%,83.38%,and 90.16%,respectively.The proposed novel technique can be used to classify all biometric medical image datasets on mobile devices. 展开更多
关键词 Retinal images ocular disorder deep transfer learning disease identification mobile device
下载PDF
A deep learning fusion model for accurate classification of brain tumours in Magnetic Resonance images
15
作者 Nechirvan Asaad Zebari Chira Nadheef Mohammed +8 位作者 Dilovan Asaad Zebari Mazin Abed Mohammed Diyar Qader Zeebaree Haydar Abdulameer Marhoon Karrar Hameed Abdulkareem Seifedine Kadry Wattana Viriyasitavat Jan Nedoma Radek Martinek 《CAAI Transactions on Intelligence Technology》 SCIE EI 2024年第4期790-804,共15页
Detecting brain tumours is complex due to the natural variation in their location, shape, and intensity in images. While having accurate detection and segmentation of brain tumours would be beneficial, current methods... Detecting brain tumours is complex due to the natural variation in their location, shape, and intensity in images. While having accurate detection and segmentation of brain tumours would be beneficial, current methods still need to solve this problem despite the numerous available approaches. Precise analysis of Magnetic Resonance Imaging (MRI) is crucial for detecting, segmenting, and classifying brain tumours in medical diagnostics. Magnetic Resonance Imaging is a vital component in medical diagnosis, and it requires precise, efficient, careful, efficient, and reliable image analysis techniques. The authors developed a Deep Learning (DL) fusion model to classify brain tumours reliably. Deep Learning models require large amounts of training data to achieve good results, so the researchers utilised data augmentation techniques to increase the dataset size for training models. VGG16, ResNet50, and convolutional deep belief networks networks extracted deep features from MRI images. Softmax was used as the classifier, and the training set was supplemented with intentionally created MRI images of brain tumours in addition to the genuine ones. The features of two DL models were combined in the proposed model to generate a fusion model, which significantly increased classification accuracy. An openly accessible dataset from the internet was used to test the model's performance, and the experimental results showed that the proposed fusion model achieved a classification accuracy of 98.98%. Finally, the results were compared with existing methods, and the proposed model outperformed them significantly. 展开更多
关键词 brain tumour deep learning feature fusion model MRI images multi‐classification
下载PDF
Automated Algorithms for Detecting and Classifying X-Ray Images of Spine Fractures
16
作者 Fayez Alfayez 《Computers, Materials & Continua》 SCIE EI 2024年第4期1539-1560,共22页
This paper emphasizes a faster digital processing time while presenting an accurate method for identifying spinefractures in X-ray pictures. The study focuses on efficiency by utilizing many methods that include pictu... This paper emphasizes a faster digital processing time while presenting an accurate method for identifying spinefractures in X-ray pictures. The study focuses on efficiency by utilizing many methods that include picturesegmentation, feature reduction, and image classification. Two important elements are investigated to reducethe classification time: Using feature reduction software and leveraging the capabilities of sophisticated digitalprocessing hardware. The researchers use different algorithms for picture enhancement, including theWiener andKalman filters, and they look into two background correction techniques. The article presents a technique forextracting textural features and evaluates three picture segmentation algorithms and three fractured spine detectionalgorithms using transformdomain, PowerDensity Spectrum(PDS), andHigher-Order Statistics (HOS) for featureextraction.With an emphasis on reducing digital processing time, this all-encompassing method helps to create asimplified system for classifying fractured spine fractures. A feature reduction program code has been built toimprove the processing speed for picture classification. Overall, the proposed approach shows great potential forsignificantly reducing classification time in clinical settings where time is critical. In comparison to other transformdomains, the texture features’ discrete cosine transform (DCT) yielded an exceptional classification rate, and theprocess of extracting features from the transform domain took less time. More capable hardware can also result inquicker execution times for the feature extraction algorithms. 展开更多
关键词 Feature reduction image classification X-ray images
下载PDF
GAN-DIRNet:A Novel Deformable Image Registration Approach for Multimodal Histological Images
17
作者 Haiyue Li Jing Xie +4 位作者 Jing Ke Ye Yuan Xiaoyong Pan Hongyi Xin Hongbin Shen 《Computers, Materials & Continua》 SCIE EI 2024年第7期487-506,共20页
Multi-modal histological image registration tasks pose significant challenges due to tissue staining operations causing partial loss and folding of tissue.Convolutional neural network(CNN)and generative adversarial ne... Multi-modal histological image registration tasks pose significant challenges due to tissue staining operations causing partial loss and folding of tissue.Convolutional neural network(CNN)and generative adversarial network(GAN)are pivotal inmedical image registration.However,existing methods often struggle with severe interference and deformation,as seen in histological images of conditions like Cushing’s disease.We argue that the failure of current approaches lies in underutilizing the feature extraction capability of the discriminator inGAN.In this study,we propose a novel multi-modal registration approach GAN-DIRNet based on GAN for deformable histological image registration.To begin with,the discriminators of two GANs are embedded as a new dual parallel feature extraction module into the unsupervised registration networks,characterized by implicitly extracting feature descriptors of specific modalities.Additionally,modal feature description layers and registration layers collaborate in unsupervised optimization,facilitating faster convergence and more precise results.Lastly,experiments and evaluations were conducted on the registration of the Mixed National Institute of Standards and Technology database(MNIST),eight publicly available datasets of histological sections and the Clustering-Registration-Classification-Segmentation(CRCS)dataset on the Cushing’s disease.Experimental results demonstrate that our proposed GAN-DIRNet method surpasses existing approaches like DIRNet in terms of both registration accuracy and time efficiency,while also exhibiting robustness across different image types. 展开更多
关键词 Histological images registration deformable registration generative adversarial network cushing’s disease machine learning computer vision
下载PDF
Integrating Transformer and Bidirectional Long Short-Term Memory for Intelligent Breast Cancer Detection from Histopathology Biopsy Images
18
作者 Prasanalakshmi Balaji Omar Alqahtani +2 位作者 Sangita Babu Mousmi Ajay Chaurasia Shanmugapriya Prakasam 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第10期443-458,共16页
Breast cancer is a significant threat to the global population,affecting not only women but also a threat to the entire population.With recent advancements in digital pathology,Eosin and hematoxylin images provide enh... Breast cancer is a significant threat to the global population,affecting not only women but also a threat to the entire population.With recent advancements in digital pathology,Eosin and hematoxylin images provide enhanced clarity in examiningmicroscopic features of breast tissues based on their staining properties.Early cancer detection facilitates the quickening of the therapeutic process,thereby increasing survival rates.The analysis made by medical professionals,especially pathologists,is time-consuming and challenging,and there arises a need for automated breast cancer detection systems.The upcoming artificial intelligence platforms,especially deep learning models,play an important role in image diagnosis and prediction.Initially,the histopathology biopsy images are taken from standard data sources.Further,the gathered images are given as input to the Multi-Scale Dilated Vision Transformer,where the essential features are acquired.Subsequently,the features are subjected to the Bidirectional Long Short-Term Memory(Bi-LSTM)for classifying the breast cancer disorder.The efficacy of the model is evaluated using divergent metrics.When compared with other methods,the proposed work reveals that it offers impressive results for detection. 展开更多
关键词 Bidirectional long short-term memory breast cancer detection feature extraction histopathology biopsy images multi-scale dilated vision transformer
下载PDF
Meibomian glands segmentation in infrared images with limited annotation
19
作者 Jia-Wen Lin Ling-Jie Lin +5 位作者 Feng Lu Tai-Chen Lai Jing Zou Lin-Ling Guo Zhi-Ming Lin Li Li 《International Journal of Ophthalmology(English edition)》 SCIE CAS 2024年第3期401-407,共7页
●AIM:To investigate a pioneering framework for the segmentation of meibomian glands(MGs),using limited annotations to reduce the workload on ophthalmologists and enhance the efficiency of clinical diagnosis.●METHODS... ●AIM:To investigate a pioneering framework for the segmentation of meibomian glands(MGs),using limited annotations to reduce the workload on ophthalmologists and enhance the efficiency of clinical diagnosis.●METHODS:Totally 203 infrared meibomian images from 138 patients with dry eye disease,accompanied by corresponding annotations,were gathered for the study.A rectified scribble-supervised gland segmentation(RSSGS)model,incorporating temporal ensemble prediction,uncertainty estimation,and a transformation equivariance constraint,was introduced to address constraints imposed by limited supervision information inherent in scribble annotations.The viability and efficacy of the proposed model were assessed based on accuracy,intersection over union(IoU),and dice coefficient.●RESULTS:Using manual labels as the gold standard,RSSGS demonstrated outcomes with an accuracy of 93.54%,a dice coefficient of 78.02%,and an IoU of 64.18%.Notably,these performance metrics exceed the current weakly supervised state-of-the-art methods by 0.76%,2.06%,and 2.69%,respectively.Furthermore,despite achieving a substantial 80%reduction in annotation costs,it only lags behind fully annotated methods by 0.72%,1.51%,and 2.04%.●CONCLUSION:An innovative automatic segmentation model is developed for MGs in infrared eyelid images,using scribble annotation for training.This model maintains an exceptionally high level of segmentation accuracy while substantially reducing training costs.It holds substantial utility for calculating clinical parameters,thereby greatly enhancing the diagnostic efficiency of ophthalmologists in evaluating meibomian gland dysfunction. 展开更多
关键词 infrared meibomian glands images meibomian gland dysfunction meibomian glands segmentation weak supervision scribbled annotation
下载PDF
Reconstruction of Knowledge and Medical Images in the Convergence of Chinese and Western Medicine:Taking “Sweet Meat” as an Example
20
作者 GU Xiaoyang 《Chinese Medicine and Culture》 2024年第3期204-212,共9页
The pancreas is neither part of the five Zang organs(五脏) nor the six Fu organs(六腑).Thus,it has received little attention in Chinese medical literature.In the late 19th century,medical missionaries in China started... The pancreas is neither part of the five Zang organs(五脏) nor the six Fu organs(六腑).Thus,it has received little attention in Chinese medical literature.In the late 19th century,medical missionaries in China started translating and introducing anatomical and physiological knowledge about the pancreas.As for the word pancreas,an early and influential translation was “sweet meat”(甜肉),proposed by Benjamin Hobson(合信).The translation “sweet meat” is not faithful to the original meaning of “pancreas”,but is a term coined by Hobson based on his personal habits,and the word “sweet” appeared by chance.However,in the decades since the term “sweet meat” became popular,Chinese medicine practitioners,such as Tang Zonghai(唐宗海),reinterpreted it by drawing new medical illustrations for “sweet meat” and giving new connotations to the word “sweet”.This discussion and interpretation of “sweet meat” in modern China,particularly among Chinese medicine professionals,is not only a dissemination and interpretation of the knowledge of “pancreas”,but also a construction of knowledge around the term “sweet meat”. 展开更多
关键词 Medical terminology Sweet meat Medical missionaries PANCREAS History of images
下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部