The deterioration of unstable rock mass raised interest in evaluating rock mass quality.However,the traditional evaluation method for the geological strength index(GSI)primarily emphasizes the rock structure and chara...The deterioration of unstable rock mass raised interest in evaluating rock mass quality.However,the traditional evaluation method for the geological strength index(GSI)primarily emphasizes the rock structure and characteristics of discontinuities.It ignores the influence of mineral composition and shows a deficiency in assessing the integrity coefficient.In this context,hyperspectral imaging and digital panoramic borehole camera technologies are applied to analyze the mineral content and integrity of rock mass.Based on the carbonate mineral content and fissure area ratio,the strength reduction factor and integrity coefficient are calculated to improve the GSI evaluation method.According to the results of mineral classification and fissure identification,the strength reduction factor and integrity coefficient increase with the depth of rock mass.The rock mass GSI calculated by the improved method is mainly concentrated between 40 and 60,which is close to the calculation results of the traditional method.The GSI error rates obtained by the two methods are mostly less than 10%,indicating the rationality of the hyperspectral-digital borehole image coupled evaluation method.Moreover,the sensitivity of the fissure area ratio(Sr)to GSI is greater than that of the strength reduction factor(a),which means the proposed GSI is suitable for rocks with significant fissure development.The improved method reduces the influence of subjective factors and provides a reliable index for the deterioration evaluation of rock mass.展开更多
Deep learning(DL)has shown its superior performance in dealing with various computer vision tasks in recent years.As a simple and effective DL model,autoencoder(AE)is popularly used to decompose hyperspectral images(H...Deep learning(DL)has shown its superior performance in dealing with various computer vision tasks in recent years.As a simple and effective DL model,autoencoder(AE)is popularly used to decompose hyperspectral images(HSIs)due to its powerful ability of feature extraction and data reconstruction.However,most existing AE-based unmixing algorithms usually ignore the spatial information of HSIs.To solve this problem,a hypergraph regularized deep autoencoder(HGAE)is proposed for unmixing.Firstly,the traditional AE architecture is specifically improved as an unsupervised unmixing framework.Secondly,hypergraph learning is employed to reformulate the loss function,which facilitates the expression of high-order similarity among locally neighboring pixels and promotes the consistency of their abundances.Moreover,L_(1/2)norm is further used to enhance abundances sparsity.Finally,the experiments on simulated data,real hyperspectral remote sensing images,and textile cloth images are used to verify that the proposed method can perform better than several state-of-the-art unmixing algorithms.展开更多
Hyperspectral remote sensing/imaging spectroscopy is a novel approach to reaching a spectrum from all the places of a huge array of spatial places so that several spectral wavelengths are utilized for making coherent ...Hyperspectral remote sensing/imaging spectroscopy is a novel approach to reaching a spectrum from all the places of a huge array of spatial places so that several spectral wavelengths are utilized for making coherent images.Hyperspectral remote sensing contains acquisition of digital images from several narrow,contiguous spectral bands throughout the visible,Thermal Infrared(TIR),Near Infrared(NIR),and Mid-Infrared(MIR)regions of the electromagnetic spectrum.In order to the application of agricultural regions,remote sensing approaches are studied and executed to their benefit of continuous and quantitativemonitoring.Particularly,hyperspectral images(HSI)are considered the precise for agriculture as they can offer chemical and physical data on vegetation.With this motivation,this article presents a novel Hurricane Optimization Algorithm with Deep Transfer Learning Driven Crop Classification(HOADTL-CC)model onHyperspectralRemote Sensing Images.The presentedHOADTL-CC model focuses on the identification and categorization of crops on hyperspectral remote sensing images.To accomplish this,the presentedHOADTL-CC model involves the design ofHOAwith capsule network(CapsNet)model for generating a set of useful feature vectors.Besides,Elman neural network(ENN)model is applied to allot proper class labels into the input HSI.Finally,glowworm swarm optimization(GSO)algorithm is exploited to fine tune the ENNparameters involved in this article.The experimental result scrutiny of the HOADTL-CC method can be tested with the help of benchmark dataset and the results are assessed under distinct aspects.Extensive comparative studies stated the enhanced performance of the HOADTL-CC model over recent approaches with maximum accuracy of 99.51%.展开更多
Spectral unmixing helps to identify different components present in the spectral mixtures which occur in the uppermost layer of the area owing to the low spatial resolution of hyperspectral images.Most spectral unmixi...Spectral unmixing helps to identify different components present in the spectral mixtures which occur in the uppermost layer of the area owing to the low spatial resolution of hyperspectral images.Most spectral unmixing methods are globally based and do not consider the spectral variability among its endmembers that occur due to illumination,atmospheric,and environmental conditions.Here,endmember bundle extraction plays a major role in overcoming the above-mentioned limitations leading to more accurate abundance fractions.Accordingly,a two-stage approach is proposed to extract endmembers through endmember bundles in hyperspectral images.The divide and conquer method is applied as the first step in subset images with only the non-redundant bands to extract endmembers using the Vertex Component Analysis(VCA)and N-FINDR algorithms.A fuzzy rule-based inference system utilizing spectral matching parameters is proposed in the second step to categorize endmembers.The endmember with the minimum error is chosen as the final endmember in each specific category.The proposed method is simple and automatically considers endmember variability in hyperspectral images.The efficiency of the proposed method is evaluated using two real hyperspectral datasets.The average spectral angle and abundance angle are used to analyze the performance measures.展开更多
Hyperspectral imaging instruments could capture detailed spatial information and rich spectral signs of observed scenes.Much spatial information and spectral signatures of hyperspectral images(HSIs)present greater pot...Hyperspectral imaging instruments could capture detailed spatial information and rich spectral signs of observed scenes.Much spatial information and spectral signatures of hyperspectral images(HSIs)present greater potential for detecting and classifying fine crops.The accurate classification of crop kinds utilizing hyperspectral remote sensing imaging(RSI)has become an indispensable application in the agricultural domain.It is significant for the prediction and growth monitoring of crop yields.Amongst the deep learning(DL)techniques,Convolution Neural Network(CNN)was the best method for classifying HSI for their incredible local contextual modeling ability,enabling spectral and spatial feature extraction.This article designs a Hybrid Multi-Strategy Aquila Optimization with a Deep Learning-Driven Crop Type Classification(HMAODL-CTC)algorithm onHSI.The proposed HMAODL-CTC model mainly intends to categorize different types of crops on HSI.To accomplish this,the presented HMAODL-CTC model initially carries out image preprocessing to improve image quality.In addition,the presented HMAODL-CTC model develops dilated convolutional neural network(CNN)for feature extraction.For hyperparameter tuning of the dilated CNN model,the HMAO algorithm is utilized.Eventually,the presented HMAODL-CTC model uses an extreme learning machine(ELM)model for crop type classification.A comprehensive set of simulations were performed to illustrate the enhanced performance of the presented HMAODL-CTC algorithm.Extensive comparison studies reported the improved performance of the presented HMAODL-CTC algorithm over other compared methods.展开更多
Hyperspectral imaging is gaining a significant role in agricultural remote sensing applications.Its data unit is the hyperspectral cube which holds spatial information in two dimensions while spectral band information...Hyperspectral imaging is gaining a significant role in agricultural remote sensing applications.Its data unit is the hyperspectral cube which holds spatial information in two dimensions while spectral band information of each pixel in the third dimension.The classification accuracy of hyperspectral images(HSI)increases significantly by employing both spatial and spectral features.For this work,the data was acquired using an airborne hyperspectral imager system which collected HSI in the visible and near-infrared(VNIR)range of 400 to 1000 nm wavelength within 180 spectral bands.The dataset is collected for nine different crops on agricultural land with a spectral resolution of 3.3 nm wavelength for each pixel.The data was cleaned from geometric distortions and stored with the class labels and annotations of global localization using the inertial navigation system.In this study,a unique pixel-based approach was designed to improve the crops'classification accuracy by using the edge-preserving features(EPF)and principal component analysis(PCA)in conjunction.The preliminary processing generated the high-dimensional EPF stack by applying the edge-preserving filters on acquired HSI.In the second step,this high dimensional stack was treated with the PCA for dimensionality reduction without losing significant spectral information.The resultant feature space(PCA-EPF)demonstrated enhanced class separability for improved crop classification with reduced dimensionality and computational cost.The support vector machines classifier was employed for multiclass classification of target crops using PCA-EPF.The classification performance evaluation was measured in terms of individual class accuracy,overall accuracy,average accuracy,and Cohen kappa factor.The proposed scheme achieved greater than 90%results for all the performance evaluation metrics.The PCA-EPF proved to be an effective attribute for crop classification using hyperspectral imaging in the VNIR range.The proposed scheme is well-suited for practical applications of crops and landfill estimations using agricultural remote sensing methods.展开更多
Most methods for classifying hyperspectral data only consider the local spatial relation-ship among samples,ignoring the important non-local topological relationship.However,the non-local topological relationship is b...Most methods for classifying hyperspectral data only consider the local spatial relation-ship among samples,ignoring the important non-local topological relationship.However,the non-local topological relationship is better at representing the structure of hyperspectral data.This paper proposes a deep learning model called Topology and semantic information fusion classification network(TSFnet)that incorporates a topology structure and semantic information transmis-sion network to accurately classify traditional Chinese medicine in hyperspectral images.TSFnet uses a convolutional neural network(CNN)to extract features and a graph convolution network(GCN)to capture potential topological relationships among different types of Chinese herbal medicines.The results show that TSFnet outperforms other state-of-the-art deep learning classification algorithms in two different scenarios of herbal medicine datasets.Additionally,the proposed TSFnet model is lightweight and can be easily deployed for mobile herbal medicine classification.展开更多
Hyperspectral images typically have high spectral resolution but low spatial resolution,which impacts the reliability and accuracy of subsequent applications,for example,remote sensingclassification and mineral identi...Hyperspectral images typically have high spectral resolution but low spatial resolution,which impacts the reliability and accuracy of subsequent applications,for example,remote sensingclassification and mineral identification.But in traditional methods via deep convolution neural net-works,indiscriminately extracting and fusing spectral and spatial features makes it challenging toutilize the differentiated information across adjacent spectral channels.Thus,we proposed a multi-branch interleaved iterative upsampling hyperspectral image super-resolution reconstruction net-work(MIIUSR)to address the above problems.We reinforce spatial feature extraction by integrat-ing detailed features from different receptive fields across adjacent channels.Furthermore,we pro-pose an interleaved iterative upsampling process during the reconstruction stage,which progres-sively fuses incremental information among adjacent frequency bands.Additionally,we add twoparallel three dimensional(3D)feature extraction branches to the backbone network to extractspectral and spatial features of varying granularity.We further enhance the backbone network’sconstruction results by leveraging the difference between two dimensional(2D)channel-groupingspatial features and 3D multi-granularity features.The results obtained by applying the proposednetwork model to the CAVE test set show that,at a scaling factor of×4,the peak signal to noiseratio,spectral angle mapping,and structural similarity are 37.310 dB,3.525 and 0.9438,respec-tively.Besides,extensive experiments conducted on the Harvard and Foster datasets demonstratethe superior potential of the proposed model in hyperspectral super-resolution reconstruction.展开更多
Graph learning,when used as a semi-supervised learning(SSL)method,performs well for classification tasks with a low label rate.We provide a graph-based batch active learning pipeline for pixel/patch neighborhood multi...Graph learning,when used as a semi-supervised learning(SSL)method,performs well for classification tasks with a low label rate.We provide a graph-based batch active learning pipeline for pixel/patch neighborhood multi-or hyperspectral image segmentation.Our batch active learning approach selects a collection of unlabeled pixels that satisfy a graph local maximum constraint for the active learning acquisition function that determines the relative importance of each pixel to the classification.This work builds on recent advances in the design of novel active learning acquisition functions(e.g.,the Model Change approach in arXiv:2110.07739)while adding important further developments including patch-neighborhood image analysis and batch active learning methods to further increase the accuracy and greatly increase the computational efficiency of these methods.In addition to improvements in the accuracy,our approach can greatly reduce the number of labeled pixels needed to achieve the same level of the accuracy based on randomly selected labeled pixels.展开更多
The adulteration concentration of palm kernel oil(PKO)in virgin coconut oil(VCO)was quantified using near-infrared(NIR)hyperspectral imaging.Nowadays,some VCO is adulterated with lower-priced PKO to reduce production ...The adulteration concentration of palm kernel oil(PKO)in virgin coconut oil(VCO)was quantified using near-infrared(NIR)hyperspectral imaging.Nowadays,some VCO is adulterated with lower-priced PKO to reduce production costs,which diminishes the quality of the VCO.This study used NIR hyperspectral imaging in the wavelength region 900-1,650 nm to create a quantitative model for the detection of PKO contaminants(0-100%)in VCO and to develop predictive mapping.The prediction equation for the adulteration of VCO with PKO was constructed using the partial least squares regression method.The best predictive model was pre-processed using the standard normal variate method,and the coefficient of determination of prediction was 0.991,the root mean square error of prediction was 2.93%,and the residual prediction deviation was 10.37.The results showed that this model could be applied for quantifying the adulteration concentration of PKO in VCO.The prediction adulteration concentration mapping of VCO with PKO was created from a calibration model that showed the color level according to the adulteration concentration in the range of 0-100%.NIR hyperspectral imaging could be clearly used to quantify the adulteration of VCO with a color level map that provides a quick,accurate,and non-destructive detection method.展开更多
By automatically learning the priors embedded in images with powerful modelling ca-pabilities,deep learning-based algorithms have recently made considerable progress in reconstructing the high-resolution hyperspectral...By automatically learning the priors embedded in images with powerful modelling ca-pabilities,deep learning-based algorithms have recently made considerable progress in reconstructing the high-resolution hyperspectral(HR-HS)image.With previously collected large-amount of external data,these methods are intuitively realised under the full supervision of the ground-truth data.Thus,the database construction in merging the low-resolution(LR)HS(LR-HS)and HR multispectral(MS)or RGB image research paradigm,commonly named as HSI SR,requires collecting corresponding training triplets:HR-MS(RGB),LR-HS and HR-HS image simultaneously,and often faces dif-ficulties in reality.The learned models with the training datasets collected simultaneously under controlled conditions may significantly degrade the HSI super-resolved perfor-mance to the real images captured under diverse environments.To handle the above-mentioned limitations,the authors propose to leverage the deep internal and self-supervised learning to solve the HSI SR problem.The authors advocate that it is possible to train a specific CNN model at test time,called as deep internal learning(DIL),by on-line preparing the training triplet samples from the observed LR-HS/HR-MS(or RGB)images and the down-sampled LR-HS version.However,the number of the training triplets extracted solely from the transformed data of the observation itself is extremely few particularly for the HSI SR tasks with large spatial upscale factors,which would result in limited reconstruction performance.To solve this problem,the authors further exploit deep self-supervised learning(DSL)by considering the observations as the unlabelled training samples.Specifically,the degradation modules inside the network were elaborated to realise the spatial and spectral down-sampling procedures for transforming the generated HR-HS estimation to the high-resolution RGB/LR-HS approximation,and then the reconstruction errors of the observations were formulated for measuring the network modelling performance.By consolidating the DIL and DSL into a unified deep framework,the authors construct a more robust HSI SR method without any prior training and have great potential of flexible adaptation to different settings per obser-vation.To verify the effectiveness of the proposed approach,extensive experiments have been conducted on two benchmark HS datasets,including the CAVE and Harvard datasets,and demonstrate the great performance gain of the proposed method over the state-of-the-art methods.展开更多
Convolutional neural network(CNN)has excellent ability to model locally contextual information.However,CNNs face challenges for descripting long-range semantic features,which will lead to relatively low classification...Convolutional neural network(CNN)has excellent ability to model locally contextual information.However,CNNs face challenges for descripting long-range semantic features,which will lead to relatively low classification accuracy of hyperspectral images.To address this problem,this article proposes an algorithm based on multiscale fusion and transformer network for hyperspectral image classification.Firstly,the low-level spatial-spectral features are extracted by multi-scale residual structure.Secondly,an attention module is introduced to focus on the more important spatialspectral information.Finally,high-level semantic features are represented and learned by a token learner and an improved transformer encoder.The proposed algorithm is compared with six classical hyperspectral classification algorithms on real hyperspectral images.The experimental results show that the proposed algorithm effectively improves the land cover classification accuracy of hyperspectral images.展开更多
The accurate identification of marine oil spills and their emulsions is of great significance for emergency response to oil spill pollution.The selection of characteristic bands with strong separability helps to reali...The accurate identification of marine oil spills and their emulsions is of great significance for emergency response to oil spill pollution.The selection of characteristic bands with strong separability helps to realize the rapid calculation of data on aircraft or in orbit,which will improve the timeliness of oil spill emergency monitoring.At the same time,the combination of spectral and spatial features can improve the accuracy of oil spill monitoring.Two ground-based experiments were designed to collect measured airborne hyperspectral data of crude oil and its emulsions,for which the multiscale superpixel level group clustering framework(MSGCF)was used to select spectral feature bands with strong separability.In addition,the double-branch dual-attention(DBDA)model was applied to identify crude oil and its emulsions.Compared with the recognition results based on original hyperspectral images,using the feature bands determined by MSGCF improved the recognition accuracy,and greatly shortened the running time.Moreover,the characteristic bands for quantifying the volume concentration of water-in-oil emulsions were determined,and a quantitative inversion model was constructed and applied to the AVIRIS image of the deepwater horizon oil spill event in 2010.This study verified the effectiveness of feature bands in identifying oil spill pollution types and quantifying concentration,laying foundation for rapid identification and quantification of marine oil spills and their emulsions on aircraft or in orbit.展开更多
Hyperspectral image classification stands as a pivotal task within the field of remote sensing,yet achieving highprecision classification remains a significant challenge.In response to this challenge,a Spectral Convol...Hyperspectral image classification stands as a pivotal task within the field of remote sensing,yet achieving highprecision classification remains a significant challenge.In response to this challenge,a Spectral Convolutional Neural Network model based on Adaptive Fick’s Law Algorithm(AFLA-SCNN)is proposed.The Adaptive Fick’s Law Algorithm(AFLA)constitutes a novel metaheuristic algorithm introduced herein,encompassing three new strategies:Adaptive weight factor,Gaussian mutation,and probability update policy.With adaptive weight factor,the algorithmcan adjust theweights according to the change in the number of iterations to improve the performance of the algorithm.Gaussianmutation helps the algorithm avoid falling into local optimal solutions and improves the searchability of the algorithm.The probability update strategy helps to improve the exploitability and adaptability of the algorithm.Within the AFLA-SCNN model,AFLA is employed to optimize two hyperparameters in the SCNN model,namely,“numEpochs”and“miniBatchSize”,to attain their optimal values.AFLA’s performance is initially validated across 28 functions in 10D,30D,and 50D for CEC2013 and 29 functions in 10D,30D,and 50D for CEC2017.Experimental results indicate AFLA’s marked performance superiority over nine other prominent optimization algorithms.Subsequently,the AFLA-SCNN model was compared with the Spectral Convolutional Neural Network model based on Fick’s Law Algorithm(FLA-SCNN),Spectral Convolutional Neural Network model based on Harris Hawks Optimization(HHO-SCNN),Spectral Convolutional Neural Network model based onDifferential Evolution(DE-SCNN),SpectralConvolutionalNeuralNetwork(SCNN)model,and SupportVector Machines(SVM)model using the Indian Pines dataset and PaviaUniversity dataset.The experimental results show that the AFLA-SCNN model outperforms other models in terms of Accuracy,Precision,Recall,and F1-score on Indian Pines and Pavia University.Among them,the Accuracy of the AFLA-SCNN model on Indian Pines reached 99.875%,and the Accuracy on PaviaUniversity reached 98.022%.In conclusion,our proposed AFLA-SCNN model is deemed to significantly enhance the precision of hyperspectral image classification.展开更多
A hybrid feature selection and classification strategy was proposed based on the simulated annealing genetic algonthrn and multiple instance learning (MIL). The band selection method was proposed from subspace decom...A hybrid feature selection and classification strategy was proposed based on the simulated annealing genetic algonthrn and multiple instance learning (MIL). The band selection method was proposed from subspace decomposition, which combines the simulated annealing algorithm with the genetic algorithm in choosing different cross-over and mutation probabilities, as well as mutation individuals. Then MIL was combined with image segmentation, clustering and support vector machine algorithms to classify hyperspectral image. The experimental results show that this proposed method can get high classification accuracy of 93.13% at small training samples and the weaknesses of the conventional methods are overcome.展开更多
To compress hyperspectral images, a low complexity discrete cosine transform (DCT)-based distributed source coding (DSC) scheme with Gray code is proposed. Unlike most of the existing DSC schemes, which utilize tr...To compress hyperspectral images, a low complexity discrete cosine transform (DCT)-based distributed source coding (DSC) scheme with Gray code is proposed. Unlike most of the existing DSC schemes, which utilize transform in spatial domain, the proposed algorithm applies transform in spectral domain. Set-partitioning-based approach is applied to reorganize DCT coefficients into waveletlike tree structure and extract the sign, refinement, and significance bitplanes. The extracted refinement bits are Gray encoded. Because of the dependency along the line dimension of hyperspectral images, low density paritycheck-(LDPC)-based Slepian-Wolf coder is adopted to implement the DSC strategy. Experimental results on airborne visible/infrared imaging spectrometer (AVIRIS) dataset show that the proposed paradigm achieves up to 6 dB improvement over DSC-based coders which apply transform in spatial domain, with significantly reduced computational complexity and memory storage.展开更多
A distinguishing characteristic of normal and cancer cells is the difference in their nuclear chromatin content and distribution.This difference can be revealed by the transmission spectra of nuclei stained with a pH-...A distinguishing characteristic of normal and cancer cells is the difference in their nuclear chromatin content and distribution.This difference can be revealed by the transmission spectra of nuclei stained with a pH-sensitive stain.Here,we used hematoxylin-eosin(HE)to stain hepatic carcinoma tissues and obtained spectral-spatial data from their nuclei using hyper-spectral microscopy.The transmission spectra of the nuclei were then used to train a support vector machine(SVM)model for cell classification.Especially,we found that the chromatin distribution in cancer cells is more uniform,because of which the correlation coefficients for the spectra at different points in their nuclei are higher.Consequently,we exploited this feature to improve the SVM model.The sensitivity and specificity for the identification of cancer cells could be increased to 99%and 98%,respectively.We also designed an image-processing method for the extraction of information from cell nuclei to automate the identification process.展开更多
Hyperspectral image super-resolution,which refers to reconstructing the high-resolution hyperspectral image from the input low-resolution observation,aims to improve the spatial resolution of the hyperspectral image,w...Hyperspectral image super-resolution,which refers to reconstructing the high-resolution hyperspectral image from the input low-resolution observation,aims to improve the spatial resolution of the hyperspectral image,which is beneficial for subsequent applications.The development of deep learning has promoted significant progress in hyperspectral image super-resolution,and the powerful expression capabilities of deep neural networks make the predicted results more reliable.Recently,several latest deep learning technologies have made the hyperspectral image super-resolution method explode.However,a comprehensive review and analysis of the latest deep learning methods from the hyperspectral image super-resolution perspective is absent.To this end,in this survey,we first introduce the concept of hyperspectral image super-resolution and classify the methods from the perspectives with or without auxiliary information.Then,we review the learning-based methods in three categories,including single hyperspectral image super-resolution,panchromatic-based hyperspectral image super-resolution,and multispectral-based hyperspectral image super-resolution.Subsequently,we summarize the commonly used hyperspectral dataset,and the evaluations for some representative methods in three categories are performed qualitatively and quantitatively.Moreover,we briefly introduce several typical applications of hyperspectral image super-resolution,including ground object classification,urban change detection,and ecosystem monitoring.Finally,we provide the conclusion and challenges in existing learning-based methods,looking forward to potential future research directions.展开更多
With limited number of labeled samples,hyperspectral image(HSI)classification is a difficult Problem in current research.The graph neural network(GNN)has emerged as an approach to semi-supervised classification,and th...With limited number of labeled samples,hyperspectral image(HSI)classification is a difficult Problem in current research.The graph neural network(GNN)has emerged as an approach to semi-supervised classification,and the application of GNN to hyperspectral images has attracted much attention.However,in the existing GNN-based methods a single graph neural network or graph filter is mainly used to extract HSI features,which does not take full advantage of various graph neural networks(graph filters).Moreover,the traditional GNNs have the problem of oversmoothing.To alleviate these shortcomings,we introduce a deep hybrid multi-graph neural network(DHMG),where two different graph filters,i.e.,the spectral filter and the autoregressive moving average(ARMA)filter,are utilized in two branches.The former can well extract the spectral features of the nodes,and the latter has a good suppression effect on graph noise.The network realizes information interaction between the two branches and takes good advantage of different graph filters.In addition,to address the problem of oversmoothing,a dense network is proposed,where the local graph features are preserved.The dense structure satisfies the needs of different classification targets presenting different features.Finally,we introduce a GraphSAGEbased network to refine the graph features produced by the deep hybrid network.Extensive experiments on three public HSI datasets strongly demonstrate that the DHMG dramatically outperforms the state-ofthe-art models.展开更多
Recently,the autoencoder(AE)based method plays a critical role in the hyperspectral anomaly detection domain.However,due to the strong generalised capacity of AE,the abnormal samples are usually reconstructed well alo...Recently,the autoencoder(AE)based method plays a critical role in the hyperspectral anomaly detection domain.However,due to the strong generalised capacity of AE,the abnormal samples are usually reconstructed well along with the normal background samples.Thus,in order to separate anomalies from the background by calculating reconstruction errors,it can be greatly beneficial to reduce the AE capability for abnormal sample reconstruction while maintaining the background reconstruction performance.A memory‐augmented autoencoder for hyperspectral anomaly detection(MAENet)is proposed to address this challenging problem.Specifically,the proposed MAENet mainly consists of an encoder,a memory module,and a decoder.First,the encoder transforms the original hyperspectral data into the low‐dimensional latent representation.Then,the latent representation is utilised to retrieve the most relevant matrix items in the memory matrix,and the retrieved matrix items will be used to replace the latent representation from the encoder.Finally,the decoder is used to reconstruct the input hyperspectral data using the retrieved memory items.With this strategy,the background can still be reconstructed well while the abnormal samples cannot.Experiments conducted on five real hyperspectral anomaly data sets demonstrate the superiority of the proposed method.展开更多
基金supported by the National Key R&D Program of China(Grant Nos.2021YFB3901403 and 2023YFC3007203).
文摘The deterioration of unstable rock mass raised interest in evaluating rock mass quality.However,the traditional evaluation method for the geological strength index(GSI)primarily emphasizes the rock structure and characteristics of discontinuities.It ignores the influence of mineral composition and shows a deficiency in assessing the integrity coefficient.In this context,hyperspectral imaging and digital panoramic borehole camera technologies are applied to analyze the mineral content and integrity of rock mass.Based on the carbonate mineral content and fissure area ratio,the strength reduction factor and integrity coefficient are calculated to improve the GSI evaluation method.According to the results of mineral classification and fissure identification,the strength reduction factor and integrity coefficient increase with the depth of rock mass.The rock mass GSI calculated by the improved method is mainly concentrated between 40 and 60,which is close to the calculation results of the traditional method.The GSI error rates obtained by the two methods are mostly less than 10%,indicating the rationality of the hyperspectral-digital borehole image coupled evaluation method.Moreover,the sensitivity of the fissure area ratio(Sr)to GSI is greater than that of the strength reduction factor(a),which means the proposed GSI is suitable for rocks with significant fissure development.The improved method reduces the influence of subjective factors and provides a reliable index for the deterioration evaluation of rock mass.
基金National Natural Science Foundation of China(No.62001098)Fundamental Research Funds for the Central Universities of Ministry of Education of China(No.2232020D-33)。
文摘Deep learning(DL)has shown its superior performance in dealing with various computer vision tasks in recent years.As a simple and effective DL model,autoencoder(AE)is popularly used to decompose hyperspectral images(HSIs)due to its powerful ability of feature extraction and data reconstruction.However,most existing AE-based unmixing algorithms usually ignore the spatial information of HSIs.To solve this problem,a hypergraph regularized deep autoencoder(HGAE)is proposed for unmixing.Firstly,the traditional AE architecture is specifically improved as an unsupervised unmixing framework.Secondly,hypergraph learning is employed to reformulate the loss function,which facilitates the expression of high-order similarity among locally neighboring pixels and promotes the consistency of their abundances.Moreover,L_(1/2)norm is further used to enhance abundances sparsity.Finally,the experiments on simulated data,real hyperspectral remote sensing images,and textile cloth images are used to verify that the proposed method can perform better than several state-of-the-art unmixing algorithms.
基金the Deanship of Scientific Research at King Khalid University for funding this work through Large Groups Project under Grant Number(25/43)Princess Nourah bint Abdulrahman University Researchers Supporting Project Number(PNURSP2022R303)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code:22UQU4340237DSR28.
文摘Hyperspectral remote sensing/imaging spectroscopy is a novel approach to reaching a spectrum from all the places of a huge array of spatial places so that several spectral wavelengths are utilized for making coherent images.Hyperspectral remote sensing contains acquisition of digital images from several narrow,contiguous spectral bands throughout the visible,Thermal Infrared(TIR),Near Infrared(NIR),and Mid-Infrared(MIR)regions of the electromagnetic spectrum.In order to the application of agricultural regions,remote sensing approaches are studied and executed to their benefit of continuous and quantitativemonitoring.Particularly,hyperspectral images(HSI)are considered the precise for agriculture as they can offer chemical and physical data on vegetation.With this motivation,this article presents a novel Hurricane Optimization Algorithm with Deep Transfer Learning Driven Crop Classification(HOADTL-CC)model onHyperspectralRemote Sensing Images.The presentedHOADTL-CC model focuses on the identification and categorization of crops on hyperspectral remote sensing images.To accomplish this,the presentedHOADTL-CC model involves the design ofHOAwith capsule network(CapsNet)model for generating a set of useful feature vectors.Besides,Elman neural network(ENN)model is applied to allot proper class labels into the input HSI.Finally,glowworm swarm optimization(GSO)algorithm is exploited to fine tune the ENNparameters involved in this article.The experimental result scrutiny of the HOADTL-CC method can be tested with the help of benchmark dataset and the results are assessed under distinct aspects.Extensive comparative studies stated the enhanced performance of the HOADTL-CC model over recent approaches with maximum accuracy of 99.51%.
文摘Spectral unmixing helps to identify different components present in the spectral mixtures which occur in the uppermost layer of the area owing to the low spatial resolution of hyperspectral images.Most spectral unmixing methods are globally based and do not consider the spectral variability among its endmembers that occur due to illumination,atmospheric,and environmental conditions.Here,endmember bundle extraction plays a major role in overcoming the above-mentioned limitations leading to more accurate abundance fractions.Accordingly,a two-stage approach is proposed to extract endmembers through endmember bundles in hyperspectral images.The divide and conquer method is applied as the first step in subset images with only the non-redundant bands to extract endmembers using the Vertex Component Analysis(VCA)and N-FINDR algorithms.A fuzzy rule-based inference system utilizing spectral matching parameters is proposed in the second step to categorize endmembers.The endmember with the minimum error is chosen as the final endmember in each specific category.The proposed method is simple and automatically considers endmember variability in hyperspectral images.The efficiency of the proposed method is evaluated using two real hyperspectral datasets.The average spectral angle and abundance angle are used to analyze the performance measures.
基金This work was supported by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2023R384)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Hyperspectral imaging instruments could capture detailed spatial information and rich spectral signs of observed scenes.Much spatial information and spectral signatures of hyperspectral images(HSIs)present greater potential for detecting and classifying fine crops.The accurate classification of crop kinds utilizing hyperspectral remote sensing imaging(RSI)has become an indispensable application in the agricultural domain.It is significant for the prediction and growth monitoring of crop yields.Amongst the deep learning(DL)techniques,Convolution Neural Network(CNN)was the best method for classifying HSI for their incredible local contextual modeling ability,enabling spectral and spatial feature extraction.This article designs a Hybrid Multi-Strategy Aquila Optimization with a Deep Learning-Driven Crop Type Classification(HMAODL-CTC)algorithm onHSI.The proposed HMAODL-CTC model mainly intends to categorize different types of crops on HSI.To accomplish this,the presented HMAODL-CTC model initially carries out image preprocessing to improve image quality.In addition,the presented HMAODL-CTC model develops dilated convolutional neural network(CNN)for feature extraction.For hyperparameter tuning of the dilated CNN model,the HMAO algorithm is utilized.Eventually,the presented HMAODL-CTC model uses an extreme learning machine(ELM)model for crop type classification.A comprehensive set of simulations were performed to illustrate the enhanced performance of the presented HMAODL-CTC algorithm.Extensive comparison studies reported the improved performance of the presented HMAODL-CTC algorithm over other compared methods.
文摘Hyperspectral imaging is gaining a significant role in agricultural remote sensing applications.Its data unit is the hyperspectral cube which holds spatial information in two dimensions while spectral band information of each pixel in the third dimension.The classification accuracy of hyperspectral images(HSI)increases significantly by employing both spatial and spectral features.For this work,the data was acquired using an airborne hyperspectral imager system which collected HSI in the visible and near-infrared(VNIR)range of 400 to 1000 nm wavelength within 180 spectral bands.The dataset is collected for nine different crops on agricultural land with a spectral resolution of 3.3 nm wavelength for each pixel.The data was cleaned from geometric distortions and stored with the class labels and annotations of global localization using the inertial navigation system.In this study,a unique pixel-based approach was designed to improve the crops'classification accuracy by using the edge-preserving features(EPF)and principal component analysis(PCA)in conjunction.The preliminary processing generated the high-dimensional EPF stack by applying the edge-preserving filters on acquired HSI.In the second step,this high dimensional stack was treated with the PCA for dimensionality reduction without losing significant spectral information.The resultant feature space(PCA-EPF)demonstrated enhanced class separability for improved crop classification with reduced dimensionality and computational cost.The support vector machines classifier was employed for multiclass classification of target crops using PCA-EPF.The classification performance evaluation was measured in terms of individual class accuracy,overall accuracy,average accuracy,and Cohen kappa factor.The proposed scheme achieved greater than 90%results for all the performance evaluation metrics.The PCA-EPF proved to be an effective attribute for crop classification using hyperspectral imaging in the VNIR range.The proposed scheme is well-suited for practical applications of crops and landfill estimations using agricultural remote sensing methods.
基金supported by the National Natural Science Foundation of China(No.62001023)Beijing Natural Science Foundation(No.JQ20021)。
文摘Most methods for classifying hyperspectral data only consider the local spatial relation-ship among samples,ignoring the important non-local topological relationship.However,the non-local topological relationship is better at representing the structure of hyperspectral data.This paper proposes a deep learning model called Topology and semantic information fusion classification network(TSFnet)that incorporates a topology structure and semantic information transmis-sion network to accurately classify traditional Chinese medicine in hyperspectral images.TSFnet uses a convolutional neural network(CNN)to extract features and a graph convolution network(GCN)to capture potential topological relationships among different types of Chinese herbal medicines.The results show that TSFnet outperforms other state-of-the-art deep learning classification algorithms in two different scenarios of herbal medicine datasets.Additionally,the proposed TSFnet model is lightweight and can be easily deployed for mobile herbal medicine classification.
基金the National Natural Science Foun-dation of China(Nos.61471263,61872267 and U21B2024)the Natural Science Foundation of Tianjin,China(No.16JCZDJC31100)Tianjin University Innovation Foundation(No.2021XZC0024).
文摘Hyperspectral images typically have high spectral resolution but low spatial resolution,which impacts the reliability and accuracy of subsequent applications,for example,remote sensingclassification and mineral identification.But in traditional methods via deep convolution neural net-works,indiscriminately extracting and fusing spectral and spatial features makes it challenging toutilize the differentiated information across adjacent spectral channels.Thus,we proposed a multi-branch interleaved iterative upsampling hyperspectral image super-resolution reconstruction net-work(MIIUSR)to address the above problems.We reinforce spatial feature extraction by integrat-ing detailed features from different receptive fields across adjacent channels.Furthermore,we pro-pose an interleaved iterative upsampling process during the reconstruction stage,which progres-sively fuses incremental information among adjacent frequency bands.Additionally,we add twoparallel three dimensional(3D)feature extraction branches to the backbone network to extractspectral and spatial features of varying granularity.We further enhance the backbone network’sconstruction results by leveraging the difference between two dimensional(2D)channel-groupingspatial features and 3D multi-granularity features.The results obtained by applying the proposednetwork model to the CAVE test set show that,at a scaling factor of×4,the peak signal to noiseratio,spectral angle mapping,and structural similarity are 37.310 dB,3.525 and 0.9438,respec-tively.Besides,extensive experiments conducted on the Harvard and Foster datasets demonstratethe superior potential of the proposed model in hyperspectral super-resolution reconstruction.
基金supported by the UC-National Lab In-Residence Graduate Fellowship Grant L21GF3606supported by a DOD National Defense Science and Engineering Graduate(NDSEG)Research Fellowship+1 种基金supported by the Laboratory Directed Research and Development program of Los Alamos National Laboratory under project numbers 20170668PRD1 and 20210213ERsupported by the NGA under Contract No.HM04762110003.
文摘Graph learning,when used as a semi-supervised learning(SSL)method,performs well for classification tasks with a low label rate.We provide a graph-based batch active learning pipeline for pixel/patch neighborhood multi-or hyperspectral image segmentation.Our batch active learning approach selects a collection of unlabeled pixels that satisfy a graph local maximum constraint for the active learning acquisition function that determines the relative importance of each pixel to the classification.This work builds on recent advances in the design of novel active learning acquisition functions(e.g.,the Model Change approach in arXiv:2110.07739)while adding important further developments including patch-neighborhood image analysis and batch active learning methods to further increase the accuracy and greatly increase the computational efficiency of these methods.In addition to improvements in the accuracy,our approach can greatly reduce the number of labeled pixels needed to achieve the same level of the accuracy based on randomly selected labeled pixels.
基金supported by the Thailand Research Fund through the Royal Golden Jubilee Ph.D.Program(PHD/0225/2561)the Faculty of Engineering,Kamphaeng Saen Campus,Kasetsart University,Thailand。
文摘The adulteration concentration of palm kernel oil(PKO)in virgin coconut oil(VCO)was quantified using near-infrared(NIR)hyperspectral imaging.Nowadays,some VCO is adulterated with lower-priced PKO to reduce production costs,which diminishes the quality of the VCO.This study used NIR hyperspectral imaging in the wavelength region 900-1,650 nm to create a quantitative model for the detection of PKO contaminants(0-100%)in VCO and to develop predictive mapping.The prediction equation for the adulteration of VCO with PKO was constructed using the partial least squares regression method.The best predictive model was pre-processed using the standard normal variate method,and the coefficient of determination of prediction was 0.991,the root mean square error of prediction was 2.93%,and the residual prediction deviation was 10.37.The results showed that this model could be applied for quantifying the adulteration concentration of PKO in VCO.The prediction adulteration concentration mapping of VCO with PKO was created from a calibration model that showed the color level according to the adulteration concentration in the range of 0-100%.NIR hyperspectral imaging could be clearly used to quantify the adulteration of VCO with a color level map that provides a quick,accurate,and non-destructive detection method.
基金Ministry of Education,Culture,Sports,Science and Technology,Grant/Award Number:20K11867。
文摘By automatically learning the priors embedded in images with powerful modelling ca-pabilities,deep learning-based algorithms have recently made considerable progress in reconstructing the high-resolution hyperspectral(HR-HS)image.With previously collected large-amount of external data,these methods are intuitively realised under the full supervision of the ground-truth data.Thus,the database construction in merging the low-resolution(LR)HS(LR-HS)and HR multispectral(MS)or RGB image research paradigm,commonly named as HSI SR,requires collecting corresponding training triplets:HR-MS(RGB),LR-HS and HR-HS image simultaneously,and often faces dif-ficulties in reality.The learned models with the training datasets collected simultaneously under controlled conditions may significantly degrade the HSI super-resolved perfor-mance to the real images captured under diverse environments.To handle the above-mentioned limitations,the authors propose to leverage the deep internal and self-supervised learning to solve the HSI SR problem.The authors advocate that it is possible to train a specific CNN model at test time,called as deep internal learning(DIL),by on-line preparing the training triplet samples from the observed LR-HS/HR-MS(or RGB)images and the down-sampled LR-HS version.However,the number of the training triplets extracted solely from the transformed data of the observation itself is extremely few particularly for the HSI SR tasks with large spatial upscale factors,which would result in limited reconstruction performance.To solve this problem,the authors further exploit deep self-supervised learning(DSL)by considering the observations as the unlabelled training samples.Specifically,the degradation modules inside the network were elaborated to realise the spatial and spectral down-sampling procedures for transforming the generated HR-HS estimation to the high-resolution RGB/LR-HS approximation,and then the reconstruction errors of the observations were formulated for measuring the network modelling performance.By consolidating the DIL and DSL into a unified deep framework,the authors construct a more robust HSI SR method without any prior training and have great potential of flexible adaptation to different settings per obser-vation.To verify the effectiveness of the proposed approach,extensive experiments have been conducted on two benchmark HS datasets,including the CAVE and Harvard datasets,and demonstrate the great performance gain of the proposed method over the state-of-the-art methods.
基金National Natural Science Foundation of China(No.62201457)Natural Science Foundation of Shaanxi Province(Nos.2022JQ-668,2022JQ-588)。
文摘Convolutional neural network(CNN)has excellent ability to model locally contextual information.However,CNNs face challenges for descripting long-range semantic features,which will lead to relatively low classification accuracy of hyperspectral images.To address this problem,this article proposes an algorithm based on multiscale fusion and transformer network for hyperspectral image classification.Firstly,the low-level spatial-spectral features are extracted by multi-scale residual structure.Secondly,an attention module is introduced to focus on the more important spatialspectral information.Finally,high-level semantic features are represented and learned by a token learner and an improved transformer encoder.The proposed algorithm is compared with six classical hyperspectral classification algorithms on real hyperspectral images.The experimental results show that the proposed algorithm effectively improves the land cover classification accuracy of hyperspectral images.
基金Supported by the National Natural Science Foundation of China(Nos.42206177,U1906217)the Shandong Provincial Natural Science Foundation(No.ZR2022QD075)the Fundamental Research Funds for the Central Universities(No.21CX06057A)。
文摘The accurate identification of marine oil spills and their emulsions is of great significance for emergency response to oil spill pollution.The selection of characteristic bands with strong separability helps to realize the rapid calculation of data on aircraft or in orbit,which will improve the timeliness of oil spill emergency monitoring.At the same time,the combination of spectral and spatial features can improve the accuracy of oil spill monitoring.Two ground-based experiments were designed to collect measured airborne hyperspectral data of crude oil and its emulsions,for which the multiscale superpixel level group clustering framework(MSGCF)was used to select spectral feature bands with strong separability.In addition,the double-branch dual-attention(DBDA)model was applied to identify crude oil and its emulsions.Compared with the recognition results based on original hyperspectral images,using the feature bands determined by MSGCF improved the recognition accuracy,and greatly shortened the running time.Moreover,the characteristic bands for quantifying the volume concentration of water-in-oil emulsions were determined,and a quantitative inversion model was constructed and applied to the AVIRIS image of the deepwater horizon oil spill event in 2010.This study verified the effectiveness of feature bands in identifying oil spill pollution types and quantifying concentration,laying foundation for rapid identification and quantification of marine oil spills and their emulsions on aircraft or in orbit.
基金Natural Science Foundation of Shandong Province,China(Grant No.ZR202111230202).
文摘Hyperspectral image classification stands as a pivotal task within the field of remote sensing,yet achieving highprecision classification remains a significant challenge.In response to this challenge,a Spectral Convolutional Neural Network model based on Adaptive Fick’s Law Algorithm(AFLA-SCNN)is proposed.The Adaptive Fick’s Law Algorithm(AFLA)constitutes a novel metaheuristic algorithm introduced herein,encompassing three new strategies:Adaptive weight factor,Gaussian mutation,and probability update policy.With adaptive weight factor,the algorithmcan adjust theweights according to the change in the number of iterations to improve the performance of the algorithm.Gaussianmutation helps the algorithm avoid falling into local optimal solutions and improves the searchability of the algorithm.The probability update strategy helps to improve the exploitability and adaptability of the algorithm.Within the AFLA-SCNN model,AFLA is employed to optimize two hyperparameters in the SCNN model,namely,“numEpochs”and“miniBatchSize”,to attain their optimal values.AFLA’s performance is initially validated across 28 functions in 10D,30D,and 50D for CEC2013 and 29 functions in 10D,30D,and 50D for CEC2017.Experimental results indicate AFLA’s marked performance superiority over nine other prominent optimization algorithms.Subsequently,the AFLA-SCNN model was compared with the Spectral Convolutional Neural Network model based on Fick’s Law Algorithm(FLA-SCNN),Spectral Convolutional Neural Network model based on Harris Hawks Optimization(HHO-SCNN),Spectral Convolutional Neural Network model based onDifferential Evolution(DE-SCNN),SpectralConvolutionalNeuralNetwork(SCNN)model,and SupportVector Machines(SVM)model using the Indian Pines dataset and PaviaUniversity dataset.The experimental results show that the AFLA-SCNN model outperforms other models in terms of Accuracy,Precision,Recall,and F1-score on Indian Pines and Pavia University.Among them,the Accuracy of the AFLA-SCNN model on Indian Pines reached 99.875%,and the Accuracy on PaviaUniversity reached 98.022%.In conclusion,our proposed AFLA-SCNN model is deemed to significantly enhance the precision of hyperspectral image classification.
文摘A hybrid feature selection and classification strategy was proposed based on the simulated annealing genetic algonthrn and multiple instance learning (MIL). The band selection method was proposed from subspace decomposition, which combines the simulated annealing algorithm with the genetic algorithm in choosing different cross-over and mutation probabilities, as well as mutation individuals. Then MIL was combined with image segmentation, clustering and support vector machine algorithms to classify hyperspectral image. The experimental results show that this proposed method can get high classification accuracy of 93.13% at small training samples and the weaknesses of the conventional methods are overcome.
基金supported by the National Natural Science Foundationof China (60702012)the Scientific Research Foundation for the Re-turned Overseas Chinese Scholars, State Education Ministry
文摘To compress hyperspectral images, a low complexity discrete cosine transform (DCT)-based distributed source coding (DSC) scheme with Gray code is proposed. Unlike most of the existing DSC schemes, which utilize transform in spatial domain, the proposed algorithm applies transform in spectral domain. Set-partitioning-based approach is applied to reorganize DCT coefficients into waveletlike tree structure and extract the sign, refinement, and significance bitplanes. The extracted refinement bits are Gray encoded. Because of the dependency along the line dimension of hyperspectral images, low density paritycheck-(LDPC)-based Slepian-Wolf coder is adopted to implement the DSC strategy. Experimental results on airborne visible/infrared imaging spectrometer (AVIRIS) dataset show that the proposed paradigm achieves up to 6 dB improvement over DSC-based coders which apply transform in spatial domain, with significantly reduced computational complexity and memory storage.
基金This paper was supported by the National Key Research and Development Program of China(2017YFB1104500)National Natural Science Foundation of China(61605062,61735005 and 11704155)+2 种基金Science and Technology Planning Project of Guangdong Province(2018B030323017)Research Project of Scientific Research Cultivation and Innovation Fund of Jinan University(11617329)Guangzhou Science and Technology Project(201903010042 and 201904010294).
文摘A distinguishing characteristic of normal and cancer cells is the difference in their nuclear chromatin content and distribution.This difference can be revealed by the transmission spectra of nuclei stained with a pH-sensitive stain.Here,we used hematoxylin-eosin(HE)to stain hepatic carcinoma tissues and obtained spectral-spatial data from their nuclei using hyper-spectral microscopy.The transmission spectra of the nuclei were then used to train a support vector machine(SVM)model for cell classification.Especially,we found that the chromatin distribution in cancer cells is more uniform,because of which the correlation coefficients for the spectra at different points in their nuclei are higher.Consequently,we exploited this feature to improve the SVM model.The sensitivity and specificity for the identification of cancer cells could be increased to 99%and 98%,respectively.We also designed an image-processing method for the extraction of information from cell nuclei to automate the identification process.
基金supported in part by the National Natural Science Foundation of China(62276192)。
文摘Hyperspectral image super-resolution,which refers to reconstructing the high-resolution hyperspectral image from the input low-resolution observation,aims to improve the spatial resolution of the hyperspectral image,which is beneficial for subsequent applications.The development of deep learning has promoted significant progress in hyperspectral image super-resolution,and the powerful expression capabilities of deep neural networks make the predicted results more reliable.Recently,several latest deep learning technologies have made the hyperspectral image super-resolution method explode.However,a comprehensive review and analysis of the latest deep learning methods from the hyperspectral image super-resolution perspective is absent.To this end,in this survey,we first introduce the concept of hyperspectral image super-resolution and classify the methods from the perspectives with or without auxiliary information.Then,we review the learning-based methods in three categories,including single hyperspectral image super-resolution,panchromatic-based hyperspectral image super-resolution,and multispectral-based hyperspectral image super-resolution.Subsequently,we summarize the commonly used hyperspectral dataset,and the evaluations for some representative methods in three categories are performed qualitatively and quantitatively.Moreover,we briefly introduce several typical applications of hyperspectral image super-resolution,including ground object classification,urban change detection,and ecosystem monitoring.Finally,we provide the conclusion and challenges in existing learning-based methods,looking forward to potential future research directions.
文摘With limited number of labeled samples,hyperspectral image(HSI)classification is a difficult Problem in current research.The graph neural network(GNN)has emerged as an approach to semi-supervised classification,and the application of GNN to hyperspectral images has attracted much attention.However,in the existing GNN-based methods a single graph neural network or graph filter is mainly used to extract HSI features,which does not take full advantage of various graph neural networks(graph filters).Moreover,the traditional GNNs have the problem of oversmoothing.To alleviate these shortcomings,we introduce a deep hybrid multi-graph neural network(DHMG),where two different graph filters,i.e.,the spectral filter and the autoregressive moving average(ARMA)filter,are utilized in two branches.The former can well extract the spectral features of the nodes,and the latter has a good suppression effect on graph noise.The network realizes information interaction between the two branches and takes good advantage of different graph filters.In addition,to address the problem of oversmoothing,a dense network is proposed,where the local graph features are preserved.The dense structure satisfies the needs of different classification targets presenting different features.Finally,we introduce a GraphSAGEbased network to refine the graph features produced by the deep hybrid network.Extensive experiments on three public HSI datasets strongly demonstrate that the DHMG dramatically outperforms the state-ofthe-art models.
基金supported in part by the National Natural Science Foundation of China under Grant 62076199in part by the Open Research Fund of Beijing Key Laboratory of Big Data Technology for Food Safety under Grant BTBD‐2020KF08Beijing Technology and Business University,and in part by the Key R&D project of Shaan'xi Province under Grant 2021GY‐027 and 2022ZDLGY01‐03.
文摘Recently,the autoencoder(AE)based method plays a critical role in the hyperspectral anomaly detection domain.However,due to the strong generalised capacity of AE,the abnormal samples are usually reconstructed well along with the normal background samples.Thus,in order to separate anomalies from the background by calculating reconstruction errors,it can be greatly beneficial to reduce the AE capability for abnormal sample reconstruction while maintaining the background reconstruction performance.A memory‐augmented autoencoder for hyperspectral anomaly detection(MAENet)is proposed to address this challenging problem.Specifically,the proposed MAENet mainly consists of an encoder,a memory module,and a decoder.First,the encoder transforms the original hyperspectral data into the low‐dimensional latent representation.Then,the latent representation is utilised to retrieve the most relevant matrix items in the memory matrix,and the retrieved matrix items will be used to replace the latent representation from the encoder.Finally,the decoder is used to reconstruct the input hyperspectral data using the retrieved memory items.With this strategy,the background can still be reconstructed well while the abnormal samples cannot.Experiments conducted on five real hyperspectral anomaly data sets demonstrate the superiority of the proposed method.