Hyperspectral imagery encompasses spectral and spatial dimensions,reflecting the material properties of objects.Its application proves crucial in search and rescue,concealed target identification,and crop growth analy...Hyperspectral imagery encompasses spectral and spatial dimensions,reflecting the material properties of objects.Its application proves crucial in search and rescue,concealed target identification,and crop growth analysis.Clustering is an important method of hyperspectral analysis.The vast data volume of hyperspectral imagery,coupled with redundant information,poses significant challenges in swiftly and accurately extracting features for subsequent analysis.The current hyperspectral feature clustering methods,which are mostly studied from space or spectrum,do not have strong interpretability,resulting in poor comprehensibility of the algorithm.So,this research introduces a feature clustering algorithm for hyperspectral imagery from an interpretability perspective.It commences with a simulated perception process,proposing an interpretable band selection algorithm to reduce data dimensions.Following this,amulti-dimensional clustering algorithm,rooted in fuzzy and kernel clustering,is developed to highlight intra-class similarities and inter-class differences.An optimized P systemis then introduced to enhance computational efficiency.This system coordinates all cells within a mapping space to compute optimal cluster centers,facilitating parallel computation.This approach diminishes sensitivity to initial cluster centers and augments global search capabilities,thus preventing entrapment in local minima and enhancing clustering performance.Experiments conducted on 300 datasets,comprising both real and simulated data.The results show that the average accuracy(ACC)of the proposed algorithm is 0.86 and the combination measure(CM)is 0.81.展开更多
The interpretability of deep learning models has emerged as a compelling area in artificial intelligence research.The safety criteria for medical imaging are highly stringent,and models are required for an explanation...The interpretability of deep learning models has emerged as a compelling area in artificial intelligence research.The safety criteria for medical imaging are highly stringent,and models are required for an explanation.However,existing convolutional neural network solutions for left ventricular segmentation are viewed in terms of inputs and outputs.Thus,the interpretability of CNNs has come into the spotlight.Since medical imaging data are limited,many methods to fine-tune medical imaging models that are popular in transfer models have been built using massive public Image Net datasets by the transfer learning method.Unfortunately,this generates many unreliable parameters and makes it difficult to generate plausible explanations from these models.In this study,we trained from scratch rather than relying on transfer learning,creating a novel interpretable approach for autonomously segmenting the left ventricle with a cardiac MRI.Our enhanced GPU training system implemented interpretable global average pooling for graphics using deep learning.The deep learning tasks were simplified.Simplification included data management,neural network architecture,and training.Our system monitored and analyzed the gradient changes of different layers with dynamic visualizations in real-time and selected the optimal deployment model.Our results demonstrated that the proposed method was feasible and efficient:the Dice coefficient reached 94.48%,and the accuracy reached 99.7%.It was found that no current transfer learning models could perform comparably to the ImageNet transfer learning architectures.This model is lightweight and more convenient to deploy on mobile devices than transfer learning models.展开更多
An algorithm named InterOpt for optimizing operational parameters is proposed based on interpretable machine learning,and is demonstrated via optimization of shale gas development.InterOpt consists of three parts:a ne...An algorithm named InterOpt for optimizing operational parameters is proposed based on interpretable machine learning,and is demonstrated via optimization of shale gas development.InterOpt consists of three parts:a neural network is used to construct an emulator of the actual drilling and hydraulic fracturing process in the vector space(i.e.,virtual environment);:the Sharpley value method in inter-pretable machine learning is applied to analyzing the impact of geological and operational parameters in each well(i.e.,single well feature impact analysis):and ensemble randomized maximum likelihood(EnRML)is conducted to optimize the operational parameters to comprehensively improve the efficiency of shale gas development and reduce the average cost.In the experiment,InterOpt provides different drilling and fracturing plans for each well according to its specific geological conditions,and finally achieves an average cost reduction of 9.7%for a case study with 104 wells.展开更多
Defining the structure characteristics of amorphous materials is one of the fundamental problems that need to be solved urgently in complex materials because of their complex structure and long-range disorder.In this ...Defining the structure characteristics of amorphous materials is one of the fundamental problems that need to be solved urgently in complex materials because of their complex structure and long-range disorder.In this study,we develop an interpretable deep learning model capable of accurately classifying amorphous configurations and characterizing their structural properties.The results demonstrate that the multi-dimensional hybrid convolutional neural network can classify the two-dimensional(2D)liquids and amorphous solids of molecular dynamics simulation.The classification process does not make a priori assumptions on the amorphous particle environment,and the accuracy is 92.75%,which is better than other convolutional neural networks.Moreover,our model utilizes the gradient-weighted activation-like mapping method,which generates activation-like heat maps that can precisely identify important structures in the amorphous configuration maps.We obtain an order parameter from the heatmap and conduct finite scale analysis of this parameter.Our findings demonstrate that the order parameter effectively captures the amorphous phase transition process across various systems.These results hold significant scientific implications for the study of amorphous structural characteristics via deep learning.展开更多
Thermoelectric and thermal materials are essential in achieving carbon neutrality. However, the high cost of lattice thermal conductivity calculations and the limited applicability of classical physical models have le...Thermoelectric and thermal materials are essential in achieving carbon neutrality. However, the high cost of lattice thermal conductivity calculations and the limited applicability of classical physical models have led to the inefficient development of thermoelectric materials. In this study, we proposed a two-stage machine learning framework with physical interpretability incorporating domain knowledge to calculate high/low thermal conductivity rapidly. Specifically, crystal graph convolutional neural network(CGCNN) is constructed to predict the fundamental physical parameters related to lattice thermal conductivity. Based on the above physical parameters, an interpretable machine learning model–sure independence screening and sparsifying operator(SISSO), is trained to predict the lattice thermal conductivity. We have predicted the lattice thermal conductivity of all available materials in the open quantum materials database(OQMD)(https://www.oqmd.org/). The proposed approach guides the next step of searching for materials with ultra-high or ultralow lattice thermal conductivity and promotes the development of new thermal insulation materials and thermoelectric materials.展开更多
The prediction of processor performance has important referencesignificance for future processors. Both the accuracy and rationality of theprediction results are required. The hierarchical belief rule base (HBRB)can i...The prediction of processor performance has important referencesignificance for future processors. Both the accuracy and rationality of theprediction results are required. The hierarchical belief rule base (HBRB)can initially provide a solution to low prediction accuracy. However, theinterpretability of the model and the traceability of the results still warrantfurther investigation. Therefore, a processor performance prediction methodbased on interpretable hierarchical belief rule base (HBRB-I) and globalsensitivity analysis (GSA) is proposed. The method can yield more reliableprediction results. Evidence reasoning (ER) is firstly used to evaluate thehistorical data of the processor, followed by a performance prediction modelwith interpretability constraints that is constructed based on HBRB-I. Then,the whale optimization algorithm (WOA) is used to optimize the parameters.Furthermore, to test the interpretability of the performance predictionprocess, GSA is used to analyze the relationship between the input and thepredicted output indicators. Finally, based on the UCI database processordataset, the effectiveness and superiority of the method are verified. Accordingto our experiments, our prediction method generates more reliable andaccurate estimations than traditional models.展开更多
A liquid launch vehicle is an important carrier in aviation,and its regular operation is essential to maintain space security.In the safety assessment of fluid launch vehicle body structure,it is necessary to ensure t...A liquid launch vehicle is an important carrier in aviation,and its regular operation is essential to maintain space security.In the safety assessment of fluid launch vehicle body structure,it is necessary to ensure that the assessmentmodel can learn self-response rules from various uncertain data and not differently to provide a traceable and interpretable assessment process.Therefore,a belief rule base with interpretability(BRB-i)assessment method of liquid launch vehicle structure safety status combines data and knowledge.Moreover,an innovative whale optimization algorithm with interpretable constraints is proposed.The experiments are carried out based on the liquid launch vehicle safety experiment platform,and the information on the safety status of the liquid launch vehicle is obtained by monitoring the detection indicators under the simulation platform.The MSEs of the proposed model are 3.8000e-03,1.3000e-03,2.1000e-03,and 1.8936e-04 for 25%,45%,65%,and 84%of the training samples,respectively.It can be seen that the proposed model also shows a better ability to handle small sample data.Meanwhile,the belief distribution of the BRB-i model output has a high fitting trend with the belief distribution of the expert knowledge settings,which indicates the interpretability of the BRB-i model.Experimental results show that,compared with other methods,the BRB-i model guarantees the model’s interpretability and the high precision of experimental results.展开更多
Major issues currently restricting the use of learning analytics are the lack of interpretability and adaptability of the machine learning models used in this domain.Interpretability makes it easy for the stakeholders...Major issues currently restricting the use of learning analytics are the lack of interpretability and adaptability of the machine learning models used in this domain.Interpretability makes it easy for the stakeholders to understand the working of these models and adaptability makes it easy to use the same model for multiple cohorts and courses in educational institutions.Recently,some models in learning analytics are constructed with the consideration of interpretability but their interpretability is not quantified.However,adaptability is not specifically considered in this domain.This paper presents a new framework based on hybrid statistical fuzzy theory to overcome these limitations.It also provides explainability in the form of rules describing the reasoning behind a particular output.The paper also discusses the system evaluation on a benchmark dataset showing promising results.The measure of explainability,fuzzy index,shows that the model is highly interpretable.This system achieves more than 82%recall in both the classification and the context adaptation stages.展开更多
If we have gone through the first generation of housing design that?pursued functional optimization, ergonomics, and circulation efficiency during the last century, now we are living in the second generation where mor...If we have gone through the first generation of housing design that?pursued functional optimization, ergonomics, and circulation efficiency during the last century, now we are living in the second generation where more advanced goals, such as universal design, ubiquitous design, sustainable design, and environment-friendly design, are emphasized. Although this second generation of design focuses upon the wellness of humans in accordance with environment, it still has the attitude that a more precisely designed home can guarantee a better life. What lacks in this approach is the freedom of the body;it needs to make its own choice as to how to use a space. Thus, it is suggested in this paper that what is important in designing a home is to provide alternatives in daily lives so as to make a full exploration of a given space. These alternatives can be made by offering residents an interpretable space where they can figure out space usages and routs in a constantly changing context. Two spatial devices are discussed in depths as a way to realize this interpretable house: room-to-room enfilade and ring spatial structure. By investigating some existing house plans, it is illustrated how they can guarantee the freedom of the body, and thus alternatives for the flexible domestic life.展开更多
Roof falls due to geological conditions are major hazards in the mining industry,causing work time loss,injuries,and fatalities.There are roof fall problems caused by high horizontal stress in several largeopening lim...Roof falls due to geological conditions are major hazards in the mining industry,causing work time loss,injuries,and fatalities.There are roof fall problems caused by high horizontal stress in several largeopening limestone mines in the eastern and midwestern United States.The typical hazard management approach for this type of roof fall hazards relies heavily on visual inspections and expert knowledge.In this context,we proposed a deep learning system for detection of the roof fall hazards caused by high horizontal stress.We used images depicting hazardous and non-hazardous roof conditions to develop a convolutional neural network(CNN)for autonomous detection of hazardous roof conditions.To compensate for limited input data,we utilized a transfer learning approach.In the transfer learning approach,an already-trained network is used as a starting point for classification in a similar domain.Results show that this approach works well for classifying roof conditions as hazardous or safe,achieving a statistical accuracy of 86.4%.This result is also compared with a random forest classifier,and the deep learning approach is more successful at classification of roof conditions.However,accuracy alone is not enough to ensure a reliable hazard management system.System constraints and reliability are improved when the features used by the network are understood.Therefore,we used a deep learning interpretation technique called integrated gradients to identify the important geological features in each image for prediction.The analysis of integrated gradients shows that the system uses the same roof features as the experts do on roof fall hazards detection.The system developed in this paper demonstrates the potential of deep learning in geotechnical hazard management to complement human experts,and likely to become an essential part of autonomous operations in cases where hazard identification heavily depends on expert knowledge.Moreover,deep learning-based systems reduce expert exposure to hazardous conditions.展开更多
The existence of time delay in complex industrial processes or dynamical systems is a common phenomenon and is a difficult problem to deal with in industrial control systems,as well as in the textile field.Accurate id...The existence of time delay in complex industrial processes or dynamical systems is a common phenomenon and is a difficult problem to deal with in industrial control systems,as well as in the textile field.Accurate identification of the time delay can greatly improve the efficiency of the design of industrial process control systems.The time delay identification methods based on mathematical modeling require prior knowledge of the structural information of the model,especially for nonlinear systems.The neural network-based identification method can predict the time delay of the system,but cannot accurately obtain the specific parameters of the time delay.Benefit from the interpretability of machine learning,a novel method for delay identification based on an interpretable regression decision tree is proposed.Utilizing the self-explanatory analysis of the decision tree model,the parameters with the highest feature importance are obtained to identify the time delay of the system.Excellent results are gained by the simulation data of linear and nonlinear control systems,and the time delay of the systems can be accurately identified.展开更多
Arrhythmias may lead to sudden cardiac death if not detected and treated in time.A supraventricular premature beat(SPB)and premature ventricular contraction(PVC)are important categories of arrhythmia disease.Recently,...Arrhythmias may lead to sudden cardiac death if not detected and treated in time.A supraventricular premature beat(SPB)and premature ventricular contraction(PVC)are important categories of arrhythmia disease.Recently,deep learning methods have been applied to the PVC/SPB heartbeats detection.However,most researchers have focused on time-domain information of the electrocardiogram and there has been a lack of exploration of the interpretability of the model.In this study,we design an interpretable and accurate PVC/SPB recognition algorithm,called the interpretable multilevel wavelet decomposition deep network(IMWDDN).Wavelet decomposition is introduced into the deep network and the squeeze and excitation(SE)-Residual block is designed for extracting time-domain and frequency-domain features.Additionally,inspired by the idea of residual learning,we construct a novel loss function for the constant updating of the multilevel wavelet decomposition parameters.Finally,the IMWDDN is evaluated on the Third China Physiological Signal Challenge Dataset and the MIT-BIH Arrhythmia database.The comparison results show IMWDDN has better detection performance with 98.51%accuracy and a 93.75%F1-macro on average,and its areas of concern are similar to those of an expert diagnosis to a certain extent.Generally,the IMWDDN has good application value in the clinical screening of PVC/SPB heartbeats.展开更多
Facing the escalating effects of climate change,it is critical to improve the prediction and understanding of the hurricane evacuation decisions made by households in order to enhance emergency management.Current stud...Facing the escalating effects of climate change,it is critical to improve the prediction and understanding of the hurricane evacuation decisions made by households in order to enhance emergency management.Current studies in this area often have relied on psychology-driven linear models,which frequently exhibited limitations in practice.The present study proposed a novel interpretable machine learning approach to predict household-level evacuation decisions by leveraging easily accessible demographic and resource-related predictors,compared to existing models that mainly rely on psychological factors.An enhanced logistic regression model(that is,an interpretable machine learning approach) was developed for accurate predictions by automatically accounting for nonlinearities and interactions(that is,univariate and bivariate threshold effects).Specifically,nonlinearity and interaction detection were enabled by low-depth decision trees,which offer transparent model structure and robustness.A survey dataset collected in the aftermath of Hurricanes Katrina and Rita,two of the most intense tropical storms of the last two decades,was employed to test the new methodology.The findings show that,when predicting the households’ evacuation decisions,the enhanced logistic regression model outperformed previous linear models in terms of both model fit and predictive capability.This outcome suggests that our proposed methodology could provide a new tool and framework for emergency management authorities to improve the prediction of evacuation traffic demands in a timely and accurate manner.展开更多
Amidst the inefficiencies of traditional job-seeking approaches in the recruitment ecosystem, the importance of automated job recommendation systems has been magnified. However, existing models optimized to maximize u...Amidst the inefficiencies of traditional job-seeking approaches in the recruitment ecosystem, the importance of automated job recommendation systems has been magnified. However, existing models optimized to maximize user clicks for general product recommendations prove inept in addressing the unique challenges of job recommendation, namely reciprocity and competition. Moreover, sparse data on online recruitment platforms can further negatively impact the performance of existing job recommendation algorithms. To counteract these limitations, we propose a bilateral heterogeneous graph-based competition iteration model. This model comprises three integral components: 1) two bilateral heterogeneous graphs for capturing multi-source information from people and jobs and alleviating data sparsity, 2) fusion strategies for synthesizing attributes and preferences to produce mutually beneficial job matches, and 3) a competition-enhancing strategy for dispersing competition realized through a two-stage optimization algorithm. Augmented by granular attention mechanisms for enhanced interpretability, the model’s efficacy, competition dispersion, and interpretability are validated through rigorous empirical evaluations on a real-world recruitment platform.展开更多
To equip data-driven dynamic chemical process models with strong interpretability,we develop a light attention–convolution–gate recurrent unit(LACG)architecture with three sub-modules—a basic module,a brand-new lig...To equip data-driven dynamic chemical process models with strong interpretability,we develop a light attention–convolution–gate recurrent unit(LACG)architecture with three sub-modules—a basic module,a brand-new light attention module,and a residue module—that are specially designed to learn the general dynamic behavior,transient disturbances,and other input factors of chemical processes,respectively.Combined with a hyperparameter optimization framework,Optuna,the effectiveness of the proposed LACG is tested by distributed control system data-driven modeling experiments on the discharge flowrate of an actual deethanization process.The LACG model provides significant advantages in prediction accuracy and model generalization compared with other models,including the feedforward neural network,convolution neural network,long short-term memory(LSTM),and attention-LSTM.Moreover,compared with the simulation results of a deethanization model built using Aspen Plus Dynamics V12.1,the LACG parameters are demonstrated to be interpretable,and more details on the variable interactions can be observed from the model parameters in comparison with the traditional interpretable model attention-LSTM.This contribution enriches interpretable machine learning knowledge and provides a reliable method with high accuracy for actual chemical process modeling,paving a route to intelligent manufacturing.展开更多
The present study extracts human-understandable insights from machine learning(ML)-based mesoscale closure in fluid-particle flows via several novel data-driven analysis approaches,i.e.,maximal information coefficient...The present study extracts human-understandable insights from machine learning(ML)-based mesoscale closure in fluid-particle flows via several novel data-driven analysis approaches,i.e.,maximal information coefficient(MIC),interpretable ML,and automated ML.It is previously shown that the solidvolume fraction has the greatest effect on the drag force.The present study aims to quantitativelyinvestigate the influence of flow properties on mesoscale drag correction(H_(d)).The MIC results showstrong correlations between the features(i.e.,slip velocity(u^(*)_(sy))and particle volume fraction(εs))and thelabel H_(d).The interpretable ML analysis confirms this conclusion,and quantifies the contribution of u^(*)_(sy),εs and gas pressure gradient to the model as 71.9%,27.2%and 0.9%,respectively.Automated ML without theneed to select the model structure and hyperparameters is used for modeling,improving the predictionaccuracy over our previous model(Zhu et al.,2020;Ouyang,Zhu,Su,&Luo,2021).展开更多
The identification of factors that may be forcing ecological observations to approach the upper boundary provides insight into potential mechanisms affecting driver-response relationships,and can help inform ecosystem...The identification of factors that may be forcing ecological observations to approach the upper boundary provides insight into potential mechanisms affecting driver-response relationships,and can help inform ecosystem management,but has rarely been explored.In this study,we propose a novel framework integrating quantile regression with interpretable machine learning.In the first stage of the framework,we estimate the upper boundary of a driver-response relationship using quantile regression.Next,we calculate“potentials”of the response variable depending on the driver,which are defined as vertical distances from the estimated upper boundary of the relationship to observations in the driver-response variable scatter plot.Finally,we identify key factors impacting the potential using a machine learning model.We illustrate the necessary steps to implement the framework using the total phosphorus(TP)-Chlorophyll a(CHL)relationship in lakes across the continental US.We found that the nitrogen to phosphorus ratio(N:P),annual average precipitation,total nitrogen(TN),and summer average air temperature were key factors impacting the potential of CHL depending on TP.We further revealed important implications of our findings for lake eutrophication management.The important role of N:P and TN on the potential highlights the co-limitation of phosphorus and nitrogen and indicates the need for dual nutrient criteria.Future wetter and/or warmer climate scenarios can decrease the potential which may reduce the efficacy of lake eutrophication management.The novel framework advances the application of quantile regression to identify factors driving observations to approach the upper boundary of driver-response relationships.展开更多
Objective:To validate two proposed coronavirus disease 2019(COVID-19)prognosis models,analyze the characteristics of different models,consider the performance of models in predicting different outcomes,and provide new...Objective:To validate two proposed coronavirus disease 2019(COVID-19)prognosis models,analyze the characteristics of different models,consider the performance of models in predicting different outcomes,and provide new insights into the development and use of artificial intelligence(AI)predictive models in clinical decision-making for COVID-19 and other diseases.Materials and Methods:We compared two proposed prediction models for COVID-19 prognosis that use a decision tree and logistic regression modeling.We evaluated the effectiveness of different model-building strategies using laboratory tests and/or clinical record data,their sensitivity and robustness to the timings of records used and the presence of missing data,and their predictive performance and capabilities in single-site and multicenter settings.Results:The predictive accuracies of the two models after retraining were improved to 93.2% and 93.9%,compared with that of the models directly used,with accuracies of 84.3% and 87.9%,indicating that the prediction models could not be used directly and require retraining based on actual data.In addition,based on the prediction model,new features obtained by model comparison and literature evidence were transferred to integrate the new models with better performance.Conclusions:Comparing the characteristics and differences of datasets used in model training,effective model verification,and a fusion of models is necessary in improving the performance of AI models.展开更多
Near real-time spatial prediction of earthquake-induced landslides(EQILs)can rapidly forecast the occurrence position of widespread landslides just after a violent earthquake;thus,EQIL prediction is very crucial to th...Near real-time spatial prediction of earthquake-induced landslides(EQILs)can rapidly forecast the occurrence position of widespread landslides just after a violent earthquake;thus,EQIL prediction is very crucial to the 72-hour‘golden window’for survivors.This work focuses on a series of earthquake events from 2008 to 2022 occurring in the Tibetan Plateau,a famous seismically-active zone,and proposes a novel interpretable self-supervised learning(ISeL)method for the near real-time spatial prediction of EQILs.This new method innovatively introduces swap noise at the unsupervised mechanism,which can improve the generalization performance and transferability of the model,and can effectively reduce false alarm and improve accuracy through supervisedfine-tuning.An interpretable module is built based on a self-attention mechanism to reveal the importance and contribution of various influencing factors to EQIL spatial distribution.Experimental results demonstrate that the ISeL model is superior to the excellent state-of-the-art machine learning and deep learning methods.Furthermore,according to the interpretable module in the ISeL method,the critical controlling and triggering factors are revealed.The ISeL method can also be applied in other earthquake-frequent regions worldwide because of its good generalization and transferability.展开更多
It is vital to establish an interpretable fault diagnosis model for critical equipment.Belief Rule Base(BRB)is an interpretable expert system gradually applied in fault diagnosis.However,the expert knowledge cannot be...It is vital to establish an interpretable fault diagnosis model for critical equipment.Belief Rule Base(BRB)is an interpretable expert system gradually applied in fault diagnosis.However,the expert knowledge cannot be utilized to establish the initial BRB accurately if there are multiple referential grades in different fault features.In addition,the interpretability of BRB-based fault diagnosis is destroyed in the optimization process,which reflects in two aspects:deviation from the initial expert judgment and over-optimization of parameters.To solve these problems,a new interpretable fault diagnosis model based on BRB and probability table,called the BRB-P,is proposed in this paper.Compared with the traditional BRB,the BRB-P constructed by the probability table is more accurate.Then,the interpretability constraints,i.e.,the credibility of expert knowledge,the penalty factor and the rule-activation factor,are inserted into the projection covariance matrix adaption evolution strategy to maintain the interpretability of BRB-P.A case study of the aerospace relay is conducted to verify the effectiveness of the proposed method.展开更多
基金Yulin Science and Technology Bureau production Project“Research on Smart Agricultural Product Traceability System”(No.CXY-2022-64)Light of West China(No.XAB2022YN10)+1 种基金The China Postdoctoral Science Foundation(No.2023M740760)Shaanxi Province Key Research and Development Plan(No.2024SF-YBXM-678).
文摘Hyperspectral imagery encompasses spectral and spatial dimensions,reflecting the material properties of objects.Its application proves crucial in search and rescue,concealed target identification,and crop growth analysis.Clustering is an important method of hyperspectral analysis.The vast data volume of hyperspectral imagery,coupled with redundant information,poses significant challenges in swiftly and accurately extracting features for subsequent analysis.The current hyperspectral feature clustering methods,which are mostly studied from space or spectrum,do not have strong interpretability,resulting in poor comprehensibility of the algorithm.So,this research introduces a feature clustering algorithm for hyperspectral imagery from an interpretability perspective.It commences with a simulated perception process,proposing an interpretable band selection algorithm to reduce data dimensions.Following this,amulti-dimensional clustering algorithm,rooted in fuzzy and kernel clustering,is developed to highlight intra-class similarities and inter-class differences.An optimized P systemis then introduced to enhance computational efficiency.This system coordinates all cells within a mapping space to compute optimal cluster centers,facilitating parallel computation.This approach diminishes sensitivity to initial cluster centers and augments global search capabilities,thus preventing entrapment in local minima and enhancing clustering performance.Experiments conducted on 300 datasets,comprising both real and simulated data.The results show that the average accuracy(ACC)of the proposed algorithm is 0.86 and the combination measure(CM)is 0.81.
基金The National Natural Science Foundation of China (62176048)provided funding for this research.
文摘The interpretability of deep learning models has emerged as a compelling area in artificial intelligence research.The safety criteria for medical imaging are highly stringent,and models are required for an explanation.However,existing convolutional neural network solutions for left ventricular segmentation are viewed in terms of inputs and outputs.Thus,the interpretability of CNNs has come into the spotlight.Since medical imaging data are limited,many methods to fine-tune medical imaging models that are popular in transfer models have been built using massive public Image Net datasets by the transfer learning method.Unfortunately,this generates many unreliable parameters and makes it difficult to generate plausible explanations from these models.In this study,we trained from scratch rather than relying on transfer learning,creating a novel interpretable approach for autonomously segmenting the left ventricle with a cardiac MRI.Our enhanced GPU training system implemented interpretable global average pooling for graphics using deep learning.The deep learning tasks were simplified.Simplification included data management,neural network architecture,and training.Our system monitored and analyzed the gradient changes of different layers with dynamic visualizations in real-time and selected the optimal deployment model.Our results demonstrated that the proposed method was feasible and efficient:the Dice coefficient reached 94.48%,and the accuracy reached 99.7%.It was found that no current transfer learning models could perform comparably to the ImageNet transfer learning architectures.This model is lightweight and more convenient to deploy on mobile devices than transfer learning models.
文摘An algorithm named InterOpt for optimizing operational parameters is proposed based on interpretable machine learning,and is demonstrated via optimization of shale gas development.InterOpt consists of three parts:a neural network is used to construct an emulator of the actual drilling and hydraulic fracturing process in the vector space(i.e.,virtual environment);:the Sharpley value method in inter-pretable machine learning is applied to analyzing the impact of geological and operational parameters in each well(i.e.,single well feature impact analysis):and ensemble randomized maximum likelihood(EnRML)is conducted to optimize the operational parameters to comprehensively improve the efficiency of shale gas development and reduce the average cost.In the experiment,InterOpt provides different drilling and fracturing plans for each well according to its specific geological conditions,and finally achieves an average cost reduction of 9.7%for a case study with 104 wells.
基金National Natural Science Foundation of China(Grant No.11702289)the Key Core Technology and Generic Technology Research and Development Project of Shanxi Province,China(Grant No.2020XXX013)the National Key Research and Development Project of China。
文摘Defining the structure characteristics of amorphous materials is one of the fundamental problems that need to be solved urgently in complex materials because of their complex structure and long-range disorder.In this study,we develop an interpretable deep learning model capable of accurately classifying amorphous configurations and characterizing their structural properties.The results demonstrate that the multi-dimensional hybrid convolutional neural network can classify the two-dimensional(2D)liquids and amorphous solids of molecular dynamics simulation.The classification process does not make a priori assumptions on the amorphous particle environment,and the accuracy is 92.75%,which is better than other convolutional neural networks.Moreover,our model utilizes the gradient-weighted activation-like mapping method,which generates activation-like heat maps that can precisely identify important structures in the amorphous configuration maps.We obtain an order parameter from the heatmap and conduct finite scale analysis of this parameter.Our findings demonstrate that the order parameter effectively captures the amorphous phase transition process across various systems.These results hold significant scientific implications for the study of amorphous structural characteristics via deep learning.
基金support of the National Natural Science Foundation of China(Grant Nos.12104356 and52250191)China Postdoctoral Science Foundation(Grant No.2022M712552)+2 种基金the Opening Project of Shanghai Key Laboratory of Special Artificial Microstructure Materials and Technology(Grant No.Ammt2022B-1)the Fundamental Research Funds for the Central Universitiessupport by HPC Platform,Xi’an Jiaotong University。
文摘Thermoelectric and thermal materials are essential in achieving carbon neutrality. However, the high cost of lattice thermal conductivity calculations and the limited applicability of classical physical models have led to the inefficient development of thermoelectric materials. In this study, we proposed a two-stage machine learning framework with physical interpretability incorporating domain knowledge to calculate high/low thermal conductivity rapidly. Specifically, crystal graph convolutional neural network(CGCNN) is constructed to predict the fundamental physical parameters related to lattice thermal conductivity. Based on the above physical parameters, an interpretable machine learning model–sure independence screening and sparsifying operator(SISSO), is trained to predict the lattice thermal conductivity. We have predicted the lattice thermal conductivity of all available materials in the open quantum materials database(OQMD)(https://www.oqmd.org/). The proposed approach guides the next step of searching for materials with ultra-high or ultralow lattice thermal conductivity and promotes the development of new thermal insulation materials and thermoelectric materials.
基金This work is supported in part by the Postdoctoral Science Foundation of China under Grant No.2020M683736in part by the Teaching reform project of higher education in Heilongjiang Province under Grant No.SJGY20210456in part by the Natural Science Foundation of Heilongjiang Province of China under Grant No.LH2021F038.
文摘The prediction of processor performance has important referencesignificance for future processors. Both the accuracy and rationality of theprediction results are required. The hierarchical belief rule base (HBRB)can initially provide a solution to low prediction accuracy. However, theinterpretability of the model and the traceability of the results still warrantfurther investigation. Therefore, a processor performance prediction methodbased on interpretable hierarchical belief rule base (HBRB-I) and globalsensitivity analysis (GSA) is proposed. The method can yield more reliableprediction results. Evidence reasoning (ER) is firstly used to evaluate thehistorical data of the processor, followed by a performance prediction modelwith interpretability constraints that is constructed based on HBRB-I. Then,the whale optimization algorithm (WOA) is used to optimize the parameters.Furthermore, to test the interpretability of the performance predictionprocess, GSA is used to analyze the relationship between the input and thepredicted output indicators. Finally, based on the UCI database processordataset, the effectiveness and superiority of the method are verified. Accordingto our experiments, our prediction method generates more reliable andaccurate estimations than traditional models.
基金This work was supported in part by the Natural Science Foundation of China under Grant 62203461 and Grant 62203365in part by the Postdoctoral Science Foundation of China under Grant No.2020M683736,in part by the Teaching Reform Project of Higher Education in Heilongjiang Province under Grant Nos.SJGY20210456 and SJGY20210457in part by the Natural Science Foundation of Heilongjiang Province of China under Grant No.LH2021F038,and in part by the Graduate Academic Innovation Project of Harbin Normal University under Grant Nos.HSDSSCX2022-17,HSDSSCX2022-18 and HSDSSCX2022-19。
文摘A liquid launch vehicle is an important carrier in aviation,and its regular operation is essential to maintain space security.In the safety assessment of fluid launch vehicle body structure,it is necessary to ensure that the assessmentmodel can learn self-response rules from various uncertain data and not differently to provide a traceable and interpretable assessment process.Therefore,a belief rule base with interpretability(BRB-i)assessment method of liquid launch vehicle structure safety status combines data and knowledge.Moreover,an innovative whale optimization algorithm with interpretable constraints is proposed.The experiments are carried out based on the liquid launch vehicle safety experiment platform,and the information on the safety status of the liquid launch vehicle is obtained by monitoring the detection indicators under the simulation platform.The MSEs of the proposed model are 3.8000e-03,1.3000e-03,2.1000e-03,and 1.8936e-04 for 25%,45%,65%,and 84%of the training samples,respectively.It can be seen that the proposed model also shows a better ability to handle small sample data.Meanwhile,the belief distribution of the BRB-i model output has a high fitting trend with the belief distribution of the expert knowledge settings,which indicates the interpretability of the BRB-i model.Experimental results show that,compared with other methods,the BRB-i model guarantees the model’s interpretability and the high precision of experimental results.
文摘Major issues currently restricting the use of learning analytics are the lack of interpretability and adaptability of the machine learning models used in this domain.Interpretability makes it easy for the stakeholders to understand the working of these models and adaptability makes it easy to use the same model for multiple cohorts and courses in educational institutions.Recently,some models in learning analytics are constructed with the consideration of interpretability but their interpretability is not quantified.However,adaptability is not specifically considered in this domain.This paper presents a new framework based on hybrid statistical fuzzy theory to overcome these limitations.It also provides explainability in the form of rules describing the reasoning behind a particular output.The paper also discusses the system evaluation on a benchmark dataset showing promising results.The measure of explainability,fuzzy index,shows that the model is highly interpretable.This system achieves more than 82%recall in both the classification and the context adaptation stages.
文摘If we have gone through the first generation of housing design that?pursued functional optimization, ergonomics, and circulation efficiency during the last century, now we are living in the second generation where more advanced goals, such as universal design, ubiquitous design, sustainable design, and environment-friendly design, are emphasized. Although this second generation of design focuses upon the wellness of humans in accordance with environment, it still has the attitude that a more precisely designed home can guarantee a better life. What lacks in this approach is the freedom of the body;it needs to make its own choice as to how to use a space. Thus, it is suggested in this paper that what is important in designing a home is to provide alternatives in daily lives so as to make a full exploration of a given space. These alternatives can be made by offering residents an interpretable space where they can figure out space usages and routs in a constantly changing context. Two spatial devices are discussed in depths as a way to realize this interpretable house: room-to-room enfilade and ring spatial structure. By investigating some existing house plans, it is illustrated how they can guarantee the freedom of the body, and thus alternatives for the flexible domestic life.
基金partially supported by the National Institute for Occupational Safety and Health,contract number 0000HCCR-2019-36403。
文摘Roof falls due to geological conditions are major hazards in the mining industry,causing work time loss,injuries,and fatalities.There are roof fall problems caused by high horizontal stress in several largeopening limestone mines in the eastern and midwestern United States.The typical hazard management approach for this type of roof fall hazards relies heavily on visual inspections and expert knowledge.In this context,we proposed a deep learning system for detection of the roof fall hazards caused by high horizontal stress.We used images depicting hazardous and non-hazardous roof conditions to develop a convolutional neural network(CNN)for autonomous detection of hazardous roof conditions.To compensate for limited input data,we utilized a transfer learning approach.In the transfer learning approach,an already-trained network is used as a starting point for classification in a similar domain.Results show that this approach works well for classifying roof conditions as hazardous or safe,achieving a statistical accuracy of 86.4%.This result is also compared with a random forest classifier,and the deep learning approach is more successful at classification of roof conditions.However,accuracy alone is not enough to ensure a reliable hazard management system.System constraints and reliability are improved when the features used by the network are understood.Therefore,we used a deep learning interpretation technique called integrated gradients to identify the important geological features in each image for prediction.The analysis of integrated gradients shows that the system uses the same roof features as the experts do on roof fall hazards detection.The system developed in this paper demonstrates the potential of deep learning in geotechnical hazard management to complement human experts,and likely to become an essential part of autonomous operations in cases where hazard identification heavily depends on expert knowledge.Moreover,deep learning-based systems reduce expert exposure to hazardous conditions.
基金Shanghai Philosophy and Social Science Program,China(No.2019BGL004)。
文摘The existence of time delay in complex industrial processes or dynamical systems is a common phenomenon and is a difficult problem to deal with in industrial control systems,as well as in the textile field.Accurate identification of the time delay can greatly improve the efficiency of the design of industrial process control systems.The time delay identification methods based on mathematical modeling require prior knowledge of the structural information of the model,especially for nonlinear systems.The neural network-based identification method can predict the time delay of the system,but cannot accurately obtain the specific parameters of the time delay.Benefit from the interpretability of machine learning,a novel method for delay identification based on an interpretable regression decision tree is proposed.Utilizing the self-explanatory analysis of the decision tree model,the parameters with the highest feature importance are obtained to identify the time delay of the system.Excellent results are gained by the simulation data of linear and nonlinear control systems,and the time delay of the systems can be accurately identified.
基金supported by the National Postdoctoral Program for Innovative Talents(Grant No.BX20230215)China Postdoctoral Science Foundation(Grant No.2023M732219)Shanghai Municipal Science and Technology Major Project(Grant No.2021SHZDZX0102)。
文摘Arrhythmias may lead to sudden cardiac death if not detected and treated in time.A supraventricular premature beat(SPB)and premature ventricular contraction(PVC)are important categories of arrhythmia disease.Recently,deep learning methods have been applied to the PVC/SPB heartbeats detection.However,most researchers have focused on time-domain information of the electrocardiogram and there has been a lack of exploration of the interpretability of the model.In this study,we design an interpretable and accurate PVC/SPB recognition algorithm,called the interpretable multilevel wavelet decomposition deep network(IMWDDN).Wavelet decomposition is introduced into the deep network and the squeeze and excitation(SE)-Residual block is designed for extracting time-domain and frequency-domain features.Additionally,inspired by the idea of residual learning,we construct a novel loss function for the constant updating of the multilevel wavelet decomposition parameters.Finally,the IMWDDN is evaluated on the Third China Physiological Signal Challenge Dataset and the MIT-BIH Arrhythmia database.The comparison results show IMWDDN has better detection performance with 98.51%accuracy and a 93.75%F1-macro on average,and its areas of concern are similar to those of an expert diagnosis to a certain extent.Generally,the IMWDDN has good application value in the clinical screening of PVC/SPB heartbeats.
基金supported by the National Science Foundation under Grant Nos.2303578,2303579, 05 27699,0838654,and 1212790by an Early-Career Research Fellowship from the Gulf Research Program of the National Academies of Sciences,Engineering,and Medicine
文摘Facing the escalating effects of climate change,it is critical to improve the prediction and understanding of the hurricane evacuation decisions made by households in order to enhance emergency management.Current studies in this area often have relied on psychology-driven linear models,which frequently exhibited limitations in practice.The present study proposed a novel interpretable machine learning approach to predict household-level evacuation decisions by leveraging easily accessible demographic and resource-related predictors,compared to existing models that mainly rely on psychological factors.An enhanced logistic regression model(that is,an interpretable machine learning approach) was developed for accurate predictions by automatically accounting for nonlinearities and interactions(that is,univariate and bivariate threshold effects).Specifically,nonlinearity and interaction detection were enabled by low-depth decision trees,which offer transparent model structure and robustness.A survey dataset collected in the aftermath of Hurricanes Katrina and Rita,two of the most intense tropical storms of the last two decades,was employed to test the new methodology.The findings show that,when predicting the households’ evacuation decisions,the enhanced logistic regression model outperformed previous linear models in terms of both model fit and predictive capability.This outcome suggests that our proposed methodology could provide a new tool and framework for emergency management authorities to improve the prediction of evacuation traffic demands in a timely and accurate manner.
基金supported by the National Natural Science Foundation of China(Grant No.72172070).
文摘Amidst the inefficiencies of traditional job-seeking approaches in the recruitment ecosystem, the importance of automated job recommendation systems has been magnified. However, existing models optimized to maximize user clicks for general product recommendations prove inept in addressing the unique challenges of job recommendation, namely reciprocity and competition. Moreover, sparse data on online recruitment platforms can further negatively impact the performance of existing job recommendation algorithms. To counteract these limitations, we propose a bilateral heterogeneous graph-based competition iteration model. This model comprises three integral components: 1) two bilateral heterogeneous graphs for capturing multi-source information from people and jobs and alleviating data sparsity, 2) fusion strategies for synthesizing attributes and preferences to produce mutually beneficial job matches, and 3) a competition-enhancing strategy for dispersing competition realized through a two-stage optimization algorithm. Augmented by granular attention mechanisms for enhanced interpretability, the model’s efficacy, competition dispersion, and interpretability are validated through rigorous empirical evaluations on a real-world recruitment platform.
基金support provided by the National Natural Science Foundation of China(22122802,22278044,and 21878028)the Chongqing Science Fund for Distinguished Young Scholars(CSTB2022NSCQ-JQX0021)the Fundamental Research Funds for the Central Universities(2022CDJXY-003).
文摘To equip data-driven dynamic chemical process models with strong interpretability,we develop a light attention–convolution–gate recurrent unit(LACG)architecture with three sub-modules—a basic module,a brand-new light attention module,and a residue module—that are specially designed to learn the general dynamic behavior,transient disturbances,and other input factors of chemical processes,respectively.Combined with a hyperparameter optimization framework,Optuna,the effectiveness of the proposed LACG is tested by distributed control system data-driven modeling experiments on the discharge flowrate of an actual deethanization process.The LACG model provides significant advantages in prediction accuracy and model generalization compared with other models,including the feedforward neural network,convolution neural network,long short-term memory(LSTM),and attention-LSTM.Moreover,compared with the simulation results of a deethanization model built using Aspen Plus Dynamics V12.1,the LACG parameters are demonstrated to be interpretable,and more details on the variable interactions can be observed from the model parameters in comparison with the traditional interpretable model attention-LSTM.This contribution enriches interpretable machine learning knowledge and provides a reliable method with high accuracy for actual chemical process modeling,paving a route to intelligent manufacturing.
基金This work was supported by the National Natural ScienceFoundation of China(No.U1862201,91834303 and 22208208)the China Postdoctoral Science Foundation(No.2022M712056)the China National Postdoctoral Program for Innovative Talents(No.BX20220205).
文摘The present study extracts human-understandable insights from machine learning(ML)-based mesoscale closure in fluid-particle flows via several novel data-driven analysis approaches,i.e.,maximal information coefficient(MIC),interpretable ML,and automated ML.It is previously shown that the solidvolume fraction has the greatest effect on the drag force.The present study aims to quantitativelyinvestigate the influence of flow properties on mesoscale drag correction(H_(d)).The MIC results showstrong correlations between the features(i.e.,slip velocity(u^(*)_(sy))and particle volume fraction(εs))and thelabel H_(d).The interpretable ML analysis confirms this conclusion,and quantifies the contribution of u^(*)_(sy),εs and gas pressure gradient to the model as 71.9%,27.2%and 0.9%,respectively.Automated ML without theneed to select the model structure and hyperparameters is used for modeling,improving the predictionaccuracy over our previous model(Zhu et al.,2020;Ouyang,Zhu,Su,&Luo,2021).
基金This research was funded by the National Natural Science Foundation of China(Nos.71761147001 and 42030707)the International Partnership Program by the Chinese Academy of Sciences(No.121311KYSB20190029)+2 种基金the Fundamental Research Fund for the Central Universities(No.20720210083)the National Science Foundation(Nos.EF-1638679,EF-1638554,EF-1638539,and EF-1638550)Any use of trade,firm,or product names is for descriptive purposes only and does not imply endorsement by the US Government.
文摘The identification of factors that may be forcing ecological observations to approach the upper boundary provides insight into potential mechanisms affecting driver-response relationships,and can help inform ecosystem management,but has rarely been explored.In this study,we propose a novel framework integrating quantile regression with interpretable machine learning.In the first stage of the framework,we estimate the upper boundary of a driver-response relationship using quantile regression.Next,we calculate“potentials”of the response variable depending on the driver,which are defined as vertical distances from the estimated upper boundary of the relationship to observations in the driver-response variable scatter plot.Finally,we identify key factors impacting the potential using a machine learning model.We illustrate the necessary steps to implement the framework using the total phosphorus(TP)-Chlorophyll a(CHL)relationship in lakes across the continental US.We found that the nitrogen to phosphorus ratio(N:P),annual average precipitation,total nitrogen(TN),and summer average air temperature were key factors impacting the potential of CHL depending on TP.We further revealed important implications of our findings for lake eutrophication management.The important role of N:P and TN on the potential highlights the co-limitation of phosphorus and nitrogen and indicates the need for dual nutrient criteria.Future wetter and/or warmer climate scenarios can decrease the potential which may reduce the efficacy of lake eutrophication management.The novel framework advances the application of quantile regression to identify factors driving observations to approach the upper boundary of driver-response relationships.
基金financially supported by the Natural Science Foundation of Beijing(No.M21012)National Natural Science Foundation of China(No.82174533)Key Technologies R and D Program of the China Academy of Chinese Medical Sciences(No.CI2021A00920).
文摘Objective:To validate two proposed coronavirus disease 2019(COVID-19)prognosis models,analyze the characteristics of different models,consider the performance of models in predicting different outcomes,and provide new insights into the development and use of artificial intelligence(AI)predictive models in clinical decision-making for COVID-19 and other diseases.Materials and Methods:We compared two proposed prediction models for COVID-19 prognosis that use a decision tree and logistic regression modeling.We evaluated the effectiveness of different model-building strategies using laboratory tests and/or clinical record data,their sensitivity and robustness to the timings of records used and the presence of missing data,and their predictive performance and capabilities in single-site and multicenter settings.Results:The predictive accuracies of the two models after retraining were improved to 93.2% and 93.9%,compared with that of the models directly used,with accuracies of 84.3% and 87.9%,indicating that the prediction models could not be used directly and require retraining based on actual data.In addition,based on the prediction model,new features obtained by model comparison and literature evidence were transferred to integrate the new models with better performance.Conclusions:Comparing the characteristics and differences of datasets used in model training,effective model verification,and a fusion of models is necessary in improving the performance of AI models.
基金funded by the National Natural Science Foundation of China(U21A2013,71874165)Opening Fund of Key Laboratory of Geological Survey and Evaluation of Ministry of Education[Grant Nos.GLAB2020ZR02,GLAB2022ZR02]+2 种基金State Key Laboratory of Biogeology and Environmental Geology[grant number GBL12107]the Fundamental Research Funds for the Central Universities,China University of Geosciences(Wuhan)[CUG2642022006]Hunan Provincial Natural Science Foundation of China[2021JC0009].
文摘Near real-time spatial prediction of earthquake-induced landslides(EQILs)can rapidly forecast the occurrence position of widespread landslides just after a violent earthquake;thus,EQIL prediction is very crucial to the 72-hour‘golden window’for survivors.This work focuses on a series of earthquake events from 2008 to 2022 occurring in the Tibetan Plateau,a famous seismically-active zone,and proposes a novel interpretable self-supervised learning(ISeL)method for the near real-time spatial prediction of EQILs.This new method innovatively introduces swap noise at the unsupervised mechanism,which can improve the generalization performance and transferability of the model,and can effectively reduce false alarm and improve accuracy through supervisedfine-tuning.An interpretable module is built based on a self-attention mechanism to reveal the importance and contribution of various influencing factors to EQIL spatial distribution.Experimental results demonstrate that the ISeL model is superior to the excellent state-of-the-art machine learning and deep learning methods.Furthermore,according to the interpretable module in the ISeL method,the critical controlling and triggering factors are revealed.The ISeL method can also be applied in other earthquake-frequent regions worldwide because of its good generalization and transferability.
基金supported by the National Natural Science Foundation of China(No.61833016)the Shaanxi Outstanding Youth Science Foundation,China(No.2020JC-34)+1 种基金the Shaanxi Science and Technology Innovation Team,China(No.2022TD-24)the Natural Science Foundation of Heilongjiang Province of China(No.LH2021F038)。
文摘It is vital to establish an interpretable fault diagnosis model for critical equipment.Belief Rule Base(BRB)is an interpretable expert system gradually applied in fault diagnosis.However,the expert knowledge cannot be utilized to establish the initial BRB accurately if there are multiple referential grades in different fault features.In addition,the interpretability of BRB-based fault diagnosis is destroyed in the optimization process,which reflects in two aspects:deviation from the initial expert judgment and over-optimization of parameters.To solve these problems,a new interpretable fault diagnosis model based on BRB and probability table,called the BRB-P,is proposed in this paper.Compared with the traditional BRB,the BRB-P constructed by the probability table is more accurate.Then,the interpretability constraints,i.e.,the credibility of expert knowledge,the penalty factor and the rule-activation factor,are inserted into the projection covariance matrix adaption evolution strategy to maintain the interpretability of BRB-P.A case study of the aerospace relay is conducted to verify the effectiveness of the proposed method.