The importance and complexity of prioritizing construction projects (PCP) in urban road network planning lead to the necessity to develop an aided decision making program (ADMP). Cost benefit ratio model and stage rol...The importance and complexity of prioritizing construction projects (PCP) in urban road network planning lead to the necessity to develop an aided decision making program (ADMP). Cost benefit ratio model and stage rolled method are chosen as the theoretical foundations of the program, and then benefit model is improved to accord with the actuality of urban traffic in China. Consequently, program flows, module functions and data structures are designed, and particularly an original data structure of road ...展开更多
In order to facilitate spare parts management,an integrated approach of BP neural network and supportability analysis(SA)was proposed to evaluate the criticality of spare parts as well as to prioritize spare parts.Inf...In order to facilitate spare parts management,an integrated approach of BP neural network and supportability analysis(SA)was proposed to evaluate the criticality of spare parts as well as to prioritize spare parts.Influential factors of prioritizing spare parts were detailedly analyzed.Framework of the integrated method was established.The modelling process based on BP neural network was presented.As the input of the neural network,the values of influential factors were determined by supportability analysis data.Based on the presented method,spare parts could be automatically prioritized after supportability analysis for a new system.A case study results showed that the new method was applicable and effective.展开更多
Digital forensics aims to uncover evidence of cybercrimes within compromised systems.These cybercrimes are often perpetrated through the deployment of malware,which inevitably leaves discernible traces within the comp...Digital forensics aims to uncover evidence of cybercrimes within compromised systems.These cybercrimes are often perpetrated through the deployment of malware,which inevitably leaves discernible traces within the compromised systems.Forensic analysts are tasked with extracting and subsequently analyzing data,termed as artifacts,from these systems to gather evidence.Therefore,forensic analysts must sift through extensive datasets to isolate pertinent evidence.However,manually identifying suspicious traces among numerous artifacts is time-consuming and labor-intensive.Previous studies addressed such inefficiencies by integrating artificial intelligence(AI)technologies into digital forensics.Despite the efforts in previous studies,artifacts were analyzed without considering the nature of the data within them and failed to prove their efficiency through specific evaluations.In this study,we propose a system to prioritize suspicious artifacts from compromised systems infected with malware to facilitate efficient digital forensics.Our system introduces a double-checking method that recognizes the nature of data within target artifacts and employs algorithms ideal for anomaly detection.The key ideas of this method are:(1)prioritize suspicious artifacts and filter remaining artifacts using autoencoder and(2)further prioritize suspicious artifacts and filter remaining artifacts using logarithmic entropy.Our evaluation demonstrates that our system can identify malicious artifacts with high accuracy and that its double-checking method is more efficient than alternative approaches.Our system can significantly reduce the time required for forensic analysis and serve as a reference for future studies.展开更多
Edge devices,due to their limited computational and storage resources,often require the use of compilers for program optimization.Therefore,ensuring the security and reliability of these compilers is of paramount impo...Edge devices,due to their limited computational and storage resources,often require the use of compilers for program optimization.Therefore,ensuring the security and reliability of these compilers is of paramount importance in the emerging field of edge AI.One widely used testing method for this purpose is fuzz testing,which detects bugs by inputting random test cases into the target program.However,this process consumes significant time and resources.To improve the efficiency of compiler fuzz testing,it is common practice to utilize test case prioritization techniques.Some researchers use machine learning to predict the code coverage of test cases,aiming to maximize the test capability for the target compiler by increasing the overall predicted coverage of the test cases.Nevertheless,these methods can only forecast the code coverage of the compiler at a specific optimization level,potentially missing many optimization-related bugs.In this paper,we introduce C-CORE(short for Clustering by Code Representation),the first framework to prioritize test cases according to their code representations,which are derived directly from the source codes.This approach avoids being limited to specific compiler states and extends to a broader range of compiler bugs.Specifically,we first train a scaled pre-trained programming language model to capture as many common features as possible from the test cases generated by a fuzzer.Using this pre-trained model,we then train two downstream models:one for predicting the likelihood of triggering a bug and another for identifying code representations associated with bugs.Subsequently,we cluster the test cases according to their code representations and select the highest-scoring test case from each cluster as the high-quality test case.This reduction in redundant testing cases leads to time savings.Comprehensive evaluation results reveal that code representations are better at distinguishing test capabilities,and C-CORE significantly enhances testing efficiency.Across four datasets,C-CORE increases the average of the percentage of faults detected(APFD)value by 0.16 to 0.31 and reduces test time by over 50% in 46% of cases.When compared to the best results from approaches using predicted code coverage,C-CORE improves the APFD value by 1.1% to 12.3% and achieves an overall time-saving of 159.1%.展开更多
China’s employment landscape looks sparse right now as many private enterprises that absorb much of the country’s workforce are succumbing to economic stagnation. While the Western world is caught in a tailspin, num...China’s employment landscape looks sparse right now as many private enterprises that absorb much of the country’s workforce are succumbing to economic stagnation. While the Western world is caught in a tailspin, numerous private enterprises, and exporters in particular, have no place to hide. Their orders are declining to a trickle, and the supply chains of many also have broken down because of a lack of cash. In such a time of gloom, how can they weather the storm through innovation and industrial upgrading? Beijing Review reporter Hu Yue interviewed several economists and entrepreneurs about this issue at the 2008 Dialogue Between Chinese Private Enterprises and Global Fortune 500 held last December in Wenzhou, Zhejiang Province.展开更多
Mobile applications usually can only access limited amount of memory. Improper use of the memory can cause memory leaks, which may lead to performance slowdowns or even cause applications to be unexpectedly killed. Al...Mobile applications usually can only access limited amount of memory. Improper use of the memory can cause memory leaks, which may lead to performance slowdowns or even cause applications to be unexpectedly killed. Although a large body of research has been devoted into the memory leak diagnosing techniques after leaks have been discovered, it is still challenging to find out the memory leak phenomena at first. Testing is the most widely used technique for failure discovery. However, traditional testing techniques are not directed for the discovery of memory leaks. They may spend lots of time on testing unlikely leaking executions and therefore can be inefficient. To address the problem, we propose a novel approach to prioritize test cases according to their likelihood to cause memory leaks in a given test suite. It firstly builds a prediction model to determine whether each test can potentially lead to memory leaks based on machine learning on selected code features. Then, for each input test case, we partly run it to get its code features and predict its likelihood to cause leaks. The most suspicious test cases will be suggested to run at first in order to reveal memory leak faults as soon as possible. Experimental evaluation on several Android applications shows that our approach is effective.展开更多
With the continuing development and improvement of genome-wide techniques, a great number of candidate genes are discovered. How to identify the most likely disease genes among a large number of candidates becomes a f...With the continuing development and improvement of genome-wide techniques, a great number of candidate genes are discovered. How to identify the most likely disease genes among a large number of candidates becomes a fundamental challenge in human health. A common view is that genes related to a specific or similar disease tend to reside in the same neighbourhood of biomolecular networks. Recently, based on such observations,many methods have been developed to tackle this challenge. In this review, we firstly introduce the concept of disease genes, their properties, and available data for identifying them. Then we review the recent computational approaches for prioritizing candidate disease genes based on Protein-Protein Interaction(PPI) networks and investigate their advantages and disadvantages. Furthermore, some pieces of existing software and network resources are summarized. Finally, we discuss key issues in prioritizing candidate disease genes and point out some future research directions.展开更多
With recent advances in genotyping and sequencing technologies,many disease susceptibility loci have been identified.However,much of the genetic heritability remains unexplained and the replication rate between indepe...With recent advances in genotyping and sequencing technologies,many disease susceptibility loci have been identified.However,much of the genetic heritability remains unexplained and the replication rate between independent studies is still low.Meanwhile,there have been increasing efforts on functional annotations of the entire human genome,such as the Encyclopedia of DNA Elements(ENCODE)project and other similar projects.It has been shown that incorporating these functional annotations to prioritize genome wide association signals may help identify true association signals.However,to our knowledge,the extent of the improvement when functional annotation data are considered has not been studied in the literature.In this article,we propose a statistical framework to estimate the improvement in replication rate with annotation data,and apply it to Crohn’s disease and DNase I hypersensitive sites.The results show that with cell line specific functional annotations,the expected replication rate is improved,but only at modest level.展开更多
Test Case Prioritization(TCP)techniques perform better than other regression test optimization techniques including Test Suite Reduction(TSR)and Test Case Selection(TCS).Many TCP techniques are available,and their per...Test Case Prioritization(TCP)techniques perform better than other regression test optimization techniques including Test Suite Reduction(TSR)and Test Case Selection(TCS).Many TCP techniques are available,and their performance is usually measured through a metric Average Percentage of Fault Detection(APFD).This metric is value-neutral because it only works well when all test cases have the same cost,and all faults have the same severity.Using APFD for performance evaluation of test case orders where test cases cost or faults severity varies is prone to produce false results.Therefore,using the right metric for performance evaluation of TCP techniques is very important to get reliable and correct results.In this paper,two value-based TCP techniques have been introduced using Genetic Algorithm(GA)including Value-Cognizant Fault Detection-Based TCP(VCFDB-TCP)and Value-Cognizant Requirements Coverage-Based TCP(VCRCB-TCP).Two novel value-based performance evaluation metrics are also introduced for value-based TCP including Average Percentage of Fault Detection per value(APFDv)and Average Percentage of Requirements Coverage per value(APRCv).Two case studies are performed to validate proposed techniques and performance evaluation metrics.The proposed GA-based techniques outperformed the existing state-of-the-art TCP techniques including Original Order(OO),Reverse Order(REV-O),Random Order(RO),and Greedy algorithm.展开更多
Although the construction of underground dams is one of the best methods to conserve water resources in arid and semi-arid regions,applying efficient methods for the selection of suitable sites for subsurface dam cons...Although the construction of underground dams is one of the best methods to conserve water resources in arid and semi-arid regions,applying efficient methods for the selection of suitable sites for subsurface dam construction remains a challenge.Due to the costly and time-consuming methods of site selection for underground dam construction,this study aimed to present a new method using geographic information systems techniques and decision-making processes.The exclusionary criteria including fault,slope,hypsometry,land use,soil,stream,geology,and chemical properties of groundwater were selected for site selection of dam construction and inappropriate regions were omitted by integration and scoring layers in ArcGIS based on the Boolean logic.Finally,appropriate sites were prioritized using the Multi-Attribute Utility Theory.According to the results of the utility coefficient,seven sites were selected as the region for underground dam construction based on all criteria and experts’opinions.The site of Nazarabad dam was the best location for underground dam construction with a utility coefficient of 0.7137 followed by sites of Akhavan with a utility coefficient of 0.4633 and Mirshamsi with a utility coefficient of 0.4083.This study proposed a new approach for the construction of the subsurface dam at the proper site and help managers and decision-makers achieve sustainable water resources with limited facilities and capital and avoid wasting national capital.展开更多
Software needs modifications and requires revisions regularly.Owing to these revisions,retesting software becomes essential to ensure that the enhancements made,have not affected its bug-free functioning.The time and ...Software needs modifications and requires revisions regularly.Owing to these revisions,retesting software becomes essential to ensure that the enhancements made,have not affected its bug-free functioning.The time and cost incurred in this process,need to be reduced by the method of test case selection and prioritization.It is observed that many nature-inspired techniques are applied in this area.African Buffalo Optimization is one such approach,applied to regression test selection and prioritization.In this paper,the proposed work explains and proves the applicability of the African Buffalo Optimization approach to test case selection and prioritization.The proposed algorithm converges in polynomial time(O(n^(2))).In this paper,the empirical evaluation of applying African Buffalo Optimization for test case prioritization is done on sample data set with multiple iterations.An astounding 62.5%drop in size and a 48.57%drop in the runtime of the original test suite were recorded.The obtained results are compared with Ant Colony Optimization.The comparative analysis indicates that African Buffalo Optimization and Ant Colony Optimization exhibit similar fault detection capabilities(80%),and a reduction in the overall execution time and size of the resultant test suite.The results and analysis,hence,advocate and encourages the use of African Buffalo Optimization in the area of test case selection and prioritization.展开更多
Regression testing is a widely used approach to confirm the correct functionality of the software in incremental development.The use of test cases makes it easier to test the ripple effect of changed requirements.Rigo...Regression testing is a widely used approach to confirm the correct functionality of the software in incremental development.The use of test cases makes it easier to test the ripple effect of changed requirements.Rigorous testingmay help in meeting the quality criteria that is based on the conformance to the requirements as given by the intended stakeholders.However,a minimized and prioritized set of test cases may reduce the efforts and time required for testingwhile focusing on the timely delivery of the software application.In this research,a technique named Test Reduce has been presented to get a minimal set of test cases based on high priority to ensure that the web applicationmeets the required quality criteria.A new technique TestReduce is proposed with a blend of genetic algorithm to find an optimized and minimal set of test cases.The ultimate objective associated with this study is to provide a technique that may solve the minimization problem of regression test cases in the case of linked requirements.In this research,the 100-Dollar prioritization approach is used to define the priority of the new requirements.展开更多
In real life,incomplete information,inaccurate data,and the preferences of decision-makers during qualitative judgment would impact the process of decision-making.As a technical instrument that can successfully handle...In real life,incomplete information,inaccurate data,and the preferences of decision-makers during qualitative judgment would impact the process of decision-making.As a technical instrument that can successfully handle uncertain information,Fermatean fuzzy sets have recently been used to solve the multi-attribute decision-making(MADM)problems.This paper proposes a Fermatean hesitant fuzzy information aggregation method to address the problem of fusion where the membership,non-membership,and priority are considered simultaneously.Combining the Fermatean hesitant fuzzy sets with Heronian Mean operators,this paper proposes the Fermatean hesitant fuzzy Heronian mean(FHFHM)operator and the Fermatean hesitant fuzzyweighted Heronian mean(FHFWHM)operator.Then,considering the priority relationship between attributes is often easier to obtain than the weight of attributes,this paper defines a new Fermatean hesitant fuzzy prioritized Heronian mean operator(FHFPHM),and discusses its elegant properties such as idempotency,boundedness and monotonicity in detail.Later,for problems with unknown weights and the Fermatean hesitant fuzzy information,aMADM approach based on prioritized attributes is proposed,which can effectively depict the correlation between attributes and avoid the influence of subjective factors on the results.Finally,a numerical example of multi-sensor electronic surveillance is applied to verify the feasibility and validity of the method proposed in this paper.展开更多
Both unit and integration testing are incredibly crucial for almost any software application because each of them operates a distinct process to examine the product.Due to resource constraints,when software is subject...Both unit and integration testing are incredibly crucial for almost any software application because each of them operates a distinct process to examine the product.Due to resource constraints,when software is subjected to modifications,the drastic increase in the count of test cases forces the testers to opt for a test optimization strategy.One such strategy is test case prioritization(TCP).Existing works have propounded various methodologies that re-order the system-level test cases intending to boost either the fault detection capabilities or the coverage efficacy at the earliest.Nonetheless,singularity in objective functions and the lack of dissimilitude among the re-ordered test sequences have degraded the cogency of their approaches.Considering such gaps and scenarios when the meteoric and continuous updations in the software make the intensive unit and integration testing process more fragile,this study has introduced a memetics-inspired methodology for TCP.The proposed structure is first embedded with diverse parameters,and then traditional steps of the shuffled-frog-leaping approach(SFLA)are followed to prioritize the test cases at unit and integration levels.On 5 standard test functions,a comparative analysis is conducted between the established algorithms and the proposed approach,where the latter enhances the coverage rate and fault detection of re-ordered test sets.Investigation results related to the mean average percentage of fault detection(APFD)confirmed that the proposed approach exceeds the memetic,basic multi-walk,PSO,and optimized multi-walk by 21.7%,13.99%,12.24%,and 11.51%,respectively.展开更多
Automation software need to be continuously updated by addressing software bugs contained in their repositories.However,bugs have different levels of importance;hence,it is essential to prioritize bug reports based on...Automation software need to be continuously updated by addressing software bugs contained in their repositories.However,bugs have different levels of importance;hence,it is essential to prioritize bug reports based on their sever-ity and importance.Manually managing the deluge of incoming bug reports faces time and resource constraints from the development team and delays the resolu-tion of critical bugs.Therefore,bug report prioritization is vital.This study pro-poses a new model for bug prioritization based on average one dependence estimator;it prioritizes bug reports based on severity,which is determined by the number of attributes.The more the number of attributes,the more the severity.The proposed model is evaluated using precision,recall,F1-Score,accuracy,G-Measure,and Matthew’s correlation coefficient.Results of the proposed model are compared with those of the support vector machine(SVM)and Naive Bayes(NB)models.Eclipse and Mozilla datasetswere used as the sources of bug reports.The proposed model improved the bug repository management and out-performed the SVM and NB models.Additionally,the proposed model used a weaker attribute independence supposition than the former models,thereby improving prediction accuracy with minimal computational cost.展开更多
Conversion of forest land to farmland in the Hyrcanian forest of northern Iran increases the nutrient input, especially the phosphorus(P) nutrient, thus impacting the water quality. Modeling the effect of forest los...Conversion of forest land to farmland in the Hyrcanian forest of northern Iran increases the nutrient input, especially the phosphorus(P) nutrient, thus impacting the water quality. Modeling the effect of forest loss on surface water quality provides valuable information for forest management. This study predicts the future impacts of forest loss between 2010 and 2040 on P loading in the Tajan River watershed at the sub-watershed level. To understand drivers of the land cover, we used Land Change Modeler(LCM) combining with the Soil Water Assessment Tool(SWAT) model to simulate the impacts of land use change on P loading. We characterized priority management areas for locating comprehensive and cost-effective management practices at the sub-watershed level. Results show that agricultural expansion has led to an intense deforestation. During the future period 2010–2040, forest area is expected to decrease by 34,739 hm^2. And the areas of pasture and agriculture are expected to increase by 7668 and 27,071 hm^2, respectively. In most sub-watersheds, P pollution will be intensified with the increase in deforestation by the year 2040. And the P concentration is expected to increase from 0.08 to 2.30 mg/L in all of sub-watersheds by the year 2040. It should be noted that the phosphorous concentration exceeds the American Public Health Association′s water quality standard of 0.2 mg/L for P in drinking water in both current and future scenarios in the Tajan River watershed. Only 30% of sub-watersheds will comply with the water quality standards by the year 2040. The finding of the present study highlights the importance of conserving forest area to maintain a stable water quality.展开更多
By analyzing the average percent of faults detected (APFD) metric and its variant versions, which are widely utilized as metrics to evaluate the fault detection efficiency of the test suite, this paper points out so...By analyzing the average percent of faults detected (APFD) metric and its variant versions, which are widely utilized as metrics to evaluate the fault detection efficiency of the test suite, this paper points out some limitations of the APFD series metrics. These limitations include APFD series metrics having inaccurate physical explanations and being unable to precisely describe the process of fault detection. To avoid the limitations of existing metrics, this paper proposes two improved metrics for evaluating fault detection efficiency of a test suite, including relative-APFD and relative-APFDc. The proposed metrics refer to both the speed of fault detection and the constraint of the testing source. The case study shows that the two proposed metrics can provide much more precise descriptions of the fault detection process and the fault detection efficiency of the test suite.展开更多
[Objective]The aim was to establish a multi-attribute decision making method and introduce its application in rice breeding.[Method]Based on the defined closeness degree among attributes,the difference degrees among a...[Objective]The aim was to establish a multi-attribute decision making method and introduce its application in rice breeding.[Method]Based on the defined closeness degree among attributes,the difference degrees among attributes were discussed.Furthermore,the weights of attributes were determined based on the difference degrees among the attributes.[Result]A multi-attribute decision making method based on difference degrees among attributes was established,the feasibility of applying it in rice breeding was also analyzed.[Conclusion]This study enriched the methods to determine attribute weights in multi-attribute decision making and provided the necessary theoretical support for selecting rice varieties scientifically and rationally.展开更多
Abstract—Focused crawlers (also known as subjectoriented crawlers), as the core part of vertical search engine, collect topic-specific web pages as many as they can to form a subject-oriented corpus for the latter ...Abstract—Focused crawlers (also known as subjectoriented crawlers), as the core part of vertical search engine, collect topic-specific web pages as many as they can to form a subject-oriented corpus for the latter data analyzing or user querying. This paper demonstrates that the popular algorithms utilized at the process of focused web crawling, basically refer to webpage analyzing algorithms and crawling strategies (prioritize the uniform resource locator (URLs) in the queue). Advantages and disadvantages of three crawling strategies are shown in the first experiment, which indicates that the best-first search with an appropriate heuristics is a smart choice for topic-oriented crawlingwhile the depth-first search is helpless in focused crawling. Besides, another experiment on comparison of improved ones (with a webpage analyzing algorithm added) is carried out to verify that crawling strategies alone are not quite efficient for focused crawling and in most cases their mutual efforts are taken into consideration. In light of the experiment results and recent researches, some points on the research tendency of focused crawler algorithms are suggested.展开更多
Patients with bronchogenic carcinoma comprise a high-risk group for coronavirus disease 2019(COVID-19),pneumonia and related complications.Symptoms of COVID-19 related pulmonary syndrome may be similar to deterioratin...Patients with bronchogenic carcinoma comprise a high-risk group for coronavirus disease 2019(COVID-19),pneumonia and related complications.Symptoms of COVID-19 related pulmonary syndrome may be similar to deteriorating symptoms encountered during bronchogenic carcinoma progression.These resemblances add further complexity for imaging assessment of bronchogenic carcinoma.Similarities between clinical and imaging findings can pose a major challenge to clinicians in distinguishing COVID-19 super-infection from evolving bronchogenic carcinoma,as the above-mentioned entities require very different therapeutic approaches.However,the goal of bronchogenic carcinoma management during the pandemic is to minimize the risk of exposing patients to COVID-19,whilst still managing all life-threatening events related to bronchogenic carcinoma.The current pandemic has forced all healthcare stakeholders to prioritize per value resources and reorganize therapeutic strategies for timely management of patients with COVID-19 related pulmonary syndrome.Processing of radiographic and computed tomography images by means of artificial intelligence techniques can facilitate triage of patients.Modified and newer therapeutic strategies for patients with bronchogenic carcinoma have been adopted by oncologists around the world for providing uncompromised care within the accepted standards and new guidelines.展开更多
文摘The importance and complexity of prioritizing construction projects (PCP) in urban road network planning lead to the necessity to develop an aided decision making program (ADMP). Cost benefit ratio model and stage rolled method are chosen as the theoretical foundations of the program, and then benefit model is improved to accord with the actuality of urban traffic in China. Consequently, program flows, module functions and data structures are designed, and particularly an original data structure of road ...
文摘In order to facilitate spare parts management,an integrated approach of BP neural network and supportability analysis(SA)was proposed to evaluate the criticality of spare parts as well as to prioritize spare parts.Influential factors of prioritizing spare parts were detailedly analyzed.Framework of the integrated method was established.The modelling process based on BP neural network was presented.As the input of the neural network,the values of influential factors were determined by supportability analysis data.Based on the presented method,spare parts could be automatically prioritized after supportability analysis for a new system.A case study results showed that the new method was applicable and effective.
基金supported by the MSIT(Ministry of Science and ICT),Korea,under the ITRC(Information Technology Research Center)support program(IITP-2024-RS-2024-00437494)supervised by the IITP(Institute for Information&Communications Technology Planning&Evaluation).
文摘Digital forensics aims to uncover evidence of cybercrimes within compromised systems.These cybercrimes are often perpetrated through the deployment of malware,which inevitably leaves discernible traces within the compromised systems.Forensic analysts are tasked with extracting and subsequently analyzing data,termed as artifacts,from these systems to gather evidence.Therefore,forensic analysts must sift through extensive datasets to isolate pertinent evidence.However,manually identifying suspicious traces among numerous artifacts is time-consuming and labor-intensive.Previous studies addressed such inefficiencies by integrating artificial intelligence(AI)technologies into digital forensics.Despite the efforts in previous studies,artifacts were analyzed without considering the nature of the data within them and failed to prove their efficiency through specific evaluations.In this study,we propose a system to prioritize suspicious artifacts from compromised systems infected with malware to facilitate efficient digital forensics.Our system introduces a double-checking method that recognizes the nature of data within target artifacts and employs algorithms ideal for anomaly detection.The key ideas of this method are:(1)prioritize suspicious artifacts and filter remaining artifacts using autoencoder and(2)further prioritize suspicious artifacts and filter remaining artifacts using logarithmic entropy.Our evaluation demonstrates that our system can identify malicious artifacts with high accuracy and that its double-checking method is more efficient than alternative approaches.Our system can significantly reduce the time required for forensic analysis and serve as a reference for future studies.
文摘Edge devices,due to their limited computational and storage resources,often require the use of compilers for program optimization.Therefore,ensuring the security and reliability of these compilers is of paramount importance in the emerging field of edge AI.One widely used testing method for this purpose is fuzz testing,which detects bugs by inputting random test cases into the target program.However,this process consumes significant time and resources.To improve the efficiency of compiler fuzz testing,it is common practice to utilize test case prioritization techniques.Some researchers use machine learning to predict the code coverage of test cases,aiming to maximize the test capability for the target compiler by increasing the overall predicted coverage of the test cases.Nevertheless,these methods can only forecast the code coverage of the compiler at a specific optimization level,potentially missing many optimization-related bugs.In this paper,we introduce C-CORE(short for Clustering by Code Representation),the first framework to prioritize test cases according to their code representations,which are derived directly from the source codes.This approach avoids being limited to specific compiler states and extends to a broader range of compiler bugs.Specifically,we first train a scaled pre-trained programming language model to capture as many common features as possible from the test cases generated by a fuzzer.Using this pre-trained model,we then train two downstream models:one for predicting the likelihood of triggering a bug and another for identifying code representations associated with bugs.Subsequently,we cluster the test cases according to their code representations and select the highest-scoring test case from each cluster as the high-quality test case.This reduction in redundant testing cases leads to time savings.Comprehensive evaluation results reveal that code representations are better at distinguishing test capabilities,and C-CORE significantly enhances testing efficiency.Across four datasets,C-CORE increases the average of the percentage of faults detected(APFD)value by 0.16 to 0.31 and reduces test time by over 50% in 46% of cases.When compared to the best results from approaches using predicted code coverage,C-CORE improves the APFD value by 1.1% to 12.3% and achieves an overall time-saving of 159.1%.
文摘China’s employment landscape looks sparse right now as many private enterprises that absorb much of the country’s workforce are succumbing to economic stagnation. While the Western world is caught in a tailspin, numerous private enterprises, and exporters in particular, have no place to hide. Their orders are declining to a trickle, and the supply chains of many also have broken down because of a lack of cash. In such a time of gloom, how can they weather the storm through innovation and industrial upgrading? Beijing Review reporter Hu Yue interviewed several economists and entrepreneurs about this issue at the 2008 Dialogue Between Chinese Private Enterprises and Global Fortune 500 held last December in Wenzhou, Zhejiang Province.
文摘Mobile applications usually can only access limited amount of memory. Improper use of the memory can cause memory leaks, which may lead to performance slowdowns or even cause applications to be unexpectedly killed. Although a large body of research has been devoted into the memory leak diagnosing techniques after leaks have been discovered, it is still challenging to find out the memory leak phenomena at first. Testing is the most widely used technique for failure discovery. However, traditional testing techniques are not directed for the discovery of memory leaks. They may spend lots of time on testing unlikely leaking executions and therefore can be inefficient. To address the problem, we propose a novel approach to prioritize test cases according to their likelihood to cause memory leaks in a given test suite. It firstly builds a prediction model to determine whether each test can potentially lead to memory leaks based on machine learning on selected code features. Then, for each input test case, we partly run it to get its code features and predict its likelihood to cause leaks. The most suspicious test cases will be suggested to run at first in order to reveal memory leak faults as soon as possible. Experimental evaluation on several Android applications shows that our approach is effective.
文摘With the continuing development and improvement of genome-wide techniques, a great number of candidate genes are discovered. How to identify the most likely disease genes among a large number of candidates becomes a fundamental challenge in human health. A common view is that genes related to a specific or similar disease tend to reside in the same neighbourhood of biomolecular networks. Recently, based on such observations,many methods have been developed to tackle this challenge. In this review, we firstly introduce the concept of disease genes, their properties, and available data for identifying them. Then we review the recent computational approaches for prioritizing candidate disease genes based on Protein-Protein Interaction(PPI) networks and investigate their advantages and disadvantages. Furthermore, some pieces of existing software and network resources are summarized. Finally, we discuss key issues in prioritizing candidate disease genes and point out some future research directions.
基金supported in part by the National Institutes of Health(R01 GM59507 and U01 HG005718)the VA Cooperative Studies Program of the Department of Veterans Affairs,Office of Research and Development
文摘With recent advances in genotyping and sequencing technologies,many disease susceptibility loci have been identified.However,much of the genetic heritability remains unexplained and the replication rate between independent studies is still low.Meanwhile,there have been increasing efforts on functional annotations of the entire human genome,such as the Encyclopedia of DNA Elements(ENCODE)project and other similar projects.It has been shown that incorporating these functional annotations to prioritize genome wide association signals may help identify true association signals.However,to our knowledge,the extent of the improvement when functional annotation data are considered has not been studied in the literature.In this article,we propose a statistical framework to estimate the improvement in replication rate with annotation data,and apply it to Crohn’s disease and DNase I hypersensitive sites.The results show that with cell line specific functional annotations,the expected replication rate is improved,but only at modest level.
文摘Test Case Prioritization(TCP)techniques perform better than other regression test optimization techniques including Test Suite Reduction(TSR)and Test Case Selection(TCS).Many TCP techniques are available,and their performance is usually measured through a metric Average Percentage of Fault Detection(APFD).This metric is value-neutral because it only works well when all test cases have the same cost,and all faults have the same severity.Using APFD for performance evaluation of test case orders where test cases cost or faults severity varies is prone to produce false results.Therefore,using the right metric for performance evaluation of TCP techniques is very important to get reliable and correct results.In this paper,two value-based TCP techniques have been introduced using Genetic Algorithm(GA)including Value-Cognizant Fault Detection-Based TCP(VCFDB-TCP)and Value-Cognizant Requirements Coverage-Based TCP(VCRCB-TCP).Two novel value-based performance evaluation metrics are also introduced for value-based TCP including Average Percentage of Fault Detection per value(APFDv)and Average Percentage of Requirements Coverage per value(APRCv).Two case studies are performed to validate proposed techniques and performance evaluation metrics.The proposed GA-based techniques outperformed the existing state-of-the-art TCP techniques including Original Order(OO),Reverse Order(REV-O),Random Order(RO),and Greedy algorithm.
文摘Although the construction of underground dams is one of the best methods to conserve water resources in arid and semi-arid regions,applying efficient methods for the selection of suitable sites for subsurface dam construction remains a challenge.Due to the costly and time-consuming methods of site selection for underground dam construction,this study aimed to present a new method using geographic information systems techniques and decision-making processes.The exclusionary criteria including fault,slope,hypsometry,land use,soil,stream,geology,and chemical properties of groundwater were selected for site selection of dam construction and inappropriate regions were omitted by integration and scoring layers in ArcGIS based on the Boolean logic.Finally,appropriate sites were prioritized using the Multi-Attribute Utility Theory.According to the results of the utility coefficient,seven sites were selected as the region for underground dam construction based on all criteria and experts’opinions.The site of Nazarabad dam was the best location for underground dam construction with a utility coefficient of 0.7137 followed by sites of Akhavan with a utility coefficient of 0.4633 and Mirshamsi with a utility coefficient of 0.4083.This study proposed a new approach for the construction of the subsurface dam at the proper site and help managers and decision-makers achieve sustainable water resources with limited facilities and capital and avoid wasting national capital.
基金This research is funded by the Deanship of Scientific Research at Umm Al-Qura University,Grant Code:22UQU4281755DSR02.
文摘Software needs modifications and requires revisions regularly.Owing to these revisions,retesting software becomes essential to ensure that the enhancements made,have not affected its bug-free functioning.The time and cost incurred in this process,need to be reduced by the method of test case selection and prioritization.It is observed that many nature-inspired techniques are applied in this area.African Buffalo Optimization is one such approach,applied to regression test selection and prioritization.In this paper,the proposed work explains and proves the applicability of the African Buffalo Optimization approach to test case selection and prioritization.The proposed algorithm converges in polynomial time(O(n^(2))).In this paper,the empirical evaluation of applying African Buffalo Optimization for test case prioritization is done on sample data set with multiple iterations.An astounding 62.5%drop in size and a 48.57%drop in the runtime of the original test suite were recorded.The obtained results are compared with Ant Colony Optimization.The comparative analysis indicates that African Buffalo Optimization and Ant Colony Optimization exhibit similar fault detection capabilities(80%),and a reduction in the overall execution time and size of the resultant test suite.The results and analysis,hence,advocate and encourages the use of African Buffalo Optimization in the area of test case selection and prioritization.
基金The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through Large Groups,Project under grant number RGP.2/49/43.
文摘Regression testing is a widely used approach to confirm the correct functionality of the software in incremental development.The use of test cases makes it easier to test the ripple effect of changed requirements.Rigorous testingmay help in meeting the quality criteria that is based on the conformance to the requirements as given by the intended stakeholders.However,a minimized and prioritized set of test cases may reduce the efforts and time required for testingwhile focusing on the timely delivery of the software application.In this research,a technique named Test Reduce has been presented to get a minimal set of test cases based on high priority to ensure that the web applicationmeets the required quality criteria.A new technique TestReduce is proposed with a blend of genetic algorithm to find an optimized and minimal set of test cases.The ultimate objective associated with this study is to provide a technique that may solve the minimization problem of regression test cases in the case of linked requirements.In this research,the 100-Dollar prioritization approach is used to define the priority of the new requirements.
文摘In real life,incomplete information,inaccurate data,and the preferences of decision-makers during qualitative judgment would impact the process of decision-making.As a technical instrument that can successfully handle uncertain information,Fermatean fuzzy sets have recently been used to solve the multi-attribute decision-making(MADM)problems.This paper proposes a Fermatean hesitant fuzzy information aggregation method to address the problem of fusion where the membership,non-membership,and priority are considered simultaneously.Combining the Fermatean hesitant fuzzy sets with Heronian Mean operators,this paper proposes the Fermatean hesitant fuzzy Heronian mean(FHFHM)operator and the Fermatean hesitant fuzzyweighted Heronian mean(FHFWHM)operator.Then,considering the priority relationship between attributes is often easier to obtain than the weight of attributes,this paper defines a new Fermatean hesitant fuzzy prioritized Heronian mean operator(FHFPHM),and discusses its elegant properties such as idempotency,boundedness and monotonicity in detail.Later,for problems with unknown weights and the Fermatean hesitant fuzzy information,aMADM approach based on prioritized attributes is proposed,which can effectively depict the correlation between attributes and avoid the influence of subjective factors on the results.Finally,a numerical example of multi-sensor electronic surveillance is applied to verify the feasibility and validity of the method proposed in this paper.
文摘Both unit and integration testing are incredibly crucial for almost any software application because each of them operates a distinct process to examine the product.Due to resource constraints,when software is subjected to modifications,the drastic increase in the count of test cases forces the testers to opt for a test optimization strategy.One such strategy is test case prioritization(TCP).Existing works have propounded various methodologies that re-order the system-level test cases intending to boost either the fault detection capabilities or the coverage efficacy at the earliest.Nonetheless,singularity in objective functions and the lack of dissimilitude among the re-ordered test sequences have degraded the cogency of their approaches.Considering such gaps and scenarios when the meteoric and continuous updations in the software make the intensive unit and integration testing process more fragile,this study has introduced a memetics-inspired methodology for TCP.The proposed structure is first embedded with diverse parameters,and then traditional steps of the shuffled-frog-leaping approach(SFLA)are followed to prioritize the test cases at unit and integration levels.On 5 standard test functions,a comparative analysis is conducted between the established algorithms and the proposed approach,where the latter enhances the coverage rate and fault detection of re-ordered test sets.Investigation results related to the mean average percentage of fault detection(APFD)confirmed that the proposed approach exceeds the memetic,basic multi-walk,PSO,and optimized multi-walk by 21.7%,13.99%,12.24%,and 11.51%,respectively.
基金This work was supported in part by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(No.NRF-2020R1A2C1013308).
文摘Automation software need to be continuously updated by addressing software bugs contained in their repositories.However,bugs have different levels of importance;hence,it is essential to prioritize bug reports based on their sever-ity and importance.Manually managing the deluge of incoming bug reports faces time and resource constraints from the development team and delays the resolu-tion of critical bugs.Therefore,bug report prioritization is vital.This study pro-poses a new model for bug prioritization based on average one dependence estimator;it prioritizes bug reports based on severity,which is determined by the number of attributes.The more the number of attributes,the more the severity.The proposed model is evaluated using precision,recall,F1-Score,accuracy,G-Measure,and Matthew’s correlation coefficient.Results of the proposed model are compared with those of the support vector machine(SVM)and Naive Bayes(NB)models.Eclipse and Mozilla datasetswere used as the sources of bug reports.The proposed model improved the bug repository management and out-performed the SVM and NB models.Additionally,the proposed model used a weaker attribute independence supposition than the former models,thereby improving prediction accuracy with minimal computational cost.
基金The Modares Tarbiat University of Iran funded this work
文摘Conversion of forest land to farmland in the Hyrcanian forest of northern Iran increases the nutrient input, especially the phosphorus(P) nutrient, thus impacting the water quality. Modeling the effect of forest loss on surface water quality provides valuable information for forest management. This study predicts the future impacts of forest loss between 2010 and 2040 on P loading in the Tajan River watershed at the sub-watershed level. To understand drivers of the land cover, we used Land Change Modeler(LCM) combining with the Soil Water Assessment Tool(SWAT) model to simulate the impacts of land use change on P loading. We characterized priority management areas for locating comprehensive and cost-effective management practices at the sub-watershed level. Results show that agricultural expansion has led to an intense deforestation. During the future period 2010–2040, forest area is expected to decrease by 34,739 hm^2. And the areas of pasture and agriculture are expected to increase by 7668 and 27,071 hm^2, respectively. In most sub-watersheds, P pollution will be intensified with the increase in deforestation by the year 2040. And the P concentration is expected to increase from 0.08 to 2.30 mg/L in all of sub-watersheds by the year 2040. It should be noted that the phosphorous concentration exceeds the American Public Health Association′s water quality standard of 0.2 mg/L for P in drinking water in both current and future scenarios in the Tajan River watershed. Only 30% of sub-watersheds will comply with the water quality standards by the year 2040. The finding of the present study highlights the importance of conserving forest area to maintain a stable water quality.
基金The National Natural Science Foundation of China(No.61300054)the Natural Science Foundation of Jiangsu Province(No.BK2011190,BK20130879)+1 种基金the Natural Science Foundation of Higher Education Institutions of Jiangsu Province(No.13KJB520018)the Science Foundation of Nanjing University of Posts&Telecommunications(No.NY212023)
文摘By analyzing the average percent of faults detected (APFD) metric and its variant versions, which are widely utilized as metrics to evaluate the fault detection efficiency of the test suite, this paper points out some limitations of the APFD series metrics. These limitations include APFD series metrics having inaccurate physical explanations and being unable to precisely describe the process of fault detection. To avoid the limitations of existing metrics, this paper proposes two improved metrics for evaluating fault detection efficiency of a test suite, including relative-APFD and relative-APFDc. The proposed metrics refer to both the speed of fault detection and the constraint of the testing source. The case study shows that the two proposed metrics can provide much more precise descriptions of the fault detection process and the fault detection efficiency of the test suite.
基金Supported by the Science Research and Development Project of Nanning City(201002030B)~~
文摘[Objective]The aim was to establish a multi-attribute decision making method and introduce its application in rice breeding.[Method]Based on the defined closeness degree among attributes,the difference degrees among attributes were discussed.Furthermore,the weights of attributes were determined based on the difference degrees among the attributes.[Result]A multi-attribute decision making method based on difference degrees among attributes was established,the feasibility of applying it in rice breeding was also analyzed.[Conclusion]This study enriched the methods to determine attribute weights in multi-attribute decision making and provided the necessary theoretical support for selecting rice varieties scientifically and rationally.
基金supported by the Research Fund for International Young Scientists of National Natural Science Foundation of China under Grant No.61550110248Tibet Autonomous Region Key Scientific Research Projects under Grant No.Z2014A18G2-13
文摘Abstract—Focused crawlers (also known as subjectoriented crawlers), as the core part of vertical search engine, collect topic-specific web pages as many as they can to form a subject-oriented corpus for the latter data analyzing or user querying. This paper demonstrates that the popular algorithms utilized at the process of focused web crawling, basically refer to webpage analyzing algorithms and crawling strategies (prioritize the uniform resource locator (URLs) in the queue). Advantages and disadvantages of three crawling strategies are shown in the first experiment, which indicates that the best-first search with an appropriate heuristics is a smart choice for topic-oriented crawlingwhile the depth-first search is helpless in focused crawling. Besides, another experiment on comparison of improved ones (with a webpage analyzing algorithm added) is carried out to verify that crawling strategies alone are not quite efficient for focused crawling and in most cases their mutual efforts are taken into consideration. In light of the experiment results and recent researches, some points on the research tendency of focused crawler algorithms are suggested.
文摘Patients with bronchogenic carcinoma comprise a high-risk group for coronavirus disease 2019(COVID-19),pneumonia and related complications.Symptoms of COVID-19 related pulmonary syndrome may be similar to deteriorating symptoms encountered during bronchogenic carcinoma progression.These resemblances add further complexity for imaging assessment of bronchogenic carcinoma.Similarities between clinical and imaging findings can pose a major challenge to clinicians in distinguishing COVID-19 super-infection from evolving bronchogenic carcinoma,as the above-mentioned entities require very different therapeutic approaches.However,the goal of bronchogenic carcinoma management during the pandemic is to minimize the risk of exposing patients to COVID-19,whilst still managing all life-threatening events related to bronchogenic carcinoma.The current pandemic has forced all healthcare stakeholders to prioritize per value resources and reorganize therapeutic strategies for timely management of patients with COVID-19 related pulmonary syndrome.Processing of radiographic and computed tomography images by means of artificial intelligence techniques can facilitate triage of patients.Modified and newer therapeutic strategies for patients with bronchogenic carcinoma have been adopted by oncologists around the world for providing uncompromised care within the accepted standards and new guidelines.