Three approaches, i.e., the harmonic analysis (HA) technique, the thermal diffusion equation and correction (TDEC) method, and the calorimetric method used to estimate ground heat flux, are evaluated by using obse...Three approaches, i.e., the harmonic analysis (HA) technique, the thermal diffusion equation and correction (TDEC) method, and the calorimetric method used to estimate ground heat flux, are evaluated by using observations from the Semi-Arid Climate and Environment Observatory of Lanzhou University (SACOL) in July, 2008. The calorimetric method, which involves soil heat flux measurement with an HFP01SC self-calibrating heat flux plate buried at a depth of 5 cm and heat storage in the soil between the plate and the surface, is here called the ITHP approach. The results show good linear relationships between the soil heat fluxes measured with the HFP01SC heat flux plate and those calculated with the HA technique and the TDEC method, respectively, at a depth of 5 cm. The soil heat fluxes calculated with the latter two methods well follow the phase measured with the HFP01SC heat flux plate. The magnitudes of the soil heat flux calculated with the HA technique and the TDEC method are close to each other, and they are about 2 percent and 6 percent larger than the measured soil heat flux, respectively, which mainly occur during the nighttime. Moreover, the ground heat fluxes calculated with the TDEC method and the HA technique are highly correlated with each other (R2= 0.97), and their difference is only about 1 percent. The TDEC-calculated ground heat flux also has a good linear relationship with the ITttP-calculated ground heat flux (R2 = 0.99), but their difference is larger (about 9 percent). Furthermore, compared to the HFP01SC direct measurements at a depth of 5 cm, the ground heat flux calculated with the HA technique, the TDEC method, and the ITHP approach can improve the surface energy budget closure by about 6 percent, 7 percent, and 6 percent at SACOL site, respectively. Therefore, the contribution of ground heat flux to the surface energy budget is very important for the semi-arid grassland over the Loess Plateau in China. Using turbulent heat fluxes with common corrections, soil heat storage between the surface and the heat flux plate can improve the surface energy budget closure by about 6 to 7 percent, resulting in a closure of 82 to 83 percent at the SACOL site.展开更多
In recent years,convolutional neural networks(CNNs)have been applied successfully in many fields.However,these deep neural models are still considered as“black box”for most tasks.One of the fundamental issues underl...In recent years,convolutional neural networks(CNNs)have been applied successfully in many fields.However,these deep neural models are still considered as“black box”for most tasks.One of the fundamental issues underlying this problem is understanding which features are most influential in image recognition tasks and how CNNs process these features.It is widely believed that CNN models combine low‐level features to form complex shapes until the object can be readily classified,however,several recent studies have argued that texture features are more important than other features.In this paper,we assume that the importance of certain features varies depending on specific tasks,that is,specific tasks exhibit feature bias.We designed two classification tasks based on human intuition to train deep neural models to identify the anticipated biases.We designed experiments comprising many tasks to test these biases in the Res Net and Dense Net models.From the results,we conclude that(1)the combined effect of certain features is typically far more influential than any single feature;(2)in different tasks,neural models can perform different biases,that is,we can design a specific task to make a neural model biased towards a specific anticipated feature.展开更多
Relation extraction between entity pairs is an increasingly critical area in natural language processing.Recently the pre-trained bidirectional encoder representation from transformer(BERT)performs excellendy on the t...Relation extraction between entity pairs is an increasingly critical area in natural language processing.Recently the pre-trained bidirectional encoder representation from transformer(BERT)performs excellendy on the text classification or sequence labelling tasks.Here,the high-level syntactic features that consider the dependency between each word and the target entities into the pre-trained language models are incorporated.Our model also utilizes the intermediate layers of BERT to acquire different levels of semantic information and designs multi-granularity features for final relation classification.Our model offers a momentous improvement over the published methods for the relation extraction on the widely used data sets.展开更多
It is a basic task to measure the similarity between two uncertain concepts in many real-life artificial intelligence applications,such as image retrieval,collaborative filtering,public opinion guidance,and so on.As a...It is a basic task to measure the similarity between two uncertain concepts in many real-life artificial intelligence applications,such as image retrieval,collaborative filtering,public opinion guidance,and so on.As an important cognitive computing model,cloud model has been used in many fields of artificial intelligence.It can realise the bidirectional cognitive transformation between qualitative concept and quantitative data based on the theory of probability and fuzzy set.The similarity measure of two uncertain concepts is a fundamental issue in cloud model theory.Popular similarity measure methods of cloud model are surveyed in this study.Their limitations are analysed in detail.Some related future research topics are proposed.展开更多
Data mining (also known as Knowledge Discovery in Databases - KDD) is defined as the nontrivial extraction of implicit, previously unknown, and potentially useful information from data. The aims and objectives of data...Data mining (also known as Knowledge Discovery in Databases - KDD) is defined as the nontrivial extraction of implicit, previously unknown, and potentially useful information from data. The aims and objectives of data mining are to discover knowledge of interest to user needs.Data mining is really a useful tool in many domains such as marketing, decision making, etc. However, some basic issues of data mining are ignored. What is data mining? What is the product of a data mining process? What are we doing in a data mining process? Is there any rule we should obey in a data mining process? In order to discover patterns and knowledge really interesting and actionable to the real world Zhang et al proposed a domain-driven human-machine-cooperated data mining process.Zhao and Yao proposed an interactive user-driven classification method using the granule network. In our work, we find that data mining is a kind of knowledge transforming process to transform knowledge from data format into symbol format. Thus, no new knowledge could be generated (born) in a data mining process. In a data mining process, knowledge is just transformed from data format, which is not understandable for human, into symbol format,which is understandable for human and easy to be used.It is similar to the process of translating a book from Chinese into English.In this translating process,the knowledge itself in the book should remain unchanged. What will be changed is the format of the knowledge only. That is, the knowledge in the English book should be kept the same as the knowledge in the Chinese one.Otherwise, there must be some mistakes in the translating proces, that is, we are transforming knowledge from one format into another format while not producing new knowledge in a data mining process. The knowledge is originally stored in data (data is a representation format of knowledge). Unfortunately, we can not read, understand, or use it, since we can not understand data. With this understanding of data mining, we proposed a data-driven knowledge acquisition method based on rough sets. It also improved the performance of classical knowledge acquisition methods. In fact, we also find that the domain-driven data mining and user-driven data mining do not conflict with our data-driven data mining. They could be integrated into domain-oriented data-driven data mining. It is just like the views of data base. Users with different views could look at different partial data of a data base. Thus, users with different tasks or objectives wish, or could discover different knowledge (partial knowledge) from the same data base. However, all these partial knowledge should be originally existed in the data base. So, a domain-oriented data-driven data mining method would help us to extract the knowledge which is really existed in a data base, and really interesting and actionable to the real world.展开更多
A dense heterogeneous cellular network can effectively increase the system capacity and enhance the network coverage.It is a key technology for the new generation of the mobile communication system.The dense deploymen...A dense heterogeneous cellular network can effectively increase the system capacity and enhance the network coverage.It is a key technology for the new generation of the mobile communication system.The dense deployment of small base stations not only improves the quality of network service,but also brings about a significant increase in network energy consumption.This paper mainly studies the energy efficiency optimization of the Macro-Femto heterogeneous cellular network.Considering the dynamic random changes of the access users in the network,the sleep process of the Femto Base Stations(FBSs)is modeled as a Semi-Markov Decision Process(SMDP)model in order to save the network energy consumption.And further,this paper gives the dynamic sleep algorithm of the FBS based on the value iteration.The simulation results show that the proposed SMDP-based adaptive sleep strategy of the FBS can effectively reduce the network energy consumption.展开更多
A rough set,first described by Polish computer scientist Zdzis?aw Pawlak,is a formal approximation of a crisp set,and it is now known as a new mathematical tool to process vague concepts.They are used for machine lear...A rough set,first described by Polish computer scientist Zdzis?aw Pawlak,is a formal approximation of a crisp set,and it is now known as a new mathematical tool to process vague concepts.They are used for machine learning,knowledge discovery,feature selection,etc.,and are applied to artificial intelligence,medical informatics,civil engineering,Kansei engineering,decision science,business administration,and so on.Especially,research on data mining using rough sets is widely spreading,and the obtained association rules are applied to the characterisation of data and decision support.展开更多
All eight possible extended rough set models in incomplete information systems are proposed.By analyzing existing extended models and technical meth-ods of rough set theory,the strategy of model extension is found to ...All eight possible extended rough set models in incomplete information systems are proposed.By analyzing existing extended models and technical meth-ods of rough set theory,the strategy of model extension is found to be suitable for processing incomplete information systems instead of filling possible values for missing attributes.After analyzing the definitions of existing extended models,a new general extended model is proposed.The new model is a generalization of indiscernibility relations,tolerance relations and non-symmetric similarity relations.Finally,suggestions for further study of rough set theory in incomplete informa-tion systems are put forward.展开更多
There are many algorithms for solving complex problems in supervised manner. However, unsupervised tasks are more common in real scenarios. Inspired by the idea of granular computing and the characteristics of human c...There are many algorithms for solving complex problems in supervised manner. However, unsupervised tasks are more common in real scenarios. Inspired by the idea of granular computing and the characteristics of human cognitive process, this paper proposes a complex tasks decomposition mechanism based on Density Peaks Clustering(DPC) to address complex tasks with an unsupervised process, which simulates the multi-granular observation and analysis of human being. Firstly, the DPC algorithm is modified to nullify its essential defects such as the difficulty of locating correct clustering centers and classifying them accurately. Then, the improved DPC algorithm is used to construct the initial decomposition solving space with multi-granularity theory. We also define subtask centers set and the granulation rules to guide the multi-granularity decomposing procedure. These rules are further used to decompose the solving space from coarse granules to the optimal fine granules with a convergent and automated process. Furthermore, comprehensive experiments are presented to verify the applicability and veracity of our proposed method in community-detection tasks with several benchmark complex social networks.The results show that our method outperforms other four state-of-the-art approaches.展开更多
基金supported by the National Natural Science Foundation of China (GrantNo. 40725015)
文摘Three approaches, i.e., the harmonic analysis (HA) technique, the thermal diffusion equation and correction (TDEC) method, and the calorimetric method used to estimate ground heat flux, are evaluated by using observations from the Semi-Arid Climate and Environment Observatory of Lanzhou University (SACOL) in July, 2008. The calorimetric method, which involves soil heat flux measurement with an HFP01SC self-calibrating heat flux plate buried at a depth of 5 cm and heat storage in the soil between the plate and the surface, is here called the ITHP approach. The results show good linear relationships between the soil heat fluxes measured with the HFP01SC heat flux plate and those calculated with the HA technique and the TDEC method, respectively, at a depth of 5 cm. The soil heat fluxes calculated with the latter two methods well follow the phase measured with the HFP01SC heat flux plate. The magnitudes of the soil heat flux calculated with the HA technique and the TDEC method are close to each other, and they are about 2 percent and 6 percent larger than the measured soil heat flux, respectively, which mainly occur during the nighttime. Moreover, the ground heat fluxes calculated with the TDEC method and the HA technique are highly correlated with each other (R2= 0.97), and their difference is only about 1 percent. The TDEC-calculated ground heat flux also has a good linear relationship with the ITttP-calculated ground heat flux (R2 = 0.99), but their difference is larger (about 9 percent). Furthermore, compared to the HFP01SC direct measurements at a depth of 5 cm, the ground heat flux calculated with the HA technique, the TDEC method, and the ITHP approach can improve the surface energy budget closure by about 6 percent, 7 percent, and 6 percent at SACOL site, respectively. Therefore, the contribution of ground heat flux to the surface energy budget is very important for the semi-arid grassland over the Loess Plateau in China. Using turbulent heat fluxes with common corrections, soil heat storage between the surface and the heat flux plate can improve the surface energy budget closure by about 6 to 7 percent, resulting in a closure of 82 to 83 percent at the SACOL site.
基金National Natural Science Foundation of China,Grant/Award Number:61936001Natural Science Foundation of Chongqing,Grant/Award Number:cstc2019jcyj-msxmX0380China Postdoctoral Science Foundation,Grant/Award Number:2021M700562。
文摘In recent years,convolutional neural networks(CNNs)have been applied successfully in many fields.However,these deep neural models are still considered as“black box”for most tasks.One of the fundamental issues underlying this problem is understanding which features are most influential in image recognition tasks and how CNNs process these features.It is widely believed that CNN models combine low‐level features to form complex shapes until the object can be readily classified,however,several recent studies have argued that texture features are more important than other features.In this paper,we assume that the importance of certain features varies depending on specific tasks,that is,specific tasks exhibit feature bias.We designed two classification tasks based on human intuition to train deep neural models to identify the anticipated biases.We designed experiments comprising many tasks to test these biases in the Res Net and Dense Net models.From the results,we conclude that(1)the combined effect of certain features is typically far more influential than any single feature;(2)in different tasks,neural models can perform different biases,that is,we can design a specific task to make a neural model biased towards a specific anticipated feature.
基金National Key Research and Development Program of China,Grant/Award Number:2016YFB1000905the State Key Ptogtam of National Nature Science Foundatioii of China,Grant/Award Numbet:61936001。
文摘Relation extraction between entity pairs is an increasingly critical area in natural language processing.Recently the pre-trained bidirectional encoder representation from transformer(BERT)performs excellendy on the text classification or sequence labelling tasks.Here,the high-level syntactic features that consider the dependency between each word and the target entities into the pre-trained language models are incorporated.Our model also utilizes the intermediate layers of BERT to acquire different levels of semantic information and designs multi-granularity features for final relation classification.Our model offers a momentous improvement over the published methods for the relation extraction on the widely used data sets.
基金This work is supported by the National Natural Science Foundation of China(no.61572091,no.61772096).
文摘It is a basic task to measure the similarity between two uncertain concepts in many real-life artificial intelligence applications,such as image retrieval,collaborative filtering,public opinion guidance,and so on.As an important cognitive computing model,cloud model has been used in many fields of artificial intelligence.It can realise the bidirectional cognitive transformation between qualitative concept and quantitative data based on the theory of probability and fuzzy set.The similarity measure of two uncertain concepts is a fundamental issue in cloud model theory.Popular similarity measure methods of cloud model are surveyed in this study.Their limitations are analysed in detail.Some related future research topics are proposed.
文摘Data mining (also known as Knowledge Discovery in Databases - KDD) is defined as the nontrivial extraction of implicit, previously unknown, and potentially useful information from data. The aims and objectives of data mining are to discover knowledge of interest to user needs.Data mining is really a useful tool in many domains such as marketing, decision making, etc. However, some basic issues of data mining are ignored. What is data mining? What is the product of a data mining process? What are we doing in a data mining process? Is there any rule we should obey in a data mining process? In order to discover patterns and knowledge really interesting and actionable to the real world Zhang et al proposed a domain-driven human-machine-cooperated data mining process.Zhao and Yao proposed an interactive user-driven classification method using the granule network. In our work, we find that data mining is a kind of knowledge transforming process to transform knowledge from data format into symbol format. Thus, no new knowledge could be generated (born) in a data mining process. In a data mining process, knowledge is just transformed from data format, which is not understandable for human, into symbol format,which is understandable for human and easy to be used.It is similar to the process of translating a book from Chinese into English.In this translating process,the knowledge itself in the book should remain unchanged. What will be changed is the format of the knowledge only. That is, the knowledge in the English book should be kept the same as the knowledge in the Chinese one.Otherwise, there must be some mistakes in the translating proces, that is, we are transforming knowledge from one format into another format while not producing new knowledge in a data mining process. The knowledge is originally stored in data (data is a representation format of knowledge). Unfortunately, we can not read, understand, or use it, since we can not understand data. With this understanding of data mining, we proposed a data-driven knowledge acquisition method based on rough sets. It also improved the performance of classical knowledge acquisition methods. In fact, we also find that the domain-driven data mining and user-driven data mining do not conflict with our data-driven data mining. They could be integrated into domain-oriented data-driven data mining. It is just like the views of data base. Users with different views could look at different partial data of a data base. Thus, users with different tasks or objectives wish, or could discover different knowledge (partial knowledge) from the same data base. However, all these partial knowledge should be originally existed in the data base. So, a domain-oriented data-driven data mining method would help us to extract the knowledge which is really existed in a data base, and really interesting and actionable to the real world.
基金This work was supported by the Program for the National Science Foundation of China(61671096)the Chongqing Research Program of Basic Science and Frontier Technology(cstc2017jcyjBX0005)+1 种基金Chongqing Science and Technology Innovation Leading Talent Support Program(CSTCCXLJRC201710)Venture and Innovation Support Program for Chongqing Overseas Returnee.
文摘A dense heterogeneous cellular network can effectively increase the system capacity and enhance the network coverage.It is a key technology for the new generation of the mobile communication system.The dense deployment of small base stations not only improves the quality of network service,but also brings about a significant increase in network energy consumption.This paper mainly studies the energy efficiency optimization of the Macro-Femto heterogeneous cellular network.Considering the dynamic random changes of the access users in the network,the sleep process of the Femto Base Stations(FBSs)is modeled as a Semi-Markov Decision Process(SMDP)model in order to save the network energy consumption.And further,this paper gives the dynamic sleep algorithm of the FBS based on the value iteration.The simulation results show that the proposed SMDP-based adaptive sleep strategy of the FBS can effectively reduce the network energy consumption.
文摘A rough set,first described by Polish computer scientist Zdzis?aw Pawlak,is a formal approximation of a crisp set,and it is now known as a new mathematical tool to process vague concepts.They are used for machine learning,knowledge discovery,feature selection,etc.,and are applied to artificial intelligence,medical informatics,civil engineering,Kansei engineering,decision science,business administration,and so on.Especially,research on data mining using rough sets is widely spreading,and the obtained association rules are applied to the characterisation of data and decision support.
基金supported by the National Natural Science Foundation of China(Grant Nos.60573068,60773113)the Program for New Century Excellent Talents in University(NCET),and the Natural Science Foundation of Chongqing of China(No.2008BA2017).
文摘All eight possible extended rough set models in incomplete information systems are proposed.By analyzing existing extended models and technical meth-ods of rough set theory,the strategy of model extension is found to be suitable for processing incomplete information systems instead of filling possible values for missing attributes.After analyzing the definitions of existing extended models,a new general extended model is proposed.The new model is a generalization of indiscernibility relations,tolerance relations and non-symmetric similarity relations.Finally,suggestions for further study of rough set theory in incomplete informa-tion systems are put forward.
基金supported by the National Natural Science Foundation of China (No. 61572091)Chongqing Postgraduate Scientific Research and Innovation Project (No. CYB16106)
文摘There are many algorithms for solving complex problems in supervised manner. However, unsupervised tasks are more common in real scenarios. Inspired by the idea of granular computing and the characteristics of human cognitive process, this paper proposes a complex tasks decomposition mechanism based on Density Peaks Clustering(DPC) to address complex tasks with an unsupervised process, which simulates the multi-granular observation and analysis of human being. Firstly, the DPC algorithm is modified to nullify its essential defects such as the difficulty of locating correct clustering centers and classifying them accurately. Then, the improved DPC algorithm is used to construct the initial decomposition solving space with multi-granularity theory. We also define subtask centers set and the granulation rules to guide the multi-granularity decomposing procedure. These rules are further used to decompose the solving space from coarse granules to the optimal fine granules with a convergent and automated process. Furthermore, comprehensive experiments are presented to verify the applicability and veracity of our proposed method in community-detection tasks with several benchmark complex social networks.The results show that our method outperforms other four state-of-the-art approaches.