When the Transformer proposed by Google in 2017,it was first used for machine translation tasks and achieved the state of the art at that time.Although the current neural machine translation model can generate high qu...When the Transformer proposed by Google in 2017,it was first used for machine translation tasks and achieved the state of the art at that time.Although the current neural machine translation model can generate high quality translation results,there are still mistranslations and omissions in the translation of key information of long sentences.On the other hand,the most important part in traditional translation tasks is the translation of key information.In the translation results,as long as the key information is translated accurately and completely,even if other parts of the results are translated incorrect,the final translation results’quality can still be guaranteed.In order to solve the problem of mistranslation and missed translation effectively,and improve the accuracy and completeness of long sentence translation in machine translation,this paper proposes a key information fused neural machine translation model based on Transformer.The model proposed in this paper extracts the keywords of the source language text separately as the input of the encoder.After the same encoding as the source language text,it is fused with the output of the source language text encoded by the encoder,then the key information is processed and input into the decoder.With incorporating keyword information from the source language sentence,the model’s performance in the task of translating long sentences is very reliable.In order to verify the effectiveness of the method of fusion of key information proposed in this paper,a series of experiments were carried out on the verification set.The experimental results show that the Bilingual Evaluation Understudy(BLEU)score of the model proposed in this paper on theWorkshop on Machine Translation(WMT)2017 test dataset is higher than the BLEU score of Transformer proposed by Google on the WMT2017 test dataset.The experimental results show the advantages of the model proposed in this paper.展开更多
Key information extraction can reduce the dimensional effects while evaluating the correct preferences of users during semantic data analysis.Currently,the classifiers are used to maximize the performance of web-page ...Key information extraction can reduce the dimensional effects while evaluating the correct preferences of users during semantic data analysis.Currently,the classifiers are used to maximize the performance of web-page recommendation in terms of precision and satisfaction.The recent method disambiguates contextual sentiment using conceptual prediction with robustness,however the conceptual prediction method is not able to yield the optimal solution.Context-dependent terms are primarily evaluated by constructing linear space of context features,presuming that if the terms come together in certain consumerrelated reviews,they are semantically reliant.Moreover,the more frequently they coexist,the greater the semantic dependency is.However,the influence of the terms that coexist with each other can be part of the frequency of the terms of their semantic dependence,as they are non-integrative and their individual meaning cannot be derived.In this work,we consider the strength of a term and the influence of a term as a combinatorial optimization,called Combinatorial Optimized Linear Space Knapsack for Information Retrieval(COLSK-IR).The COLSK-IR is considered as a knapsack problem with the total weight being the“term influence”or“influence of term”and the total value being the“term frequency”or“frequency of term”for semantic data analysis.The method,by which the term influence and the term frequency are considered to identify the optimal solutions,is called combinatorial optimizations.Thus,we choose the knapsack for performing an integer programming problem and perform multiple experiments using the linear space through combinatorial optimization to identify the possible optimum solutions.It is evident from our experimental results that the COLSK-IR provides better results than previous methods to detect strongly dependent snippets with minimum ambiguity that are related to inter-sentential context during semantic data analysis.展开更多
Conventional principal component analysis(PCA) can obtain low-dimensional representations of original data space, but the selection of principal components(PCs) based on variance is subjective, which may lead to infor...Conventional principal component analysis(PCA) can obtain low-dimensional representations of original data space, but the selection of principal components(PCs) based on variance is subjective, which may lead to information loss and poor monitoring performance. To address dimension reduction and information preservation simultaneously, this paper proposes a novel PC selection scheme named full variable expression. On the basis of the proposed relevance of variables with each principal component, key principal components can be determined.All the key principal components serve as a low-dimensional representation of the entire original variables, preserving the information of original data space without information loss. A squared Mahalanobis distance, which is introduced as the monitoring statistic, is calculated directly in the key principal component space for fault detection. To test the modeling and monitoring performance of the proposed method, a numerical example and the Tennessee Eastman benchmark are used.展开更多
Ecosystem services(ES)are the connection between nature and society,and are essential for the well-being of local communities that depend on them.In Ethiopia,church forests and the surrounding agricultural matrix supp...Ecosystem services(ES)are the connection between nature and society,and are essential for the well-being of local communities that depend on them.In Ethiopia,church forests and the surrounding agricultural matrix supply numerous ES.However,the ES delivered by both land use types have not yet been assessed simultaneously.Here we surveyed both church forests and their agricultural matrices,aiming to quantify,compare and unravel the drivers underlying tree-based ES supply,density and multifunctionality.We found that almost all church forests and half of the agricultural matrices provided high ES densities.ES multifunctionality was higher in the agricultural matrices,suggesting that people deliberately conserve or plant multifunctional tree species.Furthermore,the supply of all categories of ES was positively correlated with church forest age(p-value<0.001)in the agricultural matrix,while the extent of church forest was positively correlated with the density of all categories ecosystem services score in the church forests(p-value<0.001).Our results can be used to prioritize conservation efforts at sites that provide high levels of ES supply,ES density and ES multifunctionality,and to prioritize restoration efforts at sites with low levels thereof.展开更多
At present,there are still many problems in language teaching in rural primary schools,which will affect the quality of teaching if we don't pay much attention to them.This article focuses on the existing flaws in...At present,there are still many problems in language teaching in rural primary schools,which will affect the quality of teaching if we don't pay much attention to them.This article focuses on the existing flaws in current language teaching and provides some solutions.展开更多
Existing power grid fault diagnosis methods relyon manual experience to design diagnosis models, lack theability to extract fault knowledge, and are difficult to adaptto complex and changeable engineering sites. Consi...Existing power grid fault diagnosis methods relyon manual experience to design diagnosis models, lack theability to extract fault knowledge, and are difficult to adaptto complex and changeable engineering sites. Considering thissituation, this paper proposes a power grid fault diagnosismethod based on a deep pyramid convolutional neural networkfor the alarm information set. This approach uses the deepfeature extraction ability of the network to extract fault featureknowledge from alarm information texts and achieve end-to-endfault classification and fault device identification. First, a deeppyramid convolutional neural network model for extracting theoverall characteristics of fault events is constructed to identifyfault types. Second, a deep pyramidal convolutional neuralnetwork model for alarm information text is constructed, thetext description characteristics associated with alarm informationtexts are extracted, the key information corresponding to faultsin the alarm information set is identified, and suspicious faultydevices are selected. Then, a fault device identification strategythat integrates fault-type and time sequence priorities is proposedto identify faulty devices. Finally, the actual fault cases and thefault cases generated by the simulation are studied, and theresults verify the effectiveness and practicability of the methodpresented in this paper.展开更多
In the context of today's big data and cloud computing,the global flow of data has become a powerful driver for international economic and investment growth.The EU and the U.S.have created two different paths for ...In the context of today's big data and cloud computing,the global flow of data has become a powerful driver for international economic and investment growth.The EU and the U.S.have created two different paths for the legal regulation of the cross-border flow of personal data due to their respective historical traditions and realistic demands.The requirements for data protection have shown significant differences.The EU advocates localization of data and firmly restricts cross-border flow of personal data.The U.S.tends to protect personal data through industry self-regulation and government law enforcement.At the same time,these two paths also merge and supplement with each other.Based on this,China needs to learn from the legal regulatory paths of the EU and the US,respectively,to establish a legal idea that places equal emphasis on personal data protection and the development of the information industry.In terms of domestic law,the Cybersecurity Law of the People's Republic of China needs to be improved and supplemented by relevant supporting legislation to improve the operability of the law;the industry self-discipline guidelines should be established;and various types of cross-border data need to be classified and supervised.In terms of international law,it is necessary to participate in international cooperation based on the priority of data sovereignty and promote the signing of bilateral,multilateral agreements,and international treaties on the cross-border flow of personal data.展开更多
基金Major Science and Technology Project of Sichuan Province[No.2022YFG0315,2022YFG0174]Sichuan Gas Turbine Research Institute stability support project of China Aero Engine Group Co.,Ltd.[No.GJCZ-2019-71].
文摘When the Transformer proposed by Google in 2017,it was first used for machine translation tasks and achieved the state of the art at that time.Although the current neural machine translation model can generate high quality translation results,there are still mistranslations and omissions in the translation of key information of long sentences.On the other hand,the most important part in traditional translation tasks is the translation of key information.In the translation results,as long as the key information is translated accurately and completely,even if other parts of the results are translated incorrect,the final translation results’quality can still be guaranteed.In order to solve the problem of mistranslation and missed translation effectively,and improve the accuracy and completeness of long sentence translation in machine translation,this paper proposes a key information fused neural machine translation model based on Transformer.The model proposed in this paper extracts the keywords of the source language text separately as the input of the encoder.After the same encoding as the source language text,it is fused with the output of the source language text encoded by the encoder,then the key information is processed and input into the decoder.With incorporating keyword information from the source language sentence,the model’s performance in the task of translating long sentences is very reliable.In order to verify the effectiveness of the method of fusion of key information proposed in this paper,a series of experiments were carried out on the verification set.The experimental results show that the Bilingual Evaluation Understudy(BLEU)score of the model proposed in this paper on theWorkshop on Machine Translation(WMT)2017 test dataset is higher than the BLEU score of Transformer proposed by Google on the WMT2017 test dataset.The experimental results show the advantages of the model proposed in this paper.
文摘Key information extraction can reduce the dimensional effects while evaluating the correct preferences of users during semantic data analysis.Currently,the classifiers are used to maximize the performance of web-page recommendation in terms of precision and satisfaction.The recent method disambiguates contextual sentiment using conceptual prediction with robustness,however the conceptual prediction method is not able to yield the optimal solution.Context-dependent terms are primarily evaluated by constructing linear space of context features,presuming that if the terms come together in certain consumerrelated reviews,they are semantically reliant.Moreover,the more frequently they coexist,the greater the semantic dependency is.However,the influence of the terms that coexist with each other can be part of the frequency of the terms of their semantic dependence,as they are non-integrative and their individual meaning cannot be derived.In this work,we consider the strength of a term and the influence of a term as a combinatorial optimization,called Combinatorial Optimized Linear Space Knapsack for Information Retrieval(COLSK-IR).The COLSK-IR is considered as a knapsack problem with the total weight being the“term influence”or“influence of term”and the total value being the“term frequency”or“frequency of term”for semantic data analysis.The method,by which the term influence and the term frequency are considered to identify the optimal solutions,is called combinatorial optimizations.Thus,we choose the knapsack for performing an integer programming problem and perform multiple experiments using the linear space through combinatorial optimization to identify the possible optimum solutions.It is evident from our experimental results that the COLSK-IR provides better results than previous methods to detect strongly dependent snippets with minimum ambiguity that are related to inter-sentential context during semantic data analysis.
基金Supported by the National Natural Science Foundation of China(No.61374140)Shanghai Pujiang Program(Project No.12PJ1402200)
文摘Conventional principal component analysis(PCA) can obtain low-dimensional representations of original data space, but the selection of principal components(PCs) based on variance is subjective, which may lead to information loss and poor monitoring performance. To address dimension reduction and information preservation simultaneously, this paper proposes a novel PC selection scheme named full variable expression. On the basis of the proposed relevance of variables with each principal component, key principal components can be determined.All the key principal components serve as a low-dimensional representation of the entire original variables, preserving the information of original data space without information loss. A squared Mahalanobis distance, which is introduced as the monitoring statistic, is calculated directly in the key principal component space for fault detection. To test the modeling and monitoring performance of the proposed method, a numerical example and the Tennessee Eastman benchmark are used.
基金flnancial support from VLIR-UOS,Belgium through the VLIR-IUC Interuniversity cooperation with Bahir Dar University,Ethiopia (BDU-IUC)
文摘Ecosystem services(ES)are the connection between nature and society,and are essential for the well-being of local communities that depend on them.In Ethiopia,church forests and the surrounding agricultural matrix supply numerous ES.However,the ES delivered by both land use types have not yet been assessed simultaneously.Here we surveyed both church forests and their agricultural matrices,aiming to quantify,compare and unravel the drivers underlying tree-based ES supply,density and multifunctionality.We found that almost all church forests and half of the agricultural matrices provided high ES densities.ES multifunctionality was higher in the agricultural matrices,suggesting that people deliberately conserve or plant multifunctional tree species.Furthermore,the supply of all categories of ES was positively correlated with church forest age(p-value<0.001)in the agricultural matrix,while the extent of church forest was positively correlated with the density of all categories ecosystem services score in the church forests(p-value<0.001).Our results can be used to prioritize conservation efforts at sites that provide high levels of ES supply,ES density and ES multifunctionality,and to prioritize restoration efforts at sites with low levels thereof.
文摘At present,there are still many problems in language teaching in rural primary schools,which will affect the quality of teaching if we don't pay much attention to them.This article focuses on the existing flaws in current language teaching and provides some solutions.
基金the National Natural Science Foundation of China(51877079).
文摘Existing power grid fault diagnosis methods relyon manual experience to design diagnosis models, lack theability to extract fault knowledge, and are difficult to adaptto complex and changeable engineering sites. Considering thissituation, this paper proposes a power grid fault diagnosismethod based on a deep pyramid convolutional neural networkfor the alarm information set. This approach uses the deepfeature extraction ability of the network to extract fault featureknowledge from alarm information texts and achieve end-to-endfault classification and fault device identification. First, a deeppyramid convolutional neural network model for extracting theoverall characteristics of fault events is constructed to identifyfault types. Second, a deep pyramidal convolutional neuralnetwork model for alarm information text is constructed, thetext description characteristics associated with alarm informationtexts are extracted, the key information corresponding to faultsin the alarm information set is identified, and suspicious faultydevices are selected. Then, a fault device identification strategythat integrates fault-type and time sequence priorities is proposedto identify faulty devices. Finally, the actual fault cases and thefault cases generated by the simulation are studied, and theresults verify the effectiveness and practicability of the methodpresented in this paper.
基金This article is supported by Law and Technology Institute,Renmin University of China.All mistakes and omissions are the responsibility of the author.
文摘In the context of today's big data and cloud computing,the global flow of data has become a powerful driver for international economic and investment growth.The EU and the U.S.have created two different paths for the legal regulation of the cross-border flow of personal data due to their respective historical traditions and realistic demands.The requirements for data protection have shown significant differences.The EU advocates localization of data and firmly restricts cross-border flow of personal data.The U.S.tends to protect personal data through industry self-regulation and government law enforcement.At the same time,these two paths also merge and supplement with each other.Based on this,China needs to learn from the legal regulatory paths of the EU and the US,respectively,to establish a legal idea that places equal emphasis on personal data protection and the development of the information industry.In terms of domestic law,the Cybersecurity Law of the People's Republic of China needs to be improved and supplemented by relevant supporting legislation to improve the operability of the law;the industry self-discipline guidelines should be established;and various types of cross-border data need to be classified and supervised.In terms of international law,it is necessary to participate in international cooperation based on the priority of data sovereignty and promote the signing of bilateral,multilateral agreements,and international treaties on the cross-border flow of personal data.