In recent years,artificial intelligence technology has exhibited great potential in seismic signal recognition,setting off a new wave of research.Vast amounts of high-quality labeled data are required to develop and a...In recent years,artificial intelligence technology has exhibited great potential in seismic signal recognition,setting off a new wave of research.Vast amounts of high-quality labeled data are required to develop and apply artificial intelligence in seismology research.In this study,based on the 2013–2020 seismic cataloging reports of the China Earthquake Networks Center,we constructed an artificial intelligence seismological training dataset(“DiTing”)with the largest known total time length.Data were recorded using broadband and short-period seismometers.The obtained dataset included 2,734,748 threecomponent waveform traces from 787,010 regional seismic events,the corresponding P-and S-phase arrival time labels,and 641,025 P-wave first-motion polarity labels.All waveforms were sampled at 50 Hz and cut to a time length of 180 s starting from a random number of seconds before the occurrence of an earthquake.Each three-component waveform contained a considerable amount of descriptive information,such as the epicentral distance,back azimuth,and signal-to-noise ratios.The magnitudes of seismic events,epicentral distance,signal-to-noise ratio of P-wave data,and signal-to-noise ratio of S-wave data ranged from 0 to 7.7,0 to 330 km,–0.05 to 5.31 dB,and–0.05 to 4.73 dB,respectively.The dataset compiled in this study can serve as a high-quality benchmark for machine learning model development and data-driven seismological research on earthquake detection,seismic phase picking,first-motion polarity determination,earthquake magnitude prediction,early warning systems,and strong ground-motion prediction.Such research will further promote the development and application of artificial intelligence in seismology.展开更多
Nowcasts of strong convective precipitation and radar-based quantitative precipitation estimations have always been hot yet challenging issues in meteorological sciences.Data-driven machine learning,especially deep le...Nowcasts of strong convective precipitation and radar-based quantitative precipitation estimations have always been hot yet challenging issues in meteorological sciences.Data-driven machine learning,especially deep learning,provides a new technical approach for the quantitative estimation and forecasting of precipitation.A high-quality,large-sample,and labeled training dataset is critical for the successful application of machine-learning technology to a specific field.The present study develops a benchmark dataset that can be applied to machine learning for minutescale quantitative precipitation estimation and forecasting(QpefBD),containing 231,978 samples of 3185 heavy precipitation events that occurred in 6 provinces of central and eastern China from April to October 2016-2018.Each individual sample consists of 8 products of weather radars at 6-min intervals within the time window of the corresponding event and products of 27 physical quantities at hourly intervals that describe the atmospheric dynamic and thermodynamic conditions.Two data labels,i.e.,ground precipitation intensity and areal coverage of heavy precipitation at 6-min intervals,are also included.The present study describes the basic components of the dataset and data processing and provides metrics for the evaluation of model performance on precipitation estimation and forecasting.Based on these evaluation metrics,some simple and commonly used methods are applied to evaluate precipitation estimates and forecasts.The results can serve as the benchmark reference for the performance evaluation of machine learning models using this dataset.This paper also gives some suggestions and scenarios of the QpefBD application.We believe that the application of this benchmark dataset will promote interdisciplinary collaboration between meteorological sciences and artificial intelligence sciences,providing a new way for the identification and forecast of heavy precipitation.展开更多
With the development of artificial intelligence-related technologies such as deep learning,various organizations,including the government,are making various efforts to generate and manage big data for use in artificia...With the development of artificial intelligence-related technologies such as deep learning,various organizations,including the government,are making various efforts to generate and manage big data for use in artificial intelligence.However,it is difficult to acquire big data due to various social problems and restrictions such as personal information leakage.There are many problems in introducing technology in fields that do not have enough training data necessary to apply deep learning technology.Therefore,this study proposes a mixed contour data augmentation technique,which is a data augmentation technique using contour images,to solve a problem caused by a lack of data.ResNet,a famous convolutional neural network(CNN)architecture,and CIFAR-10,a benchmark data set,are used for experimental performance evaluation to prove the superiority of the proposed method.And to prove that high performance improvement can be achieved even with a small training dataset,the ratio of the training dataset was divided into 70%,50%,and 30%for comparative analysis.As a result of applying the mixed contour data augmentation technique,it was possible to achieve a classification accuracy improvement of up to 4.64%and high accuracy even with a small amount of data set.In addition,it is expected that the mixed contour data augmentation technique can be applied in various fields by proving the excellence of the proposed data augmentation technique using benchmark datasets.展开更多
Recently,deep learning has achieved remarkable results in fields that require human cognitive ability,learning ability,and reasoning ability.Activation functions are very important because they provide the ability of ...Recently,deep learning has achieved remarkable results in fields that require human cognitive ability,learning ability,and reasoning ability.Activation functions are very important because they provide the ability of artificial neural networks to learn complex patterns through nonlinearity.Various activation functions are being studied to solve problems such as vanishing gradients and dying nodes that may occur in the deep learning process.However,it takes a lot of time and effort for researchers to use the existing activation function in their research.Therefore,in this paper,we propose a universal activation function(UA)so that researchers can easily create and apply various activation functions and improve the performance of neural networks.UA can generate new types of activation functions as well as functions like traditional activation functions by properly adjusting three hyperparameters.The famous Convolutional Neural Network(CNN)and benchmark datasetwere used to evaluate the experimental performance of the UA proposed in this study.We compared the performance of the artificial neural network to which the traditional activation function is applied and the artificial neural network to which theUA is applied.In addition,we evaluated the performance of the new activation function generated by adjusting the hyperparameters of theUA.The experimental performance evaluation results showed that the classification performance of CNNs improved by up to 5%through the UA,although most of them showed similar performance to the traditional activation function.展开更多
针对目标场景复杂的空间布局和高光谱影像固有的空-谱信息冗余等挑战,提出了端到端的轻量化深度全局-局部知识蒸馏(Lightweight Deep Global-Local Knowledge Distillation,LDGLKD)网络。为探索空-谱特征的全局序列属性,教师模型视觉Tra...针对目标场景复杂的空间布局和高光谱影像固有的空-谱信息冗余等挑战,提出了端到端的轻量化深度全局-局部知识蒸馏(Lightweight Deep Global-Local Knowledge Distillation,LDGLKD)网络。为探索空-谱特征的全局序列属性,教师模型视觉Transformer(Vision Transformer,ViT)被用来指导轻量化学生模型进行高光谱影像场景分类。LDGLKD选择预训练的VGG16作为学生模型来提取局部细节信息,将ViT和VGG16通过知识蒸馏协同训练后,教师模型将所学习到的远程上下文关系向小规模学生模型进行传递。LDGLKD可通过知识蒸馏结合上述两种模型的优点,在欧比特高光谱影像场景分类数据集OHID-SC及公开的高光谱遥感图像数据集HSRS-SC上的最佳分类精度分别达到91.62%和97.96%。实验结果表明:LDGLKD网络具有良好的分类性能。根据欧比特珠海一号卫星提供的遥感数据构建的OHID-SC可以反映详细的地表覆盖情况,并为高光谱场景分类任务提供数据支撑。展开更多
Video tracking is a complex problem because the environment, in which video motion needs to be tracked, is widely varied based on the application and poses several constraints on the design and performance of the trac...Video tracking is a complex problem because the environment, in which video motion needs to be tracked, is widely varied based on the application and poses several constraints on the design and performance of the tracking system. Current datasets that are used to evaluate and compare video motion tracking algorithms use a cumulative performance measure without thoroughly analyzing the effect of these different constraints imposed by the environment. But it needs to analyze these constraints as parameters. The objective of this paper is to identify these parameters and define quantitative measures for these parameters to compare video datasets for motion tracking.展开更多
近年来,篡改文本图像在互联网的广泛传播为文本图像安全带来严重威胁。然而,相应的篡改文本检测(TTD,tampered text detection)方法却未得到充分的探索。TTD任务旨在定位图像中所有文本区域,同时根据纹理的真实性判断文本区域是否被篡...近年来,篡改文本图像在互联网的广泛传播为文本图像安全带来严重威胁。然而,相应的篡改文本检测(TTD,tampered text detection)方法却未得到充分的探索。TTD任务旨在定位图像中所有文本区域,同时根据纹理的真实性判断文本区域是否被篡改。与一般的文本检测任务不同,TTD任务需要进一步感知真实文本和篡改文本分类的细粒度信息。TTD任务有两个主要挑战:一方面,由于真实文本和篡改文本的纹理具有较高的相似性,仅在空域(RGB)进行纹理特征学习的篡改文本检测方法不能很好地区分两类文本;另一方面,由于检测真实文本和篡改文本的难度不同,检测模型无法平衡两类文本的学习过程,从而造成两类文本检测精度的不平衡问题。相较于空域特征,文本纹理在频域中的不连续性能够帮助网络鉴别文本实例的真伪,根据上述依据,提出基于空域和频域(RGB and frequency)关系建模的篡改文本检测方法。采用空域和频域特征提取器分别提取空域和频域特征,通过引入频域信息增强网络对篡改纹理的鉴别能力;使用全局空频域关系模块建模不同文本实例的纹理真实性关系,通过参考同幅图像中其他文本实例的空频域特征来辅助判断当前文本实例的真伪性,从而平衡真实和篡改文本检测难度,解决检测精度不平衡问题;提出一个票据篡改文本图像数据集(Tampered-SROIE)来验证上述篡改文本检测方法的有效性,该数据集包含986张图像(626张训练图像和360张测试图像)。该方法在Tampered-SROIE上的真实和篡改文本检测F值分别达到95.97%和96.80%,同时降低检测精度不平衡性1.13%。该方法从网络结构与检测策略的角度为篡改文本检测任务提供了新的解决方案,同时Tampered-SROIE为以后的篡改文本检测方法提供了评估基准。展开更多
Protein N-phosphorylation is widely present in nature and participates in various biological processes.However,current knowledge on N-phosphorylation is extremely limited compared to that on O-phosphorylation.In this ...Protein N-phosphorylation is widely present in nature and participates in various biological processes.However,current knowledge on N-phosphorylation is extremely limited compared to that on O-phosphorylation.In this study,we collected 11,710 experimentally verified N-phosphosites of 7344 proteins from 39 species and subsequently constructed the database Nphos to share up-to-date information on protein N-phosphorylation.Upon these substantial data,we characterized the sequential and structural features of protein N-phosphorylation.Moreover,after comparing hundreds of learning models,we chose and optimized gradient boosting decision tree(GBDT)models to predict three types of human N-phosphorylation,achieving mean area under the receiver operating characteristic curve(AUC)values of 90.56%,91.24%,and 92.01%for pHis,pLys,and pArg,respectively.Meanwhile,we discovered 488,825 distinct N-phosphosites in the human proteome.The models were also deployed in Nphos for interactive N-phosphosite prediction.In summary,this work provides new insights and points for both flexible and focused investigations of N-phosphorylation.It will also facilitate a deeper and more systematic understanding of protein N-phosphorylation modification by providing a data and technical foundation.Nphos is freely available at http://www.bio-add.org/Nphos/and http://ppodd.org.cn/Nphos/.展开更多
基金the National Natural Science Foundation of China(Nos.41804047 and 42111540260)Fundamental Research Funds of the Institute of Geophysics,China Earthquake Administration(NO.DQJB19A0114)the Key Research Program of the Institute of Geology and Geophysics,Chinese Academy of Sciences(No.IGGCAS-201904).
文摘In recent years,artificial intelligence technology has exhibited great potential in seismic signal recognition,setting off a new wave of research.Vast amounts of high-quality labeled data are required to develop and apply artificial intelligence in seismology research.In this study,based on the 2013–2020 seismic cataloging reports of the China Earthquake Networks Center,we constructed an artificial intelligence seismological training dataset(“DiTing”)with the largest known total time length.Data were recorded using broadband and short-period seismometers.The obtained dataset included 2,734,748 threecomponent waveform traces from 787,010 regional seismic events,the corresponding P-and S-phase arrival time labels,and 641,025 P-wave first-motion polarity labels.All waveforms were sampled at 50 Hz and cut to a time length of 180 s starting from a random number of seconds before the occurrence of an earthquake.Each three-component waveform contained a considerable amount of descriptive information,such as the epicentral distance,back azimuth,and signal-to-noise ratios.The magnitudes of seismic events,epicentral distance,signal-to-noise ratio of P-wave data,and signal-to-noise ratio of S-wave data ranged from 0 to 7.7,0 to 330 km,–0.05 to 5.31 dB,and–0.05 to 4.73 dB,respectively.The dataset compiled in this study can serve as a high-quality benchmark for machine learning model development and data-driven seismological research on earthquake detection,seismic phase picking,first-motion polarity determination,earthquake magnitude prediction,early warning systems,and strong ground-motion prediction.Such research will further promote the development and application of artificial intelligence in seismology.
基金Supported by the National Key Research and Development Program of China(2018YFC1507305)。
文摘Nowcasts of strong convective precipitation and radar-based quantitative precipitation estimations have always been hot yet challenging issues in meteorological sciences.Data-driven machine learning,especially deep learning,provides a new technical approach for the quantitative estimation and forecasting of precipitation.A high-quality,large-sample,and labeled training dataset is critical for the successful application of machine-learning technology to a specific field.The present study develops a benchmark dataset that can be applied to machine learning for minutescale quantitative precipitation estimation and forecasting(QpefBD),containing 231,978 samples of 3185 heavy precipitation events that occurred in 6 provinces of central and eastern China from April to October 2016-2018.Each individual sample consists of 8 products of weather radars at 6-min intervals within the time window of the corresponding event and products of 27 physical quantities at hourly intervals that describe the atmospheric dynamic and thermodynamic conditions.Two data labels,i.e.,ground precipitation intensity and areal coverage of heavy precipitation at 6-min intervals,are also included.The present study describes the basic components of the dataset and data processing and provides metrics for the evaluation of model performance on precipitation estimation and forecasting.Based on these evaluation metrics,some simple and commonly used methods are applied to evaluate precipitation estimates and forecasts.The results can serve as the benchmark reference for the performance evaluation of machine learning models using this dataset.This paper also gives some suggestions and scenarios of the QpefBD application.We believe that the application of this benchmark dataset will promote interdisciplinary collaboration between meteorological sciences and artificial intelligence sciences,providing a new way for the identification and forecast of heavy precipitation.
文摘With the development of artificial intelligence-related technologies such as deep learning,various organizations,including the government,are making various efforts to generate and manage big data for use in artificial intelligence.However,it is difficult to acquire big data due to various social problems and restrictions such as personal information leakage.There are many problems in introducing technology in fields that do not have enough training data necessary to apply deep learning technology.Therefore,this study proposes a mixed contour data augmentation technique,which is a data augmentation technique using contour images,to solve a problem caused by a lack of data.ResNet,a famous convolutional neural network(CNN)architecture,and CIFAR-10,a benchmark data set,are used for experimental performance evaluation to prove the superiority of the proposed method.And to prove that high performance improvement can be achieved even with a small training dataset,the ratio of the training dataset was divided into 70%,50%,and 30%for comparative analysis.As a result of applying the mixed contour data augmentation technique,it was possible to achieve a classification accuracy improvement of up to 4.64%and high accuracy even with a small amount of data set.In addition,it is expected that the mixed contour data augmentation technique can be applied in various fields by proving the excellence of the proposed data augmentation technique using benchmark datasets.
基金This work was supported by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(No.2022R1F1A1062953).
文摘Recently,deep learning has achieved remarkable results in fields that require human cognitive ability,learning ability,and reasoning ability.Activation functions are very important because they provide the ability of artificial neural networks to learn complex patterns through nonlinearity.Various activation functions are being studied to solve problems such as vanishing gradients and dying nodes that may occur in the deep learning process.However,it takes a lot of time and effort for researchers to use the existing activation function in their research.Therefore,in this paper,we propose a universal activation function(UA)so that researchers can easily create and apply various activation functions and improve the performance of neural networks.UA can generate new types of activation functions as well as functions like traditional activation functions by properly adjusting three hyperparameters.The famous Convolutional Neural Network(CNN)and benchmark datasetwere used to evaluate the experimental performance of the UA proposed in this study.We compared the performance of the artificial neural network to which the traditional activation function is applied and the artificial neural network to which theUA is applied.In addition,we evaluated the performance of the new activation function generated by adjusting the hyperparameters of theUA.The experimental performance evaluation results showed that the classification performance of CNNs improved by up to 5%through the UA,although most of them showed similar performance to the traditional activation function.
文摘Video tracking is a complex problem because the environment, in which video motion needs to be tracked, is widely varied based on the application and poses several constraints on the design and performance of the tracking system. Current datasets that are used to evaluate and compare video motion tracking algorithms use a cumulative performance measure without thoroughly analyzing the effect of these different constraints imposed by the environment. But it needs to analyze these constraints as parameters. The objective of this paper is to identify these parameters and define quantitative measures for these parameters to compare video datasets for motion tracking.
文摘近年来,篡改文本图像在互联网的广泛传播为文本图像安全带来严重威胁。然而,相应的篡改文本检测(TTD,tampered text detection)方法却未得到充分的探索。TTD任务旨在定位图像中所有文本区域,同时根据纹理的真实性判断文本区域是否被篡改。与一般的文本检测任务不同,TTD任务需要进一步感知真实文本和篡改文本分类的细粒度信息。TTD任务有两个主要挑战:一方面,由于真实文本和篡改文本的纹理具有较高的相似性,仅在空域(RGB)进行纹理特征学习的篡改文本检测方法不能很好地区分两类文本;另一方面,由于检测真实文本和篡改文本的难度不同,检测模型无法平衡两类文本的学习过程,从而造成两类文本检测精度的不平衡问题。相较于空域特征,文本纹理在频域中的不连续性能够帮助网络鉴别文本实例的真伪,根据上述依据,提出基于空域和频域(RGB and frequency)关系建模的篡改文本检测方法。采用空域和频域特征提取器分别提取空域和频域特征,通过引入频域信息增强网络对篡改纹理的鉴别能力;使用全局空频域关系模块建模不同文本实例的纹理真实性关系,通过参考同幅图像中其他文本实例的空频域特征来辅助判断当前文本实例的真伪性,从而平衡真实和篡改文本检测难度,解决检测精度不平衡问题;提出一个票据篡改文本图像数据集(Tampered-SROIE)来验证上述篡改文本检测方法的有效性,该数据集包含986张图像(626张训练图像和360张测试图像)。该方法在Tampered-SROIE上的真实和篡改文本检测F值分别达到95.97%和96.80%,同时降低检测精度不平衡性1.13%。该方法从网络结构与检测策略的角度为篡改文本检测任务提供了新的解决方案,同时Tampered-SROIE为以后的篡改文本检测方法提供了评估基准。
基金supported by the National Key R&D Program of China(Grant No.2020YFA0608300)the Technology and Engineering Center for Space Utilization,Chinese Academy of Sciences(Grant No.YYWT-0901-EXP-16)+2 种基金the Scientific Research Grant of Ningbo University(Grant No.215-432000282)the Ningbo City Top Talent Project(Grant No.215-432094250)the National Natural Science Foundation of China(Grant Nos.22107055 and 91856126).
文摘Protein N-phosphorylation is widely present in nature and participates in various biological processes.However,current knowledge on N-phosphorylation is extremely limited compared to that on O-phosphorylation.In this study,we collected 11,710 experimentally verified N-phosphosites of 7344 proteins from 39 species and subsequently constructed the database Nphos to share up-to-date information on protein N-phosphorylation.Upon these substantial data,we characterized the sequential and structural features of protein N-phosphorylation.Moreover,after comparing hundreds of learning models,we chose and optimized gradient boosting decision tree(GBDT)models to predict three types of human N-phosphorylation,achieving mean area under the receiver operating characteristic curve(AUC)values of 90.56%,91.24%,and 92.01%for pHis,pLys,and pArg,respectively.Meanwhile,we discovered 488,825 distinct N-phosphosites in the human proteome.The models were also deployed in Nphos for interactive N-phosphosite prediction.In summary,this work provides new insights and points for both flexible and focused investigations of N-phosphorylation.It will also facilitate a deeper and more systematic understanding of protein N-phosphorylation modification by providing a data and technical foundation.Nphos is freely available at http://www.bio-add.org/Nphos/and http://ppodd.org.cn/Nphos/.