Orthogonal Time Frequency and Space(OTFS) modulation is expected to provide high-speed and ultra-reliable communications for emerging mobile applications, including low-orbit satellite communications. Using the Dopple...Orthogonal Time Frequency and Space(OTFS) modulation is expected to provide high-speed and ultra-reliable communications for emerging mobile applications, including low-orbit satellite communications. Using the Doppler frequency for positioning is a promising research direction on communication and navigation integration. To tackle the high Doppler frequency and low signal-to-noise ratio(SNR) in satellite communication, this paper proposes a Red and Blue Frequency Shift Discriminator(RBFSD) based on the pseudo-noise(PN) sequence.The paper derives that the cross-correlation function on the Doppler domain exhibits the characteristic of a Sinc function. Therefore, it applies modulation onto the Delay-Doppler domain using PN sequence and adjusts Doppler frequency estimation by red-shifting or blue-shifting. Simulation results show that the performance of Doppler frequency estimation is close to the Cramér-Rao Lower Bound when the SNR is greater than -15dB. The proposed algorithm is about 1/D times less complex than the existing PN pilot sequence algorithm, where D is the resolution of the fractional Doppler.展开更多
Underwater pulse waveform recognition is an important method for underwater object detection.Most existing works focus on the application of traditional pattern recognition methods,which ignore the time-and space-vary...Underwater pulse waveform recognition is an important method for underwater object detection.Most existing works focus on the application of traditional pattern recognition methods,which ignore the time-and space-varying characteristics in sound propagation channels and cannot easily extract valuable waveform features.Sound propagation channels in seawater are time-and space-varying convolutional channels.In the extraction of the waveform features of underwater acoustic signals,the effect of high-accuracy underwater acoustic signal recognition is identified by eliminating the influence of time-and space-varying convolutional channels to the greatest extent possible.We propose a hash aggregate discriminative network(HADN),which combines hash learning and deep learning to minimize the time-and space-varying effects on convolutional channels and adaptively learns effective underwater waveform features to achieve high-accuracy underwater pulse waveform recognition.In the extraction of the hash features of acoustic signals,a discrete constraint between clusters within a hash feature class is introduced.This constraint can ensure that the influence of convolutional channels on hash features is minimized.In addition,we design a new loss function called aggregate discriminative loss(AD-loss).The use of AD-loss and softmax-loss can increase the discriminativeness of the learned hash features.Experimental results show that on pool and ocean datasets,which were collected in pools and oceans,respectively,by using acoustic collectors,the proposed HADN performs better than other comparative models in terms of accuracy and mAP.展开更多
To reduce the cost and increase the efficiency of plant genetic marker fingerprinting for variety discrimination,it is desirable to identify the optimal marker combinations.We describe a marker combination screening m...To reduce the cost and increase the efficiency of plant genetic marker fingerprinting for variety discrimination,it is desirable to identify the optimal marker combinations.We describe a marker combination screening model based on the genetic algorithm(GA)and implemented in a software tool,Loci Scan.Ratio-based variety discrimination power provided the largest optimization space among multiple fitness functions.Among GA parameters,an increase in population size and generation number enlarged optimization depth but also calculation workload.Exhaustive algorithm afforded the same optimization depth as GA but vastly increased calculation time.In comparison with two other software tools,Loci Scan accommodated missing data,reduced calculation time,and offered more fitness functions.In large datasets,the sample size of training data exerted the strongest influence on calculation time,whereas the marker size of training data showed no effect,and target marker number had limited effect on analysis speed.展开更多
Analyzing polysorbate 20(PS20)composition and the impact of each component on stability and safety is crucial due to formulation variations and individual tolerance.The similar structures and polarities of PS20 compon...Analyzing polysorbate 20(PS20)composition and the impact of each component on stability and safety is crucial due to formulation variations and individual tolerance.The similar structures and polarities of PS20 components make accurate separation,identification,and quantification challenging.In this work,a high-resolution quantitative method was developed using single-dimensional high-performance liquid chromatography(HPLC)with charged aerosol detection(CAD)to separate 18 key components with multiple esters.The separated components were characterized by ultra-high-performance liquid chromatography-quadrupole time-of-flight mass spectrometry(UHPLC-Q-TOF-MS)with an identical gradient as the HPLC-CAD analysis.The polysorbate compound database and library were expanded over 7-time compared to the commercial database.The method investigated differences in PS20 samples from various origins and grades for different dosage forms to evaluate the composition-process relationship.UHPLC-Q-TOF-MS identified 1329 to 1511 compounds in 4 batches of PS20 from different sources.The method observed the impact of 4 degradation conditions on peak components,identifying stable components and their tendencies to change.HPLC-CAD and UHPLC-Q-TOF-MS results provided insights into fingerprint differences,distinguishing quasi products.展开更多
The objective of this study is to investigate themethods for soil liquefaction discrimination. Typically, predicting soilliquefaction potential involves conducting the standard penetration test (SPT), which requires f...The objective of this study is to investigate themethods for soil liquefaction discrimination. Typically, predicting soilliquefaction potential involves conducting the standard penetration test (SPT), which requires field testing and canbe time-consuming and labor-intensive. In contrast, the cone penetration test (CPT) provides a more convenientmethod and offers detailed and continuous information about soil layers. In this study, the feature matrix based onCPT data is proposed to predict the standard penetration test blow count N. The featurematrix comprises the CPTcharacteristic parameters at specific depths, such as tip resistance qc, sleeve resistance f s, and depth H. To fuse thefeatures on the matrix, the convolutional neural network (CNN) is employed for feature extraction. Additionally,Genetic Algorithm (GA) is utilized to obtain the best combination of convolutional kernels and the number ofneurons. The study evaluated the robustness of the proposed model using multiple engineering field data sets.Results demonstrated that the proposed model outperformed conventional methods in predicting N values forvarious soil categories, including sandy silt, silty sand, and clayey silt. Finally, the proposed model was employedfor liquefaction discrimination. The liquefaction discrimination based on the predicted N values was comparedwith the measured N values, and the results showed that the discrimination results were in 75% agreement. Thestudy has important practical application value for foundation liquefaction engineering. Also, the novel methodadopted in this research provides new ideas and methods for research in related fields, which is of great academicsignificance.展开更多
Recovering high-quality inscription images from unknown and complex inscription noisy images is a challenging research issue.Different fromnatural images,character images pay more attention to stroke information.Howev...Recovering high-quality inscription images from unknown and complex inscription noisy images is a challenging research issue.Different fromnatural images,character images pay more attention to stroke information.However,existingmodelsmainly consider pixel-level informationwhile ignoring structural information of the character,such as its edge and glyph,resulting in reconstructed images with mottled local structure and character damage.To solve these problems,we propose a novel generative adversarial network(GAN)framework based on an edge-guided generator and a discriminator constructed by a dual-domain U-Net framework,i.e.,EDU-GAN.Unlike existing frameworks,the generator introduces the edge extractionmodule,guiding it into the denoising process through the attention mechanism,which maintains the edge detail of the restored inscription image.Moreover,a dual-domain U-Net-based discriminator is proposed to learn the global and local discrepancy between the denoised and the label images in both image and morphological domains,which is helpful to blind denoising tasks.The proposed dual-domain discriminator and generator for adversarial training can reduce local artifacts and keep the denoised character structure intact.Due to the lack of a real-inscription image,we built the real-inscription dataset to provide an effective benchmark for studying inscription image denoising.The experimental results show the superiority of our method both in the synthetic and real-inscription datasets.展开更多
Extracting more information and saving quantum resources are two main aims for quantum measurements.However,the optimization of strategies for these two objectives varies when discriminating between quantum states |ψ...Extracting more information and saving quantum resources are two main aims for quantum measurements.However,the optimization of strategies for these two objectives varies when discriminating between quantum states |ψ_(0)> and |ψ_(1)> through multiple measurements.In this study,we introduce a novel state discrimination model that reveals the intricate relationship between the average error rate and average copy consumption.By integrating these two crucial metrics and minimizing their weighted sum for any given weight value,our research underscores the infeasibility of simultaneously minimizing these metrics through local measurements with one-way communication.Our findings present a compelling trade-off curve,highlighting the advantages of achieving a balance between error rate and copy consumption in quantum discrimination tasks,offering valuable insights into the optimization of quantum resources while ensuring the accuracy of quantum state discrimination.展开更多
Fast neutron flux measurements with high count rates and high time resolution have important applications in equipment such as tokamaks.In this study,real-time neutron and gamma discrimination was implemented on a sel...Fast neutron flux measurements with high count rates and high time resolution have important applications in equipment such as tokamaks.In this study,real-time neutron and gamma discrimination was implemented on a self-developed 500-Msps,12-bit digitizer,and the neutron and gamma spectra were calculated directly on an FPGA.A fast neutron flux measurement system with BC-501A and EJ-309 liquid scintillator detectors was developed and a fast neutron measurement experiment was successfully performed on the HL-2 M tokamak at the Southwestern Institute of Physics,China.The experimental results demonstrated that the system obtained the neutron and gamma spectra with a time accuracy of 1 ms.At count rates of up to 1 Mcps,the figure of merit was greater than 1.05 for energies between 50 keV and 2.8 MeV.展开更多
Intelligent diagnosis driven by big data for mechanical fault is an important means to ensure the safe operation ofequipment. In these methods, deep learning-based machinery fault diagnosis approaches have received in...Intelligent diagnosis driven by big data for mechanical fault is an important means to ensure the safe operation ofequipment. In these methods, deep learning-based machinery fault diagnosis approaches have received increasingattention and achieved some results. It might lead to insufficient performance for using transfer learning alone andcause misclassification of target samples for domain bias when building deep models to learn domain-invariantfeatures. To address the above problems, a deep discriminative adversarial domain adaptation neural networkfor the bearing fault diagnosis model is proposed (DDADAN). In this method, the raw vibration data are firstlyconverted into frequency domain data by Fast Fourier Transform, and an improved deep convolutional neuralnetwork with wide first-layer kernels is used as a feature extractor to extract deep fault features. Then, domaininvariant features are learned from the fault data with correlation alignment-based domain adversarial training.Furthermore, to enhance the discriminative property of features, discriminative feature learning is embeddedinto this network to make the features compact, as well as separable between classes within the class. Finally, theperformance and anti-noise capability of the proposedmethod are evaluated using two sets of bearing fault datasets.The results demonstrate that the proposed method is capable of handling domain offset caused by differentworkingconditions and maintaining more than 97.53% accuracy on various transfer tasks. Furthermore, the proposedmethod can achieve high diagnostic accuracy under varying noise levels.展开更多
Zero-shot learning enables the recognition of new class samples by migrating models learned from semanticfeatures and existing sample features to things that have never been seen before. The problems of consistencyof ...Zero-shot learning enables the recognition of new class samples by migrating models learned from semanticfeatures and existing sample features to things that have never been seen before. The problems of consistencyof different types of features and domain shift problems are two of the critical issues in zero-shot learning. Toaddress both of these issues, this paper proposes a new modeling structure. The traditional approach mappedsemantic features and visual features into the same feature space;based on this, a dual discriminator approachis used in the proposed model. This dual discriminator approach can further enhance the consistency betweensemantic and visual features. At the same time, this approach can also align unseen class semantic features andtraining set samples, providing a portion of information about the unseen classes. In addition, a new feature fusionmethod is proposed in the model. This method is equivalent to adding perturbation to the seen class features,which can reduce the degree to which the classification results in the model are biased towards the seen classes.At the same time, this feature fusion method can provide part of the information of the unseen classes, improvingits classification accuracy in generalized zero-shot learning and reducing domain bias. The proposed method isvalidated and compared with othermethods on four datasets, and fromthe experimental results, it can be seen thatthe method proposed in this paper achieves promising results.展开更多
To detect radioactive substances with low activity levels,an anticoincidence detector and a high-purity germanium(HPGe)detector are typically used simultaneously to suppress Compton scattering background,thereby resul...To detect radioactive substances with low activity levels,an anticoincidence detector and a high-purity germanium(HPGe)detector are typically used simultaneously to suppress Compton scattering background,thereby resulting in an extremely low detection limit and improving the measurement accuracy.However,the complex and expensive hardware required does not facilitate the application or promotion of this method.Thus,a method is proposed in this study to discriminate the digital waveform of pulse signals output using an HPGe detector,whereby Compton scattering background is suppressed and a low minimum detectable activity(MDA)is achieved without using an expensive and complex anticoincidence detector and device.The electric-field-strength and energy-deposition distributions of the detector are simulated to determine the relationship between pulse shape and energy-deposition location,as well as the characteristics of energy-deposition distributions for fulland partial-energy deposition events.This relationship is used to develop a pulse-shape-discrimination algorithm based on an artificial neural network for pulse-feature identification.To accurately determine the relationship between the deposited energy of gamma(γ)rays in the detector and the deposition location,we extract four shape parameters from the pulse signals output by the detector.Machine learning is used to input the four shape parameters into the detector.Subsequently,the pulse signals are identified and classified to discriminate between partial-and full-energy deposition events.Some partial-energy deposition events are removed to suppress Compton scattering.The proposed method effectively decreases the MDA of an HPGeγ-energy dispersive spectrometer.Test results show that the Compton suppression factors for energy spectra obtained from measurements on ^(152)Eu,^(137)Cs,and ^(60)Co radioactive sources are 1.13(344 keV),1.11(662 keV),and 1.08(1332 keV),respectively,and that the corresponding MDAs are 1.4%,5.3%,and 21.6%lower,respectively.展开更多
Given the prominence and magnitude of airport incentive schemes,it is surprising that literature hitherto remains silent as to their effectiveness.In this paper,the relationship between airport incentive schemes and t...Given the prominence and magnitude of airport incentive schemes,it is surprising that literature hitherto remains silent as to their effectiveness.In this paper,the relationship between airport incentive schemes and the route development behavior of airlines is analyzed.Because of rare and often controversial findings in the extant literature regarding relevant influencing variables for attracting airlines at an airport,expert interviews are used as a complement to formulate testable hypotheses in this regard.A fixed effects regression model is used to test the hypotheses with a dataset that covers all seat capacity offered at the 22 largest German commercial airports in the week 46 from 2004 to 2011.It is found that incentives from primary choice,as well as secondary choice airports,have a significant influence on Low Cost Carriers.Furthermore,Low Cost Carriers,in general,do not leave any of both types of airports when the incentives cease.In the case of Network Carriers,no case is found where one joins a primary choice airport and receives an incentive.Insufficient data between Network Carriers and secondary choice airports in the time when incentives have ceased means that no statement can be given.展开更多
文摘Orthogonal Time Frequency and Space(OTFS) modulation is expected to provide high-speed and ultra-reliable communications for emerging mobile applications, including low-orbit satellite communications. Using the Doppler frequency for positioning is a promising research direction on communication and navigation integration. To tackle the high Doppler frequency and low signal-to-noise ratio(SNR) in satellite communication, this paper proposes a Red and Blue Frequency Shift Discriminator(RBFSD) based on the pseudo-noise(PN) sequence.The paper derives that the cross-correlation function on the Doppler domain exhibits the characteristic of a Sinc function. Therefore, it applies modulation onto the Delay-Doppler domain using PN sequence and adjusts Doppler frequency estimation by red-shifting or blue-shifting. Simulation results show that the performance of Doppler frequency estimation is close to the Cramér-Rao Lower Bound when the SNR is greater than -15dB. The proposed algorithm is about 1/D times less complex than the existing PN pilot sequence algorithm, where D is the resolution of the fractional Doppler.
基金partially supported by the National Key Research and Development Program of China(No.2018 AAA0100400)the Natural Science Foundation of Shandong Province(Nos.ZR2020MF131 and ZR2021ZD19)the Science and Technology Program of Qingdao(No.21-1-4-ny-19-nsh).
文摘Underwater pulse waveform recognition is an important method for underwater object detection.Most existing works focus on the application of traditional pattern recognition methods,which ignore the time-and space-varying characteristics in sound propagation channels and cannot easily extract valuable waveform features.Sound propagation channels in seawater are time-and space-varying convolutional channels.In the extraction of the waveform features of underwater acoustic signals,the effect of high-accuracy underwater acoustic signal recognition is identified by eliminating the influence of time-and space-varying convolutional channels to the greatest extent possible.We propose a hash aggregate discriminative network(HADN),which combines hash learning and deep learning to minimize the time-and space-varying effects on convolutional channels and adaptively learns effective underwater waveform features to achieve high-accuracy underwater pulse waveform recognition.In the extraction of the hash features of acoustic signals,a discrete constraint between clusters within a hash feature class is introduced.This constraint can ensure that the influence of convolutional channels on hash features is minimized.In addition,we design a new loss function called aggregate discriminative loss(AD-loss).The use of AD-loss and softmax-loss can increase the discriminativeness of the learned hash features.Experimental results show that on pool and ocean datasets,which were collected in pools and oceans,respectively,by using acoustic collectors,the proposed HADN performs better than other comparative models in terms of accuracy and mAP.
基金supported by the Scientific and Technological Innovation 2030 Major Project(2022ZD04019)the Science and Technology Innovation Capacity Building Project of BAAFS(KJCX20230303)+1 种基金Hainan Province Science and Technology Special Fund(ZDYF2023XDNY077)the Beijing Scholars Program(BSP041)。
文摘To reduce the cost and increase the efficiency of plant genetic marker fingerprinting for variety discrimination,it is desirable to identify the optimal marker combinations.We describe a marker combination screening model based on the genetic algorithm(GA)and implemented in a software tool,Loci Scan.Ratio-based variety discrimination power provided the largest optimization space among multiple fitness functions.Among GA parameters,an increase in population size and generation number enlarged optimization depth but also calculation workload.Exhaustive algorithm afforded the same optimization depth as GA but vastly increased calculation time.In comparison with two other software tools,Loci Scan accommodated missing data,reduced calculation time,and offered more fitness functions.In large datasets,the sample size of training data exerted the strongest influence on calculation time,whereas the marker size of training data showed no effect,and target marker number had limited effect on analysis speed.
基金financial support from the Science Research Program Project for Drug Regulation,Jiangsu Drug Administration,China(Grant No.:202207)the National Drug Standards Revision Project,China(Grant No.:2023Y41)+1 种基金the National Natural Science Foundation of China(Grant No.:22276080)the Foreign Expert Project,China(Grant No.:G2022014096L).
文摘Analyzing polysorbate 20(PS20)composition and the impact of each component on stability and safety is crucial due to formulation variations and individual tolerance.The similar structures and polarities of PS20 components make accurate separation,identification,and quantification challenging.In this work,a high-resolution quantitative method was developed using single-dimensional high-performance liquid chromatography(HPLC)with charged aerosol detection(CAD)to separate 18 key components with multiple esters.The separated components were characterized by ultra-high-performance liquid chromatography-quadrupole time-of-flight mass spectrometry(UHPLC-Q-TOF-MS)with an identical gradient as the HPLC-CAD analysis.The polysorbate compound database and library were expanded over 7-time compared to the commercial database.The method investigated differences in PS20 samples from various origins and grades for different dosage forms to evaluate the composition-process relationship.UHPLC-Q-TOF-MS identified 1329 to 1511 compounds in 4 batches of PS20 from different sources.The method observed the impact of 4 degradation conditions on peak components,identifying stable components and their tendencies to change.HPLC-CAD and UHPLC-Q-TOF-MS results provided insights into fingerprint differences,distinguishing quasi products.
基金the Center University(Grant No.B220202013)Qinglan Project of Jiangsu Province(2022).
文摘The objective of this study is to investigate themethods for soil liquefaction discrimination. Typically, predicting soilliquefaction potential involves conducting the standard penetration test (SPT), which requires field testing and canbe time-consuming and labor-intensive. In contrast, the cone penetration test (CPT) provides a more convenientmethod and offers detailed and continuous information about soil layers. In this study, the feature matrix based onCPT data is proposed to predict the standard penetration test blow count N. The featurematrix comprises the CPTcharacteristic parameters at specific depths, such as tip resistance qc, sleeve resistance f s, and depth H. To fuse thefeatures on the matrix, the convolutional neural network (CNN) is employed for feature extraction. Additionally,Genetic Algorithm (GA) is utilized to obtain the best combination of convolutional kernels and the number ofneurons. The study evaluated the robustness of the proposed model using multiple engineering field data sets.Results demonstrated that the proposed model outperformed conventional methods in predicting N values forvarious soil categories, including sandy silt, silty sand, and clayey silt. Finally, the proposed model was employedfor liquefaction discrimination. The liquefaction discrimination based on the predicted N values was comparedwith the measured N values, and the results showed that the discrimination results were in 75% agreement. Thestudy has important practical application value for foundation liquefaction engineering. Also, the novel methodadopted in this research provides new ideas and methods for research in related fields, which is of great academicsignificance.
基金supported by the Key R&D Program of Shaanxi Province,China(Grant Nos.2022GY-274,2023-YBSF-505)the National Natural Science Foundation of China(Grant No.62273273).
文摘Recovering high-quality inscription images from unknown and complex inscription noisy images is a challenging research issue.Different fromnatural images,character images pay more attention to stroke information.However,existingmodelsmainly consider pixel-level informationwhile ignoring structural information of the character,such as its edge and glyph,resulting in reconstructed images with mottled local structure and character damage.To solve these problems,we propose a novel generative adversarial network(GAN)framework based on an edge-guided generator and a discriminator constructed by a dual-domain U-Net framework,i.e.,EDU-GAN.Unlike existing frameworks,the generator introduces the edge extractionmodule,guiding it into the denoising process through the attention mechanism,which maintains the edge detail of the restored inscription image.Moreover,a dual-domain U-Net-based discriminator is proposed to learn the global and local discrepancy between the denoised and the label images in both image and morphological domains,which is helpful to blind denoising tasks.The proposed dual-domain discriminator and generator for adversarial training can reduce local artifacts and keep the denoised character structure intact.Due to the lack of a real-inscription image,we built the real-inscription dataset to provide an effective benchmark for studying inscription image denoising.The experimental results show the superiority of our method both in the synthetic and real-inscription datasets.
基金supported by the Fundamental Research Funds for the Central Universities(WK2470000035)USTC Research Funds of the Double First-Class Initiative(YD2030002007,YD2030002011)+1 种基金the National Natural Science Foundation of China(62222512,12104439,12134014,and 11974335)the Anhui Provincial Natural Science Foundation(2208085J03).
文摘Extracting more information and saving quantum resources are two main aims for quantum measurements.However,the optimization of strategies for these two objectives varies when discriminating between quantum states |ψ_(0)> and |ψ_(1)> through multiple measurements.In this study,we introduce a novel state discrimination model that reveals the intricate relationship between the average error rate and average copy consumption.By integrating these two crucial metrics and minimizing their weighted sum for any given weight value,our research underscores the infeasibility of simultaneously minimizing these metrics through local measurements with one-way communication.Our findings present a compelling trade-off curve,highlighting the advantages of achieving a balance between error rate and copy consumption in quantum discrimination tasks,offering valuable insights into the optimization of quantum resources while ensuring the accuracy of quantum state discrimination.
基金supported by the National Magnetic Confinement Fusion Program of China(No.2019YFE03020002)the National Natural Science Foundation of China(Nos.12205085 and12125502)。
文摘Fast neutron flux measurements with high count rates and high time resolution have important applications in equipment such as tokamaks.In this study,real-time neutron and gamma discrimination was implemented on a self-developed 500-Msps,12-bit digitizer,and the neutron and gamma spectra were calculated directly on an FPGA.A fast neutron flux measurement system with BC-501A and EJ-309 liquid scintillator detectors was developed and a fast neutron measurement experiment was successfully performed on the HL-2 M tokamak at the Southwestern Institute of Physics,China.The experimental results demonstrated that the system obtained the neutron and gamma spectra with a time accuracy of 1 ms.At count rates of up to 1 Mcps,the figure of merit was greater than 1.05 for energies between 50 keV and 2.8 MeV.
基金the Natural Science Foundation of Henan Province(232300420094)the Science and TechnologyResearch Project of Henan Province(222102220092).
文摘Intelligent diagnosis driven by big data for mechanical fault is an important means to ensure the safe operation ofequipment. In these methods, deep learning-based machinery fault diagnosis approaches have received increasingattention and achieved some results. It might lead to insufficient performance for using transfer learning alone andcause misclassification of target samples for domain bias when building deep models to learn domain-invariantfeatures. To address the above problems, a deep discriminative adversarial domain adaptation neural networkfor the bearing fault diagnosis model is proposed (DDADAN). In this method, the raw vibration data are firstlyconverted into frequency domain data by Fast Fourier Transform, and an improved deep convolutional neuralnetwork with wide first-layer kernels is used as a feature extractor to extract deep fault features. Then, domaininvariant features are learned from the fault data with correlation alignment-based domain adversarial training.Furthermore, to enhance the discriminative property of features, discriminative feature learning is embeddedinto this network to make the features compact, as well as separable between classes within the class. Finally, theperformance and anti-noise capability of the proposedmethod are evaluated using two sets of bearing fault datasets.The results demonstrate that the proposed method is capable of handling domain offset caused by differentworkingconditions and maintaining more than 97.53% accuracy on various transfer tasks. Furthermore, the proposedmethod can achieve high diagnostic accuracy under varying noise levels.
文摘Zero-shot learning enables the recognition of new class samples by migrating models learned from semanticfeatures and existing sample features to things that have never been seen before. The problems of consistencyof different types of features and domain shift problems are two of the critical issues in zero-shot learning. Toaddress both of these issues, this paper proposes a new modeling structure. The traditional approach mappedsemantic features and visual features into the same feature space;based on this, a dual discriminator approachis used in the proposed model. This dual discriminator approach can further enhance the consistency betweensemantic and visual features. At the same time, this approach can also align unseen class semantic features andtraining set samples, providing a portion of information about the unseen classes. In addition, a new feature fusionmethod is proposed in the model. This method is equivalent to adding perturbation to the seen class features,which can reduce the degree to which the classification results in the model are biased towards the seen classes.At the same time, this feature fusion method can provide part of the information of the unseen classes, improvingits classification accuracy in generalized zero-shot learning and reducing domain bias. The proposed method isvalidated and compared with othermethods on four datasets, and fromthe experimental results, it can be seen thatthe method proposed in this paper achieves promising results.
基金This work was supported by the National Key R&D Program of China(Nos.2022YFF0709503,2022YFB1902700,2017YFC0602101)the Key Research and Development Program of Sichuan province(No.2023YFG0347)the Key Research and Development Program of Sichuan province(No.2020ZDZX0007).
文摘To detect radioactive substances with low activity levels,an anticoincidence detector and a high-purity germanium(HPGe)detector are typically used simultaneously to suppress Compton scattering background,thereby resulting in an extremely low detection limit and improving the measurement accuracy.However,the complex and expensive hardware required does not facilitate the application or promotion of this method.Thus,a method is proposed in this study to discriminate the digital waveform of pulse signals output using an HPGe detector,whereby Compton scattering background is suppressed and a low minimum detectable activity(MDA)is achieved without using an expensive and complex anticoincidence detector and device.The electric-field-strength and energy-deposition distributions of the detector are simulated to determine the relationship between pulse shape and energy-deposition location,as well as the characteristics of energy-deposition distributions for fulland partial-energy deposition events.This relationship is used to develop a pulse-shape-discrimination algorithm based on an artificial neural network for pulse-feature identification.To accurately determine the relationship between the deposited energy of gamma(γ)rays in the detector and the deposition location,we extract four shape parameters from the pulse signals output by the detector.Machine learning is used to input the four shape parameters into the detector.Subsequently,the pulse signals are identified and classified to discriminate between partial-and full-energy deposition events.Some partial-energy deposition events are removed to suppress Compton scattering.The proposed method effectively decreases the MDA of an HPGeγ-energy dispersive spectrometer.Test results show that the Compton suppression factors for energy spectra obtained from measurements on ^(152)Eu,^(137)Cs,and ^(60)Co radioactive sources are 1.13(344 keV),1.11(662 keV),and 1.08(1332 keV),respectively,and that the corresponding MDAs are 1.4%,5.3%,and 21.6%lower,respectively.
文摘Given the prominence and magnitude of airport incentive schemes,it is surprising that literature hitherto remains silent as to their effectiveness.In this paper,the relationship between airport incentive schemes and the route development behavior of airlines is analyzed.Because of rare and often controversial findings in the extant literature regarding relevant influencing variables for attracting airlines at an airport,expert interviews are used as a complement to formulate testable hypotheses in this regard.A fixed effects regression model is used to test the hypotheses with a dataset that covers all seat capacity offered at the 22 largest German commercial airports in the week 46 from 2004 to 2011.It is found that incentives from primary choice,as well as secondary choice airports,have a significant influence on Low Cost Carriers.Furthermore,Low Cost Carriers,in general,do not leave any of both types of airports when the incentives cease.In the case of Network Carriers,no case is found where one joins a primary choice airport and receives an incentive.Insufficient data between Network Carriers and secondary choice airports in the time when incentives have ceased means that no statement can be given.