The widespread utilisation of tunnel boring machines(TBMs)in underground construction engineering requires a detailed investigation of the cutter-rock interaction.In this paper,we conduct a series of largescale standi...The widespread utilisation of tunnel boring machines(TBMs)in underground construction engineering requires a detailed investigation of the cutter-rock interaction.In this paper,we conduct a series of largescale standing rotary cutting tests on granite in conjunction with high-fidelity numerical simulations based on a particle-type discrete element method(DEM)to explore the effects of key cutting parameters on the TBM cutter performance and the distribution of cutter-rock contact stresses.The assessment results of cutter performance obtained from the cutting tests and numerical simulations reveal similar dependencies on the key cutting parameters.More specifically,the normal and rolling forces exhibit a positive correlation with penetration but are slightly influenced by the cutting radius.In contrast,the side force decreases as the cutting radius increases.Additionally,the side force shows a positive relationship with the penetration for smaller cutting radii but tends to become negative as the cutting radius increases.The cutter's relative effectiveness in rock breaking is significantly impacted by the penetration but shows little dependency on the cutting radius.Consequently,an optimal penetration is identified,leading to a low boreability index and specific energy.A combined Hertz-Weibull function is developed to fit the cutter-rock contact stress distribution obtained in DEM simulations,whereby an improved CSM(Colorado School of Mines)model is proposed by replacing the original monotonic cutting force distribution with this combined Hertz-Weibull model.The proposed model outperforms the original CSM model as demonstrated by a comparison of the estimated cutting forces with those from the tests/simulations.The findings from this work that advance our understanding of TBM cutter performance have important implications for improving the efficiency and reliability of TBM tunnelling in granite.展开更多
Bedding slope is a typical heterogeneous slope consisting of different soil/rock layers and is likely to slide along the weakest interface.Conventional slope protection methods for bedding slopes,such as retaining wal...Bedding slope is a typical heterogeneous slope consisting of different soil/rock layers and is likely to slide along the weakest interface.Conventional slope protection methods for bedding slopes,such as retaining walls,stabilizing piles,and anchors,are time-consuming and labor-and energy-intensive.This study proposes an innovative polymer grout method to improve the bearing capacity and reduce the displacement of bedding slopes.A series of large-scale model tests were carried out to verify the effectiveness of polymer grout in protecting bedding slopes.Specifically,load-displacement relationships and failure patterns were analyzed for different testing slopes with various dosages of polymer.Results show the great potential of polymer grout in improving bearing capacity,reducing settlement,and protecting slopes from being crushed under shearing.The polymer-treated slopes remained structurally intact,while the untreated slope exhibited considerable damage when subjected to loads surpassing the bearing capacity.It is also found that polymer-cemented soils concentrate around the injection pipe,forming a fan-shaped sheet-like structure.This study proves the improvement of polymer grouting for bedding slope treatment and will contribute to the development of a fast method to protect bedding slopes from landslides.展开更多
Traditional large-scale multi-objective optimization algorithms(LSMOEAs)encounter difficulties when dealing with sparse large-scale multi-objective optimization problems(SLM-OPs)where most decision variables are zero....Traditional large-scale multi-objective optimization algorithms(LSMOEAs)encounter difficulties when dealing with sparse large-scale multi-objective optimization problems(SLM-OPs)where most decision variables are zero.As a result,many algorithms use a two-layer encoding approach to optimize binary variable Mask and real variable Dec separately.Nevertheless,existing optimizers often focus on locating non-zero variable posi-tions to optimize the binary variables Mask.However,approxi-mating the sparse distribution of real Pareto optimal solutions does not necessarily mean that the objective function is optimized.In data mining,it is common to mine frequent itemsets appear-ing together in a dataset to reveal the correlation between data.Inspired by this,we propose a novel two-layer encoding learning swarm optimizer based on frequent itemsets(TELSO)to address these SLMOPs.TELSO mined the frequent terms of multiple particles with better target values to find mask combinations that can obtain better objective values for fast convergence.Experi-mental results on five real-world problems and eight benchmark sets demonstrate that TELSO outperforms existing state-of-the-art sparse large-scale multi-objective evolutionary algorithms(SLMOEAs)in terms of performance and convergence speed.展开更多
Assessment of past-climate simulations of regional climate models(RCMs)is important for understanding the reliability of RCMs when used to project future regional climate.Here,we assess the performance and discuss pos...Assessment of past-climate simulations of regional climate models(RCMs)is important for understanding the reliability of RCMs when used to project future regional climate.Here,we assess the performance and discuss possible causes of biases in a WRF-based RCM with a grid spacing of 50 km,named WRFG,from the North American Regional Climate Change Assessment Program(NARCCAP)in simulating wet season precipitation over the Central United States for a period when observational data are available.The RCM reproduces key features of the precipitation distribution characteristics during late spring to early summer,although it tends to underestimate the magnitude of precipitation.This dry bias is partially due to the model’s lack of skill in simulating nocturnal precipitation related to the lack of eastward propagating convective systems in the simulation.Inaccuracy in reproducing large-scale circulation and environmental conditions is another contributing factor.The too weak simulated pressure gradient between the Rocky Mountains and the Gulf of Mexico results in weaker southerly winds in between,leading to a reduction of warm moist air transport from the Gulf to the Central Great Plains.The simulated low-level horizontal convergence fields are less favorable for upward motion than in the NARR and hence,for the development of moist convection as well.Therefore,a careful examination of an RCM’s deficiencies and the identification of the source of errors are important when using the RCM to project precipitation changes in future climate scenarios.展开更多
Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to tr...Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to traverse vast expanse with limited computational resources.Furthermore,in the context of sparse,most variables in Pareto optimal solutions are zero,making it difficult for algorithms to identify non-zero variables efficiently.This paper is dedicated to addressing the challenges posed by SLMOPs.To start,we introduce innovative objective functions customized to mine maximum and minimum candidate sets.This substantial enhancement dramatically improves the efficacy of frequent pattern mining.In this way,selecting candidate sets is no longer based on the quantity of nonzero variables they contain but on a higher proportion of nonzero variables within specific dimensions.Additionally,we unveil a novel approach to association rule mining,which delves into the intricate relationships between non-zero variables.This novel methodology aids in identifying sparse distributions that can potentially expedite reductions in the objective function value.We extensively tested our algorithm across eight benchmark problems and four real-world SLMOPs.The results demonstrate that our approach achieves competitive solutions across various challenges.展开更多
In situ inflow and outflow permeability tests with the BAT probe at SarapuíII soft clay test site are presented.A description of the BAT permeability test is provided,discussing its advantages and shortcomings,es...In situ inflow and outflow permeability tests with the BAT probe at SarapuíII soft clay test site are presented.A description of the BAT permeability test is provided,discussing its advantages and shortcomings,especially in the case of very soft clays under low stresses.Pore pressures were monitored during probe installation and were found to be slightly lower than piezocone u2 pore pressures,consistent with the position of the filter.The role of filter tip saturation was investigated after the usual saturation procedure provided an unsatisfactory pore pressure response during probe installation.Results show that the vacuum saturation procedure provides adequate response during installation and increases the reliability of the coefficient of permeability determination in early measurements.Both inflow and outflow tests yielded similar results,indicating that careful execution of the test can lead to good test repeatability regardless of the loading condition.Various sequences of alternated inflow and outflow tests have yielded similar results,indicating that soil reconsolidation and filter clogging were negligible in the tests performed.Data are presented concerning the relationship between index parameters and the in situ coefficient of permeability for SarapuíII clay,which plot outside the range of existing databases.展开更多
Objective:Serological tests are widely used for scrub typhus diagnosis;however,their limitations are evident.This study aims to assess their practical value in clinical settings.Methods:We analyzed the data of adult p...Objective:Serological tests are widely used for scrub typhus diagnosis;however,their limitations are evident.This study aims to assess their practical value in clinical settings.Methods:We analyzed the data of adult patients with suspected scrub typhus who visited a tertiary care hospital in the Republic of Korea from September to December from 2019 to 2021.The included patients had an acute fever and at least one of the following ten secondary findings:myalgia,skin rash,eschar,headache,thrombocytopenia,increased liver enzyme levels,lymphadenopathy,hepatomegaly,splenomegaly,and pleural effusion.The diagnoses were grouped as scrub typhus or other diseases by two infectious disease physicians.Results:Among 136 patients who met the eligibility criteria,109 had scrub typhus and 27 had different diseases.Single and paired total antibodies using immunofluorescence assay(IFA),and total antibodies using immunochromatography-based rapid diagnostic testing(ICT)were measured in 98%,22%,and 75%of all patients,respectively.Confirmation using paired samples for scrub typhus was established at a median of 11[interquartile range(IQR)10-16]days following the first visit.Among the 82 admitted patients,the median admission time was 9(IQR 7-13)days.According to IFA,58(55%)patients with scrub typhus had total immunoglobulin titers≥1:320,while 23(85%)patients with other disease had titers<1:320.Positive ICT results were observed in 64(74%)patients with scrub typhus and 10(67%)patients with other diseases showed negative ICT results.Conclusions:Serological testing for scrub typhus is currently insufficient for decision-making in clinical practice.展开更多
This article introduces the concept of load aggregation,which involves a comprehensive analysis of loads to acquire their external characteristics for the purpose of modeling and analyzing power systems.The online ide...This article introduces the concept of load aggregation,which involves a comprehensive analysis of loads to acquire their external characteristics for the purpose of modeling and analyzing power systems.The online identification method is a computer-involved approach for data collection,processing,and system identification,commonly used for adaptive control and prediction.This paper proposes a method for dynamically aggregating large-scale adjustable loads to support high proportions of new energy integration,aiming to study the aggregation characteristics of regional large-scale adjustable loads using online identification techniques and feature extraction methods.The experiment selected 300 central air conditioners as the research subject and analyzed their regulation characteristics,economic efficiency,and comfort.The experimental results show that as the adjustment time of the air conditioner increases from 5 minutes to 35 minutes,the stable adjustment quantity during the adjustment period decreases from 28.46 to 3.57,indicating that air conditioning loads can be controlled over a long period and have better adjustment effects in the short term.Overall,the experimental results of this paper demonstrate that analyzing the aggregation characteristics of regional large-scale adjustable loads using online identification techniques and feature extraction algorithms is effective.展开更多
Accurate positioning is one of the essential requirements for numerous applications of remote sensing data,especially in the event of a noisy or unreliable satellite signal.Toward this end,we present a novel framework...Accurate positioning is one of the essential requirements for numerous applications of remote sensing data,especially in the event of a noisy or unreliable satellite signal.Toward this end,we present a novel framework for aircraft geo-localization in a large range that only requires a downward-facing monocular camera,an altimeter,a compass,and an open-source Vector Map(VMAP).The algorithm combines the matching and particle filter methods.Shape vector and correlation between two building contour vectors are defined,and a coarse-to-fine building vector matching(CFBVM)method is proposed in the matching stage,for which the original matching results are described by the Gaussian mixture model(GMM).Subsequently,an improved resampling strategy is designed to reduce computing expenses with a huge number of initial particles,and a credibility indicator is designed to avoid location mistakes in the particle filter stage.An experimental evaluation of the approach based on flight data is provided.On a flight at a height of 0.2 km over a flight distance of 2 km,the aircraft is geo-localized in a reference map of 11,025 km~2using 0.09 km~2aerial images without any prior information.The absolute localization error is less than 10 m.展开更多
The longitudinal dispersion of the projectile in shooting tests of two-dimensional trajectory corrections fused with fixed canards is extremely large that it sometimes exceeds the correction ability of the correction ...The longitudinal dispersion of the projectile in shooting tests of two-dimensional trajectory corrections fused with fixed canards is extremely large that it sometimes exceeds the correction ability of the correction fuse actuator.The impact point easily deviates from the target,and thus the correction result cannot be readily evaluated.However,the cost of shooting tests is considerably high to conduct many tests for data collection.To address this issue,this study proposes an aiming method for shooting tests based on small sample size.The proposed method uses the Bootstrap method to expand the test data;repeatedly iterates and corrects the position of the simulated theoretical impact points through an improved compatibility test method;and dynamically adjusts the weight of the prior distribution of simulation results based on Kullback-Leibler divergence,which to some extent avoids the real data being"submerged"by the simulation data and achieves the fusion Bayesian estimation of the dispersion center.The experimental results show that when the simulation accuracy is sufficiently high,the proposed method yields a smaller mean-square deviation in estimating the dispersion center and higher shooting accuracy than those of the three comparison methods,which is more conducive to reflecting the effect of the control algorithm and facilitating test personnel to iterate their proposed structures and algorithms.;in addition,this study provides a knowledge base for further comprehensive studies in the future.展开更多
The large-scale multi-objective optimization algorithm(LSMOA),based on the grouping of decision variables,is an advanced method for handling high-dimensional decision variables.However,in practical problems,the intera...The large-scale multi-objective optimization algorithm(LSMOA),based on the grouping of decision variables,is an advanced method for handling high-dimensional decision variables.However,in practical problems,the interaction among decision variables is intricate,leading to large group sizes and suboptimal optimization effects;hence a large-scale multi-objective optimization algorithm based on weighted overlapping grouping of decision variables(MOEAWOD)is proposed in this paper.Initially,the decision variables are perturbed and categorized into convergence and diversity variables;subsequently,the convergence variables are subdivided into groups based on the interactions among different decision variables.If the size of a group surpasses the set threshold,that group undergoes a process of weighting and overlapping grouping.Specifically,the interaction strength is evaluated based on the interaction frequency and number of objectives among various decision variables.The decision variable with the highest interaction in the group is identified and disregarded,and the remaining variables are then reclassified into subgroups.Finally,the decision variable with the strongest interaction is added to each subgroup.MOEAWOD minimizes the interactivity between different groups and maximizes the interactivity of decision variables within groups,which contributed to the optimized direction of convergence and diversity exploration with different groups.MOEAWOD was subjected to testing on 18 benchmark large-scale optimization problems,and the experimental results demonstrate the effectiveness of our methods.Compared with the other algorithms,our method is still at an advantage.展开更多
With the development of big data and social computing,large-scale group decisionmaking(LGDM)is nowmerging with social networks.Using social network analysis(SNA),this study proposes an LGDM consensus model that consid...With the development of big data and social computing,large-scale group decisionmaking(LGDM)is nowmerging with social networks.Using social network analysis(SNA),this study proposes an LGDM consensus model that considers the trust relationship among decisionmakers(DMs).In the process of consensusmeasurement:the social network is constructed according to the social relationship among DMs,and the Louvain method is introduced to classify social networks to form subgroups.In this study,the weights of each decision maker and each subgroup are computed by comprehensive network weights and trust weights.In the process of consensus improvement:A feedback mechanism with four identification and two direction rules is designed to guide the consensus of the improvement process.Based on the trust relationship among DMs,the preferences are modified,and the corresponding social network is updated to accelerate the consensus.Compared with the previous research,the proposedmodel not only allows the subgroups to be reconstructed and updated during the adjustment process,but also improves the accuracy of the adjustment by the feedbackmechanism.Finally,an example analysis is conducted to verify the effectiveness and flexibility of the proposed method.Moreover,compared with previous studies,the superiority of the proposed method in solving the LGDM problem is highlighted.展开更多
Underground salt cavern CO_(2) storage(SCCS)offers the dual benefits of enabling extensive CO_(2) storage and facilitating the utilization of CO_(2) resources while contributing the regulation of the carbon market.Its...Underground salt cavern CO_(2) storage(SCCS)offers the dual benefits of enabling extensive CO_(2) storage and facilitating the utilization of CO_(2) resources while contributing the regulation of the carbon market.Its economic and operational advantages over traditional carbon capture,utilization,and storage(CCUS)projects make SCCS a more cost-effective and flexible option.Despite the widespread use of salt caverns for storing various substances,differences exist between SCCS and traditional salt cavern energy storage in terms of gas-tightness,carbon injection,brine extraction control,long-term carbon storage stability,and site selection criteria.These distinctions stem from the unique phase change characteristics of CO_(2) and the application scenarios of SCCS.Therefore,targeted and forward-looking scientific research on SCCS is imperative.This paper introduces the implementation principles and application scenarios of SCCS,emphasizing its connections with carbon emissions,carbon utilization,and renewable energy peak shaving.It delves into the operational characteristics and economic advantages of SCCS compared with other CCUS methods,and addresses associated scientific challenges.In this paper,we establish a pressure equation for carbon injection and brine extraction,that considers the phase change characteristics of CO_(2),and we analyze the pressure during carbon injection.By comparing the viscosities of CO_(2) and other gases,SCCS’s excellent sealing performance is demonstrated.Building on this,we develop a long-term stability evaluation model and associated indices,which analyze the impact of the injection speed and minimum operating pressure on stability.Field countermeasures to ensure stability are proposed.Site selection criteria for SCCS are established,preliminary salt mine sites suitable for SCCS are identified in China,and an initial estimate of achievable carbon storage scale in China is made at over 51.8-77.7 million tons,utilizing only 20%-30%volume of abandoned salt caverns.This paper addresses key scientific and engineering challenges facing SCCS and determines crucial technical parameters,such as the operating pressure,burial depth,and storage scale,and it offers essential guidance for implementing SCCS projects in China.展开更多
We introduce a factorized Smith method(FSM)for solving large-scale highranked T-Stein equations within the banded-plus-low-rank structure framework.To effectively reduce both computational complexity and storage requi...We introduce a factorized Smith method(FSM)for solving large-scale highranked T-Stein equations within the banded-plus-low-rank structure framework.To effectively reduce both computational complexity and storage requirements,we develop techniques including deflation and shift,partial truncation and compression,as well as redesign the residual computation and termination condition.Numerical examples demonstrate that the FSM outperforms the Smith method implemented with a hierarchical HODLR structured toolkit in terms of CPU time.展开更多
Background A task assigned to space exploration satellites involves detecting the physical environment within a certain space.However,space detection data are complex and abstract.These data are not conducive for rese...Background A task assigned to space exploration satellites involves detecting the physical environment within a certain space.However,space detection data are complex and abstract.These data are not conducive for researchers'visual perceptions of the evolution and interaction of events in the space environment.Methods A time-series dynamic data sampling method for large-scale space was proposed for sample detection data in space and time,and the corresponding relationships between data location features and other attribute features were established.A tone-mapping method based on statistical histogram equalization was proposed and applied to the final attribute feature data.The visualization process is optimized for rendering by merging materials,reducing the number of patches,and performing other operations.Results The results of sampling,feature extraction,and uniform visualization of the detection data of complex types,long duration spans,and uneven spatial distributions were obtained.The real-time visualization of large-scale spatial structures using augmented reality devices,particularly low-performance devices,was also investigated.Conclusions The proposed visualization system can reconstruct the three-dimensional structure of a large-scale space,express the structure and changes in the spatial environment using augmented reality,and assist in intuitively discovering spatial environmental events and evolutionary rules.展开更多
This paper explores the performances of a finite element simulation including four concrete models applied to a full-scale reinforced concrete beam subjected to blast loading. Field test data has been used to compare ...This paper explores the performances of a finite element simulation including four concrete models applied to a full-scale reinforced concrete beam subjected to blast loading. Field test data has been used to compare model results for each case. The numerical modelling has been, carried out using the suitable code LS-DYNA. This code integrates blast load routine(CONWEP) for the explosive description and four different material models for the concrete including: Karagozian & Case Concrete, Winfrith, Continuous Surface Cap Model and Riedel-Hiermaier-Thoma models, with concrete meshing based on 10, 15, and 20 mm. Six full-scale beams were tested: four of them used for the initial calibration of the numerical model and two more tests at lower scaled distances. For calibration, field data obtained employing pressure and accelerometers transducers were compared with the results derived from the numerical simulation. Damage surfaces and the shape of rupture in the beams have been used as references for comparison. Influence of the meshing on accelerations has been put in evidence and for some models the shape and size of the damage in the beams produced maximum differences around 15%. In all cases, the variations between material and mesh models are shown and discussed.展开更多
BACKGROUND A reliable test is essential for diagnosing Helicobacter pylori(H.pylori)infection,and crucial for managing H.pylori-related diseases.Serving as an excellent method for detecting H.pylori infection,histolog...BACKGROUND A reliable test is essential for diagnosing Helicobacter pylori(H.pylori)infection,and crucial for managing H.pylori-related diseases.Serving as an excellent method for detecting H.pylori infection,histologic examination is a test that clinicians heavily rely on,especially when complemented with immunohistochemistry(IHC).Additionally,other diagnostic tests for H.pylori,such as the rapid urease test(CLO test)and stool antigen test(SA),are also highly sensitive and specific.Typically,the results of histology and other tests align with each other.However,on rare occasions,discrepancy between histopathology and other H.pylori diagnostic tests occurs.AIM To investigate the discordance between histology and other H.pylori tests,the underlying causes,and the impact on clinical management.METHODS Pathology reports of gastric biopsies were retrieved spanning August 2013 and July 2018.Reports were included in the study only if there were other H.pylori tests within seven days of the biopsy.These additional tests include CLO test,SA,and H.pylori culture.Concordance between histopathology and other tests was determined based on the consistency of results.In instances where histology results were negative while other tests were positive,the slides were retrieved for re-assessment,and the clinical chart was reviewed.RESULTS Of 1396 pathology reports were identified,each accompanied by one additional H.pylori test.The concordance rates in detecting H.pylori infection between biopsy and other tests did not exhibit significant differences based on the number of biopsy fragments.117 discrepant cases were identified.Only 20 cases(9 with CLO test and 11 with SA)had negative biopsy but positive results in other tests.Four cases initially stained with Warthin-Starry turned out to be positive for H.pylori with subsequent IHC staining.Among the remaining 16 true discrepant cases,10 patients were on proton pump inhibitors before the biopsy and/or other tests.Most patients underwent treatment,except for two who were untreated,and two patients who were lost to follow-up.CONCLUSION There are rare discrepant cases with negative biopsy but positive in SA or CLO test.Various factors may contribute to this inconsistency.Most patients in such cases had undergone treatment.展开更多
To investigate the influence of structural parameters on the performances and internal flow characteristics of partial flow pumps at a low specific speed of 10000 rpm,special attention was paid to the first and second...To investigate the influence of structural parameters on the performances and internal flow characteristics of partial flow pumps at a low specific speed of 10000 rpm,special attention was paid to the first and second stage impeller guide vanes.Moreover,the impeller blade outlet width,impeller inlet diameter,blade inclination angle,and number of blades were considered for orthogonal tests.Accordingly,nine groups of design solutions were formed,and then used as a basis for the execution of numerical simulations(CFD)aimed at obtaining the efficiency values and heads for each design solution group.The influence of impeller geometric parameters on the efficiency and head was explored,and the“weight”of each factor was obtained via a range analysis.Optimal structural parameters were finally chosen on the basis of the numerical simulation results,and the performances of the optimized model were verified accordingly(yet by means of CFD).Evidence is provided that the increase in the efficiency and head of the optimized model was 12.11%and 23.5 m,respectively,compared with those of the original model.展开更多
Eye diagnosis is a method for inspecting systemic diseases and syndromes by observing the eyes.With the development of intelligent diagnosis in traditional Chinese medicine(TCM);artificial intelligence(AI)can improve ...Eye diagnosis is a method for inspecting systemic diseases and syndromes by observing the eyes.With the development of intelligent diagnosis in traditional Chinese medicine(TCM);artificial intelligence(AI)can improve the accuracy and efficiency of eye diagnosis.However;the research on intelligent eye diagnosis still faces many challenges;including the lack of standardized and precisely labeled data;multi-modal information analysis;and artificial in-telligence models for syndrome differentiation.The widespread application of AI models in medicine provides new insights and opportunities for the research of eye diagnosis intelli-gence.This study elaborates on the three key technologies of AI models in the intelligent ap-plication of TCM eye diagnosis;and explores the implications for the research of eye diagno-sis intelligence.First;a database concerning eye diagnosis was established based on self-su-pervised learning so as to solve the issues related to the lack of standardized and precisely la-beled data.Next;the cross-modal understanding and generation of deep neural network models to address the problem of lacking multi-modal information analysis.Last;the build-ing of data-driven models for eye diagnosis to tackle the issue of the absence of syndrome dif-ferentiation models.In summary;research on intelligent eye diagnosis has great potential to be applied the surge of AI model applications.展开更多
The Sichuan-Tibet transportation corridor is prone to numerous active faults and frequent strong earthquakes.While extensive studies have individually explored the effect of active faults and strong earthquakes on dif...The Sichuan-Tibet transportation corridor is prone to numerous active faults and frequent strong earthquakes.While extensive studies have individually explored the effect of active faults and strong earthquakes on different engineering structures,their combined effect remains unclear.This research employed multiple physical model tests to investigate the dynamic response of various engineering structures,including tunnels,bridges,and embankments,under the simultaneous influence of cumulative earthquakes and stick-slip misalignment of an active fault.The prototype selected for this study was the Kanding No.2 tunnel,which crosses the Yunongxi fault zone within the Sichuan-Tibet transportation corridor.The results demonstrated that the tunnel,bridge,and embankment exhibited amplification in response to the input seismic wave,with the amplification effect gradually decreasing as the input peak ground acceleration(PGA)increased.The PGAs of different engineering structures were weakened by the fault rupture zone.Nevertheless,the misalignment of the active fault may decrease the overall stiffness of the engineering structure,leading to more severe damage,with a small contribution from seismic vibration.Additionally,the seismic vibration effect might be enlarged with the height of the engineering structure,and the tunnel is supposed to have a smaller PGA and lower dynamic earth pressure compared to bridges and embankments in strong earthquake zones crossing active faults.The findings contribute valuable insights for evaluating the dynamic response of various engineering structures crossing an active fault and provide an experimental reference for secure engineering design in the challenging conditions of the Sichuan-Tibet transportation corridor.展开更多
基金supported by the National Natural Science Foundation of China(Grant Nos.52278407 and 52378407)the China Postdoctoral Science Foundation(Grant No.2023M732670)the support by the Postdoctoral Fellowship Program of China Postdoctoral Science Foundation.
文摘The widespread utilisation of tunnel boring machines(TBMs)in underground construction engineering requires a detailed investigation of the cutter-rock interaction.In this paper,we conduct a series of largescale standing rotary cutting tests on granite in conjunction with high-fidelity numerical simulations based on a particle-type discrete element method(DEM)to explore the effects of key cutting parameters on the TBM cutter performance and the distribution of cutter-rock contact stresses.The assessment results of cutter performance obtained from the cutting tests and numerical simulations reveal similar dependencies on the key cutting parameters.More specifically,the normal and rolling forces exhibit a positive correlation with penetration but are slightly influenced by the cutting radius.In contrast,the side force decreases as the cutting radius increases.Additionally,the side force shows a positive relationship with the penetration for smaller cutting radii but tends to become negative as the cutting radius increases.The cutter's relative effectiveness in rock breaking is significantly impacted by the penetration but shows little dependency on the cutting radius.Consequently,an optimal penetration is identified,leading to a low boreability index and specific energy.A combined Hertz-Weibull function is developed to fit the cutter-rock contact stress distribution obtained in DEM simulations,whereby an improved CSM(Colorado School of Mines)model is proposed by replacing the original monotonic cutting force distribution with this combined Hertz-Weibull model.The proposed model outperforms the original CSM model as demonstrated by a comparison of the estimated cutting forces with those from the tests/simulations.The findings from this work that advance our understanding of TBM cutter performance have important implications for improving the efficiency and reliability of TBM tunnelling in granite.
基金supported by the Fujian Science Foundation for Outstanding Youth(Grant No.2023J06039)the National Natural Science Foundation of China(Grant No.41977259 and No.U2005205)Fujian Province natural resources science and technology innovation project(Grant No.KY-090000-04-2022-019)。
文摘Bedding slope is a typical heterogeneous slope consisting of different soil/rock layers and is likely to slide along the weakest interface.Conventional slope protection methods for bedding slopes,such as retaining walls,stabilizing piles,and anchors,are time-consuming and labor-and energy-intensive.This study proposes an innovative polymer grout method to improve the bearing capacity and reduce the displacement of bedding slopes.A series of large-scale model tests were carried out to verify the effectiveness of polymer grout in protecting bedding slopes.Specifically,load-displacement relationships and failure patterns were analyzed for different testing slopes with various dosages of polymer.Results show the great potential of polymer grout in improving bearing capacity,reducing settlement,and protecting slopes from being crushed under shearing.The polymer-treated slopes remained structurally intact,while the untreated slope exhibited considerable damage when subjected to loads surpassing the bearing capacity.It is also found that polymer-cemented soils concentrate around the injection pipe,forming a fan-shaped sheet-like structure.This study proves the improvement of polymer grouting for bedding slope treatment and will contribute to the development of a fast method to protect bedding slopes from landslides.
基金supported by the Scientific Research Project of Xiang Jiang Lab(22XJ02003)the University Fundamental Research Fund(23-ZZCX-JDZ-28)+5 种基金the National Science Fund for Outstanding Young Scholars(62122093)the National Natural Science Foundation of China(72071205)the Hunan Graduate Research Innovation Project(ZC23112101-10)the Hunan Natural Science Foundation Regional Joint Project(2023JJ50490)the Science and Technology Project for Young and Middle-aged Talents of Hunan(2023TJ-Z03)the Science and Technology Innovation Program of Humnan Province(2023RC1002)。
文摘Traditional large-scale multi-objective optimization algorithms(LSMOEAs)encounter difficulties when dealing with sparse large-scale multi-objective optimization problems(SLM-OPs)where most decision variables are zero.As a result,many algorithms use a two-layer encoding approach to optimize binary variable Mask and real variable Dec separately.Nevertheless,existing optimizers often focus on locating non-zero variable posi-tions to optimize the binary variables Mask.However,approxi-mating the sparse distribution of real Pareto optimal solutions does not necessarily mean that the objective function is optimized.In data mining,it is common to mine frequent itemsets appear-ing together in a dataset to reveal the correlation between data.Inspired by this,we propose a novel two-layer encoding learning swarm optimizer based on frequent itemsets(TELSO)to address these SLMOPs.TELSO mined the frequent terms of multiple particles with better target values to find mask combinations that can obtain better objective values for fast convergence.Experi-mental results on five real-world problems and eight benchmark sets demonstrate that TELSO outperforms existing state-of-the-art sparse large-scale multi-objective evolutionary algorithms(SLMOEAs)in terms of performance and convergence speed.
文摘Assessment of past-climate simulations of regional climate models(RCMs)is important for understanding the reliability of RCMs when used to project future regional climate.Here,we assess the performance and discuss possible causes of biases in a WRF-based RCM with a grid spacing of 50 km,named WRFG,from the North American Regional Climate Change Assessment Program(NARCCAP)in simulating wet season precipitation over the Central United States for a period when observational data are available.The RCM reproduces key features of the precipitation distribution characteristics during late spring to early summer,although it tends to underestimate the magnitude of precipitation.This dry bias is partially due to the model’s lack of skill in simulating nocturnal precipitation related to the lack of eastward propagating convective systems in the simulation.Inaccuracy in reproducing large-scale circulation and environmental conditions is another contributing factor.The too weak simulated pressure gradient between the Rocky Mountains and the Gulf of Mexico results in weaker southerly winds in between,leading to a reduction of warm moist air transport from the Gulf to the Central Great Plains.The simulated low-level horizontal convergence fields are less favorable for upward motion than in the NARR and hence,for the development of moist convection as well.Therefore,a careful examination of an RCM’s deficiencies and the identification of the source of errors are important when using the RCM to project precipitation changes in future climate scenarios.
基金support by the Open Project of Xiangjiang Laboratory(22XJ02003)the University Fundamental Research Fund(23-ZZCX-JDZ-28,ZK21-07)+5 种基金the National Science Fund for Outstanding Young Scholars(62122093)the National Natural Science Foundation of China(72071205)the Hunan Graduate Research Innovation Project(CX20230074)the Hunan Natural Science Foundation Regional Joint Project(2023JJ50490)the Science and Technology Project for Young and Middle-aged Talents of Hunan(2023TJZ03)the Science and Technology Innovation Program of Humnan Province(2023RC1002).
文摘Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to traverse vast expanse with limited computational resources.Furthermore,in the context of sparse,most variables in Pareto optimal solutions are zero,making it difficult for algorithms to identify non-zero variables efficiently.This paper is dedicated to addressing the challenges posed by SLMOPs.To start,we introduce innovative objective functions customized to mine maximum and minimum candidate sets.This substantial enhancement dramatically improves the efficacy of frequent pattern mining.In this way,selecting candidate sets is no longer based on the quantity of nonzero variables they contain but on a higher proportion of nonzero variables within specific dimensions.Additionally,we unveil a novel approach to association rule mining,which delves into the intricate relationships between non-zero variables.This novel methodology aids in identifying sparse distributions that can potentially expedite reductions in the objective function value.We extensively tested our algorithm across eight benchmark problems and four real-world SLMOPs.The results demonstrate that our approach achieves competitive solutions across various challenges.
文摘In situ inflow and outflow permeability tests with the BAT probe at SarapuíII soft clay test site are presented.A description of the BAT permeability test is provided,discussing its advantages and shortcomings,especially in the case of very soft clays under low stresses.Pore pressures were monitored during probe installation and were found to be slightly lower than piezocone u2 pore pressures,consistent with the position of the filter.The role of filter tip saturation was investigated after the usual saturation procedure provided an unsatisfactory pore pressure response during probe installation.Results show that the vacuum saturation procedure provides adequate response during installation and increases the reliability of the coefficient of permeability determination in early measurements.Both inflow and outflow tests yielded similar results,indicating that careful execution of the test can lead to good test repeatability regardless of the loading condition.Various sequences of alternated inflow and outflow tests have yielded similar results,indicating that soil reconsolidation and filter clogging were negligible in the tests performed.Data are presented concerning the relationship between index parameters and the in situ coefficient of permeability for SarapuíII clay,which plot outside the range of existing databases.
基金the Korea Health Technology R&D Project through the Korea Health Industry Development Institute(KHIDI),funded by the Ministry of Health&Welfare,Republic of Korea(grant no.HI22C0306).
文摘Objective:Serological tests are widely used for scrub typhus diagnosis;however,their limitations are evident.This study aims to assess their practical value in clinical settings.Methods:We analyzed the data of adult patients with suspected scrub typhus who visited a tertiary care hospital in the Republic of Korea from September to December from 2019 to 2021.The included patients had an acute fever and at least one of the following ten secondary findings:myalgia,skin rash,eschar,headache,thrombocytopenia,increased liver enzyme levels,lymphadenopathy,hepatomegaly,splenomegaly,and pleural effusion.The diagnoses were grouped as scrub typhus or other diseases by two infectious disease physicians.Results:Among 136 patients who met the eligibility criteria,109 had scrub typhus and 27 had different diseases.Single and paired total antibodies using immunofluorescence assay(IFA),and total antibodies using immunochromatography-based rapid diagnostic testing(ICT)were measured in 98%,22%,and 75%of all patients,respectively.Confirmation using paired samples for scrub typhus was established at a median of 11[interquartile range(IQR)10-16]days following the first visit.Among the 82 admitted patients,the median admission time was 9(IQR 7-13)days.According to IFA,58(55%)patients with scrub typhus had total immunoglobulin titers≥1:320,while 23(85%)patients with other disease had titers<1:320.Positive ICT results were observed in 64(74%)patients with scrub typhus and 10(67%)patients with other diseases showed negative ICT results.Conclusions:Serological testing for scrub typhus is currently insufficient for decision-making in clinical practice.
基金supported by the State Grid Science&Technology Project(5100-202114296A-0-0-00).
文摘This article introduces the concept of load aggregation,which involves a comprehensive analysis of loads to acquire their external characteristics for the purpose of modeling and analyzing power systems.The online identification method is a computer-involved approach for data collection,processing,and system identification,commonly used for adaptive control and prediction.This paper proposes a method for dynamically aggregating large-scale adjustable loads to support high proportions of new energy integration,aiming to study the aggregation characteristics of regional large-scale adjustable loads using online identification techniques and feature extraction methods.The experiment selected 300 central air conditioners as the research subject and analyzed their regulation characteristics,economic efficiency,and comfort.The experimental results show that as the adjustment time of the air conditioner increases from 5 minutes to 35 minutes,the stable adjustment quantity during the adjustment period decreases from 28.46 to 3.57,indicating that air conditioning loads can be controlled over a long period and have better adjustment effects in the short term.Overall,the experimental results of this paper demonstrate that analyzing the aggregation characteristics of regional large-scale adjustable loads using online identification techniques and feature extraction algorithms is effective.
文摘Accurate positioning is one of the essential requirements for numerous applications of remote sensing data,especially in the event of a noisy or unreliable satellite signal.Toward this end,we present a novel framework for aircraft geo-localization in a large range that only requires a downward-facing monocular camera,an altimeter,a compass,and an open-source Vector Map(VMAP).The algorithm combines the matching and particle filter methods.Shape vector and correlation between two building contour vectors are defined,and a coarse-to-fine building vector matching(CFBVM)method is proposed in the matching stage,for which the original matching results are described by the Gaussian mixture model(GMM).Subsequently,an improved resampling strategy is designed to reduce computing expenses with a huge number of initial particles,and a credibility indicator is designed to avoid location mistakes in the particle filter stage.An experimental evaluation of the approach based on flight data is provided.On a flight at a height of 0.2 km over a flight distance of 2 km,the aircraft is geo-localized in a reference map of 11,025 km~2using 0.09 km~2aerial images without any prior information.The absolute localization error is less than 10 m.
基金the National Natural Science Foundation of China(Grant No.61973033)Preliminary Research of Equipment(Grant No.9090102010305)for funding the experiments。
文摘The longitudinal dispersion of the projectile in shooting tests of two-dimensional trajectory corrections fused with fixed canards is extremely large that it sometimes exceeds the correction ability of the correction fuse actuator.The impact point easily deviates from the target,and thus the correction result cannot be readily evaluated.However,the cost of shooting tests is considerably high to conduct many tests for data collection.To address this issue,this study proposes an aiming method for shooting tests based on small sample size.The proposed method uses the Bootstrap method to expand the test data;repeatedly iterates and corrects the position of the simulated theoretical impact points through an improved compatibility test method;and dynamically adjusts the weight of the prior distribution of simulation results based on Kullback-Leibler divergence,which to some extent avoids the real data being"submerged"by the simulation data and achieves the fusion Bayesian estimation of the dispersion center.The experimental results show that when the simulation accuracy is sufficiently high,the proposed method yields a smaller mean-square deviation in estimating the dispersion center and higher shooting accuracy than those of the three comparison methods,which is more conducive to reflecting the effect of the control algorithm and facilitating test personnel to iterate their proposed structures and algorithms.;in addition,this study provides a knowledge base for further comprehensive studies in the future.
基金supported in part by the Central Government Guides Local Science and TechnologyDevelopment Funds(Grant No.YDZJSX2021A038)in part by theNational Natural Science Foundation of China under(Grant No.61806138)in part by the China University Industry-University-Research Collaborative Innovation Fund(Future Network Innovation Research and Application Project)(Grant 2021FNA04014).
文摘The large-scale multi-objective optimization algorithm(LSMOA),based on the grouping of decision variables,is an advanced method for handling high-dimensional decision variables.However,in practical problems,the interaction among decision variables is intricate,leading to large group sizes and suboptimal optimization effects;hence a large-scale multi-objective optimization algorithm based on weighted overlapping grouping of decision variables(MOEAWOD)is proposed in this paper.Initially,the decision variables are perturbed and categorized into convergence and diversity variables;subsequently,the convergence variables are subdivided into groups based on the interactions among different decision variables.If the size of a group surpasses the set threshold,that group undergoes a process of weighting and overlapping grouping.Specifically,the interaction strength is evaluated based on the interaction frequency and number of objectives among various decision variables.The decision variable with the highest interaction in the group is identified and disregarded,and the remaining variables are then reclassified into subgroups.Finally,the decision variable with the strongest interaction is added to each subgroup.MOEAWOD minimizes the interactivity between different groups and maximizes the interactivity of decision variables within groups,which contributed to the optimized direction of convergence and diversity exploration with different groups.MOEAWOD was subjected to testing on 18 benchmark large-scale optimization problems,and the experimental results demonstrate the effectiveness of our methods.Compared with the other algorithms,our method is still at an advantage.
基金The work was supported by Humanities and Social Sciences Fund of the Ministry of Education(No.22YJA630119)the National Natural Science Foundation of China(No.71971051)Natural Science Foundation of Hebei Province(No.G2021501004).
文摘With the development of big data and social computing,large-scale group decisionmaking(LGDM)is nowmerging with social networks.Using social network analysis(SNA),this study proposes an LGDM consensus model that considers the trust relationship among decisionmakers(DMs).In the process of consensusmeasurement:the social network is constructed according to the social relationship among DMs,and the Louvain method is introduced to classify social networks to form subgroups.In this study,the weights of each decision maker and each subgroup are computed by comprehensive network weights and trust weights.In the process of consensus improvement:A feedback mechanism with four identification and two direction rules is designed to guide the consensus of the improvement process.Based on the trust relationship among DMs,the preferences are modified,and the corresponding social network is updated to accelerate the consensus.Compared with the previous research,the proposedmodel not only allows the subgroups to be reconstructed and updated during the adjustment process,but also improves the accuracy of the adjustment by the feedbackmechanism.Finally,an example analysis is conducted to verify the effectiveness and flexibility of the proposed method.Moreover,compared with previous studies,the superiority of the proposed method in solving the LGDM problem is highlighted.
基金supported by the National Natural Science Foundation of China(52074046,52122403,51834003,and 52274073)the Graduate Research and Innovation Foundation of Chongqing(CYB22023)+2 种基金the Chongqing Talents Plan for Young Talents(cstc2022ycjh-bgzxm0035)Hunan Institute of Engineering(21RC025 and XJ2005)Hunan Province Education Department(21B0664).
文摘Underground salt cavern CO_(2) storage(SCCS)offers the dual benefits of enabling extensive CO_(2) storage and facilitating the utilization of CO_(2) resources while contributing the regulation of the carbon market.Its economic and operational advantages over traditional carbon capture,utilization,and storage(CCUS)projects make SCCS a more cost-effective and flexible option.Despite the widespread use of salt caverns for storing various substances,differences exist between SCCS and traditional salt cavern energy storage in terms of gas-tightness,carbon injection,brine extraction control,long-term carbon storage stability,and site selection criteria.These distinctions stem from the unique phase change characteristics of CO_(2) and the application scenarios of SCCS.Therefore,targeted and forward-looking scientific research on SCCS is imperative.This paper introduces the implementation principles and application scenarios of SCCS,emphasizing its connections with carbon emissions,carbon utilization,and renewable energy peak shaving.It delves into the operational characteristics and economic advantages of SCCS compared with other CCUS methods,and addresses associated scientific challenges.In this paper,we establish a pressure equation for carbon injection and brine extraction,that considers the phase change characteristics of CO_(2),and we analyze the pressure during carbon injection.By comparing the viscosities of CO_(2) and other gases,SCCS’s excellent sealing performance is demonstrated.Building on this,we develop a long-term stability evaluation model and associated indices,which analyze the impact of the injection speed and minimum operating pressure on stability.Field countermeasures to ensure stability are proposed.Site selection criteria for SCCS are established,preliminary salt mine sites suitable for SCCS are identified in China,and an initial estimate of achievable carbon storage scale in China is made at over 51.8-77.7 million tons,utilizing only 20%-30%volume of abandoned salt caverns.This paper addresses key scientific and engineering challenges facing SCCS and determines crucial technical parameters,such as the operating pressure,burial depth,and storage scale,and it offers essential guidance for implementing SCCS projects in China.
基金Supported partly by NSF of China(Grant No.11801163)NSF of Hunan Province(Grant Nos.2021JJ50032,2023JJ50164 and 2023JJ50165)Degree&Postgraduate Reform Project of Hunan University of Technology and Hunan Province(Grant Nos.JGYB23009 and 2024JGYB210).
文摘We introduce a factorized Smith method(FSM)for solving large-scale highranked T-Stein equations within the banded-plus-low-rank structure framework.To effectively reduce both computational complexity and storage requirements,we develop techniques including deflation and shift,partial truncation and compression,as well as redesign the residual computation and termination condition.Numerical examples demonstrate that the FSM outperforms the Smith method implemented with a hierarchical HODLR structured toolkit in terms of CPU time.
文摘Background A task assigned to space exploration satellites involves detecting the physical environment within a certain space.However,space detection data are complex and abstract.These data are not conducive for researchers'visual perceptions of the evolution and interaction of events in the space environment.Methods A time-series dynamic data sampling method for large-scale space was proposed for sample detection data in space and time,and the corresponding relationships between data location features and other attribute features were established.A tone-mapping method based on statistical histogram equalization was proposed and applied to the final attribute feature data.The visualization process is optimized for rendering by merging materials,reducing the number of patches,and performing other operations.Results The results of sampling,feature extraction,and uniform visualization of the detection data of complex types,long duration spans,and uneven spatial distributions were obtained.The real-time visualization of large-scale spatial structures using augmented reality devices,particularly low-performance devices,was also investigated.Conclusions The proposed visualization system can reconstruct the three-dimensional structure of a large-scale space,express the structure and changes in the spatial environment using augmented reality,and assist in intuitively discovering spatial environmental events and evolutionary rules.
基金This research has been conducted under SEGTRANS project,funded by the Centre for Industrial Technological Development(CDTI,Government of Spain).
文摘This paper explores the performances of a finite element simulation including four concrete models applied to a full-scale reinforced concrete beam subjected to blast loading. Field test data has been used to compare model results for each case. The numerical modelling has been, carried out using the suitable code LS-DYNA. This code integrates blast load routine(CONWEP) for the explosive description and four different material models for the concrete including: Karagozian & Case Concrete, Winfrith, Continuous Surface Cap Model and Riedel-Hiermaier-Thoma models, with concrete meshing based on 10, 15, and 20 mm. Six full-scale beams were tested: four of them used for the initial calibration of the numerical model and two more tests at lower scaled distances. For calibration, field data obtained employing pressure and accelerometers transducers were compared with the results derived from the numerical simulation. Damage surfaces and the shape of rupture in the beams have been used as references for comparison. Influence of the meshing on accelerations has been put in evidence and for some models the shape and size of the damage in the beams produced maximum differences around 15%. In all cases, the variations between material and mesh models are shown and discussed.
文摘BACKGROUND A reliable test is essential for diagnosing Helicobacter pylori(H.pylori)infection,and crucial for managing H.pylori-related diseases.Serving as an excellent method for detecting H.pylori infection,histologic examination is a test that clinicians heavily rely on,especially when complemented with immunohistochemistry(IHC).Additionally,other diagnostic tests for H.pylori,such as the rapid urease test(CLO test)and stool antigen test(SA),are also highly sensitive and specific.Typically,the results of histology and other tests align with each other.However,on rare occasions,discrepancy between histopathology and other H.pylori diagnostic tests occurs.AIM To investigate the discordance between histology and other H.pylori tests,the underlying causes,and the impact on clinical management.METHODS Pathology reports of gastric biopsies were retrieved spanning August 2013 and July 2018.Reports were included in the study only if there were other H.pylori tests within seven days of the biopsy.These additional tests include CLO test,SA,and H.pylori culture.Concordance between histopathology and other tests was determined based on the consistency of results.In instances where histology results were negative while other tests were positive,the slides were retrieved for re-assessment,and the clinical chart was reviewed.RESULTS Of 1396 pathology reports were identified,each accompanied by one additional H.pylori test.The concordance rates in detecting H.pylori infection between biopsy and other tests did not exhibit significant differences based on the number of biopsy fragments.117 discrepant cases were identified.Only 20 cases(9 with CLO test and 11 with SA)had negative biopsy but positive results in other tests.Four cases initially stained with Warthin-Starry turned out to be positive for H.pylori with subsequent IHC staining.Among the remaining 16 true discrepant cases,10 patients were on proton pump inhibitors before the biopsy and/or other tests.Most patients underwent treatment,except for two who were untreated,and two patients who were lost to follow-up.CONCLUSION There are rare discrepant cases with negative biopsy but positive in SA or CLO test.Various factors may contribute to this inconsistency.Most patients in such cases had undergone treatment.
基金National Key R&D Program of China(Grant No.2020YFC1512404).
文摘To investigate the influence of structural parameters on the performances and internal flow characteristics of partial flow pumps at a low specific speed of 10000 rpm,special attention was paid to the first and second stage impeller guide vanes.Moreover,the impeller blade outlet width,impeller inlet diameter,blade inclination angle,and number of blades were considered for orthogonal tests.Accordingly,nine groups of design solutions were formed,and then used as a basis for the execution of numerical simulations(CFD)aimed at obtaining the efficiency values and heads for each design solution group.The influence of impeller geometric parameters on the efficiency and head was explored,and the“weight”of each factor was obtained via a range analysis.Optimal structural parameters were finally chosen on the basis of the numerical simulation results,and the performances of the optimized model were verified accordingly(yet by means of CFD).Evidence is provided that the increase in the efficiency and head of the optimized model was 12.11%and 23.5 m,respectively,compared with those of the original model.
基金National Natural Science Foundation of China(82274265 and 82274588)Hunan University of Traditional Chinese Medicine Research Unveiled Marshal Programs(2022XJJB003).
文摘Eye diagnosis is a method for inspecting systemic diseases and syndromes by observing the eyes.With the development of intelligent diagnosis in traditional Chinese medicine(TCM);artificial intelligence(AI)can improve the accuracy and efficiency of eye diagnosis.However;the research on intelligent eye diagnosis still faces many challenges;including the lack of standardized and precisely labeled data;multi-modal information analysis;and artificial in-telligence models for syndrome differentiation.The widespread application of AI models in medicine provides new insights and opportunities for the research of eye diagnosis intelli-gence.This study elaborates on the three key technologies of AI models in the intelligent ap-plication of TCM eye diagnosis;and explores the implications for the research of eye diagno-sis intelligence.First;a database concerning eye diagnosis was established based on self-su-pervised learning so as to solve the issues related to the lack of standardized and precisely la-beled data.Next;the cross-modal understanding and generation of deep neural network models to address the problem of lacking multi-modal information analysis.Last;the build-ing of data-driven models for eye diagnosis to tackle the issue of the absence of syndrome dif-ferentiation models.In summary;research on intelligent eye diagnosis has great potential to be applied the surge of AI model applications.
基金supported by the National Natural Science Foundation of China(Grant Nos.41825018,41977248,42207219)the Second Tibetan Plateau Scientific Expedition and Research Program(STEP)(Grant No.2019QZKK0904)。
文摘The Sichuan-Tibet transportation corridor is prone to numerous active faults and frequent strong earthquakes.While extensive studies have individually explored the effect of active faults and strong earthquakes on different engineering structures,their combined effect remains unclear.This research employed multiple physical model tests to investigate the dynamic response of various engineering structures,including tunnels,bridges,and embankments,under the simultaneous influence of cumulative earthquakes and stick-slip misalignment of an active fault.The prototype selected for this study was the Kanding No.2 tunnel,which crosses the Yunongxi fault zone within the Sichuan-Tibet transportation corridor.The results demonstrated that the tunnel,bridge,and embankment exhibited amplification in response to the input seismic wave,with the amplification effect gradually decreasing as the input peak ground acceleration(PGA)increased.The PGAs of different engineering structures were weakened by the fault rupture zone.Nevertheless,the misalignment of the active fault may decrease the overall stiffness of the engineering structure,leading to more severe damage,with a small contribution from seismic vibration.Additionally,the seismic vibration effect might be enlarged with the height of the engineering structure,and the tunnel is supposed to have a smaller PGA and lower dynamic earth pressure compared to bridges and embankments in strong earthquake zones crossing active faults.The findings contribute valuable insights for evaluating the dynamic response of various engineering structures crossing an active fault and provide an experimental reference for secure engineering design in the challenging conditions of the Sichuan-Tibet transportation corridor.