This paper investigates the problem of collecting multidimensional data throughout time(i.e.,longitudinal studies)for the fundamental task of frequency estimation under Local Differential Privacy(LDP)guarantees.Contra...This paper investigates the problem of collecting multidimensional data throughout time(i.e.,longitudinal studies)for the fundamental task of frequency estimation under Local Differential Privacy(LDP)guarantees.Contrary to frequency estimation of a single attribute,the multidimensional aspect demands particular attention to the privacy budget.Besides,when collecting user statistics longitudinally,privacy progressively degrades.Indeed,the“multiple”settings in combination(i.e.,many attributes and several collections throughout time)impose several challenges,for which this paper proposes the first solution for frequency estimates under LDP.To tackle these issues,we extend the analysis of three state-of-the-art LDP protocols(Generalized Randomized Response–GRR,Optimized Unary Encoding–OUE,and Symmetric Unary Encoding–SUE)for both longitudinal and multidimensional data collections.While the known literature uses OUE and SUE for two rounds of sanitization(a.k.a.memoization),i.e.,L-OUE and L-SUE,respectively,we analytically and experimentally show that starting with OUE and then with SUE provides higher data utility(i.e.,L-OSUE).Also,for attributes with small domain sizes,we propose Longitudinal GRR(L-GRR),which provides higher utility than the other protocols based on unary encoding.Last,we also propose a new solution named Adaptive LDP for LOngitudinal and Multidimensional FREquency Estimates(ALLOMFREE),which randomly samples a single attribute to be sent with the whole privacy budget and adaptively selects the optimal protocol,i.e.,either L-GRR or L-OSUE.As shown in the results,ALLOMFREE consistently and considerably outperforms the state-of-the-art L-SUE and L-OUE protocols in the quality of the frequency estimates.展开更多
Massive amounts of data are acquired in modern and future information technology industries such as communication,radar,and remote sensing.The presence of large dimensionality and size in these data offers new opportu...Massive amounts of data are acquired in modern and future information technology industries such as communication,radar,and remote sensing.The presence of large dimensionality and size in these data offers new opportunities to enhance the performance of signal processing in such applications and even motivate new ones.However,the curse of dimensionality is always a challenge when processing such high-dimensional signals.In practical tasks,high-dimensional signals need to be acquired,processed,and analyzed with high accuracy,robustness,and computational efficiency.This special section aims to address these challenges,where articles attempt to develop new theories and methods that are best suited to the high dimensional nature of the signals involved,and explore modern and emerging applications in this area.展开更多
In the era of exponential growth of data availability,the architecture of systems has a trend toward high dimensionality,and directly exploiting holistic information for state inference is not always computationally a...In the era of exponential growth of data availability,the architecture of systems has a trend toward high dimensionality,and directly exploiting holistic information for state inference is not always computationally affordable.This paper proposes a novel Bayesian filtering algorithm that considers algorithmic computational cost and estimation accuracy for high-dimensional linear systems.The high-dimensional state vector is divided into several blocks to save computation resources by avoiding the calculation of error covariance with immense dimensions.After that,two sequential states are estimated simultaneously by introducing an auxiliary variable in the new probability space,mitigating the performance degradation caused by state segmentation.Moreover,the computational cost and error covariance of the proposed algorithm are analyzed analytically to show its distinct features compared with several existing methods.Simulation results illustrate that the proposed Bayesian filtering can maintain a higher estimation accuracy with reasonable computational cost when applied to high-dimensional linear systems.展开更多
This study evaluated the genetic and agronomic parameter estimates of maize under different nitrogen rates. The trial was established at the Njala Agricultural Research Centre experimental site during 2021 and 2022 in...This study evaluated the genetic and agronomic parameter estimates of maize under different nitrogen rates. The trial was established at the Njala Agricultural Research Centre experimental site during 2021 and 2022 in a split block design with three maize varieties (IWCD2, 2009EVDT, and DMR-ESR-Yellow) and seven nitrogen (0, 30, 60, 90, 120, 150 and 180 kg∙N∙ha<sup>−</sup><sup>1</sup>) rates. Findings showed that cob diameter and anthesis silking time (ASI) had intermediate heritability, ASI had high genetic advance, ASI and grain yield had high genotypic coefficient of variation (GCV), while traits with high phenotypic coefficient of variation (PCV) were plant height, ASI, grain yield, number of kernel per cob, number of kernel rows, ear length, and ear height. The PCV values were higher than GCV, indicating the influence of the environment in the studied traits. Nitrogen rates and variety significantly (p < 0.05) influenced grain yield production. Mean grain yields and economic parameter estimates increased with increasing nitrogen rates, with the 30 and 180 kg∙N∙ha<sup>−</sup><sup>1</sup> plots exhibiting the lowest and highest grain yields of 1238 kg∙ha<sup>−</sup><sup>1</sup> and 2098 kg∙ha<sup>−</sup><sup>1</sup>, respectively. Variety and nitrogen effects on partial factor productivity (PFP<sub>N</sub>), agronomic efficiency (AEN), net returns (NR), value cost ratio (VCR) and marginal return (MR) indicated that these parameters were significantly affected (p < 0.05) by these factors. The highest PFP<sub>N</sub> (41.3 kg grain kg<sup>−</sup><sup>1</sup>∙N) and AEN (29.4 kg grain kg<sup>−</sup><sup>1</sup>∙N) were obtained in the 30 kg∙N∙ha<sup>−</sup><sup>1</sup> plots, while the highest VCR (2.8) and MR (SLL 1.8 SLL<sup>−</sup><sup>1</sup> spent on N) were obtained in the 180 kg∙N∙ha<sup>−</sup><sup>1</sup>. The significant influence of variety and nitrogen on traits suggests that increasing yields and maximizing profits require use of appropriate nitrogen fertilization and improved farming practices that could be exploited for increased productivity of maize.展开更多
Ni-Fe-based oxides are among the most promising catalysts developed to date for the bottleneck oxygen evolution reaction(OER)in water electrolysis.However,understanding and mastering the synergy of Ni and Fe remain ch...Ni-Fe-based oxides are among the most promising catalysts developed to date for the bottleneck oxygen evolution reaction(OER)in water electrolysis.However,understanding and mastering the synergy of Ni and Fe remain challenging.Herein,we report that the synergy between Ni and Fe can be tailored by crystal dimensionality of Ni,Fe-contained Ruddlesden-Popper(RP)-type perovskites(La_(0.125)Sr_(0.875))n+1(Ni_(0.25)Fe_(0.75))nO3n+1(n=1,2,3),where the material with n=3 shows the best OER performance in alkaline media.Soft X-ray absorption spectroscopy spectra before and after OER reveal that the material with n=3 shows enhanced Ni/Fe-O covalency to boost the electron transfer as compared to those with n=1 and n=2.Further experimental investigations demonstrate that the Fe ion is the active site and the Ni ion is the stable site in this system,where such unique synergy reaches the optimum at n=3.Besides,as n increases,the proportion of unstable rock-salt layers accordingly decreases and the leaching of ions(especially Sr^(2+))into the electrolyte is suppressed,which induces a decrease in the leaching of active Fe ions,ultimately leading to enhanced stability.This work provides a new avenue for rational catalyst design through the dimensional strategy.展开更多
The estimation of covariance matrices is very important in many fields, such as statistics. In real applications, data are frequently influenced by high dimensions and noise. However, most relevant studies are based o...The estimation of covariance matrices is very important in many fields, such as statistics. In real applications, data are frequently influenced by high dimensions and noise. However, most relevant studies are based on complete data. This paper studies the optimal estimation of high-dimensional covariance matrices based on missing and noisy sample under the norm. First, the model with sub-Gaussian additive noise is presented. The generalized sample covariance is then modified to define a hard thresholding estimator , and the minimax upper bound is derived. After that, the minimax lower bound is derived, and it is concluded that the estimator presented in this article is rate-optimal. Finally, numerical simulation analysis is performed. The result shows that for missing samples with sub-Gaussian noise, if the true covariance matrix is sparse, the hard thresholding estimator outperforms the traditional estimate method.展开更多
Quantum Fisher information(QFI)associated with local metrology has been used to parameter estimation in open quantum systems.In this work,we calculated the QFI for a moving Unruh-DeWitt detector coupled with massless ...Quantum Fisher information(QFI)associated with local metrology has been used to parameter estimation in open quantum systems.In this work,we calculated the QFI for a moving Unruh-DeWitt detector coupled with massless scalar fields in n-dimensional spacetime,and analyzed the behavior of QFI with various parameters,such as the dimension of spacetime,evolution time,and Unruh temperature.We discovered that the QFI of state parameter decreases monotonically from 1 to 0 over time.Additionally,we noted that the QFI for small evolution times is several orders of magnitude higher than the QFI for long evolution times.We also found that the value of QFI decreases at first and then stabilizes as the Unruh temperature increases.It was observed that the QFI depends on initial state parameterθ,and Fθis the maximum forθ=0 orθ=π,Fφis the maximum forθ=π/2.We also obtain that the maximum value of QFI for state parameters varies for different spacetime dimensions with the same evolution time.展开更多
Integrable systems play a crucial role in physics and mathematics.In particular,the traditional(1+1)-dimensional and(2+1)-dimensional integrable systems have received significant attention due to the rarity of integra...Integrable systems play a crucial role in physics and mathematics.In particular,the traditional(1+1)-dimensional and(2+1)-dimensional integrable systems have received significant attention due to the rarity of integrable systems in higher dimensions.Recent studies have shown that abundant higher-dimensional integrable systems can be constructed from(1+1)-dimensional integrable systems by using a deformation algorithm.Here we establish a new(2+1)-dimensional Chen-Lee-Liu(C-L-L)equation using the deformation algorithm from the(1+1)-dimensional C-L-L equation.The new system is integrable with its Lax pair obtained by applying the deformation algorithm to that of the(1+1)-dimension.It is challenging to obtain the exact solutions for the new integrable system because the new system combines both the original C-L-L equation and its reciprocal transformation.The traveling wave solutions are derived in implicit function expression,and some asymmetry peakon solutions are found.展开更多
Onlineγ-spectrometry systems for inland waters,most of which extract samples in situ and in real time,are able to produce reliable activity concentration measurements for waterborne radionuclides only when they are d...Onlineγ-spectrometry systems for inland waters,most of which extract samples in situ and in real time,are able to produce reliable activity concentration measurements for waterborne radionuclides only when they are distributed relatively uniformly and enter into a steady-state diffusion regime in the measurement chamber.To protect residents’health and ensure the safety of the living environment,better timeliness is required for this measurement method.To address this issue,this study established a mathematical model of the online waterγ-spectrometry system so that rapid warning and activity estimates can be obtained for water under non-steady-state(NSS)conditions.In addition,the detection efficiency of the detector for radionuclides during the NSS diffusion process was determined by applying the computational fluid dynamics technique in conjunction with Monte Carlo simulations.On this basis,a method was developed that allowed the online waterγ-spectrometry system to provide rapid warning and activity concentration estimates for radionuclides in water.Subsequent analysis of the NSS-mode measurements of^(40)K radioactive solutions with different activity concentrations determined the optimum warning threshold and measurement time for producing accurate activity concentration estimates for radionuclides.The experimental results show that the proposed NSS measurement method is able to give warning and yield accurate activity concentration estimates for radionuclides 55.42 and 69.42 min after the entry of a 10 Bq/L^(40)K radioactive solution into the measurement chamber,respectively.These times are much shorter than the 90 min required by the conventional measurement method.Furthermore,the NSS measurement method allows the measurement system to give rapid(within approximately 15 min)warning when the activity concentrations of some radionuclides reach their respective limits stipulated in the Guidelines for Drinking-water Quality of the WHO,suggesting that this method considerably enhances the warning capacity of in situ online waterγ-spectrometry systems.展开更多
This paper aims to explore the application of Extreme Value Theory (EVT) in estimating the conditional extreme quantile for time-to-event outcomes by examining the functional relationship between ambulatory blood pres...This paper aims to explore the application of Extreme Value Theory (EVT) in estimating the conditional extreme quantile for time-to-event outcomes by examining the functional relationship between ambulatory blood pressure trajectories and clinical outcomes in stroke patients. The study utilizes EVT to analyze the functional connection between ambulatory blood pressure trajectories and clinical outcomes in a sample of 297 stroke patients. The 24-hour ambulatory blood pressure measurement curves for every 15 minutes are considered, acknowledging a censored rate of 40%. The findings reveal that the sample mean excess function exhibits a positive gradient above a specific threshold, confirming the heavy-tailed distribution of data in stroke patients with a positive extreme value index. Consequently, the estimated conditional extreme quantile indicates that stroke patients with higher blood pressure measurements face an elevated risk of recurrent stroke occurrence at an early stage. This research contributes to the understanding of the relationship between ambulatory blood pressure and recurrent stroke, providing valuable insights for clinical considerations and potential interventions in stroke management.展开更多
Fractal theory offers a powerful tool for the precise description and quantification of the complex pore structures in reservoir rocks,crucial for understanding the storage and migration characteristics of media withi...Fractal theory offers a powerful tool for the precise description and quantification of the complex pore structures in reservoir rocks,crucial for understanding the storage and migration characteristics of media within these rocks.Faced with the challenge of calculating the three-dimensional fractal dimensions of rock porosity,this study proposes an innovative computational process that directly calculates the three-dimensional fractal dimensions from a geometric perspective.By employing a composite denoising approach that integrates Fourier transform(FT)and wavelet transform(WT),coupled with multimodal pore extraction techniques such as threshold segmentation,top-hat transformation,and membrane enhancement,we successfully crafted accurate digital rock models.The improved box-counting method was then applied to analyze the voxel data of these digital rocks,accurately calculating the fractal dimensions of the rock pore distribution.Further numerical simulations of permeability experiments were conducted to explore the physical correlations between the rock pore fractal dimensions,porosity,and absolute permeability.The results reveal that rocks with higher fractal dimensions exhibit more complex pore connectivity pathways and a wider,more uneven pore distribution,suggesting that the ideal rock samples should possess lower fractal dimensions and higher effective porosity rates to achieve optimal fluid transmission properties.The methodology and conclusions of this study provide new tools and insights for the quantitative analysis of complex pores in rocks and contribute to the exploration of the fractal transport properties of media within rocks.展开更多
The human pose paradigm is estimated using a transformer-based multi-branch multidimensional directed the three-dimensional(3D)method that takes into account self-occlusion,badly posedness,and a lack of depth data in ...The human pose paradigm is estimated using a transformer-based multi-branch multidimensional directed the three-dimensional(3D)method that takes into account self-occlusion,badly posedness,and a lack of depth data in the per-frame 3D posture estimation from two-dimensional(2D)mapping to 3D mapping.Firstly,by examining the relationship between the movements of different bones in the human body,four virtual skeletons are proposed to enhance the cyclic constraints of limb joints.Then,multiple parameters describing the skeleton are fused and projected into a high-dimensional space.Utilizing a multi-branch network,motion features between bones and overall motion features are extracted to mitigate the drift error in the estimation results.Furthermore,the estimated relative depth is projected into 3D space,and the error is calculated against real 3D data,forming a loss function along with the relative depth error.This article adopts the average joint pixel error as the primary performance metric.Compared to the benchmark approach,the estimation findings indicate an increase in average precision of 1.8 mm within the Human3.6M sample.展开更多
The objective of reliability-based design optimization(RBDO)is to minimize the optimization objective while satisfying the corresponding reliability requirements.However,the nested loop characteristic reduces the effi...The objective of reliability-based design optimization(RBDO)is to minimize the optimization objective while satisfying the corresponding reliability requirements.However,the nested loop characteristic reduces the efficiency of RBDO algorithm,which hinders their application to high-dimensional engineering problems.To address these issues,this paper proposes an efficient decoupled RBDO method combining high dimensional model representation(HDMR)and the weight-point estimation method(WPEM).First,we decouple the RBDO model using HDMR and WPEM.Second,Lagrange interpolation is used to approximate a univariate function.Finally,based on the results of the first two steps,the original nested loop reliability optimization model is completely transformed into a deterministic design optimization model that can be solved by a series of mature constrained optimization methods without any additional calculations.Two numerical examples of a planar 10-bar structure and an aviation hydraulic piping system with 28 design variables are analyzed to illustrate the performance and practicability of the proposed method.展开更多
BACKGROUND The prevalence of non-alcoholic fatty liver(NAFLD)has increased recently.Subjects with NAFLD are known to have higher chance for renal function impairment.Many past studies used traditional multiple linear ...BACKGROUND The prevalence of non-alcoholic fatty liver(NAFLD)has increased recently.Subjects with NAFLD are known to have higher chance for renal function impairment.Many past studies used traditional multiple linear regression(MLR)to identify risk factors for decreased estimated glomerular filtration rate(eGFR).However,medical research is increasingly relying on emerging machine learning(Mach-L)methods.The present study enrolled healthy women to identify factors affecting eGFR in subjects with and without NAFLD(NAFLD+,NAFLD-)and to rank their importance.AIM To uses three different Mach-L methods to identify key impact factors for eGFR in healthy women with and without NAFLD.METHODS A total of 65535 healthy female study participants were enrolled from the Taiwan MJ cohort,accounting for 32 independent variables including demographic,biochemistry and lifestyle parameters(independent variables),while eGFR was used as the dependent variable.Aside from MLR,three Mach-L methods were applied,including stochastic gradient boosting,eXtreme gradient boosting and elastic net.Errors of estimation were used to define method accuracy,where smaller degree of error indicated better model performance.RESULTS Income,albumin,eGFR,High density lipoprotein-Cholesterol,phosphorus,forced expiratory volume in one second(FEV1),and sleep time were all lower in the NAFLD+group,while other factors were all significantly higher except for smoking area.Mach-L had lower estimation errors,thus outperforming MLR.In Model 1,age,uric acid(UA),FEV1,plasma calcium level(Ca),plasma albumin level(Alb)and T-bilirubin were the most important factors in the NAFLD+group,as opposed to age,UA,FEV1,Alb,lactic dehydrogenase(LDH)and Ca for the NAFLD-group.Given the importance percentage was much higher than the 2nd important factor,we built Model 2 by removing age.CONCLUSION The eGFR were lower in the NAFLD+group compared to the NAFLD-group,with age being was the most important impact factor in both groups of healthy Chinese women,followed by LDH,UA,FEV1 and Alb.However,for the NAFLD-group,TSH and SBP were the 5th and 6th most important factors,as opposed to Ca and BF in the NAFLD+group.展开更多
Assume that L is a non-negative self-adjoint operator on L^(2)(ℝ^(n))with its heat kernels satisfying the so-called Gaussian upper bound estimate and that X is a ball quasi-Banach function space onℝ^(n) satisfying som...Assume that L is a non-negative self-adjoint operator on L^(2)(ℝ^(n))with its heat kernels satisfying the so-called Gaussian upper bound estimate and that X is a ball quasi-Banach function space onℝ^(n) satisfying some mild assumptions.Let HX,L(ℝ^(n))be the Hardy space associated with both X and L,which is defined by the Lusin area function related to the semigroup generated by L.In this article,the authors establish various maximal function characterizations of the Hardy space HX,L(ℝ^(n))and then apply these characterizations to obtain the solvability of the related Cauchy problem.These results have a wide range of generality and,in particular,the specific spaces X to which these results can be applied include the weighted space,the variable space,the mixed-norm space,the Orlicz space,the Orlicz-slice space,and the Morrey space.Moreover,the obtained maximal function characterizations of the mixed-norm Hardy space,the Orlicz-slice Hardy space,and the Morrey-Hardy space associated with L are completely new.展开更多
The longitudinal dispersion of the projectile in shooting tests of two-dimensional trajectory corrections fused with fixed canards is extremely large that it sometimes exceeds the correction ability of the correction ...The longitudinal dispersion of the projectile in shooting tests of two-dimensional trajectory corrections fused with fixed canards is extremely large that it sometimes exceeds the correction ability of the correction fuse actuator.The impact point easily deviates from the target,and thus the correction result cannot be readily evaluated.However,the cost of shooting tests is considerably high to conduct many tests for data collection.To address this issue,this study proposes an aiming method for shooting tests based on small sample size.The proposed method uses the Bootstrap method to expand the test data;repeatedly iterates and corrects the position of the simulated theoretical impact points through an improved compatibility test method;and dynamically adjusts the weight of the prior distribution of simulation results based on Kullback-Leibler divergence,which to some extent avoids the real data being"submerged"by the simulation data and achieves the fusion Bayesian estimation of the dispersion center.The experimental results show that when the simulation accuracy is sufficiently high,the proposed method yields a smaller mean-square deviation in estimating the dispersion center and higher shooting accuracy than those of the three comparison methods,which is more conducive to reflecting the effect of the control algorithm and facilitating test personnel to iterate their proposed structures and algorithms.;in addition,this study provides a knowledge base for further comprehensive studies in the future.展开更多
With the extensive application of large-scale array antennas,the increasing number of array elements leads to the increasing dimension of received signals,making it difficult to meet the real-time requirement of direc...With the extensive application of large-scale array antennas,the increasing number of array elements leads to the increasing dimension of received signals,making it difficult to meet the real-time requirement of direction of arrival(DOA)estimation due to the computational complexity of algorithms.Traditional subspace algorithms require estimation of the covariance matrix,which has high computational complexity and is prone to producing spurious peaks.In order to reduce the computational complexity of DOA estimation algorithms and improve their estimation accuracy under large array elements,this paper proposes a DOA estimation method based on Krylov subspace and weighted l_(1)-norm.The method uses the multistage Wiener filter(MSWF)iteration to solve the basis of the Krylov subspace as an estimate of the signal subspace,further uses the measurement matrix to reduce the dimensionality of the signal subspace observation,constructs a weighted matrix,and combines the sparse reconstruction to establish a convex optimization function based on the residual sum of squares and weighted l_(1)-norm to solve the target DOA.Simulation results show that the proposed method has high resolution under large array conditions,effectively suppresses spurious peaks,reduces computational complexity,and has good robustness for low signal to noise ratio(SNR)environment.展开更多
Xinjiang Uygur Autonomous Region is a typical inland arid area in China with a sparse and uneven distribution of meteorological stations,limited access to precipitation data,and significant water scarcity.Evaluating a...Xinjiang Uygur Autonomous Region is a typical inland arid area in China with a sparse and uneven distribution of meteorological stations,limited access to precipitation data,and significant water scarcity.Evaluating and integrating precipitation datasets from different sources to accurately characterize precipitation patterns has become a challenge to provide more accurate and alternative precipitation information for the region,which can even improve the performance of hydrological modelling.This study evaluated the applicability of widely used five satellite-based precipitation products(Climate Hazards Group InfraRed Precipitation with Station(CHIRPS),China Meteorological Forcing Dataset(CMFD),Climate Prediction Center morphing method(CMORPH),Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks-Climate Data Record(PERSIANN-CDR),and Tropical Rainfall Measuring Mission Multi-satellite Precipitation Analysis(TMPA))and a reanalysis precipitation dataset(ECMWF Reanalysis v5-Land Dataset(ERA5-Land))in Xinjiang using ground-based observational precipitation data from a limited number of meteorological stations.Based on this assessment,we proposed a framework that integrated different precipitation datasets with varying spatial resolutions using a dynamic Bayesian model averaging(DBMA)approach,the expectation-maximization method,and the ordinary Kriging interpolation method.The daily precipitation data merged using the DBMA approach exhibited distinct spatiotemporal variability,with an outstanding performance,as indicated by low root mean square error(RMSE=1.40 mm/d)and high Person's correlation coefficient(CC=0.67).Compared with the traditional simple model averaging(SMA)and individual product data,although the DBMA-fused precipitation data were slightly lower than the best precipitation product(CMFD),the overall performance of DBMA was more robust.The error analysis between DBMA-fused precipitation dataset and the more advanced Integrated Multi-satellite Retrievals for Global Precipitation Measurement Final(IMERG-F)precipitation product,as well as hydrological simulations in the Ebinur Lake Basin,further demonstrated the superior performance of DBMA-fused precipitation dataset in the entire Xinjiang region.The proposed framework for solving the fusion problem of multi-source precipitation data with different spatial resolutions is feasible for application in inland arid areas,and aids in obtaining more accurate regional hydrological information and improving regional water resources management capabilities and meteorological research in these regions.展开更多
NGLY1 Deficiency is an ultra-rare autosomal recessively inherited disorder. Characteristic symptoms include among others, developmental delays, movement disorders, liver function abnormalities, seizures, and problems ...NGLY1 Deficiency is an ultra-rare autosomal recessively inherited disorder. Characteristic symptoms include among others, developmental delays, movement disorders, liver function abnormalities, seizures, and problems with tear formation. Movements are hyperkinetic and may include dysmetric, choreo-athetoid, myoclonic and dystonic movement elements. To date, there have been no quantitative reports describing arm movements of individuals with NGLY1 Deficiency. This report provides quantitative information about a series of arm movements performed by an individual with NGLY1 Deficiency and an aged-matched neurotypical participant. Three categories of arm movements were tested: 1) open ended reaches without specific end point targets;2) goal-directed reaches that included grasping an object;3) picking up small objects from a table placed in front of the participants. Arm movement kinematics were obtained with a camera-based motion analysis system and “initiation” and “maintenance” phases were identified for each movement. The combination of the two phases was labeled as a “complete” movement. Three-dimensional analysis techniques were used to quantify the movements and included hand trajectory pathlength, joint motion area, as well as hand trajectory and joint jerk cost. These techniques were required to fully characterize the movements because the NGLY1 individual was unable to perform movements only in the primary plane of progression instead producing motion across all three planes of movement. The individual with NGLY1 Deficiency was unable to pick up objects from a table or effectively complete movements requiring crossing the midline. The successfully completed movements were analyzed using the above techniques and the results of the two participants were compared statistically. Almost all comparisons revealed significant differences between the two participants, with a notable exception of the 3D initiation area as a percentage of the complete movement. The statistical tests of these measures revealed no significant differences between the two participants, possibly suggesting a common underlying motor control strategy. The 3D techniques used in this report effectively characterized arm movements of an individual with NGLY1 deficiency and can be used to provide information to evaluate the effectiveness of genetic, pharmacological, or physical rehabilitation therapies.展开更多
Machining is as old as humanity, and changes in temperature in both the machine’s internal and external environments can be of great concern as they affect the machine’s thermal stability and, thus, the machine’s d...Machining is as old as humanity, and changes in temperature in both the machine’s internal and external environments can be of great concern as they affect the machine’s thermal stability and, thus, the machine’s dimensional accuracy. This paper is a continuation of our earlier work, which aimed to analyze the effect of the internal temperature of a machine tool as the machine is put into operation and vary the external temperature, the machine floor temperature. Some experiments are carried out under controlled conditions to study how machine tool components get heated up and how this heating up affects the machine’s accuracy due to thermally induced deviations. Additionally, another angle is added by varying the machine floor temperature. The parameters mentioned above are explored in line with the overall thermal stability of the machine tool and its dimensional accuracy. A Robodrill CNC machine tool is used. The CNC was first soaked with thermal energy by gradually raising the machine floor temperature to a certain level before putting the machine in operation. The machine was monitored, and analytical methods were deplored to evaluate thermal stability. Secondly, the machine was run idle for some time under raised floor temperature before it was put into operation. Data was also collected and analyzed. It is observed that machine thermal stability can be achieved in several ways depending on how the above parameters are joggled. This paper, in conclusion, reinforces the idea of machine tool warm-up process in conjunction with a carefully analyzed and established machine floor temperature variation for the approximation of the machine tool’s thermally stability to map the long-time behavior of the machine tool.展开更多
基金supported by the Agence Nationale de la Recherche(ANR)(contract“ANR-17-EURE-0002”)by the Region of Bourgogne Franche-ComtéCADRAN Projectsupported by the European Research Council(ERC)project HYPATIA under the European Union's Horizon 2020 research and innovation programme.Grant agreement n.835294。
文摘This paper investigates the problem of collecting multidimensional data throughout time(i.e.,longitudinal studies)for the fundamental task of frequency estimation under Local Differential Privacy(LDP)guarantees.Contrary to frequency estimation of a single attribute,the multidimensional aspect demands particular attention to the privacy budget.Besides,when collecting user statistics longitudinally,privacy progressively degrades.Indeed,the“multiple”settings in combination(i.e.,many attributes and several collections throughout time)impose several challenges,for which this paper proposes the first solution for frequency estimates under LDP.To tackle these issues,we extend the analysis of three state-of-the-art LDP protocols(Generalized Randomized Response–GRR,Optimized Unary Encoding–OUE,and Symmetric Unary Encoding–SUE)for both longitudinal and multidimensional data collections.While the known literature uses OUE and SUE for two rounds of sanitization(a.k.a.memoization),i.e.,L-OUE and L-SUE,respectively,we analytically and experimentally show that starting with OUE and then with SUE provides higher data utility(i.e.,L-OSUE).Also,for attributes with small domain sizes,we propose Longitudinal GRR(L-GRR),which provides higher utility than the other protocols based on unary encoding.Last,we also propose a new solution named Adaptive LDP for LOngitudinal and Multidimensional FREquency Estimates(ALLOMFREE),which randomly samples a single attribute to be sent with the whole privacy budget and adaptively selects the optimal protocol,i.e.,either L-GRR or L-OSUE.As shown in the results,ALLOMFREE consistently and considerably outperforms the state-of-the-art L-SUE and L-OUE protocols in the quality of the frequency estimates.
文摘Massive amounts of data are acquired in modern and future information technology industries such as communication,radar,and remote sensing.The presence of large dimensionality and size in these data offers new opportunities to enhance the performance of signal processing in such applications and even motivate new ones.However,the curse of dimensionality is always a challenge when processing such high-dimensional signals.In practical tasks,high-dimensional signals need to be acquired,processed,and analyzed with high accuracy,robustness,and computational efficiency.This special section aims to address these challenges,where articles attempt to develop new theories and methods that are best suited to the high dimensional nature of the signals involved,and explore modern and emerging applications in this area.
基金supported in part by the National Key R&D Program of China(2022YFC3401303)the Natural Science Foundation of Jiangsu Province(BK20211528)the Postgraduate Research&Practice Innovation Program of Jiangsu Province(KFCX22_2300)。
文摘In the era of exponential growth of data availability,the architecture of systems has a trend toward high dimensionality,and directly exploiting holistic information for state inference is not always computationally affordable.This paper proposes a novel Bayesian filtering algorithm that considers algorithmic computational cost and estimation accuracy for high-dimensional linear systems.The high-dimensional state vector is divided into several blocks to save computation resources by avoiding the calculation of error covariance with immense dimensions.After that,two sequential states are estimated simultaneously by introducing an auxiliary variable in the new probability space,mitigating the performance degradation caused by state segmentation.Moreover,the computational cost and error covariance of the proposed algorithm are analyzed analytically to show its distinct features compared with several existing methods.Simulation results illustrate that the proposed Bayesian filtering can maintain a higher estimation accuracy with reasonable computational cost when applied to high-dimensional linear systems.
文摘This study evaluated the genetic and agronomic parameter estimates of maize under different nitrogen rates. The trial was established at the Njala Agricultural Research Centre experimental site during 2021 and 2022 in a split block design with three maize varieties (IWCD2, 2009EVDT, and DMR-ESR-Yellow) and seven nitrogen (0, 30, 60, 90, 120, 150 and 180 kg∙N∙ha<sup>−</sup><sup>1</sup>) rates. Findings showed that cob diameter and anthesis silking time (ASI) had intermediate heritability, ASI had high genetic advance, ASI and grain yield had high genotypic coefficient of variation (GCV), while traits with high phenotypic coefficient of variation (PCV) were plant height, ASI, grain yield, number of kernel per cob, number of kernel rows, ear length, and ear height. The PCV values were higher than GCV, indicating the influence of the environment in the studied traits. Nitrogen rates and variety significantly (p < 0.05) influenced grain yield production. Mean grain yields and economic parameter estimates increased with increasing nitrogen rates, with the 30 and 180 kg∙N∙ha<sup>−</sup><sup>1</sup> plots exhibiting the lowest and highest grain yields of 1238 kg∙ha<sup>−</sup><sup>1</sup> and 2098 kg∙ha<sup>−</sup><sup>1</sup>, respectively. Variety and nitrogen effects on partial factor productivity (PFP<sub>N</sub>), agronomic efficiency (AEN), net returns (NR), value cost ratio (VCR) and marginal return (MR) indicated that these parameters were significantly affected (p < 0.05) by these factors. The highest PFP<sub>N</sub> (41.3 kg grain kg<sup>−</sup><sup>1</sup>∙N) and AEN (29.4 kg grain kg<sup>−</sup><sup>1</sup>∙N) were obtained in the 30 kg∙N∙ha<sup>−</sup><sup>1</sup> plots, while the highest VCR (2.8) and MR (SLL 1.8 SLL<sup>−</sup><sup>1</sup> spent on N) were obtained in the 180 kg∙N∙ha<sup>−</sup><sup>1</sup>. The significant influence of variety and nitrogen on traits suggests that increasing yields and maximizing profits require use of appropriate nitrogen fertilization and improved farming practices that could be exploited for increased productivity of maize.
基金Guangdong Basic and Applied Basic Research Foundation,Grant/Award Number:2023A1515012878Natural Science Foundation of Anhui Province,Grant/Award Number:2008085ME134+2 种基金Australian Research Council Discovery Projects,Grant/Award Numbers:ARC DP200103315,ARC DP200103332Major Special Science and Technology Project of Anhui Province,Grant/Award Number:202103a07020007Key Research and Development Program of Anhui Province,Grant/Award Number:202104a05020057。
文摘Ni-Fe-based oxides are among the most promising catalysts developed to date for the bottleneck oxygen evolution reaction(OER)in water electrolysis.However,understanding and mastering the synergy of Ni and Fe remain challenging.Herein,we report that the synergy between Ni and Fe can be tailored by crystal dimensionality of Ni,Fe-contained Ruddlesden-Popper(RP)-type perovskites(La_(0.125)Sr_(0.875))n+1(Ni_(0.25)Fe_(0.75))nO3n+1(n=1,2,3),where the material with n=3 shows the best OER performance in alkaline media.Soft X-ray absorption spectroscopy spectra before and after OER reveal that the material with n=3 shows enhanced Ni/Fe-O covalency to boost the electron transfer as compared to those with n=1 and n=2.Further experimental investigations demonstrate that the Fe ion is the active site and the Ni ion is the stable site in this system,where such unique synergy reaches the optimum at n=3.Besides,as n increases,the proportion of unstable rock-salt layers accordingly decreases and the leaching of ions(especially Sr^(2+))into the electrolyte is suppressed,which induces a decrease in the leaching of active Fe ions,ultimately leading to enhanced stability.This work provides a new avenue for rational catalyst design through the dimensional strategy.
文摘The estimation of covariance matrices is very important in many fields, such as statistics. In real applications, data are frequently influenced by high dimensions and noise. However, most relevant studies are based on complete data. This paper studies the optimal estimation of high-dimensional covariance matrices based on missing and noisy sample under the norm. First, the model with sub-Gaussian additive noise is presented. The generalized sample covariance is then modified to define a hard thresholding estimator , and the minimax upper bound is derived. After that, the minimax lower bound is derived, and it is concluded that the estimator presented in this article is rate-optimal. Finally, numerical simulation analysis is performed. The result shows that for missing samples with sub-Gaussian noise, if the true covariance matrix is sparse, the hard thresholding estimator outperforms the traditional estimate method.
基金Project supported by the National Natural Science Foundation of China(Grant Nos.12105097 and 12035005)the Science Research Fund of the Education Department of Hunan Province,China(Grant No.23B0480).
文摘Quantum Fisher information(QFI)associated with local metrology has been used to parameter estimation in open quantum systems.In this work,we calculated the QFI for a moving Unruh-DeWitt detector coupled with massless scalar fields in n-dimensional spacetime,and analyzed the behavior of QFI with various parameters,such as the dimension of spacetime,evolution time,and Unruh temperature.We discovered that the QFI of state parameter decreases monotonically from 1 to 0 over time.Additionally,we noted that the QFI for small evolution times is several orders of magnitude higher than the QFI for long evolution times.We also found that the value of QFI decreases at first and then stabilizes as the Unruh temperature increases.It was observed that the QFI depends on initial state parameterθ,and Fθis the maximum forθ=0 orθ=π,Fφis the maximum forθ=π/2.We also obtain that the maximum value of QFI for state parameters varies for different spacetime dimensions with the same evolution time.
基金Project supported by the National Natural Science Foundation of China (Grant Nos.12275144,12235007,and 11975131)K.C.Wong Magna Fund in Ningbo University。
文摘Integrable systems play a crucial role in physics and mathematics.In particular,the traditional(1+1)-dimensional and(2+1)-dimensional integrable systems have received significant attention due to the rarity of integrable systems in higher dimensions.Recent studies have shown that abundant higher-dimensional integrable systems can be constructed from(1+1)-dimensional integrable systems by using a deformation algorithm.Here we establish a new(2+1)-dimensional Chen-Lee-Liu(C-L-L)equation using the deformation algorithm from the(1+1)-dimensional C-L-L equation.The new system is integrable with its Lax pair obtained by applying the deformation algorithm to that of the(1+1)-dimension.It is challenging to obtain the exact solutions for the new integrable system because the new system combines both the original C-L-L equation and its reciprocal transformation.The traveling wave solutions are derived in implicit function expression,and some asymmetry peakon solutions are found.
基金supported by the National Natural Science Foundation of China(No.42127807)Natural Science Foundation of Sichuan Province of China(Project No.2023NSFSC0008)+1 种基金Uranium Geology Program of China Nuclear Geology(No.202205-6)the Sichuan Science and Technology Program(No.2021JDTD0018)。
文摘Onlineγ-spectrometry systems for inland waters,most of which extract samples in situ and in real time,are able to produce reliable activity concentration measurements for waterborne radionuclides only when they are distributed relatively uniformly and enter into a steady-state diffusion regime in the measurement chamber.To protect residents’health and ensure the safety of the living environment,better timeliness is required for this measurement method.To address this issue,this study established a mathematical model of the online waterγ-spectrometry system so that rapid warning and activity estimates can be obtained for water under non-steady-state(NSS)conditions.In addition,the detection efficiency of the detector for radionuclides during the NSS diffusion process was determined by applying the computational fluid dynamics technique in conjunction with Monte Carlo simulations.On this basis,a method was developed that allowed the online waterγ-spectrometry system to provide rapid warning and activity concentration estimates for radionuclides in water.Subsequent analysis of the NSS-mode measurements of^(40)K radioactive solutions with different activity concentrations determined the optimum warning threshold and measurement time for producing accurate activity concentration estimates for radionuclides.The experimental results show that the proposed NSS measurement method is able to give warning and yield accurate activity concentration estimates for radionuclides 55.42 and 69.42 min after the entry of a 10 Bq/L^(40)K radioactive solution into the measurement chamber,respectively.These times are much shorter than the 90 min required by the conventional measurement method.Furthermore,the NSS measurement method allows the measurement system to give rapid(within approximately 15 min)warning when the activity concentrations of some radionuclides reach their respective limits stipulated in the Guidelines for Drinking-water Quality of the WHO,suggesting that this method considerably enhances the warning capacity of in situ online waterγ-spectrometry systems.
文摘This paper aims to explore the application of Extreme Value Theory (EVT) in estimating the conditional extreme quantile for time-to-event outcomes by examining the functional relationship between ambulatory blood pressure trajectories and clinical outcomes in stroke patients. The study utilizes EVT to analyze the functional connection between ambulatory blood pressure trajectories and clinical outcomes in a sample of 297 stroke patients. The 24-hour ambulatory blood pressure measurement curves for every 15 minutes are considered, acknowledging a censored rate of 40%. The findings reveal that the sample mean excess function exhibits a positive gradient above a specific threshold, confirming the heavy-tailed distribution of data in stroke patients with a positive extreme value index. Consequently, the estimated conditional extreme quantile indicates that stroke patients with higher blood pressure measurements face an elevated risk of recurrent stroke occurrence at an early stage. This research contributes to the understanding of the relationship between ambulatory blood pressure and recurrent stroke, providing valuable insights for clinical considerations and potential interventions in stroke management.
基金supported by the National Natural Science Foundation of China (Nos.52374078 and 52074043)the Fundamental Research Funds for the Central Universities (No.2023CDJKYJH021)。
文摘Fractal theory offers a powerful tool for the precise description and quantification of the complex pore structures in reservoir rocks,crucial for understanding the storage and migration characteristics of media within these rocks.Faced with the challenge of calculating the three-dimensional fractal dimensions of rock porosity,this study proposes an innovative computational process that directly calculates the three-dimensional fractal dimensions from a geometric perspective.By employing a composite denoising approach that integrates Fourier transform(FT)and wavelet transform(WT),coupled with multimodal pore extraction techniques such as threshold segmentation,top-hat transformation,and membrane enhancement,we successfully crafted accurate digital rock models.The improved box-counting method was then applied to analyze the voxel data of these digital rocks,accurately calculating the fractal dimensions of the rock pore distribution.Further numerical simulations of permeability experiments were conducted to explore the physical correlations between the rock pore fractal dimensions,porosity,and absolute permeability.The results reveal that rocks with higher fractal dimensions exhibit more complex pore connectivity pathways and a wider,more uneven pore distribution,suggesting that the ideal rock samples should possess lower fractal dimensions and higher effective porosity rates to achieve optimal fluid transmission properties.The methodology and conclusions of this study provide new tools and insights for the quantitative analysis of complex pores in rocks and contribute to the exploration of the fractal transport properties of media within rocks.
基金supported by the Medical Special Cultivation Project of Anhui University of Science and Technology(Grant No.YZ2023H2B013)the Anhui Provincial Key Research and Development Project(Grant No.2022i01020015)the Open Project of Key Laboratory of Conveyance Equipment(East China Jiaotong University),Ministry of Education(KLCE2022-01).
文摘The human pose paradigm is estimated using a transformer-based multi-branch multidimensional directed the three-dimensional(3D)method that takes into account self-occlusion,badly posedness,and a lack of depth data in the per-frame 3D posture estimation from two-dimensional(2D)mapping to 3D mapping.Firstly,by examining the relationship between the movements of different bones in the human body,four virtual skeletons are proposed to enhance the cyclic constraints of limb joints.Then,multiple parameters describing the skeleton are fused and projected into a high-dimensional space.Utilizing a multi-branch network,motion features between bones and overall motion features are extracted to mitigate the drift error in the estimation results.Furthermore,the estimated relative depth is projected into 3D space,and the error is calculated against real 3D data,forming a loss function along with the relative depth error.This article adopts the average joint pixel error as the primary performance metric.Compared to the benchmark approach,the estimation findings indicate an increase in average precision of 1.8 mm within the Human3.6M sample.
基金supported by the Innovation Fund Project of the Gansu Education Department(Grant No.2021B-099).
文摘The objective of reliability-based design optimization(RBDO)is to minimize the optimization objective while satisfying the corresponding reliability requirements.However,the nested loop characteristic reduces the efficiency of RBDO algorithm,which hinders their application to high-dimensional engineering problems.To address these issues,this paper proposes an efficient decoupled RBDO method combining high dimensional model representation(HDMR)and the weight-point estimation method(WPEM).First,we decouple the RBDO model using HDMR and WPEM.Second,Lagrange interpolation is used to approximate a univariate function.Finally,based on the results of the first two steps,the original nested loop reliability optimization model is completely transformed into a deterministic design optimization model that can be solved by a series of mature constrained optimization methods without any additional calculations.Two numerical examples of a planar 10-bar structure and an aviation hydraulic piping system with 28 design variables are analyzed to illustrate the performance and practicability of the proposed method.
基金Supported by the Kaohsiung Armed Forces General Hospital.
文摘BACKGROUND The prevalence of non-alcoholic fatty liver(NAFLD)has increased recently.Subjects with NAFLD are known to have higher chance for renal function impairment.Many past studies used traditional multiple linear regression(MLR)to identify risk factors for decreased estimated glomerular filtration rate(eGFR).However,medical research is increasingly relying on emerging machine learning(Mach-L)methods.The present study enrolled healthy women to identify factors affecting eGFR in subjects with and without NAFLD(NAFLD+,NAFLD-)and to rank their importance.AIM To uses three different Mach-L methods to identify key impact factors for eGFR in healthy women with and without NAFLD.METHODS A total of 65535 healthy female study participants were enrolled from the Taiwan MJ cohort,accounting for 32 independent variables including demographic,biochemistry and lifestyle parameters(independent variables),while eGFR was used as the dependent variable.Aside from MLR,three Mach-L methods were applied,including stochastic gradient boosting,eXtreme gradient boosting and elastic net.Errors of estimation were used to define method accuracy,where smaller degree of error indicated better model performance.RESULTS Income,albumin,eGFR,High density lipoprotein-Cholesterol,phosphorus,forced expiratory volume in one second(FEV1),and sleep time were all lower in the NAFLD+group,while other factors were all significantly higher except for smoking area.Mach-L had lower estimation errors,thus outperforming MLR.In Model 1,age,uric acid(UA),FEV1,plasma calcium level(Ca),plasma albumin level(Alb)and T-bilirubin were the most important factors in the NAFLD+group,as opposed to age,UA,FEV1,Alb,lactic dehydrogenase(LDH)and Ca for the NAFLD-group.Given the importance percentage was much higher than the 2nd important factor,we built Model 2 by removing age.CONCLUSION The eGFR were lower in the NAFLD+group compared to the NAFLD-group,with age being was the most important impact factor in both groups of healthy Chinese women,followed by LDH,UA,FEV1 and Alb.However,for the NAFLD-group,TSH and SBP were the 5th and 6th most important factors,as opposed to Ca and BF in the NAFLD+group.
基金supported by the National Key Research and Development Program of China(2020YFA0712900)the National Natural Science Foundation of China(12371093,12071197,12122102 and 12071431)+2 种基金the Key Project of Gansu Provincial National Science Foundation(23JRRA1022)the Fundamental Research Funds for the Central Universities(2233300008 and lzujbky-2021-ey18)the Innovative Groups of Basic Research in Gansu Province(22JR5RA391).
文摘Assume that L is a non-negative self-adjoint operator on L^(2)(ℝ^(n))with its heat kernels satisfying the so-called Gaussian upper bound estimate and that X is a ball quasi-Banach function space onℝ^(n) satisfying some mild assumptions.Let HX,L(ℝ^(n))be the Hardy space associated with both X and L,which is defined by the Lusin area function related to the semigroup generated by L.In this article,the authors establish various maximal function characterizations of the Hardy space HX,L(ℝ^(n))and then apply these characterizations to obtain the solvability of the related Cauchy problem.These results have a wide range of generality and,in particular,the specific spaces X to which these results can be applied include the weighted space,the variable space,the mixed-norm space,the Orlicz space,the Orlicz-slice space,and the Morrey space.Moreover,the obtained maximal function characterizations of the mixed-norm Hardy space,the Orlicz-slice Hardy space,and the Morrey-Hardy space associated with L are completely new.
基金the National Natural Science Foundation of China(Grant No.61973033)Preliminary Research of Equipment(Grant No.9090102010305)for funding the experiments。
文摘The longitudinal dispersion of the projectile in shooting tests of two-dimensional trajectory corrections fused with fixed canards is extremely large that it sometimes exceeds the correction ability of the correction fuse actuator.The impact point easily deviates from the target,and thus the correction result cannot be readily evaluated.However,the cost of shooting tests is considerably high to conduct many tests for data collection.To address this issue,this study proposes an aiming method for shooting tests based on small sample size.The proposed method uses the Bootstrap method to expand the test data;repeatedly iterates and corrects the position of the simulated theoretical impact points through an improved compatibility test method;and dynamically adjusts the weight of the prior distribution of simulation results based on Kullback-Leibler divergence,which to some extent avoids the real data being"submerged"by the simulation data and achieves the fusion Bayesian estimation of the dispersion center.The experimental results show that when the simulation accuracy is sufficiently high,the proposed method yields a smaller mean-square deviation in estimating the dispersion center and higher shooting accuracy than those of the three comparison methods,which is more conducive to reflecting the effect of the control algorithm and facilitating test personnel to iterate their proposed structures and algorithms.;in addition,this study provides a knowledge base for further comprehensive studies in the future.
基金supported by the National Basic Research Program of China。
文摘With the extensive application of large-scale array antennas,the increasing number of array elements leads to the increasing dimension of received signals,making it difficult to meet the real-time requirement of direction of arrival(DOA)estimation due to the computational complexity of algorithms.Traditional subspace algorithms require estimation of the covariance matrix,which has high computational complexity and is prone to producing spurious peaks.In order to reduce the computational complexity of DOA estimation algorithms and improve their estimation accuracy under large array elements,this paper proposes a DOA estimation method based on Krylov subspace and weighted l_(1)-norm.The method uses the multistage Wiener filter(MSWF)iteration to solve the basis of the Krylov subspace as an estimate of the signal subspace,further uses the measurement matrix to reduce the dimensionality of the signal subspace observation,constructs a weighted matrix,and combines the sparse reconstruction to establish a convex optimization function based on the residual sum of squares and weighted l_(1)-norm to solve the target DOA.Simulation results show that the proposed method has high resolution under large array conditions,effectively suppresses spurious peaks,reduces computational complexity,and has good robustness for low signal to noise ratio(SNR)environment.
基金supported by The Technology Innovation Team(Tianshan Innovation Team),Innovative Team for Efficient Utilization of Water Resources in Arid Regions(2022TSYCTD0001)the National Natural Science Foundation of China(42171269)the Xinjiang Academician Workstation Cooperative Research Project(2020.B-001).
文摘Xinjiang Uygur Autonomous Region is a typical inland arid area in China with a sparse and uneven distribution of meteorological stations,limited access to precipitation data,and significant water scarcity.Evaluating and integrating precipitation datasets from different sources to accurately characterize precipitation patterns has become a challenge to provide more accurate and alternative precipitation information for the region,which can even improve the performance of hydrological modelling.This study evaluated the applicability of widely used five satellite-based precipitation products(Climate Hazards Group InfraRed Precipitation with Station(CHIRPS),China Meteorological Forcing Dataset(CMFD),Climate Prediction Center morphing method(CMORPH),Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks-Climate Data Record(PERSIANN-CDR),and Tropical Rainfall Measuring Mission Multi-satellite Precipitation Analysis(TMPA))and a reanalysis precipitation dataset(ECMWF Reanalysis v5-Land Dataset(ERA5-Land))in Xinjiang using ground-based observational precipitation data from a limited number of meteorological stations.Based on this assessment,we proposed a framework that integrated different precipitation datasets with varying spatial resolutions using a dynamic Bayesian model averaging(DBMA)approach,the expectation-maximization method,and the ordinary Kriging interpolation method.The daily precipitation data merged using the DBMA approach exhibited distinct spatiotemporal variability,with an outstanding performance,as indicated by low root mean square error(RMSE=1.40 mm/d)and high Person's correlation coefficient(CC=0.67).Compared with the traditional simple model averaging(SMA)and individual product data,although the DBMA-fused precipitation data were slightly lower than the best precipitation product(CMFD),the overall performance of DBMA was more robust.The error analysis between DBMA-fused precipitation dataset and the more advanced Integrated Multi-satellite Retrievals for Global Precipitation Measurement Final(IMERG-F)precipitation product,as well as hydrological simulations in the Ebinur Lake Basin,further demonstrated the superior performance of DBMA-fused precipitation dataset in the entire Xinjiang region.The proposed framework for solving the fusion problem of multi-source precipitation data with different spatial resolutions is feasible for application in inland arid areas,and aids in obtaining more accurate regional hydrological information and improving regional water resources management capabilities and meteorological research in these regions.
文摘NGLY1 Deficiency is an ultra-rare autosomal recessively inherited disorder. Characteristic symptoms include among others, developmental delays, movement disorders, liver function abnormalities, seizures, and problems with tear formation. Movements are hyperkinetic and may include dysmetric, choreo-athetoid, myoclonic and dystonic movement elements. To date, there have been no quantitative reports describing arm movements of individuals with NGLY1 Deficiency. This report provides quantitative information about a series of arm movements performed by an individual with NGLY1 Deficiency and an aged-matched neurotypical participant. Three categories of arm movements were tested: 1) open ended reaches without specific end point targets;2) goal-directed reaches that included grasping an object;3) picking up small objects from a table placed in front of the participants. Arm movement kinematics were obtained with a camera-based motion analysis system and “initiation” and “maintenance” phases were identified for each movement. The combination of the two phases was labeled as a “complete” movement. Three-dimensional analysis techniques were used to quantify the movements and included hand trajectory pathlength, joint motion area, as well as hand trajectory and joint jerk cost. These techniques were required to fully characterize the movements because the NGLY1 individual was unable to perform movements only in the primary plane of progression instead producing motion across all three planes of movement. The individual with NGLY1 Deficiency was unable to pick up objects from a table or effectively complete movements requiring crossing the midline. The successfully completed movements were analyzed using the above techniques and the results of the two participants were compared statistically. Almost all comparisons revealed significant differences between the two participants, with a notable exception of the 3D initiation area as a percentage of the complete movement. The statistical tests of these measures revealed no significant differences between the two participants, possibly suggesting a common underlying motor control strategy. The 3D techniques used in this report effectively characterized arm movements of an individual with NGLY1 deficiency and can be used to provide information to evaluate the effectiveness of genetic, pharmacological, or physical rehabilitation therapies.
文摘Machining is as old as humanity, and changes in temperature in both the machine’s internal and external environments can be of great concern as they affect the machine’s thermal stability and, thus, the machine’s dimensional accuracy. This paper is a continuation of our earlier work, which aimed to analyze the effect of the internal temperature of a machine tool as the machine is put into operation and vary the external temperature, the machine floor temperature. Some experiments are carried out under controlled conditions to study how machine tool components get heated up and how this heating up affects the machine’s accuracy due to thermally induced deviations. Additionally, another angle is added by varying the machine floor temperature. The parameters mentioned above are explored in line with the overall thermal stability of the machine tool and its dimensional accuracy. A Robodrill CNC machine tool is used. The CNC was first soaked with thermal energy by gradually raising the machine floor temperature to a certain level before putting the machine in operation. The machine was monitored, and analytical methods were deplored to evaluate thermal stability. Secondly, the machine was run idle for some time under raised floor temperature before it was put into operation. Data was also collected and analyzed. It is observed that machine thermal stability can be achieved in several ways depending on how the above parameters are joggled. This paper, in conclusion, reinforces the idea of machine tool warm-up process in conjunction with a carefully analyzed and established machine floor temperature variation for the approximation of the machine tool’s thermally stability to map the long-time behavior of the machine tool.