The hesitancy fuzzy graphs(HFGs),an extension of fuzzy graphs,are useful tools for dealing with ambiguity and uncertainty in issues involving decision-making(DM).This research implements a correlation coefficient meas...The hesitancy fuzzy graphs(HFGs),an extension of fuzzy graphs,are useful tools for dealing with ambiguity and uncertainty in issues involving decision-making(DM).This research implements a correlation coefficient measure(CCM)to assess the strength of the association between HFGs in this article since CCMs have a high capacity to process and interpret data.The CCM that is proposed between the HFGs has better qualities than the existing ones.It lowers restrictions on the hesitant fuzzy elements’length and may be used to establish whether the HFGs are connected negatively or favorably.Additionally,a CCMbased attribute DM approach is built into a hesitant fuzzy environment.This article suggests the use of weighted correlation coefficient measures(WCCMs)using the CCM concept to quantify the correlation between two HFGs.The decisionmaking problems of hesitancy fuzzy preference relations(HFPRs)are considered.This research proposes a new technique for assessing the relative weights of experts based on the uncertainty of HFPRs and the correlation coefficient degree of each HFPR.This paper determines the ranking order of all alternatives and the best one by using the CCMs between each option and the ideal choice.In the meantime,the appropriate example is given to demonstrate the viability of the new strategies.展开更多
Leakages from subsea oil and gas equipment cause substantial economic losses and damage to marine ecosystem,so it is essential to locate the source of the leak.However,due to the complexity and variability of the mari...Leakages from subsea oil and gas equipment cause substantial economic losses and damage to marine ecosystem,so it is essential to locate the source of the leak.However,due to the complexity and variability of the marine environment,the signals collected by hydrophone contain a variety of noises,which makes it challenging to extract useful signals for localization.To solve this problem,a hydrophone denoising algorithm is proposed based on variational modal decomposition(VMD)with grey wolf optimization.First,the average envelope entropy is used as the fitness function of the grey wolf optimizer to find the optimal solution for the parameters K andα.Afterward,the VMD algorithm decomposes the original signal parameters to obtain the intrinsic mode functions(IMFs).Subsequently,the number of interrelationships between each IMF and the original signal was calculated,the threshold value was set,and the noise signal was removed to calculate the time difference using the valid signal obtained by reconstruction.Finally,the arrival time difference is used to locate the origin of the leak.The localization accuracy of the method in finding leaks is investigated experimentally by constructing a simulated leak test rig,and the effectiveness and feasibility of the method are verified.展开更多
The phenomenon of stochastic resonance (SR) based on the correlation coefficient in a parallel array of threshold devices is discussed. For four representative noises: the Gaussian noise, the uniform noise, the Lap...The phenomenon of stochastic resonance (SR) based on the correlation coefficient in a parallel array of threshold devices is discussed. For four representative noises: the Gaussian noise, the uniform noise, the Laplace noise and the Cauchy noise, when the signal is subthreshold, noise can improve the correlation coefficient and SR exists. The efficacy of SR can be significantly enhanced and the maximum of the correlation coefficient can dramatically approach to one as the number of the threshold devices in the parallel array increases. Two theorems are presented to prove that SR has some robustness to noises in the parallel array. These results further extend the applicability of SR in signal processing.展开更多
The complexities of the marine environment and the unique characteristics of underwater channels pose challenges in obtaining reliable signals underwater,necessitating the filtration of underwater acoustic noise.Herei...The complexities of the marine environment and the unique characteristics of underwater channels pose challenges in obtaining reliable signals underwater,necessitating the filtration of underwater acoustic noise.Herein,an underwater acoustic signal denoising method based on ensemble empirical mode decomposition(EEMD),correlation coefficient(CC),permutation entropy(PE),and wavelet threshold denoising(WTD)is proposed.Furthermore,simulation experiments are conducted using simulated and real underwater acoustic data.The experimental results reveal that the proposed denoising method outperforms other previous methods in terms of signal-to-noise ratio,root mean square error,and CC.The proposed method eliminates noise and retains valuable information in the signal.展开更多
Arc grounding faults occur frequently in the power grid with small resistance grounding neutral points.The existing arc fault identification technology only uses the fault line signal characteristics to set the identi...Arc grounding faults occur frequently in the power grid with small resistance grounding neutral points.The existing arc fault identification technology only uses the fault line signal characteristics to set the identification index,which leads to detection failure when the arc zero-off characteristic is short.To solve this problem,this paper presents an arc fault identification method by utilizing integrated signal characteristics of both the fault line and sound lines.Firstly,the waveform characteristics of the fault line and sound lines under an arc grounding fault are studied.After that,the convex hull,gradient product,and correlation coefficient index are used as the basic characteristic parameters to establish fault identification criteria.Then,the logistic regression algorithm is employed to deal with the reference samples,establish the machine discrimination model,and realize the discrimination of fault types.Finally,simulation test results and experimental results verify the accuracy of the proposed method.The comparison analysis shows that the proposed method has higher recognition accuracy,especially when the arc dissipation power is smaller than 2×10^(3) W,the zero-off period is not obvious.In conclusion,the proposed method expands the arc fault identification theory.展开更多
Plants play an essential role in matter and energy transformations and are key messengers in the carbon and energy cycle. Net primary productivity (NPP) reflects the capability of plants to transform solar energy into...Plants play an essential role in matter and energy transformations and are key messengers in the carbon and energy cycle. Net primary productivity (NPP) reflects the capability of plants to transform solar energy into photosynthesis. It is very sensible for factors affecting on vegetation variability such as climate, soils, plant characteristics and human activities. So, it can be used as an indicator of actual and potential trend of vegetation. In this study we used the actual NPP which was derived from MODIS to assess the response of NPP to climate variables in Gadarif State, from 2000 to 2010. The correlations between NPP and climate variables (temperature and precipitation) are calculated using Pearson’s Correlation Coefficient and ordinary least squares regression. The main results show the following 1) the correlation Coefficient between NPP and mean annual temperature is Somewhat negative for Feshaga, Rahd, Gadarif and Galabat areas and weakly negative in Faw area;2) the correlation Coefficient between NPP and annual total precipitation is weakly negative in Faw, Rahd and Galabat areas and somewhat negative in Galabat and Rahd areas. This study demonstrated that the correlation analysis between NPP and climate variables (precipitation and temperature) gives reliably result of NPP responses to climate variables that is clearly in a very large scale of study area.展开更多
The district cooling system (DCS) with ice storage can reduce the peak electricity demand of the business district buildings it serves, improve system efficiency, and lower operational costs. This study utilizes a mon...The district cooling system (DCS) with ice storage can reduce the peak electricity demand of the business district buildings it serves, improve system efficiency, and lower operational costs. This study utilizes a monitoring and control platform for DCS with ice storage to analyze historical parameter values related to system operation and executed operations. We assess the distribution of cooling loads among various devices within the DCS, identify operational characteristics of the system through correlation analysis and principal component analysis (PCA), and subsequently determine key parameters affecting changes in cooling loads. Accurate forecasting of cooling loads is crucial for determining optimal control strategies. The research process can be summarized briefly as follows: data preprocessing, parameter analysis, parameter selection, and validation of load forecasting performance. The study reveals that while individual devices in the system perform well, there is considerable room for improving overall system efficiency. Six principal components have been identified as input parameters for the cold load forecasting model, with each of these components having eigenvalues greater than 1 and contributing to an accumulated variance of 87.26%, and during the dimensionality reduction process, we obtained a confidence ellipse with a 95% confidence interval. Regarding cooling load forecasting, the Relative Absolute Error (RAE) value of the light gradient boosting machine (lightGBM) algorithm is 3.62%, Relative Root Mean Square Error (RRMSE) is 42.75%, and R-squared value (R<sup>2</sup>) is 92.96%, indicating superior forecasting performance compared to other commonly used cooling load forecasting algorithms. This research provides valuable insights and auxiliary guidance for data analysis and optimizing operations in practical engineering applications. .展开更多
Most edge-detection methods rely on calculating gradient derivatives of the potential field, a process that is easily affected by noise and is therefore of low stability. We propose a new edge-detection method named c...Most edge-detection methods rely on calculating gradient derivatives of the potential field, a process that is easily affected by noise and is therefore of low stability. We propose a new edge-detection method named correlation coefficient of multidirectional standard deviations(CCMS) that is solely based on statistics. First, we prove the reliability of the proposed method using a single model and then a combination of models. The proposed method is evaluated by comparing the results with those obtained by other edge-detection methods. The CCMS method offers outstanding recognition, retains the sharpness of details, and has low sensitivity to noise. We also applied the CCMS method to Bouguer anomaly data of a potash deposit in Laos. The applicability of the CCMS method is shown by comparing the inferred tectonic framework to that inferred from remote sensing(RS) data.展开更多
In order to study the temporal variations of correlations between two time series,a running correlation coefficient(RCC)could be used.An RCC is calculated for a given time window,and the window is then moved sequentia...In order to study the temporal variations of correlations between two time series,a running correlation coefficient(RCC)could be used.An RCC is calculated for a given time window,and the window is then moved sequentially through time.The current calculation method for RCCs is based on the general definition of the Pearson product-moment correlation coefficient,calculated with the data within the time window,which we call the local running correlation coefficient(LRCC).The LRCC is calculated via the two anomalies corresponding to the two local means,meanwhile,the local means also vary.It is cleared up that the LRCC reflects only the correlation between the two anomalies within the time window but fails to exhibit the contributions of the two varying means.To address this problem,two unchanged means obtained from all available data are adopted to calculate an RCC,which is called the synthetic running correlation coefficient(SRCC).When the anomaly variations are dominant,the two RCCs are similar.However,when the variations of the means are dominant,the difference between the two RCCs becomes obvious.The SRCC reflects the correlations of both the anomaly variations and the variations of the means.Therefore,the SRCCs from different time points are intercomparable.A criterion for the superiority of the RCC algorithm is that the average value of the RCC should be close to the global correlation coefficient calculated using all data.The SRCC always meets this criterion,while the LRCC sometimes fails.Therefore,the SRCC is better than the LRCC for running correlations.We suggest using the SRCC to calculate the RCCs.展开更多
Aim To study the reason of the insensitiveness of Pearson product-momentcorrelation coefficient as a similarity measure and the method to improve its sensitivity. MethodsExperimental and simulated data sets were used....Aim To study the reason of the insensitiveness of Pearson product-momentcorrelation coefficient as a similarity measure and the method to improve its sensitivity. MethodsExperimental and simulated data sets were used. Results The distribution range of the data setsinfluences the sensitivity of Pearson product-moment correlation coefficient. Weighted Pearsonproduct-moment correlation coefficient is more sensitive when the range of the data set is large.Conclusion Weighted Pearson product-moment correlation coefficient is necessary when the range ofthe data set is large.展开更多
The problems of time delay estimation of narrowband signals are presented. The disadvantages of the existing algorithms are analyzed, and a new narrowband time delay estimating algorithm based on correlation coefficie...The problems of time delay estimation of narrowband signals are presented. The disadvantages of the existing algorithms are analyzed, and a new narrowband time delay estimating algorithm based on correlation coefficient is proposed. The original time delay information is transfered into the delay between the autocorrelation and cross-correlation function, and the precise estimating result by wave-comparison is given. The algorithm proposed here is also compared with other algorithms and its advantages over other algorithms are proved. The theoretical analysis and simulation show the effectiveness of the proposed algorithm.展开更多
The running correlation coefficient(RCC)is useful for capturing temporal variations in correlations between two time series.The local running correlation coefficient(LRCC)is a widely used algorithm that directly appli...The running correlation coefficient(RCC)is useful for capturing temporal variations in correlations between two time series.The local running correlation coefficient(LRCC)is a widely used algorithm that directly applies the Pearson correlation to a time window.A new algorithm called synthetic running correlation coefficient(SRCC)was proposed in 2018 and proven to be rea-sonable and usable;however,this algorithm lacks a theoretical demonstration.In this paper,SRCC is proven theoretically.RCC is only meaningful when its values at different times can be compared.First,the global means are proven to be the unique standard quantities for comparison.SRCC is the only RCC that satisfies the comparability criterion.The relationship between LRCC and SRCC is derived using statistical methods,and SRCC is obtained by adding a constraint condition to the LRCC algorithm.Dividing the temporal fluctuations into high-and low-frequency signals reveals that LRCC only reflects the correlation of high-frequency signals;by contrast,SRCC reflects the correlations of high-and low-frequency signals simultaneously.Therefore,SRCC is the ap-propriate method for calculating RCCs.展开更多
The correlation coefficients of random variables of mechanical structures are generally chosen with experience or even ignored,which cannot actually reflect the effects of parameter uncertainties on reliability.To dis...The correlation coefficients of random variables of mechanical structures are generally chosen with experience or even ignored,which cannot actually reflect the effects of parameter uncertainties on reliability.To discuss the selection problem of the correlation coefficients from the reliability-based sensitivity point of view,the theory principle of the problem is established based on the results of the reliability sensitivity,and the criterion of correlation among random variables is shown.The values of the correlation coefficients are obtained according to the proposed principle and the reliability sensitivity problem is discussed.Numerical studies have shown the following results:(1) If the sensitivity value of correlation coefficient ρ is less than(at what magnitude 0.000 01),then the correlation could be ignored,which could simplify the procedure without introducing additional error.(2) However,as the difference between ρs,that is the most sensitive to the reliability,and ρR,that is with the smallest reliability,is less than 0.001,ρs is suggested to model the dependency of random variables.This could ensure the robust quality of system without the loss of safety requirement.(3) In the case of |Eabs|ρ0.001 and also |Erel|ρ0.001,ρR should be employed to quantify the correlation among random variables in order to ensure the accuracy of reliability analysis.Application of the proposed approach could provide a practical routine for mechanical design and manufactory to study the reliability and reliability-based sensitivity of basic design variables in mechanical reliability analysis and design.展开更多
In this article, the author obtains the large deviation principles for the empirical correlation coefficient of two Gaussian random variables X and Y. Especially, when considering two independent Gaussian random varia...In this article, the author obtains the large deviation principles for the empirical correlation coefficient of two Gaussian random variables X and Y. Especially, when considering two independent Gaussian random variables X, Y with the means EX, EY (both known), wherein the author gives two kinds of different proofs and gets the same results.展开更多
This paper presents a new approach using correlation and cross-correlation coefficients to evaluate the stiffness degradation of beams under moving load.The theoretical study of identifying defects by vibration method...This paper presents a new approach using correlation and cross-correlation coefficients to evaluate the stiffness degradation of beams under moving load.The theoretical study of identifying defects by vibration methods showed that the traditional methods derived from the vibration measurement data have not met the needs of the actual issues.We show that the correlation coefficients allow us to evaluate the degree and the effectiveness of the defects on beams.At the same time,the cross-correlation model is the basis for determining the relative position of defects.The results of this study are experimentally conducted to confirm the relationship between the correlation coefficients and the existence of the defects.In particular,the manuscript shows that the sensitivity of the correlation coefficients and cross-correlation is much higher than the parameters such as changes in stiffness(EJ)and natural frequency values(Δf).This study suggests using the above parameters to evaluate the stiffness degradation of beams by vibration measurement data in practice.展开更多
Real-life data introduce noise,uncertainty,and imprecision to statistical projects;it is advantageous to consider strategies to overcome these information expressions and processing problems.Neutrosophic(indeterminate...Real-life data introduce noise,uncertainty,and imprecision to statistical projects;it is advantageous to consider strategies to overcome these information expressions and processing problems.Neutrosophic(indeterminate)numbers can flexibly and conveniently represent the hybrid information of the partial determinacy and partial indeterminacy in an indeterminate setting,while a fuzzy multiset is a vital mathematical tool in the expression and processing of multi-valued fuzzy information with different and/or same fuzzy values.If neutrosophic numbers are introduced into fuzzy sequences in a fuzzy multiset,the introduced neutrosophic number sequences can be constructed as the neutrosophic number multiset or indeterminate fuzzy multiset.Motivated based on the idea,this study first proposes an indeterminate fuzzy multiset,where each element in a universe set can be repeated more than once with the different and/or identical indeterminate membership values.Then,we propose the parameterized correlation coefficients of indeterminate fuzzy multisets based on the de-neutrosophication of transforming indeterminate fuzzy multisets into the parameterized fuzzy multisets by a parameter(the parameterized de-neutrosophication method).Since indeterminate decision-making issues need to be handled by an indeterminate decision-making method,a group decision-making method using the weighted parameterized correlation coefficients of indeterminate fuzzy multisets is developed along with decision makers’different decision risks(small,moderate,and large risks)so as to handle multicriteria group decision-making problems in indeterminate fuzzy multiset setting.Finally,the developed group decision-making approach is used in an example on a selection problem of slope design schemes for an open-pit mine to demonstrate its usability and flexibility in the indeterminate group decision-making problem with indeterminate fuzzy multisets.展开更多
We present definitions of the correlation degree and correlation coefficient of multi-output functions. Two relationships about the correlation degree of multi-output functions are proved. One is between the correlati...We present definitions of the correlation degree and correlation coefficient of multi-output functions. Two relationships about the correlation degree of multi-output functions are proved. One is between the correlation degree and independency, the other is between the correlation degree and balance. Especially the paper discusses the correlation degree of affine multioutput functions. We demonstrate properties of the correlation coefficient of multi-output functions. One is the value range of the correlation coefficient, one is the relationship between the correlation coefficient and independency, and another is the sufficient and necessary condition that two multi-output functions are equivalent to each other.展开更多
Prediction of reservoir fracture is the key to explore fracture-type reservoir. When a shear-wave propagates in anisotropic media containing fracture,it splits into two polarized shear waves: fast shear wave and slow ...Prediction of reservoir fracture is the key to explore fracture-type reservoir. When a shear-wave propagates in anisotropic media containing fracture,it splits into two polarized shear waves: fast shear wave and slow shear wave. The polarization and time delay of the fast and slow shear wave can be used to predict the azimuth and density of fracture. The current identification method of fracture azimuth and fracture density is cross-correlation method. It is assumed that fast and slow shear waves were symmetrical wavelets after completely separating,and use the most similar characteristics of wavelets to identify fracture azimuth and density,but in the experiment the identification is poor in accuracy. Pearson correlation coefficient method is one of the methods for separating the fast wave and slow wave. This method is faster in calculating speed and better in noise immunity and resolution compared with the traditional cross-correlation method. Pearson correlation coefficient method is a non-linear problem,particle swarm optimization( PSO) is a good nonlinear global optimization method which converges fast and is easy to implement. In this study,PSO is combined with the Pearson correlation coefficient method to achieve identifying fracture property and improve the computational efficiency.展开更多
This thesis offers the general concept of coefficient of partial correlation.Starting with regres-sion analysis,the paper,by using samples,infers the general formula of expressing coefficient of partial correlation by...This thesis offers the general concept of coefficient of partial correlation.Starting with regres-sion analysis,the paper,by using samples,infers the general formula of expressing coefficient of partial correlation by way of simple correlation coefficient.展开更多
Fluidization of non-spherical particles is very common in petroleum engineering.Understanding the complex phenomenon of non-spherical particle flow is of great significance.In this paper,coupled with two-fluid model,t...Fluidization of non-spherical particles is very common in petroleum engineering.Understanding the complex phenomenon of non-spherical particle flow is of great significance.In this paper,coupled with two-fluid model,the drag coefficient correlation based on artificial neural network was applied in the simulations of a bubbling fluidized bed filled with non-spherical particles.The simulation results were compared with the experimental data from the literature.Good agreement between the experimental data and the simulation results reveals that the modified drag model can accurately capture the interaction between the gas phase and solid phase.Then,several cases of different particles,including tetrahedron,cube,and sphere,together with the nylon beads used in the model validation,were employed in the simulations to study the effect of particle shape on the flow behaviors in the bubbling fluidized bed.Particle shape affects the hydrodynamics of non-spherical particles mainly on microscale.This work can be a basis and reference for the utilization of artificial neural network in the investigation of drag coefficient correlation in the dense gas-solid two-phase flow.Moreover,the proposed drag coefficient correlation provides one more option when investigating the hydrodynamics of non-spherical particles in the gas-solid fluidized bed.展开更多
基金This research work supported and funded was provided by Vellore Institute of Technology.
文摘The hesitancy fuzzy graphs(HFGs),an extension of fuzzy graphs,are useful tools for dealing with ambiguity and uncertainty in issues involving decision-making(DM).This research implements a correlation coefficient measure(CCM)to assess the strength of the association between HFGs in this article since CCMs have a high capacity to process and interpret data.The CCM that is proposed between the HFGs has better qualities than the existing ones.It lowers restrictions on the hesitant fuzzy elements’length and may be used to establish whether the HFGs are connected negatively or favorably.Additionally,a CCMbased attribute DM approach is built into a hesitant fuzzy environment.This article suggests the use of weighted correlation coefficient measures(WCCMs)using the CCM concept to quantify the correlation between two HFGs.The decisionmaking problems of hesitancy fuzzy preference relations(HFPRs)are considered.This research proposes a new technique for assessing the relative weights of experts based on the uncertainty of HFPRs and the correlation coefficient degree of each HFPR.This paper determines the ranking order of all alternatives and the best one by using the CCMs between each option and the ideal choice.In the meantime,the appropriate example is given to demonstrate the viability of the new strategies.
基金financially supported by the National Key Research and Development Program of China(Grant No.2022YFC2806102)the National Natural Science Foundation of China(Grant Nos.52171287,52325107)+2 种基金High Tech Ship Research Project of Ministry of Industry and Information Technology(Grant Nos.2023GXB01-05-004-03,GXBZH2022-293)the Science Foundation for Distinguished Young Scholars of Shandong Province(Grant No.ZR2022JQ25)the Taishan Scholars Project(Grant No.tsqn201909063)。
文摘Leakages from subsea oil and gas equipment cause substantial economic losses and damage to marine ecosystem,so it is essential to locate the source of the leak.However,due to the complexity and variability of the marine environment,the signals collected by hydrophone contain a variety of noises,which makes it challenging to extract useful signals for localization.To solve this problem,a hydrophone denoising algorithm is proposed based on variational modal decomposition(VMD)with grey wolf optimization.First,the average envelope entropy is used as the fitness function of the grey wolf optimizer to find the optimal solution for the parameters K andα.Afterward,the VMD algorithm decomposes the original signal parameters to obtain the intrinsic mode functions(IMFs).Subsequently,the number of interrelationships between each IMF and the original signal was calculated,the threshold value was set,and the noise signal was removed to calculate the time difference using the valid signal obtained by reconstruction.Finally,the arrival time difference is used to locate the origin of the leak.The localization accuracy of the method in finding leaks is investigated experimentally by constructing a simulated leak test rig,and the effectiveness and feasibility of the method are verified.
文摘The phenomenon of stochastic resonance (SR) based on the correlation coefficient in a parallel array of threshold devices is discussed. For four representative noises: the Gaussian noise, the uniform noise, the Laplace noise and the Cauchy noise, when the signal is subthreshold, noise can improve the correlation coefficient and SR exists. The efficacy of SR can be significantly enhanced and the maximum of the correlation coefficient can dramatically approach to one as the number of the threshold devices in the parallel array increases. Two theorems are presented to prove that SR has some robustness to noises in the parallel array. These results further extend the applicability of SR in signal processing.
基金Supported by the National Natural Science Foundation of China(No.62033011)Science and Technology Project of Hebei Province(No.216Z1704G,No.20310401D)。
文摘The complexities of the marine environment and the unique characteristics of underwater channels pose challenges in obtaining reliable signals underwater,necessitating the filtration of underwater acoustic noise.Herein,an underwater acoustic signal denoising method based on ensemble empirical mode decomposition(EEMD),correlation coefficient(CC),permutation entropy(PE),and wavelet threshold denoising(WTD)is proposed.Furthermore,simulation experiments are conducted using simulated and real underwater acoustic data.The experimental results reveal that the proposed denoising method outperforms other previous methods in terms of signal-to-noise ratio,root mean square error,and CC.The proposed method eliminates noise and retains valuable information in the signal.
基金This work was supported in part by the Natural Science Foundation of Henan Province,and the specific grant number is 232300420301。
文摘Arc grounding faults occur frequently in the power grid with small resistance grounding neutral points.The existing arc fault identification technology only uses the fault line signal characteristics to set the identification index,which leads to detection failure when the arc zero-off characteristic is short.To solve this problem,this paper presents an arc fault identification method by utilizing integrated signal characteristics of both the fault line and sound lines.Firstly,the waveform characteristics of the fault line and sound lines under an arc grounding fault are studied.After that,the convex hull,gradient product,and correlation coefficient index are used as the basic characteristic parameters to establish fault identification criteria.Then,the logistic regression algorithm is employed to deal with the reference samples,establish the machine discrimination model,and realize the discrimination of fault types.Finally,simulation test results and experimental results verify the accuracy of the proposed method.The comparison analysis shows that the proposed method has higher recognition accuracy,especially when the arc dissipation power is smaller than 2×10^(3) W,the zero-off period is not obvious.In conclusion,the proposed method expands the arc fault identification theory.
文摘Plants play an essential role in matter and energy transformations and are key messengers in the carbon and energy cycle. Net primary productivity (NPP) reflects the capability of plants to transform solar energy into photosynthesis. It is very sensible for factors affecting on vegetation variability such as climate, soils, plant characteristics and human activities. So, it can be used as an indicator of actual and potential trend of vegetation. In this study we used the actual NPP which was derived from MODIS to assess the response of NPP to climate variables in Gadarif State, from 2000 to 2010. The correlations between NPP and climate variables (temperature and precipitation) are calculated using Pearson’s Correlation Coefficient and ordinary least squares regression. The main results show the following 1) the correlation Coefficient between NPP and mean annual temperature is Somewhat negative for Feshaga, Rahd, Gadarif and Galabat areas and weakly negative in Faw area;2) the correlation Coefficient between NPP and annual total precipitation is weakly negative in Faw, Rahd and Galabat areas and somewhat negative in Galabat and Rahd areas. This study demonstrated that the correlation analysis between NPP and climate variables (precipitation and temperature) gives reliably result of NPP responses to climate variables that is clearly in a very large scale of study area.
文摘The district cooling system (DCS) with ice storage can reduce the peak electricity demand of the business district buildings it serves, improve system efficiency, and lower operational costs. This study utilizes a monitoring and control platform for DCS with ice storage to analyze historical parameter values related to system operation and executed operations. We assess the distribution of cooling loads among various devices within the DCS, identify operational characteristics of the system through correlation analysis and principal component analysis (PCA), and subsequently determine key parameters affecting changes in cooling loads. Accurate forecasting of cooling loads is crucial for determining optimal control strategies. The research process can be summarized briefly as follows: data preprocessing, parameter analysis, parameter selection, and validation of load forecasting performance. The study reveals that while individual devices in the system perform well, there is considerable room for improving overall system efficiency. Six principal components have been identified as input parameters for the cold load forecasting model, with each of these components having eigenvalues greater than 1 and contributing to an accumulated variance of 87.26%, and during the dimensionality reduction process, we obtained a confidence ellipse with a 95% confidence interval. Regarding cooling load forecasting, the Relative Absolute Error (RAE) value of the light gradient boosting machine (lightGBM) algorithm is 3.62%, Relative Root Mean Square Error (RRMSE) is 42.75%, and R-squared value (R<sup>2</sup>) is 92.96%, indicating superior forecasting performance compared to other commonly used cooling load forecasting algorithms. This research provides valuable insights and auxiliary guidance for data analysis and optimizing operations in practical engineering applications. .
基金supported by the National Hi-Tech Research and Development Program of China(863 Program)(No.2006AA06Z107)the National Natural Science Foundation of China(No.40930314)
文摘Most edge-detection methods rely on calculating gradient derivatives of the potential field, a process that is easily affected by noise and is therefore of low stability. We propose a new edge-detection method named correlation coefficient of multidirectional standard deviations(CCMS) that is solely based on statistics. First, we prove the reliability of the proposed method using a single model and then a combination of models. The proposed method is evaluated by comparing the results with those obtained by other edge-detection methods. The CCMS method offers outstanding recognition, retains the sharpness of details, and has low sensitivity to noise. We also applied the CCMS method to Bouguer anomaly data of a potash deposit in Laos. The applicability of the CCMS method is shown by comparing the inferred tectonic framework to that inferred from remote sensing(RS) data.
基金supported by the Key Program of the National Natural Science Foundation of China (No. 41330960)the Global Change Research Program of China (No. 2015CB953900)
文摘In order to study the temporal variations of correlations between two time series,a running correlation coefficient(RCC)could be used.An RCC is calculated for a given time window,and the window is then moved sequentially through time.The current calculation method for RCCs is based on the general definition of the Pearson product-moment correlation coefficient,calculated with the data within the time window,which we call the local running correlation coefficient(LRCC).The LRCC is calculated via the two anomalies corresponding to the two local means,meanwhile,the local means also vary.It is cleared up that the LRCC reflects only the correlation between the two anomalies within the time window but fails to exhibit the contributions of the two varying means.To address this problem,two unchanged means obtained from all available data are adopted to calculate an RCC,which is called the synthetic running correlation coefficient(SRCC).When the anomaly variations are dominant,the two RCCs are similar.However,when the variations of the means are dominant,the difference between the two RCCs becomes obvious.The SRCC reflects the correlations of both the anomaly variations and the variations of the means.Therefore,the SRCCs from different time points are intercomparable.A criterion for the superiority of the RCC algorithm is that the average value of the RCC should be close to the global correlation coefficient calculated using all data.The SRCC always meets this criterion,while the LRCC sometimes fails.Therefore,the SRCC is better than the LRCC for running correlations.We suggest using the SRCC to calculate the RCCs.
文摘Aim To study the reason of the insensitiveness of Pearson product-momentcorrelation coefficient as a similarity measure and the method to improve its sensitivity. MethodsExperimental and simulated data sets were used. Results The distribution range of the data setsinfluences the sensitivity of Pearson product-moment correlation coefficient. Weighted Pearsonproduct-moment correlation coefficient is more sensitive when the range of the data set is large.Conclusion Weighted Pearson product-moment correlation coefficient is necessary when the range ofthe data set is large.
基金supported partly by the National Natural Science Foundation of China(6037208130570475)the Education Ministry Doctoral Degree Foundation of China(20050141025).
文摘The problems of time delay estimation of narrowband signals are presented. The disadvantages of the existing algorithms are analyzed, and a new narrowband time delay estimating algorithm based on correlation coefficient is proposed. The original time delay information is transfered into the delay between the autocorrelation and cross-correlation function, and the precise estimating result by wave-comparison is given. The algorithm proposed here is also compared with other algorithms and its advantages over other algorithms are proved. The theoretical analysis and simulation show the effectiveness of the proposed algorithm.
基金This study was supported by the National Natural Sci-ence Foundation of China(Nos.41976022,41941012)the Major Scientific and Technological Innovation Projects of Shandong Province(No.2018SDKJ0104-1).
文摘The running correlation coefficient(RCC)is useful for capturing temporal variations in correlations between two time series.The local running correlation coefficient(LRCC)is a widely used algorithm that directly applies the Pearson correlation to a time window.A new algorithm called synthetic running correlation coefficient(SRCC)was proposed in 2018 and proven to be rea-sonable and usable;however,this algorithm lacks a theoretical demonstration.In this paper,SRCC is proven theoretically.RCC is only meaningful when its values at different times can be compared.First,the global means are proven to be the unique standard quantities for comparison.SRCC is the only RCC that satisfies the comparability criterion.The relationship between LRCC and SRCC is derived using statistical methods,and SRCC is obtained by adding a constraint condition to the LRCC algorithm.Dividing the temporal fluctuations into high-and low-frequency signals reveals that LRCC only reflects the correlation of high-frequency signals;by contrast,SRCC reflects the correlations of high-and low-frequency signals simultaneously.Therefore,SRCC is the ap-propriate method for calculating RCCs.
基金supported by Changjiang Scholars and Innovative Research Team in University of China (Grant No. IRT0816)Key National Science & Technology Special Project on "High-Grade CNC Machine Tools and Basic Manufacturing Equipments" of China (Grant No. 2010ZX04014-014)+1 种基金National Natural Science Foundation of China (Grant No. 50875039)Key Projects in National Science & Technology Pillar Program during the 11th Five-year Plan Period of China (Grant No. 2009BAG12A02-A07-2)
文摘The correlation coefficients of random variables of mechanical structures are generally chosen with experience or even ignored,which cannot actually reflect the effects of parameter uncertainties on reliability.To discuss the selection problem of the correlation coefficients from the reliability-based sensitivity point of view,the theory principle of the problem is established based on the results of the reliability sensitivity,and the criterion of correlation among random variables is shown.The values of the correlation coefficients are obtained according to the proposed principle and the reliability sensitivity problem is discussed.Numerical studies have shown the following results:(1) If the sensitivity value of correlation coefficient ρ is less than(at what magnitude 0.000 01),then the correlation could be ignored,which could simplify the procedure without introducing additional error.(2) However,as the difference between ρs,that is the most sensitive to the reliability,and ρR,that is with the smallest reliability,is less than 0.001,ρs is suggested to model the dependency of random variables.This could ensure the robust quality of system without the loss of safety requirement.(3) In the case of |Eabs|ρ0.001 and also |Erel|ρ0.001,ρR should be employed to quantify the correlation among random variables in order to ensure the accuracy of reliability analysis.Application of the proposed approach could provide a practical routine for mechanical design and manufactory to study the reliability and reliability-based sensitivity of basic design variables in mechanical reliability analysis and design.
文摘In this article, the author obtains the large deviation principles for the empirical correlation coefficient of two Gaussian random variables X and Y. Especially, when considering two independent Gaussian random variables X, Y with the means EX, EY (both known), wherein the author gives two kinds of different proofs and gets the same results.
文摘This paper presents a new approach using correlation and cross-correlation coefficients to evaluate the stiffness degradation of beams under moving load.The theoretical study of identifying defects by vibration methods showed that the traditional methods derived from the vibration measurement data have not met the needs of the actual issues.We show that the correlation coefficients allow us to evaluate the degree and the effectiveness of the defects on beams.At the same time,the cross-correlation model is the basis for determining the relative position of defects.The results of this study are experimentally conducted to confirm the relationship between the correlation coefficients and the existence of the defects.In particular,the manuscript shows that the sensitivity of the correlation coefficients and cross-correlation is much higher than the parameters such as changes in stiffness(EJ)and natural frequency values(Δf).This study suggests using the above parameters to evaluate the stiffness degradation of beams by vibration measurement data in practice.
文摘Real-life data introduce noise,uncertainty,and imprecision to statistical projects;it is advantageous to consider strategies to overcome these information expressions and processing problems.Neutrosophic(indeterminate)numbers can flexibly and conveniently represent the hybrid information of the partial determinacy and partial indeterminacy in an indeterminate setting,while a fuzzy multiset is a vital mathematical tool in the expression and processing of multi-valued fuzzy information with different and/or same fuzzy values.If neutrosophic numbers are introduced into fuzzy sequences in a fuzzy multiset,the introduced neutrosophic number sequences can be constructed as the neutrosophic number multiset or indeterminate fuzzy multiset.Motivated based on the idea,this study first proposes an indeterminate fuzzy multiset,where each element in a universe set can be repeated more than once with the different and/or identical indeterminate membership values.Then,we propose the parameterized correlation coefficients of indeterminate fuzzy multisets based on the de-neutrosophication of transforming indeterminate fuzzy multisets into the parameterized fuzzy multisets by a parameter(the parameterized de-neutrosophication method).Since indeterminate decision-making issues need to be handled by an indeterminate decision-making method,a group decision-making method using the weighted parameterized correlation coefficients of indeterminate fuzzy multisets is developed along with decision makers’different decision risks(small,moderate,and large risks)so as to handle multicriteria group decision-making problems in indeterminate fuzzy multiset setting.Finally,the developed group decision-making approach is used in an example on a selection problem of slope design schemes for an open-pit mine to demonstrate its usability and flexibility in the indeterminate group decision-making problem with indeterminate fuzzy multisets.
文摘We present definitions of the correlation degree and correlation coefficient of multi-output functions. Two relationships about the correlation degree of multi-output functions are proved. One is between the correlation degree and independency, the other is between the correlation degree and balance. Especially the paper discusses the correlation degree of affine multioutput functions. We demonstrate properties of the correlation coefficient of multi-output functions. One is the value range of the correlation coefficient, one is the relationship between the correlation coefficient and independency, and another is the sufficient and necessary condition that two multi-output functions are equivalent to each other.
文摘Prediction of reservoir fracture is the key to explore fracture-type reservoir. When a shear-wave propagates in anisotropic media containing fracture,it splits into two polarized shear waves: fast shear wave and slow shear wave. The polarization and time delay of the fast and slow shear wave can be used to predict the azimuth and density of fracture. The current identification method of fracture azimuth and fracture density is cross-correlation method. It is assumed that fast and slow shear waves were symmetrical wavelets after completely separating,and use the most similar characteristics of wavelets to identify fracture azimuth and density,but in the experiment the identification is poor in accuracy. Pearson correlation coefficient method is one of the methods for separating the fast wave and slow wave. This method is faster in calculating speed and better in noise immunity and resolution compared with the traditional cross-correlation method. Pearson correlation coefficient method is a non-linear problem,particle swarm optimization( PSO) is a good nonlinear global optimization method which converges fast and is easy to implement. In this study,PSO is combined with the Pearson correlation coefficient method to achieve identifying fracture property and improve the computational efficiency.
文摘This thesis offers the general concept of coefficient of partial correlation.Starting with regres-sion analysis,the paper,by using samples,infers the general formula of expressing coefficient of partial correlation by way of simple correlation coefficient.
基金the financial support by the National Natural Science Foundation of China(Grant No.51706055).
文摘Fluidization of non-spherical particles is very common in petroleum engineering.Understanding the complex phenomenon of non-spherical particle flow is of great significance.In this paper,coupled with two-fluid model,the drag coefficient correlation based on artificial neural network was applied in the simulations of a bubbling fluidized bed filled with non-spherical particles.The simulation results were compared with the experimental data from the literature.Good agreement between the experimental data and the simulation results reveals that the modified drag model can accurately capture the interaction between the gas phase and solid phase.Then,several cases of different particles,including tetrahedron,cube,and sphere,together with the nylon beads used in the model validation,were employed in the simulations to study the effect of particle shape on the flow behaviors in the bubbling fluidized bed.Particle shape affects the hydrodynamics of non-spherical particles mainly on microscale.This work can be a basis and reference for the utilization of artificial neural network in the investigation of drag coefficient correlation in the dense gas-solid two-phase flow.Moreover,the proposed drag coefficient correlation provides one more option when investigating the hydrodynamics of non-spherical particles in the gas-solid fluidized bed.