To capture the presence of speech embedded in nonspeech events and background noise in shortwave non-cooperative communication, an algorithm for speech-stream detection in noisy environments is presented based on Empi...To capture the presence of speech embedded in nonspeech events and background noise in shortwave non-cooperative communication, an algorithm for speech-stream detection in noisy environments is presented based on Empirical Mode Decomposition (EMD) and statistical properties of higher-order cumulants of speech signals. With the EMD, the noise signals can be decomposed into different numbers of IMFs. Then, the fourth-order cumulant ( FOC ) can be used to extract the desired feature of statistical properties for IMF components. Since the higher-order eumulants are blind for Gaussian signals, the proposed method is especially effective regarding the problem of speech-stream detection, where the speech signal is distorted by Gaussian noise. With the self-adaptive decomposition by EMD, the proposed method can also work well for non-Gaussian noise. The experiments show that the proposed algorithm can suppress different noise types with different SNRs, and the algorithm is robust in real signal tests.展开更多
Statistics of order 2 (variance, auto and cross-correlation functions, auto and cross-power spectra) and 3 (skewness, auto and cross-bicorrelation functions, auto and cross-bispectra) are used to analyze the wave-part...Statistics of order 2 (variance, auto and cross-correlation functions, auto and cross-power spectra) and 3 (skewness, auto and cross-bicorrelation functions, auto and cross-bispectra) are used to analyze the wave-particle interaction in space plasmas. The signals considered here are medium scale electron density irregularities and ELF/ULF electrostatic turbulence. Nonlinearities are mainly observed in the ELF range. They are independently pointed out in time series associated with fluctuations in electronic density and in time series associated with the measurement of one electric field component. Peaks in cross-bicorrelation function and in mutual information clearly show that, in well delimited frequency bands, the wave-particle interactions are nonlinear above a certain level of fluctuations. The way the energy is transferred within the frequencies of density fluctuations is indicated by a bi-spectra analysis.展开更多
Healthcare decisions are based on scientific evidence obtained from medical studies by gathering data and analyzing it to obtain the best results. When analyzing data, biostatistics is a powerful tool, but healthcare ...Healthcare decisions are based on scientific evidence obtained from medical studies by gathering data and analyzing it to obtain the best results. When analyzing data, biostatistics is a powerful tool, but healthcare professionals lack knowledge in this field. This lack of knowledge can manifest itself in situations such as choosing the wrong statistical test for the right situation or applying a statistical test without checking its assumptions, leading to inaccurate results and misleading conclusions. With the help of this “narrative review”, the aim is to bring biostatistics closer to healthcare professionals by answering certain questions: how to describe the distribution of data? how to assess the normality of data? how to transform data? and how to choose between nonparametric and parametric tests? Through this work, our hope is that the reader will be able to choose the right test for the right situation, in order to obtain the most accurate results.展开更多
The development of defect prediction plays a significant role in improving software quality. Such predictions are used to identify defective modules before the testing and to minimize the time and cost. The software w...The development of defect prediction plays a significant role in improving software quality. Such predictions are used to identify defective modules before the testing and to minimize the time and cost. The software with defects negatively impacts operational costs and finally affects customer satisfaction. Numerous approaches exist to predict software defects. However, the timely and accurate software bugs are the major challenging issues. To improve the timely and accurate software defect prediction, a novel technique called Nonparametric Statistical feature scaled QuAdratic regressive convolution Deep nEural Network (SQADEN) is introduced. The proposed SQADEN technique mainly includes two major processes namely metric or feature selection and classification. First, the SQADEN uses the nonparametric statistical Torgerson–Gower scaling technique for identifying the relevant software metrics by measuring the similarity using the dice coefficient. The feature selection process is used to minimize the time complexity of software fault prediction. With the selected metrics, software fault perdition with the help of the Quadratic Censored regressive convolution deep neural network-based classification. The deep learning classifier analyzes the training and testing samples using the contingency correlation coefficient. The softstep activation function is used to provide the final fault prediction results. To minimize the error, the Nelder–Mead method is applied to solve non-linear least-squares problems. Finally, accurate classification results with a minimum error are obtained at the output layer. Experimental evaluation is carried out with different quantitative metrics such as accuracy, precision, recall, F-measure, and time complexity. The analyzed results demonstrate the superior performance of our proposed SQADEN technique with maximum accuracy, sensitivity and specificity by 3%, 3%, 2% and 3% and minimum time and space by 13% and 15% when compared with the two state-of-the-art methods.展开更多
This paper proposes a set of nonparametric statistical tools for analyzing the system resilience of civil structures and infrastructure and its migration upon changes in critical system parameters.The work is founded ...This paper proposes a set of nonparametric statistical tools for analyzing the system resilience of civil structures and infrastructure and its migration upon changes in critical system parameters.The work is founded on the classic theoretic framework that system resilience is defined in multiple dimensions for a constructed system.Consequentially,system resilience can lose its parametric form as a random variable,falling into the realm of nonparametric statistics.With this nonparametric shift,traditional distribution-based statistics are ineffective in characterizing the migration of system resilience due to the variation of system parameters.Three statistical tools are proposed under the nonparametric statistical resilience analysis(npSRA)framework,including nonparametric copula-based sensitivity analysis,two-sample resilience test analysis,and a novel tool for resilience attenuation analysis.To demonstrate the use of this framework,we focus on electric distribution systems,commonly found in many urban,suburban,and rural areas and vulnerable to tropical storms.A novel procedure for considering resourcefulness parameters in the socioeconomic space is proposed.Numerical results reveal the complex sta-tistical relations between the distributions of system resilience,physical aging,and socioeconomic parameters for the power distribution system.The proposed resilience distance computing and resilience attenuation anal-ysis further suggests two proper nonparametric distance metrics,the Earth Moving Distance(EMD)metric and the Cramévon Mises(CVM)metric,for characterizing the migration of system resilience for electric distribution systems.展开更多
It is well known that the nonparametric estimation of the regression function is highly sensitive to the presence of even a small proportion of outliers in the data.To solve the problem of typical observations when th...It is well known that the nonparametric estimation of the regression function is highly sensitive to the presence of even a small proportion of outliers in the data.To solve the problem of typical observations when the covariates of the nonparametric component are functional,the robust estimates for the regression parameter and regression operator are introduced.The main propose of the paper is to consider data-driven methods of selecting the number of neighbors in order to make the proposed processes fully automatic.We use thek Nearest Neighbors procedure(kNN)to construct the kernel estimator of the proposed robust model.Under some regularity conditions,we state consistency results for kNN functional estimators,which are uniform in the number of neighbors(UINN).Furthermore,a simulation study and an empirical application to a real data analysis of octane gasoline predictions are carried out to illustrate the higher predictive performances and the usefulness of the kNN approach.展开更多
We explore two observable nonclassical properties of quantum states generated by repeatedly operating annihilationthen-creation(AC) and creation-then-annihilation(CA) on the coherent state, respectively, such as h...We explore two observable nonclassical properties of quantum states generated by repeatedly operating annihilationthen-creation(AC) and creation-then-annihilation(CA) on the coherent state, respectively, such as higher-order subPoissonian statistics and higher-order squeezing-enhanced effect. The corresponding analytical expressions are derived in detail depending on m. By numerically comparing those quantum properties, it is found that these states above have very different nonclassical properties and nonclassicality is exhibited more strongly after AC operation than after CA operation.展开更多
Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Kn...Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.展开更多
Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Kn...Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.展开更多
Normality testing is a fundamental hypothesis test in the statistical analysis of key biological indicators of diabetes.If this assumption is violated,it may cause the test results to deviate from the true value,leadi...Normality testing is a fundamental hypothesis test in the statistical analysis of key biological indicators of diabetes.If this assumption is violated,it may cause the test results to deviate from the true value,leading to incorrect inferences and conclusions,and ultimately affecting the validity and accuracy of statistical inferences.Considering this,the study designs a unified analysis scheme for different data types based on parametric statistical test methods and non-parametric test methods.The data were grouped according to sample type and divided into discrete data and continuous data.To account for differences among subgroups,the conventional chi-squared test was used for discrete data.The normal distribution is the basis of many statistical methods;if the data does not follow a normal distribution,many statistical methods will fail or produce incorrect results.Therefore,before data analysis and modeling,the data were divided into normal and non-normal groups through normality testing.For normally distributed data,parametric statistical methods were used to judge the differences between groups.For non-normal data,non-parametric tests were employed to improve the accuracy of the analysis.Statistically significant indicators were retained according to the significance index P-value of the statistical test or corresponding statistics.These indicators were then combined with relevant medical background to further explore the etiology leading to the occurrence or transformation of diabetes status.展开更多
Profile monitoring is used to check the stability of the quality of a product over time when the product quality is best represented by a function at each time point.However,most previous monitoring approaches have no...Profile monitoring is used to check the stability of the quality of a product over time when the product quality is best represented by a function at each time point.However,most previous monitoring approaches have not considered that the argument values may vary from profile to profile,which is common in practice.A novel nonparametric control scheme based on profile error is proposed for monitoring nonlinear profiles with varied argument values.The proposed scheme uses the metrics of profile error as the statistics to construct the control charts.More details about the design of this nonparametric scheme are also discussed.The monitoring performance of the combined control scheme is compared with that of alternative nonparametric methods via simulation.Simulation studies show that the combined scheme is effective in detecting parameter error and is sensitive to small shifts in the process.In addition,due to the properties of the charting statistics,the out-of-control signal can provide diagnostic information for the users.Finally,the implementation steps of the proposed monitoring scheme are given and applied for monitoring the blade manufacturing process.With the application in blade manufacturing of aircraft engines,the proposed nonparametric control scheme is effective,interpretable,and easy to apply.展开更多
For magnetotelluric sounding (MT), many processing methods based on power spectrum have put forward lots of hypotheses, such as MT signals are Gaussian, linear and minimum-phase. If practical signals do not satisfy th...For magnetotelluric sounding (MT), many processing methods based on power spectrum have put forward lots of hypotheses, such as MT signals are Gaussian, linear and minimum-phase. If practical signals do not satisfy these requirements, the results will have a few problems as follows. Firstly, when signals are non-linear and non-Gaussian, the information of the earth contained in the MT signals cannot be sufficiently extracted; Secondly, when signals are non-Gaussian and non-minimum phase, the processed results cannot reflect the minimum phase characteristics of the signals. Hence, it is necessary for us to do further research on characteristics of MT signals (YAO, SUN, 1999; LI, CHENG, 2002; Nikias, Petropulu, 1993; ZHANG, 1996). Otherwise, we cannot judge the reliability of the processed results based on power spectrum.……展开更多
Higher-order almost cyclostationary complex processes are complex random signals with almost periodically time-varying statistics, which is important to the research of non-Gaussian signals in information system. In t...Higher-order almost cyclostationary complex processes are complex random signals with almost periodically time-varying statistics, which is important to the research of non-Gaussian signals in information system. In tins paper, smoothed polyperiodograms are proposed for related to cyclic polyspectral estimation and are shown to be consistent and asymptotically complex normal. Asymptotic covariance expressions are derived along with their computable forms.展开更多
Estimation of the bivariate survival function under the competing risks caseis considered.We give an explicit formula for the estimator from a decomposition of thebivariate survival function based on competing risks,w...Estimation of the bivariate survival function under the competing risks caseis considered.We give an explicit formula for the estimator from a decomposition of thebivariate survival function based on competing risks,which is almost sure consistent.展开更多
The high penetration of distributed generation(DG)has set up a challenge for energy management and consequently for the monitoring and assessment of power quality(PQ).Besides,there are new types of disturbances owing ...The high penetration of distributed generation(DG)has set up a challenge for energy management and consequently for the monitoring and assessment of power quality(PQ).Besides,there are new types of disturbances owing to the uncontrolled connections of non-linear loads.The stochastic behaviour triggers the need for new holistic indicators which also deal with big data of PQ in terms of compression and scalability so as to extract the useful information regarding different network states and the prevailing PQ disturbances for future risk assessment and energy management systems.Permanent and continuous monitoring would guarantee the report to claim for damages and to assess the risk of PQ distortions.In this context,we propose a measurement method that postulates the use of two-dimensional(2D)diagrams based on higher-order statistics(HOSs)and a previous voltage quality index that assesses the voltage supply waveform in a continous monitoring campaign.Being suitable for both PQ and reliability applications,the results conclude that the inclusion of HOS measurements in the industrial metrological reports helps characterize the deviations of the voltage supply waveform,extracting the individual customers’pattern fingerprint,and compressing the data from both time and spatial aspects.The method allows a continuous and robust performance needed in the SG framework.Consequently,the method can be used by an average consumer as a probabilistic method to assess the risk of PQ deviations in site characterization.展开更多
In recent years, sub-synchronous oscillation accidents caused by wind power integration have received extensive attention. The recorded constant-amplitude waveforms can be induced by either linear or nonlinear oscilla...In recent years, sub-synchronous oscillation accidents caused by wind power integration have received extensive attention. The recorded constant-amplitude waveforms can be induced by either linear or nonlinear oscillation mechanisms. Hence, the nonlinear behavior needs to be distinguished prior to choosing the analysis method. Since the 1960s, the higher-order statistics(HOS) theory has become a powerful tool for the detection of nonlinear behavior(DNB) in production quality control wherein it has mainly been applied to mechanical condition monitoring and fault diagnosis. This study focuses on the hard limiters of the voltage source converter(VSC) control systems in the wind farms and attempts to detect the nonlinear behavior caused by bi-or uni-lateral saturation hard limiting using the HOS analysis. First, the conventional describing function is extended to obtain the detailed frequency domain information on the bi-and uni-lateral saturation hard limiting. Furthermore, the bi-and tri-spectra are introduced as the HOS, which are extended into bi-and tri-coherence spectra to eliminate the effects of the linear parts on the harmonic characteristics of hard limiting in the VSC control system, respectively. The effectiveness of the HOS in the DNB and the classification of the hard-limiting types is proven, and its detailed derivation and estimation procedure is presented. Finally, the quadratic and cubic phase coupling in the signals is illustrated, and the performance of the proposed method is evaluated and discussed.展开更多
The new signature of liquid-gas phase transition has been well indicated by the higher-order fluctuations of the largest fragment charge,but the uncertainties of critical temperatures based on this signature have not ...The new signature of liquid-gas phase transition has been well indicated by the higher-order fluctuations of the largest fragment charge,but the uncertainties of critical temperatures based on this signature have not been revealed.This study extracts the critical temperatures of liquid-gas phase transition in nuclear reactions and investigates their uncertainties.Utilizing the isospin-dependent quantum molecular dynamics model in conjunction with the statistical model GEMINI enables us to describe the dynamical path from the initial to the final state.An isotope thermometer and a quantum fluctuation thermometer are employed to extract the nuclear temperature.The higher-order fluctuations of the largest fragment charge and critical temperatures are studied in^(124)Sn+^(120)Sn collisions ranging from 400 to 1000 MeV/nucleon and^(124)Sn+AZ collisions at 600 MeV/nucleon.Observations revealed that the pseudo-critical point is robustly indicated by the higher-order fluctuations of the largest fragment charge.The critical temperatures extracted by the isotope thermometer are relatively consistent,with an uncertainty of 15%,while those obtained by the quantum fluctuation thermometer are heavily influenced by the incident energy and mass number of target nuclei.The excitation energy E∗and bound charge Zbound are used for event-sorting.These two ensembles represent the statistical properties of the initial and final states of the system,respectively.The initial-final correlations of statistical properties might lead to two phenomena.First,the size distribution of the largest fragment at the pseudo-critical point based on the Zbound ensemble is wide,while that based on E∗ensemble exhibits bimodality,which is a typical characteristic in the liquid-gas coexistence of a finite system.Second,the temperature at the pseudo-critical point based on the Zbound ensemble is higher than that based on the E∗ensemble.Furthermore,the projectile-like system exhibits a significant dynamical effect in its evolution path from the initial to final state,closely associated with the fluctuation of critical temperature.展开更多
基金Sponsored by the National Natural Science Foundation of China(Grant No.60475016)the Foundational Research Fund of Harbin Engineering University (Grant No.HEUF04092)
文摘To capture the presence of speech embedded in nonspeech events and background noise in shortwave non-cooperative communication, an algorithm for speech-stream detection in noisy environments is presented based on Empirical Mode Decomposition (EMD) and statistical properties of higher-order cumulants of speech signals. With the EMD, the noise signals can be decomposed into different numbers of IMFs. Then, the fourth-order cumulant ( FOC ) can be used to extract the desired feature of statistical properties for IMF components. Since the higher-order eumulants are blind for Gaussian signals, the proposed method is especially effective regarding the problem of speech-stream detection, where the speech signal is distorted by Gaussian noise. With the self-adaptive decomposition by EMD, the proposed method can also work well for non-Gaussian noise. The experiments show that the proposed algorithm can suppress different noise types with different SNRs, and the algorithm is robust in real signal tests.
文摘Statistics of order 2 (variance, auto and cross-correlation functions, auto and cross-power spectra) and 3 (skewness, auto and cross-bicorrelation functions, auto and cross-bispectra) are used to analyze the wave-particle interaction in space plasmas. The signals considered here are medium scale electron density irregularities and ELF/ULF electrostatic turbulence. Nonlinearities are mainly observed in the ELF range. They are independently pointed out in time series associated with fluctuations in electronic density and in time series associated with the measurement of one electric field component. Peaks in cross-bicorrelation function and in mutual information clearly show that, in well delimited frequency bands, the wave-particle interactions are nonlinear above a certain level of fluctuations. The way the energy is transferred within the frequencies of density fluctuations is indicated by a bi-spectra analysis.
文摘Healthcare decisions are based on scientific evidence obtained from medical studies by gathering data and analyzing it to obtain the best results. When analyzing data, biostatistics is a powerful tool, but healthcare professionals lack knowledge in this field. This lack of knowledge can manifest itself in situations such as choosing the wrong statistical test for the right situation or applying a statistical test without checking its assumptions, leading to inaccurate results and misleading conclusions. With the help of this “narrative review”, the aim is to bring biostatistics closer to healthcare professionals by answering certain questions: how to describe the distribution of data? how to assess the normality of data? how to transform data? and how to choose between nonparametric and parametric tests? Through this work, our hope is that the reader will be able to choose the right test for the right situation, in order to obtain the most accurate results.
文摘The development of defect prediction plays a significant role in improving software quality. Such predictions are used to identify defective modules before the testing and to minimize the time and cost. The software with defects negatively impacts operational costs and finally affects customer satisfaction. Numerous approaches exist to predict software defects. However, the timely and accurate software bugs are the major challenging issues. To improve the timely and accurate software defect prediction, a novel technique called Nonparametric Statistical feature scaled QuAdratic regressive convolution Deep nEural Network (SQADEN) is introduced. The proposed SQADEN technique mainly includes two major processes namely metric or feature selection and classification. First, the SQADEN uses the nonparametric statistical Torgerson–Gower scaling technique for identifying the relevant software metrics by measuring the similarity using the dice coefficient. The feature selection process is used to minimize the time complexity of software fault prediction. With the selected metrics, software fault perdition with the help of the Quadratic Censored regressive convolution deep neural network-based classification. The deep learning classifier analyzes the training and testing samples using the contingency correlation coefficient. The softstep activation function is used to provide the final fault prediction results. To minimize the error, the Nelder–Mead method is applied to solve non-linear least-squares problems. Finally, accurate classification results with a minimum error are obtained at the output layer. Experimental evaluation is carried out with different quantitative metrics such as accuracy, precision, recall, F-measure, and time complexity. The analyzed results demonstrate the superior performance of our proposed SQADEN technique with maximum accuracy, sensitivity and specificity by 3%, 3%, 2% and 3% and minimum time and space by 13% and 15% when compared with the two state-of-the-art methods.
基金supported by the National Science Foundation(NSF)under Award Number IIA-1355406.
文摘This paper proposes a set of nonparametric statistical tools for analyzing the system resilience of civil structures and infrastructure and its migration upon changes in critical system parameters.The work is founded on the classic theoretic framework that system resilience is defined in multiple dimensions for a constructed system.Consequentially,system resilience can lose its parametric form as a random variable,falling into the realm of nonparametric statistics.With this nonparametric shift,traditional distribution-based statistics are ineffective in characterizing the migration of system resilience due to the variation of system parameters.Three statistical tools are proposed under the nonparametric statistical resilience analysis(npSRA)framework,including nonparametric copula-based sensitivity analysis,two-sample resilience test analysis,and a novel tool for resilience attenuation analysis.To demonstrate the use of this framework,we focus on electric distribution systems,commonly found in many urban,suburban,and rural areas and vulnerable to tropical storms.A novel procedure for considering resourcefulness parameters in the socioeconomic space is proposed.Numerical results reveal the complex sta-tistical relations between the distributions of system resilience,physical aging,and socioeconomic parameters for the power distribution system.The proposed resilience distance computing and resilience attenuation anal-ysis further suggests two proper nonparametric distance metrics,the Earth Moving Distance(EMD)metric and the Cramévon Mises(CVM)metric,for characterizing the migration of system resilience for electric distribution systems.
文摘It is well known that the nonparametric estimation of the regression function is highly sensitive to the presence of even a small proportion of outliers in the data.To solve the problem of typical observations when the covariates of the nonparametric component are functional,the robust estimates for the regression parameter and regression operator are introduced.The main propose of the paper is to consider data-driven methods of selecting the number of neighbors in order to make the proposed processes fully automatic.We use thek Nearest Neighbors procedure(kNN)to construct the kernel estimator of the proposed robust model.Under some regularity conditions,we state consistency results for kNN functional estimators,which are uniform in the number of neighbors(UINN).Furthermore,a simulation study and an empirical application to a real data analysis of octane gasoline predictions are carried out to illustrate the higher predictive performances and the usefulness of the kNN approach.
基金Project supported by the National Natural Science Foundation of China(Grant Nos.11447002 and 11447202)the Natural Science Foundation of Jiangxi Province of China(Grant No.20151BAB202013)the Research Foundation for Changzhou Institute of Modern Optoelectronic Technology of China(Grant No.CZGY15)
文摘We explore two observable nonclassical properties of quantum states generated by repeatedly operating annihilationthen-creation(AC) and creation-then-annihilation(CA) on the coherent state, respectively, such as higher-order subPoissonian statistics and higher-order squeezing-enhanced effect. The corresponding analytical expressions are derived in detail depending on m. By numerically comparing those quantum properties, it is found that these states above have very different nonclassical properties and nonclassicality is exhibited more strongly after AC operation than after CA operation.
文摘Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.
文摘Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.
基金National Natural Science Foundation of China(No.12271261)Postgraduate Research and Practice Innovation Program of Jiangsu Province,China(Grant No.SJCX230368)。
文摘Normality testing is a fundamental hypothesis test in the statistical analysis of key biological indicators of diabetes.If this assumption is violated,it may cause the test results to deviate from the true value,leading to incorrect inferences and conclusions,and ultimately affecting the validity and accuracy of statistical inferences.Considering this,the study designs a unified analysis scheme for different data types based on parametric statistical test methods and non-parametric test methods.The data were grouped according to sample type and divided into discrete data and continuous data.To account for differences among subgroups,the conventional chi-squared test was used for discrete data.The normal distribution is the basis of many statistical methods;if the data does not follow a normal distribution,many statistical methods will fail or produce incorrect results.Therefore,before data analysis and modeling,the data were divided into normal and non-normal groups through normality testing.For normally distributed data,parametric statistical methods were used to judge the differences between groups.For non-normal data,non-parametric tests were employed to improve the accuracy of the analysis.Statistically significant indicators were retained according to the significance index P-value of the statistical test or corresponding statistics.These indicators were then combined with relevant medical background to further explore the etiology leading to the occurrence or transformation of diabetes status.
基金supported by National Natural Science Foundation of China (Grant No. 70931004,Grant No. 70802043)
文摘Profile monitoring is used to check the stability of the quality of a product over time when the product quality is best represented by a function at each time point.However,most previous monitoring approaches have not considered that the argument values may vary from profile to profile,which is common in practice.A novel nonparametric control scheme based on profile error is proposed for monitoring nonlinear profiles with varied argument values.The proposed scheme uses the metrics of profile error as the statistics to construct the control charts.More details about the design of this nonparametric scheme are also discussed.The monitoring performance of the combined control scheme is compared with that of alternative nonparametric methods via simulation.Simulation studies show that the combined scheme is effective in detecting parameter error and is sensitive to small shifts in the process.In addition,due to the properties of the charting statistics,the out-of-control signal can provide diagnostic information for the users.Finally,the implementation steps of the proposed monitoring scheme are given and applied for monitoring the blade manufacturing process.With the application in blade manufacturing of aircraft engines,the proposed nonparametric control scheme is effective,interpretable,and easy to apply.
基金National Natural Science Foundation of China (40274039).
文摘For magnetotelluric sounding (MT), many processing methods based on power spectrum have put forward lots of hypotheses, such as MT signals are Gaussian, linear and minimum-phase. If practical signals do not satisfy these requirements, the results will have a few problems as follows. Firstly, when signals are non-linear and non-Gaussian, the information of the earth contained in the MT signals cannot be sufficiently extracted; Secondly, when signals are non-Gaussian and non-minimum phase, the processed results cannot reflect the minimum phase characteristics of the signals. Hence, it is necessary for us to do further research on characteristics of MT signals (YAO, SUN, 1999; LI, CHENG, 2002; Nikias, Petropulu, 1993; ZHANG, 1996). Otherwise, we cannot judge the reliability of the processed results based on power spectrum.……
文摘Higher-order almost cyclostationary complex processes are complex random signals with almost periodically time-varying statistics, which is important to the research of non-Gaussian signals in information system. In tins paper, smoothed polyperiodograms are proposed for related to cyclic polyspectral estimation and are shown to be consistent and asymptotically complex normal. Asymptotic covariance expressions are derived along with their computable forms.
文摘Estimation of the bivariate survival function under the competing risks caseis considered.We give an explicit formula for the estimator from a decomposition of thebivariate survival function based on competing risks,which is almost sure consistent.
基金This work was supported by the Spanish Ministry of Science and Innovation(Statal Agency for Research),the EU(AEI/FEDER/UE)via project PID2019-108953RB-C21 Strategies for Aggregated Generation of Photovoltaic Plants:Energy and Meteorological Operational Data(SAGPVEMOD),the precedent TEC2016-77632-C3-3-R.
文摘The high penetration of distributed generation(DG)has set up a challenge for energy management and consequently for the monitoring and assessment of power quality(PQ).Besides,there are new types of disturbances owing to the uncontrolled connections of non-linear loads.The stochastic behaviour triggers the need for new holistic indicators which also deal with big data of PQ in terms of compression and scalability so as to extract the useful information regarding different network states and the prevailing PQ disturbances for future risk assessment and energy management systems.Permanent and continuous monitoring would guarantee the report to claim for damages and to assess the risk of PQ distortions.In this context,we propose a measurement method that postulates the use of two-dimensional(2D)diagrams based on higher-order statistics(HOSs)and a previous voltage quality index that assesses the voltage supply waveform in a continous monitoring campaign.Being suitable for both PQ and reliability applications,the results conclude that the inclusion of HOS measurements in the industrial metrological reports helps characterize the deviations of the voltage supply waveform,extracting the individual customers’pattern fingerprint,and compressing the data from both time and spatial aspects.The method allows a continuous and robust performance needed in the SG framework.Consequently,the method can be used by an average consumer as a probabilistic method to assess the risk of PQ deviations in site characterization.
基金supported by the State Grid Guide Project(No.5108-202218030A-1-1-ZN)。
文摘In recent years, sub-synchronous oscillation accidents caused by wind power integration have received extensive attention. The recorded constant-amplitude waveforms can be induced by either linear or nonlinear oscillation mechanisms. Hence, the nonlinear behavior needs to be distinguished prior to choosing the analysis method. Since the 1960s, the higher-order statistics(HOS) theory has become a powerful tool for the detection of nonlinear behavior(DNB) in production quality control wherein it has mainly been applied to mechanical condition monitoring and fault diagnosis. This study focuses on the hard limiters of the voltage source converter(VSC) control systems in the wind farms and attempts to detect the nonlinear behavior caused by bi-or uni-lateral saturation hard limiting using the HOS analysis. First, the conventional describing function is extended to obtain the detailed frequency domain information on the bi-and uni-lateral saturation hard limiting. Furthermore, the bi-and tri-spectra are introduced as the HOS, which are extended into bi-and tri-coherence spectra to eliminate the effects of the linear parts on the harmonic characteristics of hard limiting in the VSC control system, respectively. The effectiveness of the HOS in the DNB and the classification of the hard-limiting types is proven, and its detailed derivation and estimation procedure is presented. Finally, the quadratic and cubic phase coupling in the signals is illustrated, and the performance of the proposed method is evaluated and discussed.
基金Supported by the National Natural Science Foundation of China (11875328, 12075327)the Key Laboratory of Nuclear Data foundation (JCKY2022201C157)the Fundamental Research Funds for the Central Universities, Sun Yat-sen University (22lgqb39)。
文摘The new signature of liquid-gas phase transition has been well indicated by the higher-order fluctuations of the largest fragment charge,but the uncertainties of critical temperatures based on this signature have not been revealed.This study extracts the critical temperatures of liquid-gas phase transition in nuclear reactions and investigates their uncertainties.Utilizing the isospin-dependent quantum molecular dynamics model in conjunction with the statistical model GEMINI enables us to describe the dynamical path from the initial to the final state.An isotope thermometer and a quantum fluctuation thermometer are employed to extract the nuclear temperature.The higher-order fluctuations of the largest fragment charge and critical temperatures are studied in^(124)Sn+^(120)Sn collisions ranging from 400 to 1000 MeV/nucleon and^(124)Sn+AZ collisions at 600 MeV/nucleon.Observations revealed that the pseudo-critical point is robustly indicated by the higher-order fluctuations of the largest fragment charge.The critical temperatures extracted by the isotope thermometer are relatively consistent,with an uncertainty of 15%,while those obtained by the quantum fluctuation thermometer are heavily influenced by the incident energy and mass number of target nuclei.The excitation energy E∗and bound charge Zbound are used for event-sorting.These two ensembles represent the statistical properties of the initial and final states of the system,respectively.The initial-final correlations of statistical properties might lead to two phenomena.First,the size distribution of the largest fragment at the pseudo-critical point based on the Zbound ensemble is wide,while that based on E∗ensemble exhibits bimodality,which is a typical characteristic in the liquid-gas coexistence of a finite system.Second,the temperature at the pseudo-critical point based on the Zbound ensemble is higher than that based on the E∗ensemble.Furthermore,the projectile-like system exhibits a significant dynamical effect in its evolution path from the initial to final state,closely associated with the fluctuation of critical temperature.