As the increasing popularity and complexity of Web applications and the emergence of their new characteristics, the testing and maintenance of large, complex Web applications are becoming more complex and difficult. W...As the increasing popularity and complexity of Web applications and the emergence of their new characteristics, the testing and maintenance of large, complex Web applications are becoming more complex and difficult. Web applications generally contain lots of pages and are used by enormous users. Statistical testing is an effective way of ensuring their quality. Web usage can be accurately described by Markov chain which has been proved to be an ideal model for software statistical testing. The results of unit testing can be utilized in the latter stages, which is an important strategy for bottom-to-top integration testing, and the other improvement of extended Markov chain model (EMM) is to present the error type vector which is treated as a part of page node. this paper also proposes the algorithm for generating test cases of usage paths. Finally, optional usage reliability evaluation methods and an incremental usability regression testing model for testing and evaluation are presented. Key words statistical testing - evaluation for Web usability - extended Markov chain model (EMM) - Web log mining - reliability evaluation CLC number TP311. 5 Foundation item: Supported by the National Defence Research Project (No. 41315. 9. 2) and National Science and Technology Plan (2001BA102A04-02-03)Biography: MAO Cheng-ying (1978-), male, Ph.D. candidate, research direction: software testing. Research direction: advanced database system, software testing, component technology and data mining.展开更多
In this article, the unit root test for AR(p) model with GARCH errors is considered. The Dickey-Fuller test statistics are rewritten in the form of self-normalized sums, and the asymptotic distribution of the test s...In this article, the unit root test for AR(p) model with GARCH errors is considered. The Dickey-Fuller test statistics are rewritten in the form of self-normalized sums, and the asymptotic distribution of the test statistics is derived under the weak conditions.展开更多
Virtual testability demonstration test has many advantages,such as low cost,high efficiency,low risk and few restrictions.It brings new requirements to the fault sample generation.A fault sample simulation approach fo...Virtual testability demonstration test has many advantages,such as low cost,high efficiency,low risk and few restrictions.It brings new requirements to the fault sample generation.A fault sample simulation approach for virtual testability demonstration test based on stochastic process theory is proposed.First,the similarities and differences of fault sample generation between physical testability demonstration test and virtual testability demonstration test are discussed.Second,it is pointed out that the fault occurrence process subject to perfect repair is renewal process.Third,the interarrival time distribution function of the next fault event is given.Steps and flowcharts of fault sample generation are introduced.The number of faults and their occurrence time are obtained by statistical simulation.Finally,experiments are carried out on a stable tracking platform.Because a variety of types of life distributions and maintenance modes are considered and some assumptions are removed,the sample size and structure of fault sample simulation results are more similar to the actual results and more reasonable.The proposed method can effectively guide the fault injection in virtual testability demonstration test.展开更多
This paper describes the statistical methods of the comparison of the incidence or mortality rates in cancer registry and descriptive epidemiology, and the features of microcomputer program (CANTEST) which was designe...This paper describes the statistical methods of the comparison of the incidence or mortality rates in cancer registry and descriptive epidemiology, and the features of microcomputer program (CANTEST) which was designed to perform the methods. The program was written in IBM BASIC language. Using the program CANTEST we presented here the user can do several statistical tests or estimations as follow: 1. the comparison of the adjusted rates which were calculated by directly or indirectly standardized methods, 2. the calculation of the slope of regression line for testing the linear trends of the adjusted rates, 3. the estimation of the 95% or 99%conndence intervals of the directly adjusted rates, of the cumulative rates (0-64 and 0-74), and of the cumulative risk. Several examples are presented for testing the performances of the program.展开更多
We are very grateful for the letter written by Dr Lange,and indeed apologize for the mistakes noted in the word-ing of our text regarding statistical analysis.This wasdue to changes carried out while revising the manu...We are very grateful for the letter written by Dr Lange,and indeed apologize for the mistakes noted in the word-ing of our text regarding statistical analysis.This wasdue to changes carried out while revising the manuscriptat the request of reviewers,whom we thank for,point-ing out several issues that were actually similar to thosenoted by Dr.Lange.Unfortunately,we were unable todescribe and discuss our findings properly in the context展开更多
Search-based statistical structural testing(SBSST)is a promising technique that uses automated search to construct input distributions for statistical structural testing.It has been proved that a simple search algorit...Search-based statistical structural testing(SBSST)is a promising technique that uses automated search to construct input distributions for statistical structural testing.It has been proved that a simple search algorithm,for example,the hill-climber is able to optimize an input distribution.However,due to the noisy fitness estimation of the minimum triggering probability among all cover elements(Tri-Low-Bound),the existing approach does not show a satisfactory efficiency.Constructing input distributions to satisfy the Tri-Low-Bound criterion requires an extensive computation time.Tri-Low-Bound is considered a strong criterion,and it is demonstrated to sustain a high fault-detecting ability.This article tries to answer the following question:if we use a relaxed constraint that significantly reduces the time consumption on search,can the optimized input distribution still be effective in faultdetecting ability?In this article,we propose a type of criterion called fairnessenhanced-sum-of-triggering-probability(p-L1-Max).The criterion utilizes the sum of triggering probabilities as the fitness value and leverages a parameter p to adjust the uniformness of test data generation.We conducted extensive experiments to compare the computation time and the fault-detecting ability between the two criteria.The result shows that the 1.0-L1-Max criterion has the highest efficiency,and it is more practical to use than the Tri-Low-Bound criterion.To measure a criterion’s fault-detecting ability,we introduce a definition of expected faults found in the effective test set size region.To measure the effective test set size region,we present a theoretical analysis of the expected faults found with respect to various test set sizes and use the uniform distribution as a baseline to derive the effective test set size region’s definition.展开更多
We investigate redshift distributions of three long burst samples, with the first sample containing 131 long bursts with observed redshifts, the second including 220 long bursts with pseudo-redshifts calculated by the...We investigate redshift distributions of three long burst samples, with the first sample containing 131 long bursts with observed redshifts, the second including 220 long bursts with pseudo-redshifts calculated by the variability-luminosity relation, and the third including 1194 long bursts with pseudo-redshifls calculated by the lag-luminosity relation, respectively. In the redshift range 0-1 the Kolmogorov-Smirnov probability of the observed redshift distribution and that of the variability-luminosity relation is large. In the redshift ranges 1-2, 2-3, 3-6.3 and 0-37, the Kolmogorov-Smirnov probabilities of the redshift distribution from lag-luminosity relation and the observed redshift distribution are also large. For the GRBs, which appear both in the two pseudo-redshift burst samples, the KS probability of the pseudo-redshift distribution from the lag-luminosity relation and the observed reshift distribution is 0.447, which is very large. Based on these results, some conclusions are drawn: i) the V-Liso relation might be more believable than the τ-Liso relation in low redshift ranges and the τ-Liso relation might be more real than the V-Liso relation in high redshift ranges; ii) if we do not consider the redshift ranges, the τ-Liso relation might be more physical and intrinsical than the V-Liso relation.展开更多
Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Kn...Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.展开更多
Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Kn...Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.展开更多
Boundary recognition is an important research of natural language processing, and it provides a basis for the application of Chinese word segmentation, chunk analysis, named entity recognition, etc. Based on ambiguity...Boundary recognition is an important research of natural language processing, and it provides a basis for the application of Chinese word segmentation, chunk analysis, named entity recognition, etc. Based on ambiguity in boundary recognition of Chinese punctuation marks, this paper proposes grammar testing methods for boundary recognition of slight-pause marks and then calculates the annotation consistency of these methods. The statistical results show that grammar testing methods can greatly improve the annotation consistency of slight-pause marks boundary recognition. The consistency during the second time is 0.030 3 higher than during the first, which will help guarantee the consistency of large-scale corpus annotation and improve the quality of corpus annotation.展开更多
Normality testing is a fundamental hypothesis test in the statistical analysis of key biological indicators of diabetes.If this assumption is violated,it may cause the test results to deviate from the true value,leadi...Normality testing is a fundamental hypothesis test in the statistical analysis of key biological indicators of diabetes.If this assumption is violated,it may cause the test results to deviate from the true value,leading to incorrect inferences and conclusions,and ultimately affecting the validity and accuracy of statistical inferences.Considering this,the study designs a unified analysis scheme for different data types based on parametric statistical test methods and non-parametric test methods.The data were grouped according to sample type and divided into discrete data and continuous data.To account for differences among subgroups,the conventional chi-squared test was used for discrete data.The normal distribution is the basis of many statistical methods;if the data does not follow a normal distribution,many statistical methods will fail or produce incorrect results.Therefore,before data analysis and modeling,the data were divided into normal and non-normal groups through normality testing.For normally distributed data,parametric statistical methods were used to judge the differences between groups.For non-normal data,non-parametric tests were employed to improve the accuracy of the analysis.Statistically significant indicators were retained according to the significance index P-value of the statistical test or corresponding statistics.These indicators were then combined with relevant medical background to further explore the etiology leading to the occurrence or transformation of diabetes status.展开更多
Many fields,such as neuroscience,are experiencing the vast prolife ration of cellular data,underscoring the need fo r organizing and interpreting large datasets.A popular approach partitions data into manageable subse...Many fields,such as neuroscience,are experiencing the vast prolife ration of cellular data,underscoring the need fo r organizing and interpreting large datasets.A popular approach partitions data into manageable subsets via hierarchical clustering,but objective methods to determine the appropriate classification granularity are missing.We recently introduced a technique to systematically identify when to stop subdividing clusters based on the fundamental principle that cells must differ more between than within clusters.Here we present the corresponding protocol to classify cellular datasets by combining datadriven unsupervised hierarchical clustering with statistical testing.These general-purpose functions are applicable to any cellular dataset that can be organized as two-dimensional matrices of numerical values,including molecula r,physiological,and anatomical datasets.We demonstrate the protocol using cellular data from the Janelia MouseLight project to chara cterize morphological aspects of neurons.展开更多
This study investigates the relationships between agricultural spot markets and external uncertainties through multifractal detrending moving-average cross-correlation analysis(MF-X-DMA).The dataset contains the Grain...This study investigates the relationships between agricultural spot markets and external uncertainties through multifractal detrending moving-average cross-correlation analysis(MF-X-DMA).The dataset contains the Grains&Oilseeds Index(GOI)and its five subindices for wheat,maize,soyabeans,rice,and barley.Moreover,we use three uncertainty proxies,namely,economic policy uncertainty(EPU),geopolitical risk(GPR),and Volatility Index(VIX).We observe multifractal cross-correlations between agricultural markets and uncertainties.Furthermore,statistical tests reveal that maize has intrinsic joint multifractality with all the uncertainty proxies,highly sensitive to external shocks.Additionally,intrinsic multifractality among GOI-GPR,wheat-GPR,and soyabeans-VIX is illustrated.However,other series have apparent multifractal crosscorrelations with high probabilities.Moreover,our analysis suggests that among the three types of external uncertainties,GPR has the strongest association with grain prices,excluding maize and soyabeans.展开更多
In this paper,based on the observation data of air temperature during 1951-2009 in Shenyang,the interannual and interdecadal variation of annual average temperature,maximum and minimum temperature in Shenyang were con...In this paper,based on the observation data of air temperature during 1951-2009 in Shenyang,the interannual and interdecadal variation of annual average temperature,maximum and minimum temperature in Shenyang were conducted the statistical analysis by means of linear trend estimation and mutation detection by using Mann-Kendall method.As was demonstrated in the results,the annual average temperature,maximum and minimum temperature in Shenyang showed an upward trend,whose linear tendency rate was 0.231,0.181 and 0.218 respectively.The increment trend of annual average temperature,maximum and minimum temperature was extremely clear.The increase in minimum temperature was more significant than that in mean temperature and maximum temperature.The abrupt change point of annual mean temperature in Shenyang appeared in 1981;the abrupt change point of annual mean maximum temperature appeared in 1994;the annual mean minimum temperature underwent mutation in 1978.展开更多
Receiver Autonomous Integrity Monitoring (RAIM) is a software algorithm available in some GPS receivers which gives an indication if the position solution given by the GPS receiver is suitable to use. The detail alg...Receiver Autonomous Integrity Monitoring (RAIM) is a software algorithm available in some GPS receivers which gives an indication if the position solution given by the GPS receiver is suitable to use. The detail algorithm of the parity space method of RAIM technique is presented. Using FDI and FDE methods, the simulations of RAIM performance have been done in three different phases independently with respect to the bias of the fault satellite. Case study of simulation results is discussed and each performance of RAIM is analyzed. According to the analysis of simulation results, the parity space method of RAIM can meet the integrity requirements for nonprecision, terminal and enroute flight phase. It also indicates that the results of performance of FDE are better than that of FDI.展开更多
This paper discusses the nonlinearity of fish acoustic signals by using the surrogate data method. We compare the difference of three test statistics - time-irreversibility Trey, correlation dimension D2 and auto mutu...This paper discusses the nonlinearity of fish acoustic signals by using the surrogate data method. We compare the difference of three test statistics - time-irreversibility Trey, correlation dimension D2 and auto mutual information function I between the original data and the surrogate data. We come to the conclusion that there exists nonlinearity in the fish acoustic signals and there exist deterministic nonlinear components; therefore nonlinear dynamic theory can be used to analyze fish acoustic signals.展开更多
“Human-elephant conflict(HEC)”,the alarming issue,in present day context has attracted the attention of environmentalists and policy makers.The rising conflict between human beings and wild elephants is common in Bu...“Human-elephant conflict(HEC)”,the alarming issue,in present day context has attracted the attention of environmentalists and policy makers.The rising conflict between human beings and wild elephants is common in Buxa Tiger Reserve(BTR)and its adjoining area in West Bengal State,India,making the area volatile.People’s attitudes towards elephant conservation activity are very crucial to get rid of HEC,because people’s proximity with wild elephants’habitat can trigger the occurrence of HEC.The aim of this study is to conduct an in-depth investigation about the association of people’s attitudes towards HEC with their locational,demographic,and socio-economic characteristics in BTR and its adjoining area by using Pearson’s bivariate chi-square test and binary logistic regression analysis.BTR is one of the constituent parts of Eastern Doors Elephant Reserve(EDER).We interviewed 500 respondents to understand their perceptions to HEC and investigated their locational,demographic,and socio-economic characteristics including location of village,gender,age,ethnicity,religion,caste,poverty level,education level,primary occupation,secondary occupation,household type,and source of firewood.The results indicate that respondents who are living in enclave forest villages(EFVs),peripheral forest villages(PFVs),corridor village(CVs),or forest and corridor villages(FCVs),mainly males,at the age of 18–48 years old,engaged with agriculture occupation,and living in kancha and mixed houses,have more likelihood to witness HEC.Besides,respondents who are illiterate or at primary education level are more likely to regard elephant as a main problematic animal around their villages and refuse to participate in elephant conservation activity.For the sake of a sustainable environment for both human beings and wildlife,people’s attitudes towards elephants must be friendly in a more prudent way,so that the two communities can live in harmony.展开更多
Generalized method of moments based on probability generating function is considered. Estimation and model testing are unified using this approach which also leads to distribution free chi-square tests. The estimation...Generalized method of moments based on probability generating function is considered. Estimation and model testing are unified using this approach which also leads to distribution free chi-square tests. The estimation methods developed are also related to estimation methods based on generalized estimating equations but with the advantage of having statistics for model testing. The methods proposed overcome numerical problems often encountered when the probability mass functions have no closed forms which prevent the use of maximum likelihood (ML) procedures and in general, ML procedures do not lead to distribution free model testing statistics.展开更多
According to the need of project design for offshore engineering and coastal engineering, this paper statistically analyses the annual extreme data of waves acquired at 19 observation stations along the coast of China...According to the need of project design for offshore engineering and coastal engineering, this paper statistically analyses the annual extreme data of waves acquired at 19 observation stations along the coast of China. Five kinds of distribution curves are adopted: Pearson III (P-III), Log-Extreme I (LE), Log-Normal(LN), Weibull(W) and Exponential Γ(EΓ) to check the adaptability to the long-term distribution of annual extreme of wave in the China Sea areas. The New Curve Fitting Method (NFIT) method and Probability Weighted Moments (PWM) method are used to estimate the distribution parameters and thereby to derive the design wave parameters with different return periods at 19 observation stations. The test results show that by combining EΓ distribution and NFIT parameter estimation and optimum seeking by computer, the design wave parameters can be estimated with high accuracy, high speed and high efficiency, and the randomness of the estimated results can be avoided.展开更多
Finding a suitable solution to an optimization problem designed in science is a major challenge.Therefore,these must be addressed utilizing proper approaches.Based on a random search space,optimization algorithms can ...Finding a suitable solution to an optimization problem designed in science is a major challenge.Therefore,these must be addressed utilizing proper approaches.Based on a random search space,optimization algorithms can find acceptable solutions to problems.Archery Algorithm(AA)is a new stochastic approach for addressing optimization problems that is discussed in this study.The fundamental idea of developing the suggested AA is to imitate the archer’s shooting behavior toward the target panel.The proposed algorithm updates the location of each member of the population in each dimension of the search space by a member randomly marked by the archer.The AA is mathematically described,and its capacity to solve optimization problems is evaluated on twenty-three distinct types of objective functions.Furthermore,the proposed algorithm’s performance is compared vs.eight approaches,including teaching-learning based optimization,marine predators algorithm,genetic algorithm,grey wolf optimization,particle swarm optimization,whale optimization algorithm,gravitational search algorithm,and tunicate swarm algorithm.According to the simulation findings,the AA has a good capacity to tackle optimization issues in both unimodal and multimodal scenarios,and it can give adequate quasi-optimal solutions to these problems.The analysis and comparison of competing algorithms’performance with the proposed algorithm demonstrates the superiority and competitiveness of the AA.展开更多
文摘As the increasing popularity and complexity of Web applications and the emergence of their new characteristics, the testing and maintenance of large, complex Web applications are becoming more complex and difficult. Web applications generally contain lots of pages and are used by enormous users. Statistical testing is an effective way of ensuring their quality. Web usage can be accurately described by Markov chain which has been proved to be an ideal model for software statistical testing. The results of unit testing can be utilized in the latter stages, which is an important strategy for bottom-to-top integration testing, and the other improvement of extended Markov chain model (EMM) is to present the error type vector which is treated as a part of page node. this paper also proposes the algorithm for generating test cases of usage paths. Finally, optional usage reliability evaluation methods and an incremental usability regression testing model for testing and evaluation are presented. Key words statistical testing - evaluation for Web usability - extended Markov chain model (EMM) - Web log mining - reliability evaluation CLC number TP311. 5 Foundation item: Supported by the National Defence Research Project (No. 41315. 9. 2) and National Science and Technology Plan (2001BA102A04-02-03)Biography: MAO Cheng-ying (1978-), male, Ph.D. candidate, research direction: software testing. Research direction: advanced database system, software testing, component technology and data mining.
基金National Natural Science Foundation of China(1047112610671176).
文摘In this article, the unit root test for AR(p) model with GARCH errors is considered. The Dickey-Fuller test statistics are rewritten in the form of self-normalized sums, and the asymptotic distribution of the test statistics is derived under the weak conditions.
基金National Natural Science Foundation of China(51105369)
文摘Virtual testability demonstration test has many advantages,such as low cost,high efficiency,low risk and few restrictions.It brings new requirements to the fault sample generation.A fault sample simulation approach for virtual testability demonstration test based on stochastic process theory is proposed.First,the similarities and differences of fault sample generation between physical testability demonstration test and virtual testability demonstration test are discussed.Second,it is pointed out that the fault occurrence process subject to perfect repair is renewal process.Third,the interarrival time distribution function of the next fault event is given.Steps and flowcharts of fault sample generation are introduced.The number of faults and their occurrence time are obtained by statistical simulation.Finally,experiments are carried out on a stable tracking platform.Because a variety of types of life distributions and maintenance modes are considered and some assumptions are removed,the sample size and structure of fault sample simulation results are more similar to the actual results and more reasonable.The proposed method can effectively guide the fault injection in virtual testability demonstration test.
文摘This paper describes the statistical methods of the comparison of the incidence or mortality rates in cancer registry and descriptive epidemiology, and the features of microcomputer program (CANTEST) which was designed to perform the methods. The program was written in IBM BASIC language. Using the program CANTEST we presented here the user can do several statistical tests or estimations as follow: 1. the comparison of the adjusted rates which were calculated by directly or indirectly standardized methods, 2. the calculation of the slope of regression line for testing the linear trends of the adjusted rates, 3. the estimation of the 95% or 99%conndence intervals of the directly adjusted rates, of the cumulative rates (0-64 and 0-74), and of the cumulative risk. Several examples are presented for testing the performances of the program.
文摘We are very grateful for the letter written by Dr Lange,and indeed apologize for the mistakes noted in the word-ing of our text regarding statistical analysis.This wasdue to changes carried out while revising the manuscriptat the request of reviewers,whom we thank for,point-ing out several issues that were actually similar to thosenoted by Dr.Lange.Unfortunately,we were unable todescribe and discuss our findings properly in the context
基金Publication of this article in an open access journal was funded by the Portland State University Library’s Open Access Fund.
文摘Search-based statistical structural testing(SBSST)is a promising technique that uses automated search to construct input distributions for statistical structural testing.It has been proved that a simple search algorithm,for example,the hill-climber is able to optimize an input distribution.However,due to the noisy fitness estimation of the minimum triggering probability among all cover elements(Tri-Low-Bound),the existing approach does not show a satisfactory efficiency.Constructing input distributions to satisfy the Tri-Low-Bound criterion requires an extensive computation time.Tri-Low-Bound is considered a strong criterion,and it is demonstrated to sustain a high fault-detecting ability.This article tries to answer the following question:if we use a relaxed constraint that significantly reduces the time consumption on search,can the optimized input distribution still be effective in faultdetecting ability?In this article,we propose a type of criterion called fairnessenhanced-sum-of-triggering-probability(p-L1-Max).The criterion utilizes the sum of triggering probabilities as the fitness value and leverages a parameter p to adjust the uniformness of test data generation.We conducted extensive experiments to compare the computation time and the fault-detecting ability between the two criteria.The result shows that the 1.0-L1-Max criterion has the highest efficiency,and it is more practical to use than the Tri-Low-Bound criterion.To measure a criterion’s fault-detecting ability,we introduce a definition of expected faults found in the effective test set size region.To measure the effective test set size region,we present a theoretical analysis of the expected faults found with respect to various test set sizes and use the uniform distribution as a baseline to derive the effective test set size region’s definition.
基金supported by the National Natural Science Foundation of China(NSFC, No. 10473023)Scientific Research Fund of the Sichuan Provincial Education Department,the K. C. Wong Education Foundation (Hong Kong)the Jiangsu Planned Projects for PostdoctoralResearch Funds.
文摘We investigate redshift distributions of three long burst samples, with the first sample containing 131 long bursts with observed redshifts, the second including 220 long bursts with pseudo-redshifts calculated by the variability-luminosity relation, and the third including 1194 long bursts with pseudo-redshifls calculated by the lag-luminosity relation, respectively. In the redshift range 0-1 the Kolmogorov-Smirnov probability of the observed redshift distribution and that of the variability-luminosity relation is large. In the redshift ranges 1-2, 2-3, 3-6.3 and 0-37, the Kolmogorov-Smirnov probabilities of the redshift distribution from lag-luminosity relation and the observed redshift distribution are also large. For the GRBs, which appear both in the two pseudo-redshift burst samples, the KS probability of the pseudo-redshift distribution from the lag-luminosity relation and the observed reshift distribution is 0.447, which is very large. Based on these results, some conclusions are drawn: i) the V-Liso relation might be more believable than the τ-Liso relation in low redshift ranges and the τ-Liso relation might be more real than the V-Liso relation in high redshift ranges; ii) if we do not consider the redshift ranges, the τ-Liso relation might be more physical and intrinsical than the V-Liso relation.
文摘Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.
文摘Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.
基金Supported by the National Natural Science Foundation of China(61373108)Humanities and Social Science Foundation of Ministry of Education of China(16YJCZH004)the Major Projects of the National Social Science Foundation of China(11&ZD189)
文摘Boundary recognition is an important research of natural language processing, and it provides a basis for the application of Chinese word segmentation, chunk analysis, named entity recognition, etc. Based on ambiguity in boundary recognition of Chinese punctuation marks, this paper proposes grammar testing methods for boundary recognition of slight-pause marks and then calculates the annotation consistency of these methods. The statistical results show that grammar testing methods can greatly improve the annotation consistency of slight-pause marks boundary recognition. The consistency during the second time is 0.030 3 higher than during the first, which will help guarantee the consistency of large-scale corpus annotation and improve the quality of corpus annotation.
基金National Natural Science Foundation of China(No.12271261)Postgraduate Research and Practice Innovation Program of Jiangsu Province,China(Grant No.SJCX230368)。
文摘Normality testing is a fundamental hypothesis test in the statistical analysis of key biological indicators of diabetes.If this assumption is violated,it may cause the test results to deviate from the true value,leading to incorrect inferences and conclusions,and ultimately affecting the validity and accuracy of statistical inferences.Considering this,the study designs a unified analysis scheme for different data types based on parametric statistical test methods and non-parametric test methods.The data were grouped according to sample type and divided into discrete data and continuous data.To account for differences among subgroups,the conventional chi-squared test was used for discrete data.The normal distribution is the basis of many statistical methods;if the data does not follow a normal distribution,many statistical methods will fail or produce incorrect results.Therefore,before data analysis and modeling,the data were divided into normal and non-normal groups through normality testing.For normally distributed data,parametric statistical methods were used to judge the differences between groups.For non-normal data,non-parametric tests were employed to improve the accuracy of the analysis.Statistically significant indicators were retained according to the significance index P-value of the statistical test or corresponding statistics.These indicators were then combined with relevant medical background to further explore the etiology leading to the occurrence or transformation of diabetes status.
基金supported in part by NIH grants R01NS39600,U01MH114829RF1MH128693(to GAA)。
文摘Many fields,such as neuroscience,are experiencing the vast prolife ration of cellular data,underscoring the need fo r organizing and interpreting large datasets.A popular approach partitions data into manageable subsets via hierarchical clustering,but objective methods to determine the appropriate classification granularity are missing.We recently introduced a technique to systematically identify when to stop subdividing clusters based on the fundamental principle that cells must differ more between than within clusters.Here we present the corresponding protocol to classify cellular datasets by combining datadriven unsupervised hierarchical clustering with statistical testing.These general-purpose functions are applicable to any cellular dataset that can be organized as two-dimensional matrices of numerical values,including molecula r,physiological,and anatomical datasets.We demonstrate the protocol using cellular data from the Janelia MouseLight project to chara cterize morphological aspects of neurons.
基金supported by the National Social Science Fund Major Projects(22&ZD160).
文摘This study investigates the relationships between agricultural spot markets and external uncertainties through multifractal detrending moving-average cross-correlation analysis(MF-X-DMA).The dataset contains the Grains&Oilseeds Index(GOI)and its five subindices for wheat,maize,soyabeans,rice,and barley.Moreover,we use three uncertainty proxies,namely,economic policy uncertainty(EPU),geopolitical risk(GPR),and Volatility Index(VIX).We observe multifractal cross-correlations between agricultural markets and uncertainties.Furthermore,statistical tests reveal that maize has intrinsic joint multifractality with all the uncertainty proxies,highly sensitive to external shocks.Additionally,intrinsic multifractality among GOI-GPR,wheat-GPR,and soyabeans-VIX is illustrated.However,other series have apparent multifractal crosscorrelations with high probabilities.Moreover,our analysis suggests that among the three types of external uncertainties,GPR has the strongest association with grain prices,excluding maize and soyabeans.
基金Supported by the Infrastructure Project of China Meteorological Administration(CMA) in 2010~~
文摘In this paper,based on the observation data of air temperature during 1951-2009 in Shenyang,the interannual and interdecadal variation of annual average temperature,maximum and minimum temperature in Shenyang were conducted the statistical analysis by means of linear trend estimation and mutation detection by using Mann-Kendall method.As was demonstrated in the results,the annual average temperature,maximum and minimum temperature in Shenyang showed an upward trend,whose linear tendency rate was 0.231,0.181 and 0.218 respectively.The increment trend of annual average temperature,maximum and minimum temperature was extremely clear.The increase in minimum temperature was more significant than that in mean temperature and maximum temperature.The abrupt change point of annual mean temperature in Shenyang appeared in 1981;the abrupt change point of annual mean maximum temperature appeared in 1994;the annual mean minimum temperature underwent mutation in 1978.
文摘Receiver Autonomous Integrity Monitoring (RAIM) is a software algorithm available in some GPS receivers which gives an indication if the position solution given by the GPS receiver is suitable to use. The detail algorithm of the parity space method of RAIM technique is presented. Using FDI and FDE methods, the simulations of RAIM performance have been done in three different phases independently with respect to the bias of the fault satellite. Case study of simulation results is discussed and each performance of RAIM is analyzed. According to the analysis of simulation results, the parity space method of RAIM can meet the integrity requirements for nonprecision, terminal and enroute flight phase. It also indicates that the results of performance of FDE are better than that of FDI.
文摘This paper discusses the nonlinearity of fish acoustic signals by using the surrogate data method. We compare the difference of three test statistics - time-irreversibility Trey, correlation dimension D2 and auto mutual information function I between the original data and the surrogate data. We come to the conclusion that there exists nonlinearity in the fish acoustic signals and there exist deterministic nonlinear components; therefore nonlinear dynamic theory can be used to analyze fish acoustic signals.
文摘“Human-elephant conflict(HEC)”,the alarming issue,in present day context has attracted the attention of environmentalists and policy makers.The rising conflict between human beings and wild elephants is common in Buxa Tiger Reserve(BTR)and its adjoining area in West Bengal State,India,making the area volatile.People’s attitudes towards elephant conservation activity are very crucial to get rid of HEC,because people’s proximity with wild elephants’habitat can trigger the occurrence of HEC.The aim of this study is to conduct an in-depth investigation about the association of people’s attitudes towards HEC with their locational,demographic,and socio-economic characteristics in BTR and its adjoining area by using Pearson’s bivariate chi-square test and binary logistic regression analysis.BTR is one of the constituent parts of Eastern Doors Elephant Reserve(EDER).We interviewed 500 respondents to understand their perceptions to HEC and investigated their locational,demographic,and socio-economic characteristics including location of village,gender,age,ethnicity,religion,caste,poverty level,education level,primary occupation,secondary occupation,household type,and source of firewood.The results indicate that respondents who are living in enclave forest villages(EFVs),peripheral forest villages(PFVs),corridor village(CVs),or forest and corridor villages(FCVs),mainly males,at the age of 18–48 years old,engaged with agriculture occupation,and living in kancha and mixed houses,have more likelihood to witness HEC.Besides,respondents who are illiterate or at primary education level are more likely to regard elephant as a main problematic animal around their villages and refuse to participate in elephant conservation activity.For the sake of a sustainable environment for both human beings and wildlife,people’s attitudes towards elephants must be friendly in a more prudent way,so that the two communities can live in harmony.
文摘Generalized method of moments based on probability generating function is considered. Estimation and model testing are unified using this approach which also leads to distribution free chi-square tests. The estimation methods developed are also related to estimation methods based on generalized estimating equations but with the advantage of having statistics for model testing. The methods proposed overcome numerical problems often encountered when the probability mass functions have no closed forms which prevent the use of maximum likelihood (ML) procedures and in general, ML procedures do not lead to distribution free model testing statistics.
基金This paper is financially supported by the Ministry of Water Conservancy and Electric Power,P.R.China
文摘According to the need of project design for offshore engineering and coastal engineering, this paper statistically analyses the annual extreme data of waves acquired at 19 observation stations along the coast of China. Five kinds of distribution curves are adopted: Pearson III (P-III), Log-Extreme I (LE), Log-Normal(LN), Weibull(W) and Exponential Γ(EΓ) to check the adaptability to the long-term distribution of annual extreme of wave in the China Sea areas. The New Curve Fitting Method (NFIT) method and Probability Weighted Moments (PWM) method are used to estimate the distribution parameters and thereby to derive the design wave parameters with different return periods at 19 observation stations. The test results show that by combining EΓ distribution and NFIT parameter estimation and optimum seeking by computer, the design wave parameters can be estimated with high accuracy, high speed and high efficiency, and the randomness of the estimated results can be avoided.
基金The research was supported by the Excellence Project PrF UHK No.2208/2021-2022,University of Hradec Kralove,Czech Republic.
文摘Finding a suitable solution to an optimization problem designed in science is a major challenge.Therefore,these must be addressed utilizing proper approaches.Based on a random search space,optimization algorithms can find acceptable solutions to problems.Archery Algorithm(AA)is a new stochastic approach for addressing optimization problems that is discussed in this study.The fundamental idea of developing the suggested AA is to imitate the archer’s shooting behavior toward the target panel.The proposed algorithm updates the location of each member of the population in each dimension of the search space by a member randomly marked by the archer.The AA is mathematically described,and its capacity to solve optimization problems is evaluated on twenty-three distinct types of objective functions.Furthermore,the proposed algorithm’s performance is compared vs.eight approaches,including teaching-learning based optimization,marine predators algorithm,genetic algorithm,grey wolf optimization,particle swarm optimization,whale optimization algorithm,gravitational search algorithm,and tunicate swarm algorithm.According to the simulation findings,the AA has a good capacity to tackle optimization issues in both unimodal and multimodal scenarios,and it can give adequate quasi-optimal solutions to these problems.The analysis and comparison of competing algorithms’performance with the proposed algorithm demonstrates the superiority and competitiveness of the AA.