首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Dabuxilatu Wang 《Statistics》2013,47(2):167-181
Some asymptotic properties of point estimation with n-dimensional fuzzy data with respect to a special L 2-metric ρ are investigated in this article. It is shown that the collection of all n-dimensional fuzzy data endowed with the ρ-metric is a complete and separable space. Some criterions for point estimation in such fuzzy environments are proposed, and the sample mean, variance and covariance with n-dimensional fuzzy data under these criteria are further studied.  相似文献   

2.
Coppi et al. [7 R. Coppi, P. D'Urso, and P. Giordani, Fuzzy and possibilistic clustering for fuzzy data, Comput. Stat. Data Anal. 56 (2012), pp. 915927. doi: 10.1016/j.csda.2010.09.013[Crossref], [Web of Science ®] [Google Scholar]] applied Yang and Wu's [20 M.-S. Yang and K.-L. Wu, Unsupervised possibilistic clustering, Pattern Recognit. 30 (2006), pp. 521. doi: 10.1016/j.patcog.2005.07.005[Crossref], [Web of Science ®] [Google Scholar]] idea to propose a possibilistic k-means (PkM) clustering algorithm for LR-type fuzzy numbers. The memberships in the objective function of PkM no longer need to satisfy the constraint in fuzzy k-means that of a data point across classes sum to one. However, the clustering performance of PkM depends on the initializations and weighting exponent. In this paper, we propose a robust clustering method based on a self-updating procedure. The proposed algorithm not only solves the initialization problems but also obtains a good clustering result. Several numerical examples also demonstrate the effectiveness and accuracy of the proposed clustering method, especially the robustness to initial values and noise. Finally, three real fuzzy data sets are used to illustrate the superiority of this proposed algorithm.  相似文献   

3.
In healthcare studies, count data sets measured with covariates often exhibit heterogeneity and contain extreme values. To analyse such count data sets, we use a finite mixture of regression model framework and investigate a robust estimation approach, called the L2E [D.W. Scott, On fitting and adapting of density estimates, Comput. Sci. Stat. 30 (1998), pp. 124–133], to estimate the parameters. The L2E is based on an integrated L2 distance between parametric conditional and true conditional mass functions. In addition to studying the theoretical properties of the L2E estimator, we compare the performance of L2E with the maximum likelihood (ML) estimator and a minimum Hellinger distance (MHD) estimator via Monte Carlo simulations for correctly specified and gross-error contaminated mixture of Poisson regression models. These show that the L2E is a viable robust alternative to the ML and MHD estimators. More importantly, we use the L2E to perform a comprehensive analysis of a Western Australia hospital inpatient obstetrical length of stay (LOS) (in days) data that contains extreme values. It is shown that the L2E provides a two-component Poisson mixture regression fit to the LOS data which is better than those based on the ML and MHD estimators. The L2E fit identifies admission type as a significant covariate that profiles the predominant subpopulation of normal-stayers as planned patients and the small subpopulation of long-stayers as emergency patients.  相似文献   

4.
Process capability indices are widely used to evaluate the performance of processes in the manufacturing industry. Over the years, the issues have been investigated extensively. Some articles have studied them with fuzzy estimation. However, it seems that no article has proposed a version of triangular fuzzy numbers for critical value to test the process capability. In this article, we use Buckley's approach (2003 Buckley , J. J. ( 2003 ). Fuzzy Probabilities: New Approach and Application . Heidelberg : Physica-Verlag .[Crossref] [Google Scholar]) to construct the triangular fuzzy numbers for C pl and C pu , especially, the triangular fuzzy numbers for critical values are derived to execute the fuzzy hypothesis testing for C pl and C pu . Some numerical examples are taken to illustrate the proposed methodology.  相似文献   

5.
Process capability indices (PCIs) provide numerical measures on whether a process conforms to the defined manufacturing capability prerequisite. These have been successfully applied by companies to compete with and to lead high-profit markets by evaluating the quality and productivity performance. The PCI Cp compares the output of a process to the specification limits (SLs) by forming the ratio of the width between the process SLs with the width of the natural tolerance limits which is measured by six process standard deviation units. As another common PCI, Cpm incorporates two variation components which are variation to the process mean and deviation of the process mean from the target. A meaningful generalized version of above PCIs is introduced in this paper which is able to handle in a fuzzy environment. These generalized PCIs are able to measure the capability of a fuzzy-valued process in producing products on the basis of a fuzzy quality. Fast computing formulas for the generalized PCIs are computed for normal and symmetric triangular fuzzy observations, where the fuzzy quality is defined by linear and exponential fuzzy SLs. A practical example is presented to show the performance of proposed indices.  相似文献   

6.
We propose a new distribution, the so-called beta-Weibull geometric distribution, whose failure rate function can be decreasing, increasing or an upside-down bathtub. This distribution contains special sub-models the exponential geometric [K. Adamidis and S. Loukas, A lifetime distribution with decreasing failure rate, Statist. Probab. Lett. 39 (1998), pp. 35–42], beta exponential [S. Nadarajah and S. Kotz, The exponentiated type distributions, Acta Appl. Math. 92 (2006), pp. 97–111; The beta exponential distribution, Reliab. Eng. Syst. Saf. 91 (2006), pp. 689–697], Weibull geometric [W. Barreto-Souza, A.L. de Morais, and G.M. Cordeiro, The Weibull-geometric distribution, J. Stat. Comput. Simul. 81 (2011), pp. 645–657], generalized exponential geometric [R.B. Silva, W. Barreto-Souza, and G.M. Cordeiro, A new distribution with decreasing, increasing and upside-down bathtub failure rate, Comput. Statist. Data Anal. 54 (2010), pp. 935–944; G.O. Silva, E.M.M. Ortega, and G.M. Cordeiro, The beta modified Weibull distribution, Lifetime Data Anal. 16 (2010), pp. 409–430] and beta Weibull [S. Nadarajah, G.M. Cordeiro, and E.M.M. Ortega, General results for the Kumaraswamy-G distribution, J. Stat. Comput. Simul. (2011). DOI: 10.1080/00949655.2011.562504] distributions, among others. The density function can be expressed as a mixture of Weibull density functions. We derive expansions for the moments, generating function, mean deviations and Rénvy entropy. The parameters of the proposed model are estimated by maximum likelihood. The model fitting using envelops was conducted. The proposed distribution gives a good fit to the ozone level data in New York.  相似文献   

7.
Process capability indices are numerical tools that quantify how well a process can meet customer requirements, specifications or engineering tolerances. Fuzzy logic is incorporated to deal imprecise, incomplete data along with uncertainty. This paper develops two fuzzy methods for measuring the process capability in simple linear profiles for the circumstances in which lower and upper specification limits are imprecise. To guide practitioners, numerical example is provided.  相似文献   

8.
Under a two-parameter exponential distribution, this study constructs the generalized lower confidence limit of the lifetime performance index CL based on type-II right-censored data. The confidence limit has to be numerically obtained; however, the required computations are simple and straightforward. Confidence limits of CL computed under the generalized paradigm are compared with those of CL computed under the classical paradigm, citing an illustrative example with real data and two examples with simulated data, to demonstrate the merits and advantages of the proposed generalized variable method over the classical method.  相似文献   

9.
Outlier detection has always been of interest for researchers and data miners. It has been well researched in different knowledge and application domains. This study aims at exploring the correctly identifying outliers using most commonly applied statistics. We evaluate the performance of AO, IO, LS, and TC as vulnerability to spurious outliers by means of empirical level of significance (ELS), power of the test indicating the sensitivity of the statistical tests in detecting changes and the vulnerability to masking of outliers in terms of misspecification frequencies are determined. We have observed that the sampling distribution of test statistic ηtp; tp = AO,?IO,?LS,?TC in case of AR(1) model is connected with the values of n and φ. The sampling distribution of ηTC is less concentrated than the sampling distribution of ηAO, ηIO, and ηLS. In AR(1) process, empirical critical values for 1%, 5%, and 10% upper percentiles are found to be higher than those generally used. We have also found the evidence that the test statistics for transient change (TC) needs to be revisited as the test statistics ηTC is found to be eclipsed by ηAO,?ηLS and ηIO at different δ values. TC keeps on confusing with IO and AO, and at extreme δ values it just gets equal to AO and LS.  相似文献   

10.
Tests for the equality of variances are of interest in many areas such as quality control, agricultural production systems, experimental education, pharmacology, biology, as well as a preliminary to the analysis of variance, dose–response modelling or discriminant analysis. The literature is vast. Traditional non-parametric tests are due to Mood, Miller and Ansari–Bradley. A test which usually stands out in terms of power and robustness against non-normality is the W50 Brown and Forsythe [Robust tests for the equality of variances, J. Am. Stat. Assoc. 69 (1974), pp. 364–367] modification of the Levene test [Robust tests for equality of variances, in Contributions to Probability and Statistics, I. Olkin, ed., Stanford University Press, Stanford, 1960, pp. 278–292]. This paper deals with the two-sample scale problem and in particular with Levene type tests. We consider 10 Levene type tests: the W50, the M50 and L50 tests [G. Pan, On a Levene type test for equality of two variances, J. Stat. Comput. Simul. 63 (1999), pp. 59–71], the R-test [R.G. O'Brien, A general ANOVA method for robust tests of additive models for variances, J. Am. Stat. Assoc. 74 (1979), pp. 877–880], as well as the bootstrap and permutation versions of the W50, L50 and R tests. We consider also the F-test, the modified Fligner and Killeen [Distribution-free two-sample tests for scale, J. Am. Stat. Assoc. 71 (1976), pp. 210–213] test, an adaptive test due to Hall and Padmanabhan [Adaptive inference for the two-sample scale problem, Technometrics 23 (1997), pp. 351–361] and the two tests due to Shoemaker [Tests for differences in dispersion based on quantiles, Am. Stat. 49(2) (1995), pp. 179–182; Interquantile tests for dispersion in skewed distributions, Commun. Stat. Simul. Comput. 28 (1999), pp. 189–205]. The aim is to identify the effective methods for detecting scale differences. Our study is different with respect to the other ones since it is focused on resampling versions of the Levene type tests, and many tests considered here have not ever been proposed and/or compared. The computationally simplest test found robust is W50. Higher power, while preserving robustness, is achieved by considering the resampling version of Levene type tests like the permutation R-test (recommended for normal- and light-tailed distributions) and the bootstrap L50 test (recommended for heavy-tailed and skewed distributions). Among non-Levene type tests, the best one is the adaptive test due to Hall and Padmanabhan.  相似文献   

11.
The quality characteristics, which are known as attributes, cannot be conveniently and numerically represented. Generally, the attribute data can be regarded as the fuzzy data, which are ubiquitous in the manufacturing process and cannot be measured precisely and often be collected by visual inspection. In this paper, we construct a p control chart for monitoring the fraction of nonconforming items in the process in which fuzzy sample data are collected from the manufacturing process. The resolution identity – a well-known theorem in the fuzzy set theory – is invoked to construct the control limits of fuzzy-p control charts using fuzzy data. In order to determine whether the plotted imprecise fraction of nonconforming items is within the fuzzy lower and upper control limits, we also propose a ranking method for a set of fuzzy numbers. Using the fuzzy-p control charts and the proposed acceptability function to classify the manufacturing process allows the decision-maker to make linguistic decisions such as rather in control or rather out of control. A practical example is provided to describe the applicability of the fuzzy set theory to a conventional p control chart.  相似文献   

12.
In this paper, we consider the bootstrap procedure for the augmented Dickey–Fuller (ADF) unit root test by implementing the modified divergence information criterion (MDIC, Mantalos et al. [An improved divergence information criterion for the determination of the order of an AR process, Commun. Statist. Comput. Simul. 39(5) (2010a), pp. 865–879; Forecasting ARMA models: A comparative study of information criteria focusing on MDIC, J. Statist. Comput. Simul. 80(1) (2010b), pp. 61–73]) for the selection of the optimum number of lags in the estimated model. The asymptotic distribution of the resulting bootstrap ADF/MDIC test is established and its finite sample performance is investigated through Monte-Carlo simulations. The proposed bootstrap tests are found to have finite sample sizes that are generally much closer to their nominal values, than those tests that rely on other information criteria, like the Akaike information criterion [H. Akaike, Information theory and an extension of the maximum likelihood principle, in Proceedings of the 2nd International Symposium on Information Theory, B.N. Petrov and F. Csáki, eds., Akademiai Kaido, Budapest, 1973, pp. 267–281]. The simulations reveal that the proposed procedure is quite satisfactory even for models with large negative moving average coefficients.  相似文献   

13.
In a recent paper (J. Statist. Comput. Simul., 1995, Vol.53, pp. 195–203) P. A. Wright proposed a new process capability index Cs which generalizes the Pearn-Kotz-Johnson’s index Cpmk by taking into account the skewness (in addition to deviation of the mean from tliCrntarget already incorporated in Cpmk ). The purpose of this article is to study the consistency and asymptotics of an estimate ?s of Cs The asymptotic distribution provides an insight into some desirable properties of the estimate which are not apparent from its original definition  相似文献   

14.
We deal with a general class of extreme-value regression models introduced by Barreto-Souza and Vasconcellos [Bias and skewness in a general extreme-value regression model, Comput. Statist. Data Anal. 55 (2011), pp. 1379–1393]. Our goal is to derive an adjusted likelihood ratio statistic that is approximately distributed as χ2 with a high degree of accuracy. Although the adjusted statistic requires more computational effort than its unadjusted counterpart, it is shown that the adjustment term has a simple compact form that can be easily implemented in standard statistical software. Further, we compare the finite-sample performance of the three classical tests (likelihood ratio, Wald, and score), the gradient test that has been recently proposed by Terrell [The gradient statistic, Comput. Sci. Stat. 34 (2002), pp. 206–215], and the adjusted likelihood ratio test obtained in this article. Our simulations favour the latter. Applications of our results are presented.  相似文献   

15.
We investigate by simulation how the wild bootstrap and pairs bootstrap perform in t and F tests of regression parameters in the stochastic regression model, where explanatory variables are stochastic and not given and there exists no heteroskedasticity. The wild bootstrap procedure due to Davidson and Flachaire [The wild bootstrap, tamed at last, Working paper, IER#1000, Queen's University, 2001] with restricted residuals works best but its dominance is not strong compared to the result of Flachaire [Bootstrapping heteroskedastic regression models: wild bootstrap vs. pairs bootstrap, Comput. Statist. Data Anal. 49 (2005), pp. 361–376] in the fixed regression model where explanatory variables are fixed and there exists heteroskedasticity.  相似文献   

16.
Classical time-series theory assumes values of the response variable to be ‘crisp’ or ‘precise’, which is quite often violated in reality. However, forecasting of such data can be carried out through fuzzy time-series analysis. This article presents an improved method of forecasting based on LR fuzzy sets as membership functions. As an illustration, the methodology is employed for forecasting India's total foodgrain production. For the data under consideration, superiority of proposed method over other competing methods is demonstrated in respect of modelling and forecasting on the basis of mean square error and average relative error criteria. Finally, out-of-sample forecasts are also obtained.  相似文献   

17.
Doubly truncated data appear in a number of applications, including astronomy and survival analysis. For double-truncated data, the lifetime T is observable only when UTV, where U and V are the left-truncated and right-truncated time, respectively. In some situations, the lifetime T also suffers interval censoring. Using the EM algorithm of Turnbull [The empirical distribution function with arbitrarily grouped censored and truncated data, J. R. Stat. Soc. Ser. B 38 (1976), pp. 290–295] and iterative convex minorant algorithm [P. Groeneboom and J.A. Wellner, Information Bounds and Nonparametric Maximum Likelihood Estimation, Birkhäuser, Basel, 1992], we study the performance of the nonparametric maximum-likelihood estimates (NPMLEs) of the distribution function of T. Simulation results indicate that the NPMLE performs adequately for the finite sample.  相似文献   

18.
For any continuous baseline G distribution [G.M. Cordeiro and M. de Castro, A new family of generalized distributions, J. Statist. Comput. Simul. 81 (2011), pp. 883–898], proposed a new generalized distribution (denoted here with the prefix ‘Kw-G’ (Kumaraswamy-G)) with two extra positive parameters. They studied some of its mathematical properties and presented special sub-models. We derive a simple representation for the Kw-G density function as a linear combination of exponentiated-G distributions. Some new distributions are proposed as sub-models of this family, for example, the Kw-Chen [Z.A. Chen, A new two-parameter lifetime distribution with bathtub shape or increasing failure rate function, Statist. Probab. Lett. 49 (2000), pp. 155–161], Kw-XTG [M. Xie, Y. Tang, and T.N. Goh, A modified Weibull extension with bathtub failure rate function, Reliab. Eng. System Safety 76 (2002), pp. 279–285] and Kw-Flexible Weibull [M. Bebbington, C.D. Lai, and R. Zitikis, A flexible Weibull extension, Reliab. Eng. System Safety 92 (2007), pp. 719–726]. New properties of the Kw-G distribution are derived which include asymptotes, shapes, moments, moment generating function, mean deviations, Bonferroni and Lorenz curves, reliability, Rényi entropy and Shannon entropy. New properties of the order statistics are investigated. We discuss the estimation of the parameters by maximum likelihood. We provide two applications to real data sets and discuss a bivariate extension of the Kw-G distribution.  相似文献   

19.
This paper is concerned with the application of artificial neural networks (ANNs) to a practical, difficult and high-dimensional classification problem, discrimination between selected under-water sounds. The application provides for a particular comparison of the relative performance of time-delay as opposed to fully connected network architectures, in the analysis of temporal data. More originally, suggestions are given for adapting the conventional backpropagation algorithm to give greater robustness to mis-classification errors in the training examples—a particular problem with underwater sound data and one which may arise in other realistic applications of ANNs. An informal comparison is made between the generalisation performance of various architectures in classifying real dolphin sounds when networks are trained using the conventional least squares minimisation norm, L 2, that of least absolute deviation, L 1, and that of the Huber criterion, which involves a mixture of both L 1 and L 2. The results suggest that L 1 and Huber may provide performance gains. In order to evaluate these robust adjustments more formally under controlled conditions, an experiment is then conducted using simulated dolphin sounds with known levels of random noise and misclassification error. Here, the results are more ambiguous and significant interactions are indicated which raise issues for future research.  相似文献   

20.
In this paper, a generalization of inverted exponential distribution is considered as a lifetime model [A.M. Abouammoh and A.M. Alshingiti, Reliability estimation of generalized inverted exponential distribution, J. Statist. Comput. Simul. 79(11) (2009), pp. 1301–1315]. Its reliability characteristics and important distributional properties are discussed. Maximum likelihood estimation of the two parameters involved along with reliability and failure rate functions are derived. The method of least square estimation of parameters is also studied here. In view of cost and time constraints, type II progressively right censored sampling scheme has been used. For illustration of the performance of the estimates, a Monte Carlo simulation study is carried out. Finally, a real data example is given to show the practical applications of the paper.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号