首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Abstract

One of the most important factors in building and changing communication mechanisms in social networks is considering features of the members of social networks. Most of the existing methods in network monitoring don’t consider effects of features in network formation mechanisms and others don’t lead to reliable results when the features abound or when there are correlations among them. In this article, we combined two methods principal component analysis (PCA) and likelihood method to monitor the underlying network model when the features of individuals abound and when some of them have high correlations with each other.  相似文献   

2.
Multivariate control charts are used to monitor stochastic processes for changes and unusual observations. Hotelling's T2 statistic is calculated for each new observation and an out‐of‐control signal is issued if it goes beyond the control limits. However, this classical approach becomes unreliable as the number of variables p approaches the number of observations n, and impossible when p exceeds n. In this paper, we devise an improvement to the monitoring procedure in high‐dimensional settings. We regularise the covariance matrix to estimate the baseline parameter and incorporate a leave‐one‐out re‐sampling approach to estimate the empirical distribution of future observations. An extensive simulation study demonstrates that the new method outperforms the classical Hotelling T2 approach in power, and maintains appropriate false positive rates. We demonstrate the utility of the method using a set of quality control samples collected to monitor a gas chromatography–mass spectrometry apparatus over a period of 67 days.  相似文献   

3.
In in most cases, the distribution of communications is unknown and one may summarize social network communications with categorical attributes in a contingency table. Due to the categorical nature of the data and a large number of features, there are many parameters to be considered and estimated in the model. Hence, the accuracy of estimators decreases. To overcome the problem of high dimensionality and unknown communications distribution, multiple correspondence analysis is used to reduce the number of parameters. Then the rescaled data are studied in a Dirichlet model in which the parameters should be estimated. Moreover, two control charts, Hotelling’s T2 and multivariate exponentially weighted moving average (MEWMA), are developed to monitor the parameters of the Dirichlet distribution. The performance of the proposed method is evaluated through simulation studies in terms of average run length criterion. Finally, the proposed method is applied to a real case.  相似文献   

4.
We consider the problem of evaluation of the probability that all elements of a multivariate normally distributed vector have non-negative coordinates; this probability is called the non-centred orthant probability. The necessity for the evaluation of this probability arises frequently in statistics. The probability is defined by the integral of the probability density function. However, direct numerical integration is not practical. In this article, a method is proposed for the computation of the probability. The method involves the evaluation of a measure on a unit sphere surface in p-dimensional space that satisfies conditions derived from a covariance matrix. The required computational time for the p-dimensional problem is proportional to p2·2p?1, and it increases at a rate that is lower than that in the case of the existing method.  相似文献   

5.
A variant of a sexual Gallon–Watson process is considered. At each generation the population is partitioned among n‘hosts’ (population patches) and individual members mate at random only with others within the same host. This is appropriate for many macroparasite systems, and at low parasite loads it gives rise to a depressed rate of reproduction relative to an asexual system, due to the possibility that females are unmated. It is shown that stochasticity mitigates against this effect, so that for small initial populations the probability of ultimate extinction (the complement of an ‘epidemic’) displays a tradeoff as a function of n between the strength of fluctuations which overcome this ‘mating’ probability, and the probability of the subpopulation in one host being ‘rescued’ by that in another. Complementary approximations are developed for the extinction probability: an asymptotically exact approximation at large n, and for small n a short‐time probability that is exact in the limit where the mean number of offspring per parent is large.  相似文献   

6.
Social network monitoring consists of monitoring changes in networks with the aim of detecting significant ones and attempting to identify assignable cause(s) contributing to the occurrence of a change. This paper proposes a method that helps to overcome some of the weaknesses of the existing methods. A Poisson regression model for the probability of the number of communications between network members as a function of vertex attributes is constructed. Multivariate exponentially weighted moving average (MEWMA) and multivariate cumulative sum (MCUSUM) control charts are used to monitor the network formation process. The results indicate more efficient performance for the MEWMA chart in identifying significant changes.  相似文献   

7.
This paper presents a new model that monitors the basic network formation mechanisms via the attributes through time. It considers the issue of joint modeling of longitudinal inflated (0, 1)-support continuous and inflated count response variables. For joint model of mentioned response variables, a correlated generalized linear mixed model is studied. The fraction response is inflated in two points k and l (k < l) and a k and l inflated beta distribution is introduced to use as its distribution. Also, the count response is inflated in zero and we use some members of zero-inflated power series distributions, hurdle-at-zero, members of zero-inflated double power series distributions and zero-inflated generalized Poisson distribution as our count response distribution. A full likelihood-based approach is used to yield maximum likelihood estimates of the model parameters and the model is applied to a real social network obtained from an observational study where the rate of the ith node’s responsiveness to the jth node and the number of arrows or edges with some specific characteristics from the ith node to the jth node are the correlated inflated (0, 1)-support continuous and inflated count response variables, respectively. The effect of the sender and receiver positions in an office environment on the responses are investigated simultaneously.  相似文献   

8.
We consider the Gibbs sampler as a tool for generating an absolutely continuous probability measure ≥ on Rd. When an appropriate irreducibility condition is satisfied, the Gibbs Markov chain (Xn;n ≥ 0) converges in total variation to its target distribution ≥. Sufficient conditions for geometric convergence have been given by various authors. Here we illustrate, by means of simple examples, how slow the convergence can be. In particular, we show that given a sequence of positive numbers decreasing to zero, say (bn;n ≥ 1), one can construct an absolutely continuous probability measure ≥ on Rd which is such that the total variation distance between ≥ and the distribution of Xn, converges to 0 at a rate slower than that of the sequence (bn;n ≥ 1). This can even be done in such a way that ≥ is the uniform distribution over a bounded connected open subset of Rd. Our results extend to hit-and-run samplers with direction distributions having supports with symmetric gaps.  相似文献   

9.
The discrete stable family constitutes an interesting two-parameter model of distributions on the non-negative integers with a Paretian tail. The practical use of the discrete stable distribution is inhibited by the lack of an explicit expression for its probability function. Moreover, the distribution does not possess moments of any order. Therefore, the usual tools—such as the maximum-likelihood method or even the moment method—are not feasible for parameter estimation. However, the probability generating function of the discrete stable distribution is available in a simple form. Hence, we initially explore the application of some existing estimation procedures based on the empirical probability generating function. Subsequently, we propose a new estimation method by minimizing a suitable weighted L 2-distance between the empirical and the theoretical probability generating functions. In addition, we provide a goodness-of-fit statistic based on the same distance.  相似文献   

10.
EMPIRICAL LIKELIHOOD-BASED KERNEL DENSITY ESTIMATION   总被引:2,自引:0,他引:2  
This paper considers the estimation of a probability density function when extra distributional information is available (e.g. the mean of the distribution is known or the variance is a known function of the mean). The standard kernel method cannot exploit such extra information systematically as it uses an equal probability weight n-1 at each data point. The paper suggests using empirical likelihood to choose the probability weights under constraints formulated from the extra distributional information. An empirical likelihood-based kernel density estimator is given by replacing n-1 by the empirical likelihood weights, and has these advantages: it makes systematic use of the extra information, it is able to reflect the extra characteristics of the density function, and its variance is smaller than that of the standard kernel density estimator.  相似文献   

11.
This article considers fixed effects (FE) estimation for linear panel data models under possible model misspecification when both the number of individuals, n, and the number of time periods, T, are large. We first clarify the probability limit of the FE estimator and argue that this probability limit can be regarded as a pseudo-true parameter. We then establish the asymptotic distributional properties of the FE estimator around the pseudo-true parameter when n and T jointly go to infinity. Notably, we show that the FE estimator suffers from the incidental parameters bias of which the top order is O(T? 1), and even after the incidental parameters bias is completely removed, the rate of convergence of the FE estimator depends on the degree of model misspecification and is either (nT)? 1/2 or n? 1/2. Second, we establish asymptotically valid inference on the (pseudo-true) parameter. Specifically, we derive the asymptotic properties of the clustered covariance matrix (CCM) estimator and the cross-section bootstrap, and show that they are robust to model misspecification. This establishes a rigorous theoretical ground for the use of the CCM estimator and the cross-section bootstrap when model misspecification and the incidental parameters bias (in the coefficient estimate) are present. We conduct Monte Carlo simulations to evaluate the finite sample performance of the estimators and inference methods, together with a simple application to the unemployment dynamics in the U.S.  相似文献   

12.
Control charts have been used effectively for years to monitor processes and detect abnormal behaviors. However, most control charts require a specific distribution to establish their control limits. The bootstrap method is a nonparametric technique that does not rely on the assumption of a parametric distribution of the observed data. Although the bootstrap technique has been used to develop univariate control charts to monitor a single process, no effort has been made to integrate the effectiveness of the bootstrap technique with multivariate control charts. In the present study, we propose a bootstrap-based multivariate T 2 control chart that can efficiently monitor a process when the distribution of observed data is nonnormal or unknown. A simulation study was conducted to evaluate the performance of the proposed control chart and compare it with a traditional Hotelling's T 2 control chart and the kernel density estimation (KDE)-based T 2 control chart. The results showed that the proposed chart performed better than the traditional T 2 control chart and performed comparably with the KDE-based T 2 control chart. Furthermore, we present a case study to demonstrate the applicability of the proposed control chart to real situations.  相似文献   

13.
A Bayesian analysis is provided for the Wilcoxon signed-rank statistic (T+). The Bayesian analysis is based on a sign-bias parameter φ on the (0, 1) interval. For the case of a uniform prior probability distribution for φ and for small sample sizes (i.e., 6 ? n ? 25), values for the statistic T+ are computed that enable probabilistic statements about φ. For larger sample sizes, approximations are provided for the asymptotic likelihood function P(T+|φ) as well as for the posterior distribution P(φ|T+). Power analyses are examined both for properly specified Gaussian sampling and for misspecified non Gaussian models. The new Bayesian metric has high power efficiency in the range of 0.9–1 relative to a standard t test when there is Gaussian sampling. But if the sampling is from an unknown and misspecified distribution, then the new statistic still has high power; in some cases, the power can be higher than the t test (especially for probability mixtures and heavy-tailed distributions). The new Bayesian analysis is thus a useful and robust method for applications where the usual parametric assumptions are questionable. These properties further enable a way to do a generic Bayesian analysis for many non Gaussian distributions that currently lack a formal Bayesian model.  相似文献   

14.
In this paper we consider the issue of constructing retrospective T 2 control chart limits so as to control the overall probability of a false alarm at a specified value. We describe an exact method for constructing the control limits for retrospective examination. We then consider Bonferroni-adjustments to Alt's control limit and to the standard x 2 control limit as alternatives to the exact limit since it is computationally cumbersome to find the exact limit. We present the results of some simulation experiments that are carried out to compare the performance of these control limits. The results indicate that the Bonferroni-adjusted Alt's control limit performs better that the Bonferroni-adjusted x 2 control limit. Furthermore, it appears that the Bonferroni-adjusted Alt's control limit is more than adequate for controlling the overall false alarm probability at a specified value.  相似文献   

15.
The classical D-optimality principle in regression design may be motivated by a desire to maximize the coverage probability of a fixed-volume confidence ellipsoid on the regression parameters. When the fitted model is exactly correct, this amounts to minimizing the determinant of the covariance matrix of the estimators. We consider an analogue of this problem, under the approximately linear model E[y|x] = θTz(x) + f(x). The nonlinear disturbance f(x) is essentially unknown, and the experimenter fits only to the linear part of the response. The resulting bias affects the coverage probability of the confidence ellipsoid on θ. We study the construction of designs which maximize the minimum coverage probability as f varies over a certain class. Explicit designs are given in the case that the fitted response surface is a plane.  相似文献   

16.
This paper deals with the problem of estimating the volume of the excursion set of a function f:ℝ d →ℝ above a given threshold, under a probability measure on ℝ d that is assumed to be known. In the industrial world, this corresponds to the problem of estimating a probability of failure of a system. When only an expensive-to-simulate model of the system is available, the budget for simulations is usually severely limited and therefore classical Monte Carlo methods ought to be avoided. One of the main contributions of this article is to derive SUR (stepwise uncertainty reduction) strategies from a Bayesian formulation of the problem of estimating a probability of failure. These sequential strategies use a Gaussian process model of f and aim at performing evaluations of f as efficiently as possible to infer the value of the probability of failure. We compare these strategies to other strategies also based on a Gaussian process model for estimating a probability of failure.  相似文献   

17.
ABSTRACT

In Bayesian theory, calculating a posterior probability distribution is highly important but typically difficult. Therefore, some methods have been proposed to deal with such problem, among which, the most popular one is the asymptotic expansions of posterior distributions. In this paper, we propose an alternative approach, named a random weighting method, for scaled posterior distributions, and give an ideal convergence rate, o(n( ? 1/2)), which serves as the theoretical guarantee for methods of numerical simulations.  相似文献   

18.
Kumaraswamy [Generalized probability density-function for double-bounded random-processes, J. Hydrol. 462 (1980), pp. 79–88] introduced a distribution for double-bounded random processes with hydrological applications. For the first time, based on this distribution, we describe a new family of generalized distributions (denoted with the prefix ‘Kw’) to extend the normal, Weibull, gamma, Gumbel, inverse Gaussian distributions, among several well-known distributions. Some special distributions in the new family such as the Kw-normal, Kw-Weibull, Kw-gamma, Kw-Gumbel and Kw-inverse Gaussian distribution are discussed. We express the ordinary moments of any Kw generalized distribution as linear functions of probability weighted moments (PWMs) of the parent distribution. We also obtain the ordinary moments of order statistics as functions of PWMs of the baseline distribution. We use the method of maximum likelihood to fit the distributions in the new class and illustrate the potentiality of the new model with an application to real data.  相似文献   

19.
Asieh Abtahi 《Statistics》2013,47(1):126-140
There are so many proposals in construction skewed distributions, and it is worth finding an overall class which covers all of these proposals. We introduce a new unified representation of multivariate skewed distributions. We will show that this new unified multivariate form of skewed distributions includes all of the continuous multivariate skewed distributions in the literature. This new unified representation is based on the multivariate probability integral transformation and can be decomposed into one factor that is original multivariate symmetric probability density function (pdf) f on ? k and skewed factor defined by a pdf p on [0, 1] k . This decomposition leads us to prove some useful properties of this new unified form. Stochastic representations and basic properties of this new form are also investigated in this article. Our work is motivated by considering the different skewing mechanisms which lead to different skewed distributions and show that all of these common-used distributions can be viewed as a new unified form.  相似文献   

20.
Empirical likelihood has attracted much attention in the literature as a nonparametric method. A recent paper by Lu & Peng (2002) [Likelihood based confidence intervals for the tail index. Extremes 5, 337–352] applied this method to construct a confidence interval for the tail index of a heavy‐tailed distribution. It turns out that the empirical likelihood method, as well as other likelihood‐based methods, performs better than the normal approximation method in terms of coverage probability. However, when the sample size is small, the confidence interval computed using the χ2 approximation has a serious undercoverage problem. Motivated by Tsao (2004) [A new method of calibration for the empirical loglikelihood ratio. Statist. Probab. Lett. 68, 305–314], this paper proposes a new method of calibration, which corrects the undercoverage problem.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号