首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   680篇
  免费   15篇
管理学   11篇
民族学   3篇
丛书文集   4篇
理论方法论   2篇
综合类   50篇
社会学   8篇
统计学   617篇
  2023年   2篇
  2021年   3篇
  2020年   7篇
  2019年   10篇
  2018年   17篇
  2017年   22篇
  2016年   14篇
  2015年   19篇
  2014年   18篇
  2013年   260篇
  2012年   40篇
  2011年   19篇
  2010年   20篇
  2009年   17篇
  2008年   32篇
  2007年   19篇
  2006年   6篇
  2005年   9篇
  2004年   8篇
  2003年   16篇
  2002年   6篇
  2001年   10篇
  2000年   9篇
  1999年   6篇
  1998年   7篇
  1997年   6篇
  1996年   2篇
  1995年   6篇
  1994年   2篇
  1993年   5篇
  1992年   4篇
  1991年   5篇
  1990年   2篇
  1989年   4篇
  1988年   5篇
  1987年   3篇
  1986年   2篇
  1985年   8篇
  1984年   11篇
  1983年   11篇
  1982年   8篇
  1981年   2篇
  1980年   1篇
  1979年   2篇
  1978年   2篇
  1977年   5篇
  1976年   1篇
  1975年   2篇
排序方式: 共有695条查询结果,搜索用时 15 毫秒
1.
Proportional hazards are a common assumption when designing confirmatory clinical trials in oncology. This assumption not only affects the analysis part but also the sample size calculation. The presence of delayed effects causes a change in the hazard ratio while the trial is ongoing since at the beginning we do not observe any difference between treatment arms, and after some unknown time point, the differences between treatment arms will start to appear. Hence, the proportional hazards assumption no longer holds, and both sample size calculation and analysis methods to be used should be reconsidered. The weighted log‐rank test allows a weighting for early, middle, and late differences through the Fleming and Harrington class of weights and is proven to be more efficient when the proportional hazards assumption does not hold. The Fleming and Harrington class of weights, along with the estimated delay, can be incorporated into the sample size calculation in order to maintain the desired power once the treatment arm differences start to appear. In this article, we explore the impact of delayed effects in group sequential and adaptive group sequential designs and make an empirical evaluation in terms of power and type‐I error rate of the of the weighted log‐rank test in a simulated scenario with fixed values of the Fleming and Harrington class of weights. We also give some practical recommendations regarding which methodology should be used in the presence of delayed effects depending on certain characteristics of the trial.  相似文献   
2.
Comparison of Four New General Classes of Search Designs   总被引:1,自引:0,他引:1  
A factor screening experiment identifies a few important factors from a large list of factors that potentially influence the response. If a list consists of m factors each at three levels, a design is a subset of all possible 3 m runs. This paper considers the problem of finding designs with small numbers of runs, using the search linear model introduced in Srivastava (1975). The paper presents four new general classes of these 'search designs', each with 2 m −1 runs, which permit, at most, two important factors out of m factors to be searched for and identified. The paper compares the designs for 4 ≤ m ≤ 10, using arithmetic and geometric means of the determinants, traces and maximum characteristic roots of particular matrices. Two of the designs are found to be superior in all six criteria studied. The four designs are identical for m = 3 and this design is an optimal design in the class of all search designs under the six criteria. The four designs are also identical for m = 4 under some row and column permutations.  相似文献   
3.
The problem considered is that of finding an optimum measurement schedule to estimate population parameters in a nonlinear model when the patient effects are random. The paper presents examples of the use of sensitivity functions, derived from the General Equivalence Theorem for D-optimality, in the construction of optimum population designs for such schedules. With independent observations, the theorem applies to the potential inclusion of a single observation. However, in population designs the observations are correlated and the theorem applies to the inclusion of an additional measurement schedule. In one example, three groups of patients of differing size are subject to distinct schedules. Numerical, as opposed to analytical, calculation of the sensitivity function is advocated. The required covariances of the observations are found by simulation.  相似文献   
4.
For a wide variety of applications, experiments are based on units ordered over time or space. Models for these experiments generally may include one or more of: correlations, systematic trends, carryover effects and interference effects. Since the standard optimal block designs may not be efficient in these situations, orthogonal arrays of type I and type II, which were introduced in 1961 by C.R. Rao [Combinatorial arrangements analogous to orthogonal arrays, Sankhya A 23 (1961) 283–286], have been recently used to construct optimal and efficient designs for many of these experiments. Results in this area are unified and the salient features are outlined.  相似文献   
5.
Complete and partial diallel cross designs are examined as to their construction and robustness against the loss of a block of observations. A simple generalized inverse is found for the information matrix of the line effects, which allows evaluation of expressions for the variances of the line-effect differences with and without the missing block. A-efficiencies, based on average variances of the elementary contrasts of the line-effects, suggest that these designs are fairly robust. The loss of efficiency is generally less than 10%, but it is shown that specific comparisons might suffer a loss of efficiency of as much as 40%.  相似文献   
6.
To explore the projection efficiency of a design, Tsai, et al [2000. Projective three-level main effects designs robust to model uncertainty. Biometrika 87, 467–475] introduced the Q criterion to compare three-level main-effects designs for quantitative factors that allow the consideration of interactions in addition to main effects. In this paper, we extend their method and focus on the case in which experimenters have some prior knowledge, in advance of running the experiment, about the probabilities of effects being non-negligible. A criterion which incorporates experimenters’ prior beliefs about the importance of each effect is introduced to compare orthogonal, or nearly orthogonal, main effects designs with robustness to interactions as a secondary consideration. We show that this criterion, exploiting prior information about model uncertainty, can lead to more appropriate designs reflecting experimenters’ prior beliefs.  相似文献   
7.
This paper studies optimum designs for linear models when the errors are heteroscedastic. Sufficient conditions are given in order to obtainD-, A- andE-optimum designs for a complete regression model from partial optimum designs for some sub-parameters. A result about optimality for a complete model from the optimality for the submodels is included. Supported by Junta de Andalucía, research group FQM244.  相似文献   
8.
Summary This paper solves some D-optimal design problems for certain Generalized Linear Models where the mean depends on two parameters and two explanatory variables. In all of the cases considered the support point of the optimal designs are found to be independent of the unknown parameters. While in some cases the optimal design measures are given by two points with equal weights, in others the support is given by three point with weights depending on the unknown parameters, hence the designs are locally optimal in general. Empirical results on the efficiency of the locally optimal designs are also given. Some of the designs found can also be used for planning D-optimal experiments for the normal linear model, where the mean must be positive. This research was carried out in part at University College, London as an M.Sc. project. Thanks are due to Prof. I. Ford (University of Glasgow) and Prof. A. Giovagnoli (University of Perugia) for their valuable suggestions and critical observations.  相似文献   
9.
Sample selection in radiocarbon dating   总被引:1,自引:0,他引:1  
Archaeologists working on the island of O'ahu, Hawai'i, use radiocarbon dating of samples of organic matter found trapped in fish-pond sediments to help them to learn about the chronology of the construction and use of the aquicultural systems created by the Polynesians. At one particular site, Loko Kuwili, 25 organic samples were obtained and funds were available to date an initial nine. However, on calibration to the calendar scale, the radiocarbon determinations provided date estimates that had very large variances. As a result, major issues of chronology remained unresolved and the archaeologists were faced with the prospect of another expensive programme of radiocarbon dating. This paper presents results of research that tackles the problems associated with selecting samples from those which are still available. Building on considerable recent research that utilizes Markov chain Monte Carlo methods to aid archaeologists in their radiocarbon calibration and interpretation, we adopt the standard Bayesian framework of risk functions, which allows us to assess the optimal samples to be sent for dating. Although rather computer intensive, our algorithms are simple to implement within the Bayesian radiocarbon framework that is already in place and produce results that are capable of direct interpretation by the archaeologists. By dating just three more samples from Loko Kuwili the expected variance on the date of greatest interest could be substantially reduced.  相似文献   
10.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号