首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 78 毫秒
1.
In this paper, we develop a family of bivariate beta distributions that encapsulate both positive and negative correlations, and which can be of general interest for Bayesian inference. We then invoke a use of these bivariate distributions in two contexts. The first is diagnostic testing in medicine, threat detection and signal processing. The second is system survivability assessment, relevant to engineering reliability and to survival analysis in biomedicine. In diagnostic testing, one encounters two parameters that characterize the efficacy of the testing mechanism: test sensitivity and test specificity. These tend to be adversarial when their values are interpreted as utilities. In system survivability, the parameters of interest are the component reliabilities, whose values when interpreted as utilities tend to exhibit co‐operative (amiable) behavior. Besides probability modeling and Bayesian inference, this paper has a foundational import. Specifically, it advocates a conceptual change in how one may think about reliability and survival analysis. The philosophical writings of de Finetti, Kolmogorov, Popper and Savage, when brought to bear on these topics constitute the essence of this change. Its consequence is that we have at hand a defensible framework for invoking Bayesian inferential methods in diagnostics, reliability and survival analysis. Another consequence is a deeper appreciation of the judgment of independent lifetimes. Specifically, we make the important point that independent lifetimes entail at a minimum, a two‐stage hierarchical construction.  相似文献   

2.
p‐Values are commonly transformed to lower bounds on Bayes factors, so‐called minimum Bayes factors. For the linear model, a sample‐size adjusted minimum Bayes factor over the class of g‐priors on the regression coefficients has recently been proposed (Held & Ott, The American Statistician 70(4), 335–341, 2016). Here, we extend this methodology to a logistic regression to obtain a sample‐size adjusted minimum Bayes factor for 2 × 2 contingency tables. We then study the relationship between this minimum Bayes factor and two‐sided p‐values from Fisher's exact test, as well as less conservative alternatives, with a novel parametric regression approach. It turns out that for all p‐values considered, the maximal evidence against the point null hypothesis is inversely related to the sample size. The same qualitative relationship is observed for minimum Bayes factors over the more general class of symmetric prior distributions. For the p‐values from Fisher's exact test, the minimum Bayes factors do on average not tend to the large‐sample bound as the sample size becomes large, but for the less conservative alternatives, the large‐sample behaviour is as expected.  相似文献   

3.
This paper reviews the problems of project management as reported by 1,879 individuals employed in the construction, manufacturing, process, and service industries in North‐West Europe over 23 years. Data obtained by a written questionnaire confirms that organizational relationships need the most attention. The data exhibits no trends over the 23 years. It indicates that many of the problems of project management are “institutional”; in other words, they are due to organizational cultures rather than inherent in project work. The evidence supports the view that poor discipline at the front‐end of projects results in too much “fire‐fighting” in project execution to overcome problems that could have been prevented. Prevention of these institutional problems requires action at the level of corporate governance, above the level of project management.  相似文献   

4.
In this paper, we use the local influence method to study a vector autoregressive model under Students t‐distributions. We present the maximum likelihood estimators and the information matrix. We establish the normal curvature diagnostics for the vector autoregressive model under three usual perturbation schemes for identifying possible influential observations. The effectiveness of the proposed diagnostics is examined by a simulation study, followed by our data analysis using the model to fit the weekly log returns of Chevron stock and the Standard & Poor's 500 Index as an application.  相似文献   

5.
We offer an exposition of modern higher order likelihood inference and introduce software to implement this in a quite general setting. The aim is to make more accessible an important development in statistical theory and practice. The software, implemented in an R package, requires only that the user provide code to compute the likelihood function and to specify extra‐likelihood aspects of the model, such as stopping rule or censoring model, through a function generating a dataset under the model. The exposition charts a narrow course through the developments, intending thereby to make these more widely accessible. It includes the likelihood ratio approximation to the distribution of the maximum likelihood estimator, that is the p? formula, and the transformation of this yielding a second‐order approximation to the distribution of the signed likelihood ratio test statistic, based on a modified signed likelihood ratio statistic r?. This follows developments of Barndorff‐Nielsen and others. The software utilises the approximation to required Jacobians as developed by Skovgaard, which is included in the exposition. Several examples of using the software are provided.  相似文献   

6.
The focus of this article is modeling the magnitude and duration of monotone periods of log‐returns. For this, we propose a new bivariate law assuming that the probabilistic framework over the magnitude and duration is based on the joint distribution of (X,N), where N is geometric distributed and X is the sum of an identically distributed sequence of inverse‐Gaussian random variables independent of N. In this sense, X and N represent the magnitude and duration of the log‐returns, respectively, and the magnitude comes from an infinite mixture of inverse‐Gaussian distributions. This new model is named bivariate inverse‐Gaussian geometric ( in short) law. We provide statistical properties of the model and explore stochastic representations. In particular, we show that the is infinitely divisible, and with this, an induced Lévy process is proposed and studied in some detail. Estimation of the parameters is performed via maximum likelihood, and Fisher's information matrix is obtained. An empirical illustration to the log‐returns of Tyco International stock demonstrates the superior performance of the law compared to an existing model. We expect that the proposed law can be considered as a powerful tool in the modeling of log‐returns and other episodes analyses such as water resources management, risk assessment, and civil engineering projects.  相似文献   

7.
In the case of two independent samples, it turns out that among the procedures taken in consideration, BOSCHLOO'S technique of raising the nominal level in the standard conditional test as far as admissible performs best in terms of power against almost all alternatives. The computational burden entailed in exact sample size calculation is comparatively modest for both the uniformly most powerful unbiased randomized and the conservative non‐randomized version of the exact Fisher‐type test. Computing these values yields a pair of bounds enclosing the exact sample size required for the Boschloo test, and it seems reasonable to replace the exact value with the middle of the corresponding interval. Comparisons between these mid‐N estimates and the fully exact sample sizes lead to the conclusion that the extra computational effort required for obtaining the latter is mostly dispensable. This holds also true in the case of paired binary data (McNemar setting). In the latter, the level‐corrected score test turns out to be almost as powerful as the randomized uniformly most powerful unbiased test and should be preferred to the McNemar–Boschloo test. The mid‐N rule provides a fairly tight upper bound to the exact sample size for the score test for paired proportions.  相似文献   

8.
We review some first‐order and higher‐order asymptotic techniques for M‐estimators, and we study their stability in the presence of data contaminations. We show that the estimating function (ψ) and its derivative with respect to the parameter play a central role. We discuss in detail the first‐order Gaussian density approximation, saddlepoint density approximation, saddlepoint test, tail area approximation via the Lugannani–Rice formula and empirical saddlepoint density approximation (a technique related to the empirical likelihood method). For all these asymptotics, we show that a bounded ψ (in the Euclidean norm) and a bounded (e.g. in the Frobenius norm) yield stable inference in the presence of data contamination. We motivate and illustrate our findings by theoretical and numerical examples about the benchmark case of one‐dimensional location model.  相似文献   

9.
Social and economic studies are often implemented as complex survey designs. For example, multistage, unequal probability sampling designs utilised by federal statistical agencies are typically constructed to maximise the efficiency of the target domain level estimator (e.g. indexed by geographic area) within cost constraints for survey administration. Such designs may induce dependence between the sampled units; for example, with employment of a sampling step that selects geographically indexed clusters of units. A sampling‐weighted pseudo‐posterior distribution may be used to estimate the population model on the observed sample. The dependence induced between coclustered units inflates the scale of the resulting pseudo‐posterior covariance matrix that has been shown to induce under coverage of the credibility sets. By bridging results across Bayesian model misspecification and survey sampling, we demonstrate that the scale and shape of the asymptotic distributions are different between each of the pseudo‐maximum likelihood estimate (MLE), the pseudo‐posterior and the MLE under simple random sampling. Through insights from survey‐sampling variance estimation and recent advances in computational methods, we devise a correction applied as a simple and fast postprocessing step to Markov chain Monte Carlo draws of the pseudo‐posterior distribution. This adjustment projects the pseudo‐posterior covariance matrix such that the nominal coverage is approximately achieved. We make an application to the National Survey on Drug Use and Health as a motivating example and we demonstrate the efficacy of our scale and shape projection procedure on synthetic data on several common archetypes of survey designs.  相似文献   

10.
This paper concerns a class of model selection criteria based on cross‐validation techniques and estimative predictive densities. Both the simple or leave‐one‐out and the multifold or leave‐m‐out cross‐validation procedures are considered. These cross‐validation criteria define suitable estimators for the expected Kullback–Liebler risk, which measures the expected discrepancy between the fitted candidate model and the true one. In particular, we shall investigate the potential bias of these estimators, under alternative asymptotic regimes for m. The results are obtained within the general context of independent, but not necessarily identically distributed, observations and by assuming that the candidate model may not contain the true distribution. An application to the class of normal regression models is also presented, and simulation results are obtained in order to gain some further understanding on the behavior of the estimators.  相似文献   

11.
In many manufacturing and service industries, the quality department of the organization works continuously to ensure that the mean or location of the process is close to the target value. In order to understand the process, it is necessary to provide numerical statements of the processes that are being investigated. That is why the researcher needs to check the validity of the hypotheses that are concerned with some physical phenomena. It is usually assumed that the collected data behave well. However, sometimes the data may contain outliers. The presence of one or more outliers might seriously distort the statistical inference. Since the sample mean is very sensitive to outliers, this research will use the smooth adaptive (SA) estimator to estimate the population mean. The SA estimator will be used to construct testing procedures, called smooth adaptive test (SA test), for testing various null hypotheses. A Monte Carlo study is used to simulate the values of the probability of a Type I error and the power of the SA test. This is accomplished by constructing confidence intervals of the process mean by using the SA estimator and bootstrap methods. The SA test will be compared with other tests such as the normal test, t test and a nonparametric statistical method, namely, the Wilcoxon signed-rank test. Also, the cases with and without outliers will be considered. For the right-skewed distributions, the SA test is the best choice. When the population is a right-skewed distribution with one outlier, the SA test controls the probability of a Type I error better than other tests and is recommended.  相似文献   

12.
During the last three decades, integer‐valued autoregressive process of order p [or INAR(p)] based on different operators have been proposed as a natural, intuitive and maybe efficient model for integer‐valued time‐series data. However, this literature is surprisingly mute on the usefulness of the standard AR(p) process, which is otherwise meant for continuous‐valued time‐series data. In this paper, we attempt to explore the usefulness of the standard AR(p) model for obtaining coherent forecasting from integer‐valued time series. First, some advantages of this standard Box–Jenkins's type AR(p) process are discussed. We then carry out our some simulation experiments, which show the adequacy of the proposed method over the available alternatives. Our simulation results indicate that even when samples are generated from INAR(p) process, Box–Jenkins's model performs as good as the INAR(p) processes especially with respect to mean forecast. Two real data sets have been employed to study the expediency of the standard AR(p) model for integer‐valued time‐series data.  相似文献   

13.
A test statistic is considered for testing a hypothesis for the mean vector for multivariate data, when the dimension of the vector, p, may exceed the number of vectors, n, and the underlying distribution need not necessarily be normal. With n,p→∞, and under mild assumptions, but without assuming any relationship between n and p, the statistic is shown to asymptotically follow a chi‐square distribution. A by product of the paper is the approximate distribution of a quadratic form, based on the reformulation of the well‐known Box's approximation, under high‐dimensional set up. Using a classical limit theorem, the approximation is further extended to an asymptotic normal limit under the same high dimensional set up. The simulation results, generated under different parameter settings, are used to show the accuracy of the approximation for moderate n and large p.  相似文献   

14.
Empirical count data are often zero‐inflated and overdispersed. Currently, there is no software package that allows adequate imputation of these data. We present multiple‐imputation routines for these kinds of count data based on a Bayesian regression approach or alternatively based on a bootstrap approach that work as add‐ons for the popular multiple imputation by chained equations (mice ) software in R (van Buuren and Groothuis‐Oudshoorn , Journal of Statistical Software, vol. 45, 2011, p. 1). We demonstrate in a Monte Carlo simulation that our procedures are superior to currently available count data procedures. It is emphasized that thorough modeling is essential to obtain plausible imputations and that model mis‐specifications can bias parameter estimates and standard errors quite noticeably. Finally, the strengths and limitations of our procedures are discussed, and fruitful avenues for future theory and software development are outlined.  相似文献   

15.
W. L. Pearn  Chien-Wei Wu 《Metrika》2005,61(2):221-234
Process capability indices have been proposed in the manufacturing industry to provide numerical measures on process reproduction capability, which are effective tools for quality assurance and guidance for process improvement. In process capability analysis, the usual practice for testing capability indices from sample data are based on traditional distribution frequency approach. Bayesian statistical techniques are an alternative to the frequency approach. Shiau, Chiang and Hung (1999) applied Bayesian method to index Cpm and the index Cpk but under the restriction that the process mean μ equals to the midpoint of the two specification limits, m. We note that this restriction is a rather impractical assumption for most factory applications, since in this case Cpk will reduce to Cp. In this paper, we consider testing the most popular capability index Cpk for general situation – no restriction on the process mean based on Bayesian approach. The results obtained are more general and practical for real applications. We derive the posterior probability, p, for which the process under investigation is capable and propose accordingly a Bayesian procedure for capability testing. To make this Bayesian procedure practical for in-plant applications, we tabulate the minimum values of Ĉpk for which the posterior probability p reaches desirable confidence levels with various pre-specified capability levels.  相似文献   

16.
The reappraisal of the post‐Soviet landscape is in danger of overlooking two of its most important elements: firstly, the mass modernist housing that was more extensive here than probably anywhere else; and secondly, the post‐1989 capitalist context of property speculation, office development and decay. These routinely missed landscapes constitute the very things travelled through on the way to utopian, if ruined, monuments, such as those documented in Frédéric Chaubin's CCCP — Cosmic Communist Constructions Photographed. When visited, the surroundings of these structures turn out to be at least as interesting as the photogenic modernist monument itself. This essay is an account of a visit to one of the most architecturally contemporary of these structures — the Park of Memory crematoria in the Ukrainian capital, Kiev, designed by Abraham Miletsky in 1974. In Chaubin's photographs, the curling concrete volumes of the Park's central crematoria are flamboyant, fantastical and self‐referential, the very ‘iconic’ architecture that many post‐Soviet capitals would like to have in order to attract tourists. There is a lot more going on in the surrounding city than what is typically recorded in its visual representations, however, as discussed in this essay. Such monuments are not mute, and cannot be severed from their surroundings.  相似文献   

17.
Survey statisticians use either approximate or optimisation‐based methods to stratify finite populations. Examples of the former are the cumrootf (Dalenius & Hodges, 1957 ) and geometric (Gunning & Horgan, 2004 ) methods, while examples of the latter are Sethi ( 1963 ) and Kozak ( 2004 ) algorithms. The approximate procedures result in inflexible stratum boundaries; this lack of flexibility results in non‐optimal boundaries. On the other hand, optimisation‐based methods provide stratum boundaries that can simultaneously account for (i) a chosen allocation scheme, (ii) overall sample size or required reliability of the estimator of a studied parameter and (iii) presence or absence of a take‐all stratum. Given these additional conditions, optimisation‐based methods will result in optimal boundaries. The only disadvantage of these methods is their complexity. However, in the second decade of 21st century, this complexity does not actually pose a problem. We illustrate how these two groups of methods differ by comparing their efficiency for two artificial populations and a real population. Our final point is that statistical offices should prefer optimisation‐based over approximate stratification methods; such a decision will help them either save much public money or, if funds are already allocated to a survey, result in more precise estimates of national statistics.  相似文献   

18.
This article reveals how newcomers weave their own threads into the fabric of urban infrastructure. Entangling their own with other urban assemblages, newcomers generate multi‐layered dynamics situationally in order to render possible the lives to which they aspire. They forge openings where there seemed none before and keep negative potentialities in check. To offer an ethnography of how the Senegalese presence in Rio de Janeiro has grown dynamically between 2014 and 2019, I draw analytical strength from the double meaning of agencement: the action of interweaving varied socio‐material components—agencer—so that they work together well, and the resulting assemblage of social and material components. Two case studies act as a starting point: how Senegalese came to inhabit an urban architectural landmark and how they regularize their residence status. Their transformative power of city‐making is generated both through the mutual intertwining of a dahira, a religious group of Senegalese migrants, and a diasporic Senegalese association and through the ways in which the Senegalese interweave themselves and their institutionalized collective forms with ever more socio‐material components of the urban space. Beyond the better‐known transnational embeddedness of the Senegalese, their complex infrastructuring practices upon arrival become constitutive of new urban realities, moulding the city fabric of which they are becoming part.  相似文献   

19.
Typical data that arise from surveys, experiments, and observational studies include continuous and discrete variables. In this article, we study the interdependence among a mixed (continuous, count, ordered categorical, and binary) set of variables via graphical models. We propose an ?1‐penalized extended rank likelihood with an ascent Monte Carlo expectation maximization approach for the copula Gaussian graphical models and establish near conditional independence relations and zero elements of a precision matrix. In particular, we focus on high‐dimensional inference where the number of observations are in the same order or less than the number of variables under consideration. To illustrate how to infer networks for mixed variables through conditional independence, we consider two datasets: one in the area of sports and the other concerning breast cancer.  相似文献   

20.
The asymptotic approach and Fisher's exact approach have often been used for testing the association between two dichotomous variables. The asymptotic approach may be appropriate to use in large samples but is often criticized for being associated with unacceptable high actual type I error rates for small to medium sample sizes. Fisher's exact approach suffers from conservative type I error rates and low power. For these reasons, a number of exact unconditional approaches have been proposed, which have been seen to be generally more powerful than exact conditional counterparts. We consider the traditional unconditional approach based on maximization and compare it to our presented approach, which is based on estimation and maximization. We extend the unconditional approach based on estimation and maximization to designs with the total sum fixed. The procedures based on the Pearson chi‐square, Yates's corrected, and likelihood ratio test statistics are evaluated with regard to actual type I error rates and powers. A real example is used to illustrate the various testing procedures. The unconditional approach based on estimation and maximization performs well, having an actual level much closer to the nominal level. The Pearson chi‐square and likelihood ratio test statistics work well with this efficient unconditional approach. This approach is generally more powerful than the other p‐value calculation methods in the scenarios considered.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号