共查询到20条相似文献,搜索用时 0 毫秒
1.
Felix Famoye 《Statistica Neerlandica》2010,64(1):112-124
In this paper, a new bivariate generalized Poisson distribution (GPD) that allows any type of correlation is defined and studied. The marginal distributions of the bivariate model are the univariate GPDs. The parameters of the bivariate distribution are estimated by using the moment and maximum likelihood methods. Some test statistics are discussed and one numerical data set is used to illustrate the applications of the bivariate model. 相似文献
2.
The applied econometrics of bivariate count data predominantly focus on a bivariate Poisson density with a correlation structure that is very restrictive. The main limitation is that this bivariate distribution excludes zero and negative correlation. This paper introduces a new model which allows for a more flexible correlation structure. To this end the joint density is decomposed by means of the multiplication rule in marginal and conditional densities. Simulation experiments and an application of the model to recreational data are presented. 相似文献
3.
Extensions of the Cox proportional hazards model for survival data are studied where allowance is made for unobserved heterogeneity and for correlation between the life times of several individuals. The extended models are frailty models inspired by Y ashin et al. (1995). Estimation is carried out using the EM algorithm. Inference is discussed and potential applications are outlined, in particular to statistical research in human genetics using twin data or adoption data, aimed at separating the effects of genetic and environmental factors on mortality. 相似文献
4.
Empirical count data are often zero‐inflated and overdispersed. Currently, there is no software package that allows adequate imputation of these data. We present multiple‐imputation routines for these kinds of count data based on a Bayesian regression approach or alternatively based on a bootstrap approach that work as add‐ons for the popular multiple imputation by chained equations (mice ) software in R (van Buuren and Groothuis‐Oudshoorn , Journal of Statistical Software, vol. 45, 2011, p. 1). We demonstrate in a Monte Carlo simulation that our procedures are superior to currently available count data procedures. It is emphasized that thorough modeling is essential to obtain plausible imputations and that model mis‐specifications can bias parameter estimates and standard errors quite noticeably. Finally, the strengths and limitations of our procedures are discussed, and fruitful avenues for future theory and software development are outlined. 相似文献
5.
A statistical test for the degree of overdispersion of count data time series based on the empirical version of the (Poisson) index of dispersion is considered. The test design relies on asymptotic properties of this index of dispersion, which in turn have been analyzed for time series stemming from a compound Poisson (Poisson‐stopped sum) INAR(1) model. This approach is extended to the popular Poisson INARCH(1) model, which exhibits unconditional overdispersion but has an (equidispersed) conditional Poisson distribution. The asymptotic distribution of the index of dispersion if applied to time series stemming from such a model is derived. These results allow us to investigate the ability of the dispersion test to discriminate between Poisson INAR(1) and INARCH(1) models. Furthermore, the question is considered if the index of dispersion could be used to test the null of a Poisson INARCH(1) model against the alternative of an INARCH(1) model with additional conditional overdispersion. 相似文献
6.
Tamás Rudas 《Metrika》1999,50(2):163-172
A measure of the fit of a statistical model can be obtained by estimating the relative size of the largest fraction of the population where a distribution belonging to the model may be valid. This is the mixture index of fit that was suggested for models for contingency tables by Rudas, Clogg, Lindsay (1994) and it is extended here for models involving continuous observations. In particular, the approach is applied to regression models with normal and uniform error structures. Best fit, as measured by the mixture index of fit, is obtained with minimax estimation of the regression parameters. Therefore, whenever minimax estimation is used for these problems, the mixture index of fit provides a natural approach for measuring model fit and for variable selection. Received: September 1997 相似文献
7.
Comparing occurrence rates of events of interest in science, business, and medicine is an important topic. Because count data are often under‐reported, we desire to account for this error in the response when constructing interval estimators. In this article, we derive a Bayesian interval for the difference of two Poisson rates when counts are potentially under‐reported. The under‐reporting causes a lack of identifiability. Here, we use informative priors to construct a credible interval for the difference of two Poisson rate parameters with under‐reported data. We demonstrate the efficacy of our new interval estimates using a real data example. We also investigate the performance of our newly derived Bayesian approach via simulation and examine the impact of various informative priors on the new interval. 相似文献
8.
Fit is generally conceptualized as a dynamic construct, but most research on person-environment fit has focused on fit in the current moment. We addressed this oversight by examining the dynamic relationships among person-job (PJ) fit, demand-ability (DA) fit, need-supply (NS) fit, and employee work attitudes over time using a three-wave survey design over a 12-month period. Results from 168 employees revealed that change in PJ fit was significantly related to changes in job satisfaction and affective organizational commitment. In addition, DA and NS fit changes were significantly and indirectly associated with job satisfaction and commitment changes through PJ fit change. We also found that increases in job demands and employee abilities significantly decreased DA fit, and increases in employee needs significantly decreased NS fit whereas increases in job supplies significantly increased NS fit. Finally, we examined age as an important moderator for employees’ reactions to PJ fit changes, and found that younger employees reacted more strongly to increases/decreases in PJ fit than did older employees. 相似文献
9.
10.
作为维系员工与组织关系的心理纽带,员工心理契约的主观期望与实际感知的吻合程度直接影响到员工对于组织的态度与行为。以往文献中对员工心理契约主观期望与实际感知吻合程度的研究缺乏系统性。本文引入心理契约契合度概念,对员工心理契约主观期望与实际感知吻合程度进行测量,提出了构建心理契约契合度评价指标体系的思路及测量方法,最后结合某军工研究所某研发团队的实际情况,对成员心理契约契合度进行了评价。该研究为组织衡量员工心理契约主观期望与实际感知一致性的度量提供了科学依据,增强了心理契约在实际管理工作中的应用性。 相似文献
11.
Forecast error is not only caused by the randomness of the data-generating process but also by the uncertainty due to estimated model parameters. We investigate these different sources of forecast error for a popular type of count process, the Poisson first-order integer-valued autoregressive (INAR(1)) process. However, many of our analytical derivations also hold for the more general family of conditional linear AR(1) (CLAR(1)) processes. In addition, results from a simulation study are presented, to verify and complement our asymptotic approximations. 相似文献
12.
价格离散是衡量市场整合的重要指标,为全国统一大市场研究提供新的切入点。通过采集同一商品在不同平台商家的日度价格数据,对中国价格离散程度进行测度。结果表明:(1)中国线上市场的价格离散程度总体处于较高水平,且不同类别商品间存在显著异质性。基于此构建的高频价格离散指数(iPDI)发现,指数越低,说明市场统一程度越高,反之统一程度越低。(2)商品价格方差分解发现,价格离散主要源自跨平台竞争,并且构建实证模型,得出跨平台竞争对价格离散呈倒U型关系。(3)将价格离散的微观测度结果应用于宏观层面,基于价格离散特征编制的居民消费价格指数(iCPI_pd),更准确地反映商品价格的波动特征。 相似文献
13.
Helmert(赫尔墨特)方差分量估计作为近代平差随机模型的验后估计,可以准确给出各类观测量之间的权比,提高平差结果的可靠性。文章就该模型在不同观测类型平差数据中的算法进行了探讨,在实际应用中具有重要的参考价值。 相似文献
14.
Tobias Rydén 《Metrika》1998,47(1):119-145
For a recursive maximum-likelihood estimator with step lengths decaying as 1/n, an adaptive matrix needs to be incorporated to obtain asymptotic efficiency. Ideally, this matrix should be chosen as the inverse Fisher information matrix, which is usually very difficult to compute for incomplete data models. In this paper we give conditions under which the observed information can be incorporated into the recursive procedure to yield an efficient estimator, and we also investigate the finite sample properties of these estimators by simulation. 相似文献
15.
《International Journal of Forecasting》2020,36(3):1116-1127
As the volume and complexity of data continues to grow, more attention is being focused on solving so-called big data problems. One field where this focus is pertinent is credit card fraud detection. Model selection approaches can identify key predictors for preventing fraud. Stagewise Selection is a classic model selection technique that has experienced a revitalized interest due to its computational simplicity and flexibility. Over a sequence of simple learning steps, stagewise techniques build a sequence of candidate models that is less greedy than the stepwise approach.This paper introduces a new stochastic stagewise technique that integrates a sub-sampling approach into the stagewise framework, yielding a simple tool for model selection when working with big data. Simulation studies demonstrate the proposed technique offers a reasonable trade off between computational cost and predictive performance. We apply the proposed approach to synthetic credit card fraud data to demonstrate the technique’s application. 相似文献
16.
This paper proposes a general framework for the analysis of survey data with missing observations. The approach presented here treats missing data as an unavoidable feature of any survey of the human population and aims at incorporating the unobserved part of the data into the analysis rather than trying to avoid it or make up for it. To handle coverage error and unit non-response, the true distribution is modeled as a mixture of an observable and of an unobservable component. Generally, for the unobserved component, its relative size (the no-observation rate) and its distribution are not known. It is assumed that the goal of the analysis is to assess the fit of a statistical model, and for this purpose the mixture index of fit is used. The mixture index of fit does not postulate that the statistical model of interest is able to account for the entire population rather, that it may only describe a fraction of it. This leads to another mixture representation of the true distribution, with one component from the statistical model of interest and another unrestricted one. Inference with respect to the fit of the model, with missing data taken into account, is obtained by equating these two mixtures and asking, for different no-observation rates, what is the largest fraction of the population where the statistical model may hold. A statistical model is deemed relevant for the population, if it may account for a large enough fraction of the population, assuming the true (if known) or a sufficiently small or a realistic no-observation rate. 相似文献
17.
Abstract. A large number of different Pseudo- R 2 measures for some common limited dependent variable models are surveyed. Measures include those based solely on the maximized likelihoods with and without the restriction that slope coefficients are zero, those which require further calculations based on parameter estimates of the coefficients and variances and those that are based solely on whether the qualitative predictions of the model are correct or not. The theme of the survey is that while there is no obvious criterion for choosing which Pseudo- R 2 to use, if the estimation is in the context of an underlying latent dependent variable model, a case can be made for basing the choice on the strength of the numerical relationship to the OLS- R 2 in the latent dependent variable. As such an OLS- R 2 can be known in a Monte Carlo simulation, we summarize Monte Carlo results for some important latent dependent variable models (binary probit, ordinal probit and Tobit) and find that a Pseudo- R 2 measure due to McKelvey and Zavoina scores consistently well under our criterion. We also very briefly discuss Pseudo- R 2 measures for count data, for duration models and for prediction-realization tables. 相似文献
18.
Eugene Demidenko 《Revue internationale de statistique》2007,75(1):96-113
We compare five methods for parameter estimation of a Poisson regression model for clustered data: (1) ordinary (naive) Poisson regression (OP), which ignores intracluster correlation, (2) Poisson regression with fixed cluster‐specific intercepts (FI), (3) a generalized estimating equations (GEE) approach with an equi‐correlation matrix, (4) an exact generalized estimating equations (EGEE) approach with an exact covariance matrix, and (5) maximum likelihood (ML). Special attention is given to the simplest case of the Poisson regression with a cluster‐specific intercept random when the asymptotic covariance matrix is obtained in closed form. We prove that methods 1–5, except GEE, produce the same estimates of slope coefficients for balanced data (an equal number of observations in each cluster and the same vectors of covariates). All five methods lead to consistent estimates of slopes but have different efficiency for unbalanced data design. It is shown that the FI approach can be derived as a limiting case of maximum likelihood when the cluster variance increases to infinity. Exact asymptotic covariance matrices are derived for each method. In terms of asymptotic efficiency, the methods split into two groups: OP & GEE and EGEE & FI & ML. Thus, contrary to the existing practice, there is no advantage in using GEE because it is substantially outperformed by EGEE and FI. In particular, EGEE does not require integration and is easy to compute with the asymptotic variances of the slope estimates close to those of the ML. 相似文献
19.
Bahadır Yüzbaşı Mohammad Arashi S. Ejaz Ahmed 《Revue internationale de statistique》2020,88(1):229-251
In this study, we suggest pretest and shrinkage methods based on the generalised ridge regression estimation that is suitable for both multicollinear and high-dimensional problems. We review and develop theoretical results for some of the shrinkage estimators. The relative performance of the shrinkage estimators to some penalty methods is compared and assessed by both simulation and real-data analysis. We show that the suggested methods can be accounted as good competitors to regularisation techniques, by means of a mean squared error of estimation and prediction error. A thorough comparison of pretest and shrinkage estimators based on the maximum likelihood method to the penalty methods. In this paper, we extend the comparison outlined in his work using the least squares method for the generalised ridge regression. 相似文献
20.
We develop a generalized method of moments (GMM) estimator for the distribution of a variable where summary statistics are available only for intervals of the random variable. Without individual data, one cannot calculate the weighting matrix for the GMM estimator. Instead, we propose a simulated weighting matrix based on a first-step consistent estimate. When the functional form of the underlying distribution is unknown, we estimate it using a simple yet flexible maximum entropy density. Our Monte Carlo simulations show that the proposed maximum entropy density is able to approximate various distributions extremely well. The two-step GMM estimator with a simulated weighting matrix improves the efficiency of the one-step GMM considerably. We use this method to estimate the U.S. income distribution and compare these results with those based on the underlying raw income data. 相似文献