首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 9 毫秒
1.
A condition is given by which optimal normal theory methods, such as the maximum likelihood methods, are robust against violation of the normality assumption in a general linear structural equation model. Specifically, the estimators and the goodness of fit test are robust. The estimator is efficient within some defined class, and its standard errors can be obtained by a correction formula applied to the inverse of the information matrix. Some special models, like the factor analysis model and path models, are discussed in more detail. A method for evaluating the robustness condition is given.  相似文献   

2.
The main purpose of this paper is to unify and extend the existing theory of 'estimated zeroes' in log-linear and logit models. To this end it is shown that every generalized linear model (GLM) can be embedded in a larger model with a compact parameter space and a continuous likelihood (a 'CGLM'). Clearly in a CGLM the maximum likelihood estimate (MLE) always exists, easing a major data analysis problem. In the mean-value parametrization, the construction of the CGLM is remarkably simple; except in a rather pathological and rare case, the estimated expected values are always finite., In the β-parametrization however, the compactification is more complex; the MLE need not correspond with a finite β, as is well known for estimated zeros in log-linear models. The boundary distributions of CGLMs are classified in four categories: 'Inadmissible', 'degenerate', 'Chentsov', and 'constrained'. For a large class of GLMs, including all GLMs with canonical link functions and probit models, the MLE in the corresponding CGLM exists and is unique. Even stronger, the likelihood has no other local maxima. We give equivalent algebraic and geometric conditions (in the vein of Haberman (1974, 1977) and Albert and Anderson (1984) respectively), necessary for the existence of the MLE in the GLM corresponding to a finite β. For a large class of GLMs these conditions are also sufficient. Even for log-linear models this seams to be a new result.  相似文献   

3.
Typical data that arise from surveys, experiments, and observational studies include continuous and discrete variables. In this article, we study the interdependence among a mixed (continuous, count, ordered categorical, and binary) set of variables via graphical models. We propose an ?1‐penalized extended rank likelihood with an ascent Monte Carlo expectation maximization approach for the copula Gaussian graphical models and establish near conditional independence relations and zero elements of a precision matrix. In particular, we focus on high‐dimensional inference where the number of observations are in the same order or less than the number of variables under consideration. To illustrate how to infer networks for mixed variables through conditional independence, we consider two datasets: one in the area of sports and the other concerning breast cancer.  相似文献   

4.
This article discusses modelling strategies for repeated measurements of multiple response variables. Such data arise in the context of categorical variables where one can select more than one of the categories as the response. We consider each of the multiple responses as a binary outcome and use a marginal (or population‐averaged) modelling approach to analyse its means. Generalized estimating equations are used to account for different correlation structures, both over time and between items. We also discuss an alternative approach using a generalized linear mixed model with conditional interpretations. We illustrate the methods using data from a panel study in Australia called the Household, Income, and Labour Dynamics Survey.  相似文献   

5.
Survival models allowing for random effects (e.g., frailty models) have been widely used for analyzing clustered time-to-event data. Accelerated failure time (AFT) models with random effects are useful alternatives to frailty models. Because survival times are directly modeled, interpretation of the fixed and random effects is straightforward. Moreover, the fixed effect estimates are robust against various violations of the assumed model. In this paper, we propose a penalized h-likelihood (HL) procedure for variable selection of fixed effects in the AFT random-effect models. For the purpose of variable selection, we consider three penalty functions, namely, least absolute shrinkage and selection operator (LASSO), smoothly clipped absolute deviation (SCAD), and HL. We demonstrate via simulation studies that the proposed variable selection procedure is robust against the misspecification of the assumed model. The proposed method is illustrated using data from a bladder cancer clinical trial.  相似文献   

6.
The construction of an importance density for partially non‐Gaussian state space models is crucial when simulation methods are used for likelihood evaluation, signal extraction, and forecasting. The method of efficient importance sampling is successful in this respect, but we show that it can be implemented in a computationally more efficient manner using standard Kalman filter and smoothing methods. Efficient importance sampling is generally applicable for a wide range of models, but it is typically a custom‐built procedure. For the class of partially non‐Gaussian state space models, we present a general method for efficient importance sampling. Our novel method makes the efficient importance sampling methodology more accessible because it does not require the computation of a (possibly) complicated density kernel that needs to be tracked for each time period. The new method is illustrated for a stochastic volatility model with a Student's t distribution.  相似文献   

7.
In the presence of heteroskedasticity, conventional test statistics based on the ordinary least squares (OLS) estimator lead to incorrect inference results for the linear regression model. Given that heteroskedasticity is common in cross-sectional data, the test statistics based on various forms of heteroskedasticity-consistent covariance matrices (HCCMs) have been developed in the literature. In contrast to the standard linear regression model, heteroskedasticity is a more serious problem for spatial econometric models, generally causing inconsistent extremum estimators of model coefficients. This paper investigates the finite sample properties of the heteroskedasticity-robust generalized method of moments estimator (RGMME) for a spatial econometric model with an unknown form of heteroskedasticity. In particular, it develops various HCCM-type corrections to improve the finite sample properties of the RGMME and the conventional Wald test. The Monte Carlo results indicate that the HCCM-type corrections can produce more accurate results for inference on model parameters and the impact effects estimates in small samples.  相似文献   

8.
J. Ahmadi  N. R. Arghami 《Metrika》2001,53(3):195-206
In this article, we establish some general results concerning the comparison of the amount of the Fisher information contained in n record values with the Fisher information contained in n iid observations from the original distribution. Some common distributions are classified according to this criterion. We also propose some methods of estimation based on record values. The results may be of interest in some life testing problems. Received: September 1999  相似文献   

9.
For Poisson inverse Gaussian regression models, it is very complicated to obtain the influence measures based on the traditional method, because the associated likelihood function involves intractable expressions, such as the modified Bessel function. In this paper, the EM algorithm is employed as a basis to derive diagnostic measures for the models by treating them as a mixed Poisson regression with the weights from the inverse Gaussian distributions. Several diagnostic measures are obtained in both case-deletion model and local influence analysis, based on the conditional expectation of the complete-data log-likelihood function in the EM algorithm. Two numerical examples are given to illustrate the results.  相似文献   

10.
Phylogenetic trees are types of networks that describe the temporal relationship between individuals, species, or other units that are subject to evolutionary diversification. Many phylogenetic trees are constructed from molecular data that is often only available for extant species, and hence they lack all or some of the branches that did not make it into the present. This feature makes inference on the diversification process challenging. For relatively simple diversification models, analytical or numerical methods to compute the likelihood exist, but these do not work for more realistic models in which the likelihood depends on properties of the missing lineages. In this article, we study a general class of species diversification models, and we provide an expectation-maximization framework in combination with a uniform sampling scheme to perform maximum likelihood estimation of the parameters of the diversification process.  相似文献   

11.
This paper presents a method for fitting a copula‐driven generalized linear mixed models. For added flexibility, the skew‐normal copula is adopted for fitting. The correlation matrix of the skew‐normal copula is used to capture the dependence structure within units, while the fixed and random effects coefficients are estimated through the mean of the copula. For estimation, a Monte Carlo expectation–maximization algorithm is developed. Simulations are shown alongside a real data example from the Framingham Heart Study.  相似文献   

12.
In this paper we consider the problem of estimating nonparametric panel data models with fixed effects. We introduce an iterative nonparametric kernel estimator. We also extend the estimation method to the case of a semiparametric partially linear fixed effects model. To determine whether a parametric, semiparametric or nonparametric model is appropriate, we propose test statistics to test between the three alternatives in practice. We further propose a test statistic for testing the null hypothesis of random effects against fixed effects in a nonparametric panel data regression model. Simulations are used to examine the finite sample performance of the proposed estimators and the test statistics.  相似文献   

13.
For modelling the effect of crossed, fixed factors on the response variable in balanced designs with nested stratifications, a generalized linear mixed model is proposed. This model is based on a set of quasi-likelihood assumptions which imply quadratic variance functions. From these variance functions, deviances are obtained to quantify the variation per stratification. The effects of the fixed factors will be tested, an dispersion components will be estimated. The practical use of the model is illustrated by reanalysing a soldering failures problem.  相似文献   

14.
In this study Variance-Gamma (VG) and Normal-Inverse Gaussian (NIG) distributions are compared with the benchmark of generalized hyperbolic distribution in terms of their fit to the empirical distribution of high-frequency stock market index returns in China. First, we estimate the considered models in a Markov regime switching framework for the identification of different volatility regimes. Second, the goodness-of-fit results are compared at different time scales of log-returns. Third, the goodness-of-fit results are validated through bootstrapping experiments. Our results show that as the time scale of log-returns decrease NIG model outperforms the VG model consistently and the difference between the goodness-of-fit statistics increase. For high-frequency Chinese index returns, NIG model is more robust and provides a better fit to the empirical distributions of returns at different time scales.  相似文献   

15.
We introduce a new family of network models, called hierarchical network models, that allow us to represent in an explicit manner the stochastic dependence among the dyads (random ties) of the network. In particular, each member of this family can be associated with a graphical model defining conditional independence clauses among the dyads of the network, called the dependency graph. Every network model with dyadic independence assumption can be generalized to construct members of this new family. Using this new framework, we generalize the Erdös–Rényi and the β models to create hierarchical Erdös–Rényi and β models. We describe various methods for parameter estimation, as well as simulation studies for models with sparse dependency graphs.  相似文献   

16.
This article treats the analysis of 'time-series–cross-section' (TSCS) data. Such data consists of repeated observations on a series of fixed units. Examples of such data are annual observations on the political economy of OECD nations in the post-war era. TSCS data is distinguished from 'panel' data, in that asymptotics are in the number of repeated observations, not the number of units.
The article begins by treating the complications of TSCS data in an 'old-fashioned' manner, that is, as a nuisance which causes estimation difficulties. It claims that TSCS data should be analyzed via ordinary least squares with 'panel correct standard errors' rather than generalized least squares methods. Dynamics should be modeled via a lagged dependent variable or, if appropriate, a single equation error correction model.
The article then treats more modern issues, in particular, the modeling of spatial effects and heterogeneity. It also claims that heterogeneity should be assessed with 'panel cross-validation' as well as more standard tests. The article concludes with a discussion of estimation in the presence of a binary dependent variable.  相似文献   

17.
In-depth data analysis plus statistical modeling can produce inferentialcausal models. Their creation thus combines aspects of analysis by close inspection,that is, reason analysis and cross-tabular analysis, with statistical analysis procedures,especially those that are special cases of the generalized linear model (McCullaghand Nelder, 1989; Agresti, 1996; Lindsey, 1997). This paper explores some of the roots of this combined method and suggests some new directions. An exercise clarifies some limitations of classic reason analysis by showing how the cross tabulation of variables with controls for test factors may produce better inferences. Then, given the cross tabulation of several variables, by explicating Coleman effect parameters, logistic regressions, and Poisson log-linear models, it shows how generalized linear models provide appropriate measures of effects and tests of statistical significance. Finally, to address a weakness of reason analysis, a case-control design is proposed and an example is developed.  相似文献   

18.
19.
In this article, we propose a new identifiability condition by using the logarithmic calibration for the distortion measurement error models, where neither the response variable nor the covariates can be directly observed but are measured with multiplicative measurement errors. Under the logarithmic calibration, the direct-plug-in estimators of parameters and empirical likelihood based confidence intervals are proposed, and we studied the asymptotic properties of the proposed estimators. For the hypothesis testing of parameter, a restricted estimator under the null hypothesis and a test statistic are proposed. The asymptotic properties for the restricted estimator and test statistic are established. Simulation studies demonstrate the performance of the proposed procedure and a real example is analyzed to illustrate its practical usage.  相似文献   

20.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号