首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Our paper estimates the effect of US internal migration on wage growth for young men between their first and second job. Our analysis of migration extends previous research by: (i) exploiting the distance-based measures of migration in the National Longitudinal Surveys of Youth 1979 (NLSY79); (ii) allowing the effect of migration to differ by schooling level and (iii) using propensity score matching to estimate the average treatment effect on the treated (ATET) for movers and (iv) using local average treatment effect (LATE) estimators with covariates to estimate the average treatment effect (ATE) and ATET for compliers.We believe the Conditional Independence Assumption (CIA) is reasonable for our matching estimators since the NLSY79 provides a relatively rich array of variables on which to match. Our matching methods are based on local linear, local cubic, and local linear ridge regressions. Local linear and local ridge regression matching produce relatively similar point estimates and standard errors, while local cubic regression matching badly over-fits the data and provides very noisy estimates.We use the bootstrap to calculate standard errors. Since the validity of the bootstrap has not been investigated for the matching estimators we use, and has been shown to be invalid for nearest neighbor matching estimators, we conduct a Monte Carlo study on the appropriateness of using the bootstrap to calculate standard errors for local linear regression matching. The data generating processes in our Monte Carlo study are relatively rich and calibrated to match our empirical models or to test the sensitivity of our results to the choice of parameter values. The estimated standard errors from the bootstrap are very close to those from the Monte Carlo experiments, which lends support to our using the bootstrap to calculate standard errors in our setting.From the matching estimators we find a significant positive effect of migration on the wage growth of college graduates, and a marginally significant negative effect for high school dropouts. We do not find any significant effects for other educational groups or for the overall sample. Our results are generally robust to changes in the model specification and changes in our distance-based measure of migration. We find that better data matters; if we use a measure of migration based on moving across county lines, we overstate the number of moves, while if we use a measure based on moving across state lines, we understate the number of moves. Further, using either the county or state measures leads to much less precise estimates.We also consider semi-parametric LATE estimators with covariates (Frölich 2007), using two sets of instrumental variables. We precisely estimate the proportion of compliers in our data, but because we have a small number of compliers, we cannot obtain precise LATE estimates.  相似文献   

2.
This paper investigates the method of matching regarding two crucial implementation choices: the distance measure and the type of algorithm. We implement optimal full matching—a fully efficient algorithm—and present a framework for statistical inference. The implementation uses data from the NLSY79 to study the effect of college education on earnings. We find that decisions regarding the matching algorithm depend on the structure of the data: In the case of strong selection into treatment and treatment effect heterogeneity a full matching seems preferable. If heterogeneity is weak, pair matching suffices. Copyright © 2007 John Wiley & Sons, Ltd.  相似文献   

3.
We investigate the finite sample properties of a large number of estimators for the average treatment effect on the treated that are suitable when adjustment for observed covariates is required, like inverse probability weighting, kernel and other variants of matching, as well as different parametric models. The simulation design used is based on real data usually employed for the evaluation of labour market programmes in Germany. We vary several dimensions of the design that are of practical importance, like sample size, the type of the outcome variable, and aspects of the selection process. We find that trimming individual observations with too much weight as well as the choice of tuning parameters are important for all estimators. A conclusion from our simulations is that a particular radius matching estimator combined with regression performs best overall, in particular when robustness to misspecifications of the propensity score and different types of outcome variables is considered an important property.  相似文献   

4.
This paper proposes new ?1‐penalized quantile regression estimators for panel data, which explicitly allows for individual heterogeneity associated with covariates. Existing fixed‐effects estimators can potentially suffer from three limitations which are overcome by the proposed approach: (i) incidental parameters bias in nonlinear models with large N and small T ; (ii) lack of efficiency; and (iii) inability to estimate the effects of time‐invariant regressors. We conduct Monte Carlo simulations to assess the small‐sample performance of the new estimators and provide comparisons of new and existing penalized estimators in terms of quadratic loss. We apply the technique to an empirical example of the estimation of consumer preferences for nutrients from a demand model using a large transaction‐level dataset of household food purchases. We show that preferences for nutrients vary across the conditional distribution of expenditure and across genders, and emphasize the importance of fully capturing consumer heterogeneity in demand modeling. Copyright © 2016 John Wiley & Sons, Ltd.  相似文献   

5.
We use administrative records in Israel to identify the treatment effect of vocational training for the unemployed on the duration of unemployment. We randomize training using data on the availability of courses in and around the time of becoming unemployed. This assumes that the timing of entry into unemployment is random, and course availability around the time of entry is independent of current labor market conditions. According to naïve estimators, which ignore self-selectivity, training reduces unemployment durations. The same applies to linear (IV) estimators, which control for self-selectivity. However, the opposite is found when non-linear IV estimators are used, i.e. training prolongs unemployment durations. Sign reversal depends on how the first stage is specified. A non-nested test indicates that the nonlinear estimate of the treatment effect is preferable to its linear counterpart, in which case training prolongs unemployment spells.  相似文献   

6.
In this paper we introduce the Random Recursive Partitioning (RRP) matching method. RRP generates a proximity matrix which might be useful in econometric applications like average treatment effect estimation. RRP is a Monte Carlo method that randomly generates non‐empty recursive partitions of the data and evaluates the proximity between two observations as the empirical frequency they fall in a same cell of these random partitions over all Monte Carlo replications. From the proximity matrix it is possible to derive both graphical and analytical tools to evaluate the extent of the common support between data sets. The RRP method is “honest” in that it does not match observations “at any cost”: if data sets are separated, the method clearly states it. The match obtained with RRP is invariant under monotonic transformation of the data. Average treatment effect estimators derived from the proximity matrix seem to be competitive compared to more commonly used estimators. RRP method does not require a particular structure of the data and for this reason it can be applied when distances like Mahalanobis or Euclidean are not suitable, in the presence of missing data or when the estimated propensity score is too sensitive to model specifications. Copyright © 2008 John Wiley & Sons, Ltd.  相似文献   

7.
Abstract

This paper considers the problem of prediction in a panel data regression model with spatial autocorrelation in the context of a simple demand equation for liquor. This is based on a panel of 43 states over the period 1965–1994. The spatial autocorrelation due to neighbouring states and the individual heterogeneity across states is taken explicitly into account. We compare the performance of several predictors of the states’ demand for liquor for 1 year and 5 years ahead. The estimators whose predictions are compared include OLS, fixed effects ignoring spatial correlation, fixed effects with spatial correlation, random-effects GLS estimator ignoring spatial correlation and random-effects estimator accounting for the spatial correlation. Based on RMSE forecast performance, estimators that take into account spatial correlation and heterogeneity across the states perform the best for forecasts 1 year ahead. However, for forecasts 2–5 years ahead, estimators that take into account the heterogeneity across the states yield the best forecasts.  相似文献   

8.
Fixed and Random Effects in Stochastic Frontier Models   总被引:5,自引:1,他引:5  
Received stochastic frontier analyses with panel data have relied on traditional fixed and random effects models. We propose extensions that circumvent two shortcomings of these approaches. The conventional panel data estimators assume that technical or cost inefficiency is time invariant. Second, the fixed and random effects estimators force any time invariant cross unit heterogeneity into the same term that is being used to capture the inefficiency. Inefficiency measures in these models may be picking up heterogeneity in addition to or even instead of inefficiency. A fixed effects model is extended to the stochastic frontier model using results that specifically employ the nonlinear specification. The random effects model is reformulated as a special case of the random parameters model. The techniques are illustrated in applications to the U.S. banking industry and a cross country comparison of the efficiency of health care delivery.JEL classification: C1, C4  相似文献   

9.
The aim of this paper is to convey to a wider audience of applied statisticians that nonparametric (matching) estimation methods can be a very convenient tool to overcome problems with endogenous control variables. In empirical research one is often interested in the causal effect of a variable X on some outcome variable Y . With observational data, i.e. in the absence of random assignment, the correlation between X and Y generally does not reflect the treatment effect but is confounded by differences in observed and unobserved characteristics. Econometricians often use two different approaches to overcome this problem of confounding by other characteristics. First, controlling for observed characteristics, often referred to as selection on observables, or instrumental variables regression, usually with additional control variables. Instrumental variables estimation is probably the most important estimator in applied work. In many applications, these control variables are themselves correlated with the error term, making ordinary least squares and two-stage least squares inconsistent. The usual solution is to search for additional instrumental variables for these endogenous control variables, which is often difficult. We argue that nonparametric methods help to reduce the number of instruments needed. In fact, we need only one instrument whereas with conventional approaches one may need two, three or even more instruments for consistency. Nonparametric matching estimators permit     consistent estimation without the need for (additional) instrumental variables and permit arbitrary functional forms and treatment effect heterogeneity.  相似文献   

10.
We contrast the forecasting performance of alternative panel estimators, divided into three main groups: homogeneous, heterogeneous and shrinkage/Bayesian. Via a series of Monte Carlo simulations, the comparison is performed using different levels of heterogeneity and cross sectional dependence, alternative panel structures in terms of T and N and the specification of the dynamics of the error term. To assess the predictive performance, we use traditional measures of forecast accuracy (Theil’s U statistics, RMSE and MAE), the Diebold–Mariano test, and Pesaran and Timmerman’s statistic on the capability of forecasting turning points. The main finding of our analysis is that when the level of heterogeneity is high, shrinkage/Bayesian estimators are preferred, whilst when there is low or mild heterogeneity, homogeneous estimators have the best forecast accuracy.  相似文献   

11.
《Journal of econometrics》2005,124(2):335-361
This paper discusses estimation of nonparametric models whose regressor vectors consist of a vector of exogenous variables and a univariate discrete endogenous regressor with finite support. Both identification and estimators are derived from a transform of the model that evaluates the nonparametric structural function via indicator functions in the support of the discrete regressor. A two-step estimator is proposed where the first step constitutes nonparametric estimation of the instrument and the second step is a nonparametric version of two-stage least squares. Linear functionals of the model are shown to be asymptotically normal, and a consistent estimator of the asymptotic covariance matrix is described. For the binary endogenous regressor case, it is shown that one functional of the model is a conditional (on covariates) local average treatment effect, that permits both unobservable and observable heterogeneity in treatments. Finite sample properties of the estimators from a Monte Carlo simulation study illustrate the practicability of the proposed estimators.  相似文献   

12.
Under a conditional mean restriction Das et al. (2003) considered nonparametric estimation of sample selection models. However, their method can only identify the outcome regression function up to a constant. In this paper we strengthen the conditional mean restriction to a symmetry restriction under which selection biases due to selection on unobservables can be eliminated through proper matching of propensity scores; consequently we are able to identify and obtain consistent estimators for the average treatment effects and the structural regression functions. The results from a simulation study suggest that our estimators perform satisfactorily.  相似文献   

13.
The present paper shows that cross-section demeaning with respect to time fixed effects is more useful than commonly appreciated, in that it enables consistent and asymptotically normal estimation of interactive effects models with heterogeneous slope coefficients when the number of time periods, T, is small and only the number of cross-sectional units, N, is large. This is important when using OLS but also when using more sophisticated estimators of interactive effects models whose validity does not require demeaning, a point that to the best of our knowledge has not been made before in the literature. As an illustration, we consider the problem of estimating the average treatment effect in the presence of unobserved time-varying heterogeneity. Gobillon and Magnac (2016) recently considered this problem. They employed a principal components-based approach designed to deal with general unobserved heterogeneity, which does not require fixed effects demeaning. The approach does, however, require that T is large, which is typically not the case in practice, and the results reported here confirm that the performance can be extremely poor in small-T samples. The exception is when the approach is applied to data that have been demeaned with respect to fixed effects.  相似文献   

14.
Recent literature on panel data emphasizes the importance of accounting for time-varying unobservable individual effects, which may stem from either omitted individual characteristics or macro-level shocks that affect each individual unit differently. In this paper, we propose a simple specification test of the null hypothesis that the individual effects are time-invariant against the alternative that they are time-varying. Our test is an application of Hausman (1978) testing procedure and can be used for any generalized linear model for panel data that admits a sufficient statistic for the individual effect. This is a wide class of models which includes the Gaussian linear model and a variety of nonlinear models typically employed for discrete or categorical outcomes. The basic idea of the test is to compare two alternative estimators of the model parameters based on two different formulations of the conditional maximum likelihood method. Our approach does not require assumptions on the distribution of unobserved heterogeneity, nor it requires the latter to be independent of the regressors in the model. We investigate the finite sample properties of the test through a set of Monte Carlo experiments. Our results show that the test performs well, with small size distortions and good power properties. We use a health economics example based on data from the Health and Retirement Study to illustrate the proposed test.  相似文献   

15.
Estimating the casual effect of smoking on birth outcomes is difficult since omitted (unobserved) variables are likely to be correlated with a mother's decision to smoke. While some previous work has dealt with this endogeneity problem by using instrumental variables, this paper instead attempts to estimate the smoking effect from panel data (i.e., data on mothers with multiple births). Panel data sets are constructed with matching algorithms applied to federal natality data. The fixed effects regressions, which control for individual heterogeneity, yield significantly different results from ordinary least squares and previous instrumental variable approaches. The potential inconsistency caused by ‘false matches’ and other violations of the fixed effects strict exogeneity assumption are considered. Copyright © 2006 John Wiley & Sons, Ltd.  相似文献   

16.
We analyse the choices of 399 contestants in the Australian version of the television game show Deal or No Deal. We calculate risk aversion bounds for each contestant, revealing considerable heterogeneity. We then estimate a structural stochastic choice model that captures the dynamic decision problem faced by contestants. To address individual heterogeneity, we nest the dynamic problem within the settings of both a random effects and a random coefficients probit model. Our structural model produces plausible estimates of risk aversion, confirms the role of individual heterogeneity and suggests that a model of stochastic choice is indeed appropriate. We find mixed evidence of greater risk aversion by females. We also examine generalizations to expected utility theory, finding that the rank‐dependent utility model adds non‐negligible explanatory power and indicates optimism in probability weighting. Finally, we test, but are unable to confirm, the existence of an endowment effect for lotteries. Copyright © 2009 John Wiley & Sons, Ltd.  相似文献   

17.
This paper deals with a special case of estimation with grouped data, where the dependent variable is only available for groups, whereas the endogenous regressor(s) is available at the individual level. By estimating the first stage using the available individual data, and then estimating the second stage at the aggregate level, it might be possible to gain efficiency relative to the OLS and 2SLS estimators that use only grouped data. We term this the mixed-2SLS estimator (M2SLS). The M2SLS estimator is consistent and asymptotically normal. We also provide a test of efficiency of M2SLS relative to OLS and “2SLS” estimators.  相似文献   

18.
We propose new information criteria for impulse response function matching estimators (IRFMEs). These estimators yield sampling distributions of the structural parameters of dynamic stochastic general equilibrium (DSGE) models by minimizing the distance between sample and theoretical impulse responses. First, we propose an information criterion to select only the responses that produce consistent estimates of the true but unknown structural parameters: the Valid Impulse Response Selection Criterion (VIRSC). The criterion is especially useful for mis-specified models. Second, we propose a criterion to select the impulse responses that are most informative about DSGE model parameters: the Relevant Impulse Response Selection Criterion (RIRSC). These criteria can be used in combination to select the subset of valid impulse response functions with minimal dimension that yields asymptotically efficient estimators. The criteria are general enough to apply to impulse responses estimated by VARs, local projections, and simulation methods. We show that the use of our criteria significantly affects estimates and inference about key parameters of two well-known new Keynesian DSGE models. Monte Carlo evidence indicates that the criteria yield gains in terms of finite sample bias as well as offering tests statistics whose behavior is better approximated by the first order asymptotic theory. Thus, our criteria improve existing methods used to implement IRFMEs.  相似文献   

19.
This paper explores the impact of copyrights on firm value and on the demand for firm output. Using panel data on franchise value and ticket sales from the National Football League over the 1991–2000 period, we analyze the effect of copyrights (in this case, team logos) using several parametric estimators, the Arellano and Bond [1991. Some tests of specification for panel data: Monte Carlo evidence and an application to employment equations. Review of Economic Studies 58, 277–297] dynamic panel data estimator, and a semi-non-parametric method based on difference-in-differences propensity score matching. We find a negative effect of logo changes on franchise value that is robust across multiple specifications. In addition, logo changes also appear to have a moderate positive, albeit not particularly robust, impact on ticket sales.  相似文献   

20.
When some of the regressors in a panel data model are correlated with the random individual effects, the random effect (RE) estimator becomes inconsistent while the fixed effect (FE) estimator is consistent. Depending on the various degree of such correlation, we can combine the RE estimator and FE estimator to form a combined estimator which can be better than each of the FE and RE estimators. In this paper, we are interested in whether the combined estimator may be used to form a combined forecast to improve upon the RE forecast (forecast made using the RE estimator) and the FE forecast (forecast using the FE estimator) in out-of-sample forecasting. Our simulation experiment shows that the combined forecast does dominate the FE forecast for all degrees of endogeneity in terms of mean squared forecast errors (MSFE), demonstrating that the theoretical results of the risk dominance for the in-sample estimation carry over to the out-of-sample forecasting. It also shows that the combined forecast can reduce MSFE relative to the RE forecast for moderate to large degrees of endogeneity and for large degrees of heterogeneity in individual effects.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号