全文获取类型
收费全文 | 975篇 |
免费 | 16篇 |
国内免费 | 1篇 |
专业分类
财政金融 | 194篇 |
工业经济 | 40篇 |
计划管理 | 322篇 |
经济学 | 174篇 |
综合类 | 31篇 |
运输经济 | 39篇 |
旅游经济 | 7篇 |
贸易经济 | 125篇 |
农业经济 | 26篇 |
经济概况 | 33篇 |
信息产业经济 | 1篇 |
出版年
2024年 | 2篇 |
2023年 | 13篇 |
2022年 | 12篇 |
2021年 | 25篇 |
2020年 | 37篇 |
2019年 | 40篇 |
2018年 | 23篇 |
2017年 | 32篇 |
2016年 | 35篇 |
2015年 | 19篇 |
2014年 | 45篇 |
2013年 | 103篇 |
2012年 | 55篇 |
2011年 | 74篇 |
2010年 | 44篇 |
2009年 | 72篇 |
2008年 | 48篇 |
2007年 | 49篇 |
2006年 | 37篇 |
2005年 | 30篇 |
2004年 | 35篇 |
2003年 | 26篇 |
2002年 | 23篇 |
2001年 | 16篇 |
2000年 | 22篇 |
1999年 | 10篇 |
1998年 | 8篇 |
1997年 | 8篇 |
1996年 | 6篇 |
1995年 | 5篇 |
1994年 | 3篇 |
1993年 | 5篇 |
1992年 | 2篇 |
1991年 | 4篇 |
1990年 | 2篇 |
1989年 | 2篇 |
1988年 | 3篇 |
1987年 | 4篇 |
1986年 | 2篇 |
1985年 | 4篇 |
1984年 | 5篇 |
1983年 | 1篇 |
1982年 | 1篇 |
排序方式: 共有992条查询结果,搜索用时 15 毫秒
11.
The Bank of England first acquired a macroeconomic model of the UK economy in early 1973, and used it for forecasting in June and July of that year. The initial model was obtained from the London Business School (LBS), but the last 14 years or so have, on the part of both the Bank and the LBS, led to developments which now make the models no closer to each other than to other large scale models of the UK. This article describes the structure and central properties of the current version of the Bank model, which has some 663 variables, 134 of which are modelled by behavioural equations, 153 by technical equations, 212 identities and 164 exogenous variables. In order to provide a transitional step between the kind of models with which most macroeconomists are familiar and the full scale version of the Bank model, one part of the article presents a very simple stylized version of the full model. This is a stepping stone to the full model which is described and analysed on a sectoral basis in Part 3, with a complete listing of equations in Part 4. Some of the simulation properties, and hence full model dynamic responses, are considered in Part 5. 相似文献
12.
An analysis of the process generating de facto standards in the PC spreadsheet software market 总被引:1,自引:0,他引:1
This paper develops a model to analyze the emergence of de facto or market defined compatibility standards in the market for PC spreadsheet software over the period 1982–1988. The model is capable of integrating diverse fragments of empirical evidence and a number of important theoretical building blocks, in particular the analysis of gateways between different versions of the same package, product preannouncements, and diverse consumer tastes towards intrinsic quality and network externalities. The model also explores the implications of different functional forms for the relationship between installed base and the value of network externalities. The paper finds that at least some enhancements to the basic model of standards have to be incorporated to offer a reasonable approximation to developments in the PC spreadsheet software market. The simplest model of de facto standards is not able to describe developments in this market. 相似文献
13.
Zusammenfassung Die Frage, warum bestimmte Informationen oder Werbemittel mehr und andere weniger Überzeugungskraft aufweisen, beschäftigt die Marketingforschung und verwandte Gebiete schon seit geraumer Zeit. Eine dieser Thematik zuzuordnende Forschungsrichtung, die in den letzten zwanzig Jahren eine Forschungstradition entwickelte, ist die Imagery-Forschung. Autoren, die sich dieser Forschungsrichtung zuwenden, erklären die Wirkung von Informationen damit, dass die Elemente in dieser Information Gedächtnisinhalte oder Imaginationen (Fantasien bzw. Vorstellungen) bei den Rezipienten auslösen, die ihrerseits die Bewertung des relevanten Meinungs- oder Werbeobjekts beeinflussen. In diesem Beitrag wird zunächst dargestellt, welche Hypothesen im Mittelpunkt der neueren Imagery-Forschung stehen. Der Nutzen dieser Überlegungen besteht zum Beispiel für die Werbepraxis darin, dass konkrete Hinweise für die Werbegestaltung gegeben werden. Anschließend wird der Stand der empirischen Forschung zu diesen Hypothesen vorgestellt. Hier zeigt sich, dass die empirischen Erkenntnisse hinter dem Stand der theoretischen Forschung zurückgeblieben sind. Am Ende dieser Abhandlung werden Vorschläge unterbreitet, wie die theoretischen Überlegungen einer weitergehenden Analyse unterzogen werden können. JEL classifications M31, M37 相似文献
14.
The problem of comparing the precisions of two instruments using repeated measurements can be cast as an extension of the Pitman-Morgan problem of testing equality of variances of a bivariate normal distribution. Hawkins (1981) decomposes the hypothesis of equal variances in this model into two subhypotheses for which simple tests exist. For the overall hypothesis he proposes to combine the tests of the subhypotheses using Fisher's method and empirically compares the component tests and their combination with the likelihood ratio test. In this paper an attempt is made to resolve some discrepancies and puzzling conclusions in Hawkins's study and to propose simple modifications.
The new tests are compared to the tests discussed by Hawkins and to each other both in terms of the finite sample power (estimated by Monte Carlo simulation) and theoretically in terms of asymptotic relative efficiencies. 相似文献
The new tests are compared to the tests discussed by Hawkins and to each other both in terms of the finite sample power (estimated by Monte Carlo simulation) and theoretically in terms of asymptotic relative efficiencies. 相似文献
15.
It is generally acknowledged that the growth rate of output, the seasonal pattern, and the business cycle are best estimated
simultaneously. To achieve this, we develop an unobserved component time series model for seasonally unadjusted US GDP. Our
model incorporates a Markov switching regime to produce periods of expansion and recession, both of which are characterized
by different underlying growth rates. Although both growth rates are time-varying, they are assumed to be cointegrated. The
analysis is Bayesian, which fully accounts for all sources of uncertainty. Comparison with results from a similar model for
seasonally adjusted data indicates that the seasonal adjustment of the data significantly alters several aspects of the full
model.
First Version Received: January 2001/Final Version Received: February 2002
Send offprint requests to: Rob Luginbuhl?Correspondence to: Rob Luginbuhl 相似文献
16.
A Closer Look at the Relation between GARCH and Stochastic Autoregressive Volatility 总被引:1,自引:0,他引:1
We show that, for three common SARV models, fitting a minimummean square linear filter is equivalent to fitting a GARCH model.This suggests that GARCH models may be useful for filtering,forecasting, and parameter estimation in stochastic volatilitysettings. To investigate, we use simulations to evaluate howthe three SARV models and their associated GARCH filters performunder controlled conditions and then we use daily currency andequity index returns to evaluate how the models perform in arisk management application. Although the GARCH models produceless precise forecasts than the SARV models in the simulations,it is not clear that the performance differences are large enoughto be economically meaningful. Consistent with this view, wefind that the GARCH and SARV models perform comparably in testsof conditional value-at-risk estimates using the actual data. 相似文献
17.
In this paper, we study the family of renewal shot-noise processes. The Feynmann–Kac formula is obtained based on the piecewise deterministic Markov process theory and the martingale methodology. We then derive the Laplace transforms of the conditional moments and asymptotic moments of the processes. In general, by inverting the Laplace transforms, the asymptotic moments and the first conditional moments can be derived explicitly; however, other conditional moments may need to be estimated numerically. As an example, we develop a very efficient and general algorithm of Monte Carlo exact simulation for estimating the second conditional moments. The results can be then easily transformed to the counterparts of discounted aggregate claims for insurance applications, and we apply the first two conditional moments for the actuarial net premium calculation. Similarly, they can also be applied to credit risk and reliability modelling. Numerical examples with four distribution choices for interarrival times are provided to illustrate how the models can be implemented. 相似文献
18.
In this article, we analyze export sophistication based on a large panel dataset (2001–2015; 101 countries) and using various estimation algorithms. Using Monte Carlo simulations, we evaluate the bias properties of estimators and show that GMM-type estimators outperform instrumental-variable and fixed-effects estimators. Based on our analysis we document that GDP per capita and the size of the economy exhibit significant and positive effects on export sophistication; weak institutional quality exhibits negative effect. We also show that export sophistication is path-dependent and stable even during a major economic crisis, which is especially important for emerging and developing economies. 相似文献
19.
The construction of an importance density for partially non‐Gaussian state space models is crucial when simulation methods are used for likelihood evaluation, signal extraction, and forecasting. The method of efficient importance sampling is successful in this respect, but we show that it can be implemented in a computationally more efficient manner using standard Kalman filter and smoothing methods. Efficient importance sampling is generally applicable for a wide range of models, but it is typically a custom‐built procedure. For the class of partially non‐Gaussian state space models, we present a general method for efficient importance sampling. Our novel method makes the efficient importance sampling methodology more accessible because it does not require the computation of a (possibly) complicated density kernel that needs to be tracked for each time period. The new method is illustrated for a stochastic volatility model with a Student's t distribution. 相似文献
20.
Enders-Granger方法在协整检验中的应用研究 总被引:4,自引:0,他引:4
刘汉中 《数量经济技术经济研究》2007,24(8):137-144
本文将协整检验由传统的线性协整检验扩展到线性协整检验和阈值协整检验,并在Enders和Granger(1998)方法的基础上提出了一个新的检验协整是否存在的Sup-F和Sup-F*统计量。通过MC仿真研究发现:在线性协整下,ADF方法比Sup-F法具有更高的检验势,但在"持久性"较强时,Sup-F检验比ADF检验法具有更高的检验势;Sup-F统计量在Three-Regime的阈值协整检验中比ADF法有更高的检验势;Sup-F*在检验协整(包括线性协整和阈值协整)时都具有较低的检验势;随着在不同Regime中自回归系数差距的增大(非对称程度增大),sup-F统计量的检验势提高很快,且比ADF法的检验势高。 相似文献