排序方式: 共有2条查询结果,搜索用时 0 毫秒
1
1.
Statistical offices are responsible for publishing accurate statistical information about many different aspects of society. This task is complicated considerably by the fact that data collected by statistical offices generally contain errors. These errors have to be corrected before reliable statistical information can be published. This correction process is referred to as statistical data editing. Traditionally, data editing was mainly an interactive activity with the aim to correct all data in every detail. For that reason the data editing process was both expensive and time-consuming. To improve the efficiency of the editing process it can be partly automated. One often divides the statistical data editing process into the error localisation step and the imputation step. In this article we restrict ourselves to discussing the former step, and provide an assessment, based on personal experience, of several selected algorithms for automatically solving the error localisation problem for numerical (continuous) data. Our article can be seen as an extension of the overview article by Liepins, Garfinkel & Kunnathur (1982). All algorithms we discuss are based on the (generalised) Fellegi–Holt paradigm that says that the data of a record should be made to satisfy all edits by changing the fewest possible (weighted) number of fields. The error localisation problem may have several optimal solutions for a record. In contrast to what is common in the literature, most of the algorithms we describe aim to find all optimal solutions rather than just one. As numerical data mostly occur in business surveys, the described algorithms are mainly suitable for business surveys and less so for social surveys. For four algorithms we compare the computing times on six realistic data sets as well as their complexity. 相似文献
2.
Guowei Hua T.C.E. Cheng Shouyang Wang 《International Journal of Production Economics》2011,131(2):568-574
This paper considers the problem of siting p new facilities of an entering firm to a competitive market so as to maximize the market share captured from competitors per unit cost. We first formulate the problem as a mixed 0-1 fractional programming model, in which we incorporate the fixed cost and transportation cost. The model can deal with the case where some demand nodes have two or more possible closest servers. We then re-formulate the problem as a 0-1 mixed integer linear program. We use a one-opt heuristic algorithm based on the Teitz-Bart method to obtain feasible solutions and compare them with the optimal solutions obtained by a branch-and-bound algorithm. We conduct computational experiments to evaluate the two algorithms. The results show that both algorithms can solve the model efficiently and the model is integer-friendly. We discuss other computational results and provide managerial insights. 相似文献
1