Evaluation (not validation) of quantitative models

被引:132
|
作者
Oreskes, N [1 ]
机构
[1] NYU, Gallatin Sch Individualized Study, New York, NY USA
关键词
model evaluation; model validation; quantitative models;
D O I
10.1289/ehp.98106s61453
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
The present regulatory climate has led to increasing demands for scientists to attest to the predictive reliability of numerical simulation models used to help set public policy, a process frequently referred to as model validation. But while model validation may reveal useful information, this paper argues that it is not possible to demonstrate the predictive reliability of any model of a complex natural system in advance of its actual use. Ail models embed uncertainties, and these uncertainties can and frequently do undermine predictive reliability. In the case of lead in the environment, we may categorize model uncertainties as theoretical, empirical, parametrical, and temporal. Theoretical uncertainties are aspects of the system that are not fully understood, such as the biokinetic pathways of lead metabolism Empirical uncertainties are aspects of the system that are difficult (or impossible) to measure, such as actual lead ingestion by an individual child. Parametrical uncertainties arise when complexities in the system are simplified to provide manageable model input, such as representing longitudinal lead exposure by cross-sectional measurements. Temporal uncertainties arise from the assumption that systems are stable in time. A model may also be conceptually flawed. The Ptolemaic system of astronomy is a historical example of a model that was empirically adequate but based on a wrong conceptualization. Yet had it been computerized-and had the word then existed-its users would have had every right to call it validated. Thus, rather than talking about strategies for validation, we should be talking about means of evaluation. That is not to say that language alone will solve our problems or that the problems of model evaluation are primarily linguistic. The uncertainties inherent in large, complex models will not go away simply because we change the way we talk about them. But this is precisely the point: calling a model validated does not make it valid. Modelers and policymakers must continue to work toward finding effective ways to evaluate and judge the quality of their models, and to develop appropriate terminology to communicate these judgments to the public whose health and safety may be at stake.
引用
收藏
页码:1453 / 1460
页数:8
相关论文
共 50 条
  • [1] Quantitative Validation of Formal Domain Models
    Iliasov, Alexei
    Romanovsky, Alexander
    Laibinis, Linas
    201919TH IEEE INTERNATIONAL SYMPOSIUM ON HIGH ASSURANCE SYSTEMS ENGINEERING (HASE 2019), 2019, : 17 - 24
  • [2] Experimental validation for quantitative protein network models
    Nishizuka, Satoshi
    Spurrier, Brett
    CURRENT OPINION IN BIOTECHNOLOGY, 2008, 19 (01) : 41 - 49
  • [3] Establishment and validation of psoriasis evaluation models
    Hu, Yibo
    Jiang, Ling
    Lei, Li
    Luo, Liping
    Guo, Haoran
    Zhou, Ying
    Huang, Jinhua
    Chen, Jing
    Zeng, Qinghai
    FUNDAMENTAL RESEARCH, 2022, 2 (01): : 166 - 176
  • [4] Quantitative validation of deterministic models: When is a model acceptable?
    Scholten, H
    van der Tol, MWM
    PROCEEDINGS OF THE 1998 SUMMER COMPUTER SIMULATION CONFERENCE: SIMULATION AND MODELING TECHNOLOGY FOR THE TWENTY-FIRST CENTURY, 1998, : 404 - 409
  • [5] Validation schemes for tropical cyclone quantitative precipitation forecasts: Evaluation of operational models for US landfalling cases
    Marchok, Timothy
    Rogers, Robert
    Tuleya, Robert
    WEATHER AND FORECASTING, 2007, 22 (04) : 726 - 746
  • [6] Methodology of quantitative evaluation of texture models
    Hernandez, Jorge
    Prieto, Flavio
    DYNA-COLOMBIA, 2007, 74 (151): : 115 - 124
  • [7] EVALUATION OF RECENT QUANTITATIVE MAGNETOSPHERIC MODELS
    WALKER, RJ
    TRANSACTIONS-AMERICAN GEOPHYSICAL UNION, 1975, 56 (09): : 615 - 615
  • [8] Quantitative validation of a general competency composite assessment evaluation
    Reisdorff, EJ
    Carlson, DJ
    Reeves, M
    Walker, G
    Hayes, OW
    Reynolds, B
    ACADEMIC EMERGENCY MEDICINE, 2004, 11 (08) : 881 - 884
  • [9] EVALUATION OF NUMERICAL POUNDING MODELS WITH EXPERIMENTAL VALIDATION
    Khatiwada, S.
    Chouw, N.
    Butterworth, J. W.
    BULLETIN OF THE NEW ZEALAND SOCIETY FOR EARTHQUAKE ENGINEERING, 2013, 46 (03): : 117 - 130
  • [10] An evaluation of the bootstrap for model validation in mixture models
    Jaki, Thomas
    Su, Ting-Li
    Kim, Minjung
    Van Horn, M. Lee
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2018, 47 (04) : 1028 - 1038