为了正常的体验网站,请在浏览器设置里面开启Javascript功能!

Describing the uncertainties in experimental results

2012-10-08 15页 pdf 1MB 66阅读

用户头像

is_422594

暂无简介

举报
Describing the uncertainties in experimental results Describing the Uncertainties in Experimental Results Robert J. Moffat Professor of Mechanical Engineering, Stanford University, Stanford, California I I It is no longer acceptable, in most circles, to present experimental results without describing the unc...
Describing the uncertainties in experimental results
Describing the Uncertainties in Experimental Results Robert J. Moffat Professor of Mechanical Engineering, Stanford University, Stanford, California I I It is no longer acceptable, in most circles, to present experimental results without describing the uncertainties involved. Besides its obvious role in pub- lishing, uncertainty analysis provides the experimenter a rational way of eval- uating the significance of the scatter on repeated trials. This can be a powerful tool in locating the source of trouble in a misbehaving experiment. To the user of the data, a statement (by the experimenter) of the range within which the results of the present experiment might have fallen by chance alone is of great help in deciding whether the present data agree with past results or dif- fer from them. These benefits can be realized only if both the experimenter and the reader understand what an uncertainty analysis is, what it can do (and cannot do), and how to interpret its results. This paper begins with a general description of the sources of errors in en- gineering measurements and the relationship between error and uncertainty. Then the path of an uncertainty analysis is traced from its first step, identify- ing the intended true value of a measurement, through the quantitative estima- tion of the individual errors, to the end objective--the interpretation and re- porting of the results. The basic mathematics of both single-sample and multiple-sample analysis are presented, as well as a technique for numerically executing uncertainty analyses when computerized data interpretation is in- volved. The material presented in this paper covers the method of describing the uncertainties in an engineering experiment and the necessary background ma- terial. Keywords: experimental uncertainty, error analysis, single-sample analysis, multiple-sample analysis, system errors INTRODUCTION The error in a measurement is usually defined as the difference between its true value and the measured value. This definition is clear but not very helpful: the only real situations in which we even claim to know both the true value and the measured value are those in which we are calibrating or "qualifying" an experiment against baseline data or against one of the basic conservation laws of engineering, In most situations, we cannot talk very confidently about what the error in a measurement is, we can only talk about what it might be--about the limits that we feel bound the possible error. The term "uncertainty" is used to refer to "a possible value that an error may have." Kline and McClintock [1] attribute this definition to Airy [2], and it still seems an appropriate and valuable concept. The terms "uncertainty interval" and "uncer- tainty" are commonly used interchangeably, and they will be so used in this discussion, both referring to the interval around the measured value within which the true value is believed to lie. The term "uncertainty analysis" refers to the process of estimating how great an effect the uncertainties in the individual measure- ments have on the calculated result. There is more to uncertainty analysis than just drawing error bars on a plot of the data or calculating the root-mean-square deviation of data from a curve fit, or (worse yet) calculating the mean absolute value of that deviation. Those are simply tech- niques for providing some measure of the scatter in the results and provide no way of judging whether or not the observed scatter was "reasonable"; that is where uncertainty analysis comes in. Uncertainty analysis began as the statistical interpretation of the errors in well-replicated experimental results. It became apparent during the early 1950s that many important engineering experi- ments could not be repeated enough times to provide useful statistical information, for reasons of economy or pressure of time. A rational way to use the framework of statistical inference to estimate the uncertainty in these single-sample experiments was described by Kline and McClintock [1] and still forms the basis for this branch of the an. Over the years, single-sample uncertainty analysis has been used more in research experiments than in production, and that association has led to its evolution as a diagnostic tool for the development of experiments. As a consequence, single-sample uncertainty analysis presents two uncertainty measures that are particularly useful during the planning and debugging stages of experiments, in addition to the usual "overall uncertainty." The distinction between single-sample and multiple sample Address correspondence to Professor Robert J. Moffat, Mechanical Engineering Department, Stanford University, Stanford, CA 94305. Experimental Thermal and Fluid Science 1988; 1:3-17 ©1988 by Elsevier Science Publishing Co., Inc., 52 Vanderbilt Avenue, New York, NY 10017 0894-1777/88/$3.50 4 R. J . Moffat analysis hinges on whether or not a " large" or a "smal l" number of independent data points are taken at each test point and on how the data are handled. In this era of high-speed digital data acquisition, the issue of independence takes on more subtle overtones than it had in the early 1950s when the term was coined. Consider, for example, a measuring system capable of acquiring data at 100 kHz, a system readily available under current technology. Applied to an experiment whose output varied at approximately 10 mHz, this 100 kHz sample rate would produce about 100 readings in a 1 ms period. These can each be regarded as independent measures of the process, assuming the time between consecutive readings to be large compared with the autocorrelation time of the signal. This set of 100 readings would be, then, a multiple-sample set of observations. The same equipment, applied to a system whose output varied at 1 Hz, would produce a single-sample measure of the process, tainted, perhaps, by a multiple-sample contribution from the high- frequency random errors present in the measuring system. Classification of a given experiment as single-sample or multiple-sample must take into account not only the number of observations made at each test point but also the data-sampling rate and the spectrum of significant frequencies in the process being studied and whether the data are averaged before processing or processed before averaging. Single-sample uncertainty analysis has been described in the engineering literature by the works of Kline and McClintock [ 1] and Moffat [3, 4]. The techniques of multiple-sample analysis are described by Abernethy and Thompson [5] and summarized by Abernethy et al [6] and by ANSI/ASME PTC 19.1-1985 |7]. Both have the same final objective (to estimate the effect of the accumulated measurement uncertainties on the accuracy of the result), even though somewhat different procedures are required. Either the Nth-order uncertainty interval of single-sample analysis or the U0.95 interval of multiple-sample analysis answers the question: How close to this result does the true value probably lie. assuming that these are the right equations to use, that all the important variables have been included, and that the test situation is representative? In addition to this final result, each method produces auxiliary information about the experiment, mainly useful as diagnostics, during the developmental phase of an experiment or in monitoring its "health" during a long series of runs. Single-sample uncer- tainty analysis generates two diagnostics: the zeroth-order uncer- tainty, which evaluates the contribution to the total uncertainty introduced by the measuring system, and the first-order uncer- tainty, which predicts the scatter that should be observed on repeated trials with the same equipment and the same instruments. Multiple-sample uncertainty analysis produces estimates of the total fixed error and the total random error in the result. An uncertainty estimate of either type is only as good as the equation(s) it is based on. If those equations are incomplete and do not acknowledge all the significant factors that affect the result, o, if falsely low values are used for the component uncertainties, then the analysis will underestimate the uncertainty in the result. On the other hand, if the component uncertainties are exagger- ated, then the analysis will overestimate the uncertainty. The most common, and most visible, use of uncertainty analysis is in reporting results to the technical community through publications, but it must be noted here that it has far broader uses. During the shakedown period of an experiment, it is a powerful diagnostic tool in seeking out the sources of residual error. In the early stages of an experiment, for example, when comparing the first results from a new test rig with those from an existing baseline set, the most frequent question is: Does the difference 1 see mean that the new results are really different, or is this difference just a consequence of the uncertainties in my measure. ments? Uncertainty analysis provides clear, unambiguous guid- ance: If the observed difference exceeds zero by more than the expected uncertainty interval for the difference, then the observed difference is probably significant. Even earlier in an experiment, uncertainty analysis can be used to help choose the most reliable technique for a given measurement or to identify the critically important instruments in a system (and thereby determine where expensive instruments are needed!). THE BAS IC MATHEMATICS This section introduces the root-sum-square (RSS) combination, the basic form used for combining uncertainty contributions in both single-sample and multiple-sample analyses. In this section, the term 3Xi refers to the uncertainty in Xi in a general and nonspecific way: whatever is being dealt with at the moment (tor example, fixed errors, random errors, or uncertainties). Describing One Variable Consider a variable Xi, which has a known uncertainty 6X,. The form for representing this variable and its uncertainty is X,=Xi(measured)+_6Xi (20:1) (1~ This statement should be interpreted to mean the lollowing: • The best estimate of X, is Xi (measured) • There is an uncertainty in Xi that may be as large as +_ bX~_ • The odds are 20 to 1 against the uncertainty of Xi being larger than +_ 8X,. The value of Xi (measured) represents the observation in a single-sample experiment or the mean of a set of N observations in a multiple-sample experiment. The value of 6Xi represents 20 for a single-sample analysis, where o is the standard deviation of the population of possible measurements from which the single sample Xi was taken. For multiple-sample experiments, 6Xi can have three meanings. It may represent tS~N)/q-N for random error components, where ScN I is the standard deviation of the set of N observations used to calculate the mean value Xi and t is the Student's t statistic appropriate for the number of samples N and the confidence level desired. It may represent the bias limit for fixed errors (this interpretation implicitly requires that the bias limit be estimated at 20:1 odds). Finally, 8Xi may represent U95, the overall uncer- tainty in X,. The Student's t multiplier is a number, always larger than 2.0, that allows one to use S~.~ rather than o in estimating the uncertainty in the mean of a set. The Student's t statistic is tabulated in most statistical reference books under that name The result R of the experiment is assumed to be calculated from a set of measurements using a data interpretation program (b~ hand or by computer) represented by R=R(XI, X2, A'~, "- - , Xx) ~2i The objective is to express the uncertainty in the calculated result at the same odds as were used in estimating the uncertainties in the measurements. This issue was taken up by Kline and McClintock [1], who showed that the uncertainty in a computed result could be estimated with good accuracy using a root-sum square combination of the effects of each of the individual inputs and that the RSS operation preserved the odds. The effect of the uncertainty in a single measurement on the Uncertainties in Experimental Results 5 calculated result, if only that one measurement were in error would be fR _ OR Xi- -~i fxi (3) The partial derivative of R with respect to Xi is the sensitivity coefficient for the result R with respect to the measurement Xi. When several independent variables are used in the function R, the individual terms are combined by a root-sum-square method. fR = 6Xi (4) This is the basic equation of uncertainty analysis. Each term represents the contribution made by the uncertainty in one variable, fXi, to the overall uncertainty in the result, fiR. Each term has the same form: the partial derivative of R with respect to Xi multiplied by the uncertainty interval for that variable. The estimated uncertainty in the result has the same probability of encompassing the true value of the result as the uncertainties in the individual measurements have of encompassing their true values. Equation (4) applies as long as [1] 1. Each of the measurements was independent 2. Repeated observations of each measurement, if made, would display Gaussian distributions 3. The uncertainty in each measurement was initially expressed at the same odds The data interpretation program may be simple enough that all the partial derivatives can be evaluated analytically, or it may be too complex for that and require direct computer analysis; the procedures are the same in either case. A general technique for computerized uncertainty analysis is given in the next section. The following comments might be helpful in "hand" analysis; they are mainly aimed at simplifying the task. In most situations, the overall uncertainty in a given result is dominated by only a few of its terms. Terms in the uncertainty equation that are smaller than the largest term by a factor of 3 or more can usually be ignored. This is a natural consequence of the RSS combination: Small terms have very small effects. There are exceptions, of course, where there are many terms of approxi- mately the same size, but in general that is not the case. In many applications, the uncertainty estimate is wanted as a fraction of reading, rather than in engineering units. While this can always be calculated, starting from the results of the general form in Eq. (4), it is sometimes possible to do the calculation of relative uncertainty directly. In particular, whenever the equation describing the result is a pure "product form," such as Eq. (5), or can be put into that form, then the relative uncertainty can be found directly. That is, if , b ~. ~ (5) R=XlX2X 3 .. X M then fR [( fXl~2 ( fX2~2 _ ( fXm' l l /2 -~= a--~-i ; + b X2 /i " "+ m Xm ) (6) This is a natural and convenient approach in situations where the uncertainties of the component measurements are described in terms of percent of reading and the result is needed in the same terms. The exponent of Xi becomes its sensitivity coefficient. COMPUTERIZED UNCERTAINTY ANALYS IS When R is calculated using a large-scale computer program or involves operations that are difficult to differentiate (e.g., table look-ups or numerical integrations), the operations represented by Eqs. (3) and (4) either cannot or will not be done by hand. In most cases, it is not practical to write a separate computer program for the evaluation of uncertainties, both from the standpoint of complexity and because of the difficulty of ensuring that the uncertainty code is updated each time the main data interpretation code is revised. For these more complex experiments, the data interpretation program itself can be used to generate the uncer- tainty analysis, by sequentially perturbing the input values and accumulating the individual uncertainty contributions. This direct computer-executed uncertainty analysis can be accomplished by sequentially perturbing the inputs according to the following procedure [4]: 1. Calculate the result R for the recorded data. Identify the value as R0 and store it. 2. For i = 1 to N, where N is the number of variables in R: Increase the value of the ith variable, Xi, by its uncertainty interval, fXi, and calculate the result, Ri+ , using the aug- mented value of the ith variable with all other variables at their recorded (nominal) values. Find the difference Ri+ - Ro and store it as Ci+, the contribution to the uncertainty of R caused by the ith variable, assuming a positive excursion. If the result R is likely to be a strongly nonlinear function of Xi, including consideration of the size of its uncertainty interval, then also calculate Ci- using Ri- - Ro. The present recommendation would be to use the average of the absolute values of Ci+ and Ci- as the working value of Ci, but that recommendation must be regarded as tentative, since no definitive analysis has been done to investigate the issue. Next i. 3. The uncertainty in the result is the root-sum-square of the Ci. A primary advantage of this method is that the actual working data interpretation program itself is used in the assessment of uncertainties. Thus, each time the program is modified, the modifications are automatically incorporated into the uncertainty calculations. The process of determining the uncertainty in the computed heat transfer coefficient from its basic measurements is illustrated in Fig. 1. The method of sequential perturbations can also be applied to data interpretation programs of intermediate difficulty (too large to do by hand but not large enough to warrant a dedicated program) using spreadsheets on a personal computer, as pointed out by Catz [8]. THE SOURCES OF UNCERTAINT IES The uncertainty attributed to a measurement is an estimate of the possible residual error in that measurement after all proposed corrections have been made. Although this discussion of the sources of uncertainties will begin with a discussion of errors, the distinction between the original error and the residual error must be clearly maintained. The uncertainty is determined by the residual error after correction, not the original error. An error source is usually categorized as "fixed" or "random" depending on whether the error it introduces is steady or changes during the time of one complete experiment. The errors themsel- ves are called bias or precision errors, and the precision error is presumed to behave randomly, with a zero mean. Both the bias and precision are presumed to represent stationary statistical properties of a Gaussian distributed data set. Such a description tacitly assumes that each observation of the error is independent of its preceding observation--that the er ror source in question has no systematic variation. This is true only 6 R. J . Moffat Figure 1. The method of sequen- tial perturbation for calculating uncertainty intervals from the data reduction program. INPUT - Q",T~,T2,&(~",6T~,&T2 ho= i Run "O" Run 1 Run 2 Run 3 (Base Case) (Q"-Loop) (T~-Loop) (T2-Loop) Q".T~.T2 ~-~ (Q"+aQ"),T,,T2 ~ Q",(T,~T~).T2 ~. .~ Q".T,.(T2+6T2) 1 Q,, T~ - T~ il I h d Q"+bQ" T~ - "I"2 a,, h,,~ - (TI +6T,) T2 a,, h'T~- TI (T2+&T2) 2 2 2 6h__ { ( 0h &Q" + ~ 0T2 6TI ) ( ah ' TTT ) ,,=, when the interval between observations is longer than the autocorrelation time of the error source. For most engineering experiments, a third category of error is necessary: a category for errors that change during an experiment, but not randomly. These will be described here as "variable but deterministic." For example, the radiation error of
/
本文档为【Describing the uncertainties in experimental results】,请使用软件OFFICE或WPS软件打开。作品中的文字与图均可以修改和编辑, 图片更改请在作品中右键图片并更换,文字修改请直接点击文字进行修改,也可以新增和删除文档中的内容。
[版权声明] 本站所有资料为用户分享产生,若发现您的权利被侵害,请联系客服邮件isharekefu@iask.cn,我们尽快处理。 本作品所展示的图片、画像、字体、音乐的版权可能需版权方额外授权,请谨慎使用。 网站提供的党政主题相关内容(国旗、国徽、党徽..)目的在于配合国家政策宣传,仅限个人学习分享使用,禁止用于任何广告和商用目的。

历史搜索

    清空历史搜索