We first published this glossary in April 1999 along with our article "Accuracy: Know What You're Getting." The glossary and article are timeless pieces, well-suited for our anniversary look back at the content we've produced during Control Design's 15 years.
Accuracy: The degree of conformity of an indicated value to a recognized accepted standard value.1 Accuracy for instruments is normally stated in terms of error (±.05% of upper range value [URV], ±1% of span, ±0.5% of reading, ±3/4 degree, etc.).
Accuracy can also be stated in terms of bias and precision errors.2,3 Looking at Figure 1, the shift of the bullet holes from the bull's-eye is the bias and the tightness of the bullet pattern is the precision error.
Bias may be known or unknown. An example of a known bias is the deviation of a calibration standard from a National Institute of Standards and Technology (NIST)/National Bureau of Standards (NBS) reference. Large known biases are normally calibrated out. Small known biases are normally compensated out.
Examples of unknown biases include human error, installation effects, environmental disturbances, etc.2
Precision errors are considered statistically random. They can be stated as the product of the measurement's standard deviation and Student T distribution, which will provide an error specification to the 95% confidence level.
Also random are errors specified for transmitters, calculation devices, constant uncertainty, recorders, input/output devices, etc. Bias and precision errors can be individually combined using the root sum square (RSS) method, then the bias and precision errors can be combined:
E = total probable error,
B = total probable bias errors,
b = bias errors,
P = total probable precision error, and
p = precision errors.
Absolute accuracy: How close a measurement is to the NIST/NBS standard (the "golden ruler"). The accuracy traceability pyramid is shown in Figure 2.
Conformity: The maximum deviation of a calibration curve (average of upscale and downscale readings) from a specified characteristic curve.1 Conformity can be independent (best fit), zero, or terminal-based. This spec is commonly used as a measure of how close an instrument converts a non- linear input signal to a linear output signal.
Deadband: The range through which an input signal may be varied upon reversal of direction without initiating an observable change in the output signal.1
Drift or stability error: The undesired change in output over a specified period of time for a constant input under specified reference operating conditions.1
Dynamic error: The error resulting from the difference between the reading of an instrument and the actual value during a change in the actual value. Instrument damping contributes to this error as does measurement and transport deadtime, and the significance depends on the process time constant. Dynamic error must be considered when designing safety systems—if your instrumentation system cannot measure a developing hazardous condition in time for the safety system to react, the process may not get to a safe state in a timely manner.
EMI/RFI errors: Errors due to electromagnetic or radio frequency interference.
Filter error: Error caused by the improper application of a filter on the signal. This error can also be caused by improper settings in exception reporting and compression algorithms.
Hysteresis: The dependence of the output, for a given excursion of the input, upon the history of prior excursions and the direction of the current traverse.1
Influence errors: Errors due to operating conditions deviating from base or reference conditions. Typically specified as effects on the zero and span, these errors reflect the instrument's capacity to compensate for variations in operating conditions. Influence errors are often significant contributors to the overall error of an instrument.