Back to the basics of electrical measurements

Scientific Devices Australia
Tuesday, 12 July, 2011


Accurate measurements are central to virtually every scientific and engineering discipline, but all too often measurement science gets little attention in the undergraduate curriculum. Even those who received a thorough grounding in measurement fundamentals as undergraduates can be forgiven if they’ve forgotten some of the details.

This article is intended to refresh those fading memories or to bring those who want to learn more about making good quality measurements up to speed.

But what exactly does ‘good quality measurement’ mean? Although it can mean a variety of things, one of the most important is the ability to create a test set-up that’s suitable for the purpose intended.

Let’s start with a typical test scenario that involves measuring some characteristics of a device or material. This can range from a very simple set-up, such as using a benchtop digital multimeter to measure resistance values, to more complex systems that involve fixturing, special cabling, etc.

When determining the required performance of the system, that is, the required measurement accuracies, tolerances, speed etc, one must include not only the performance of the measurement instrument but also the limitations imposed by and the effects of the cabling, connectors, test fixture and even the environment under which tests will be carried out.

When considering a specific measurement instrument for an application, the specification or data sheet is the first place to look for information on its performance and how that will limit the results. However, data sheets are not always easy to interpret because they typically use specialised terminology.

Also, one can’t always determine if a piece of test equipment will meet the requirements of the application simply by looking at its specifications. For example, the characteristics of the material or device under test may have a significant impact on measurement quality. The cabling, switching hardware and the test fixture, if required, can also affect the test results.

The process of designing and characterising the performance of any test set-up can be broken down into a four-step process. Following this process will greatly increase the chances of building a system that meets requirements and eliminates unpleasant and expensive surprises.

The first step, before specifying a piece of equipment, is to define the system’s required measurement performance. Defining the required level of performance involves understanding the specialised terminology like resolution, accuracy, repeatability, rise time, sensitivity and many others.

Resolution is the smallest portion of the signal that can actually be observed. It is determined by the analog-to-digital converter in the measurement device. There are several ways to characterise resolution - bits, digits, counts, etc. The more bits or digits there are, the greater the device’s resolution.

The resolution of most benchtop instruments is specified in digits, such as a 6½-digit DMM. Be aware that the ½-digit terminology means that the most significant digit has less than a full range of 0 to 9. As a general rule, ½digit implies the most significant digit can have the values 0, 1, or 2.

In contrast, data acquisition boards are often specified by the number of bits their A/D converters have.

  • 12-bit A/D - 4096 counts - approx 3½ digits;
  • 16-bit A/D - 65,536 counts - approx 4½ digits;
  • 18-bit A/D - 262,144 counts - approx 5½ digits;
  • 22-bit A/D - 4,194,304 counts - approx 6½ digits;
  • 25-bit A/D - 33,554,304 counts - approx 7½ digits;
  • 28-bit A/D - 268,435,456 counts - approx 8½ digits.

Although the terms sensitivity and accuracy are often considered synonymous, they do not mean the same thing. Sensitivity refers to the smallest change in the measurement that can be detected and is specified in units of the measured value, such as volts, ohms, amps, degrees, etc.

The sensitivity of an instrument is equal to its lowest range divided by the resolution. Therefore, the sensitivity of a 16-bit A/D based on a 2 V scale is two divided by 65536 or 30 microvolts. A variety of instruments is optimised for making highly sensitive measurements, including nanovoltmeters, picoammeters, electrometers and high-resolution DMMs. Here are some examples of how to calculate the sensitivity for A/Ds of varying levels of resolution:

  • 3½ digits (2000) on 2 V range = 1 mV;
  • 4½ digits (20000) on 2Ω range = 100 mΩ;
  • 16-bit (65536) A/D on 2 V range = 30 µV;
  • 8½ digits on 200 mV range = 1 nV.

Now that we have a better understanding of sensitivity, what do we mean when describing the accuracy of an instrument? In fact, there are two types of accuracy to consider, namely absolute accuracy and relative accuracy.

Absolute accuracy indicates the closeness of agreement between the result of a measurement and its true value, as traceable to an accepted national or international standard value. Devices are typically calibrated by comparing them to a known standard value. Most countries have their own standards institute where national standards are kept.

The drift of an instrument refers to its ability to retain its calibration over time. Relative accuracy refers to the extent to which a measurement accurately reflects the relationship between an unknown and a locally established reference value.

The implications of these terms are demonstrated by the challenge of ensuring the absolute accuracy of a temperature measurement of 100.00°C to ±0.01° versus measuring a change in temperature of 0.01°C.

Measuring the change is far easier than ensuring absolute accuracy to this tolerance, and, often, that is all that an application requires. For example, in product testing, it is often important to measure the heat rise accurately (for example, in a power supply), but it really doesn’t matter if it’s at exactly 25.00°C ambient.

Repeatability is the ability to measure the same input to the same value over and over again. Ideally, the repeatability of measurements should be better than the accuracy. If repeatability is high and the sources of error are known and quantified, then high resolution and repeatable measurements are often acceptable for many applications. Such measurements may have high relative accuracy with low absolute accuracy.

The second step gets into the actual process of designing the measurement system, including the selection of equipment and fixtures etc. As mentioned previously, interpreting a data sheet to determine which specifications are relevant to a system can be daunting, so let’s look at some of the most important specs included:

  • Accuracy. Keithley normally expresses its accuracy specifications in two parts, namely as a proportion of the value being measured and a proportion of the scale that the measurement is on, for example ± (gain error + offset error). This can be expressed as ± (% reading + % range) or ± (ppm of reading + ppm of range). The range in Figure 1 is represented by FS or ‘full scale’.

 
Figure 1.

For example, the specification for the company’s 2000 6½-digit multimeter, when measuring voltage on the 1 V range, states an accuracy of 30 ppm of the reading + 7 ppm of range. The green box represents the offset error, which is expressed either as a percentage of the range or ppm of the range. Figure 2 illustrates the gain error, which is expressed either as a % of the reading or ppm of the reading.

When carrying out a reading, we can expect the error to be anywhere within the purple and green areas of the graph. Accuracy specs for high-quality measurement devices can be given for 24 hours, 90 days, one year, two years, or even five years from the time of last calibration. Basic accuracy specs often assume usage within 90 days of calibration.

 
Figure 2.

  • Temperature coefficient. Accuracy specs are normally guaranteed within a specific temperature range; for example, the 2000 DMM’s guaranteed range is 23°C, ±5°C. If carrying out measurements in an environment where temperatures are outside this range, it’s necessary to add a temperature-related error. This becomes especially difficult if the ambient temperatures vary considerably.
  • Instrumentation error. Some measurement errors are a result of the instrument itself. As we have already discussed, instrument error or accuracy specifications always require two components: a proportion of the measured value, sometimes called gain error, and an offset value specified as a portion of full range.

Let’s look as the different instrument specifications for measuring the same value. In this example, we are trying to measure 0.5 V on the 2 V range, using a lesser quality DMM. Using the specifications, we can see that the uncertainty, or accuracy, will be ±350 mV. In abbreviated specs, frequently only the gain error is provided. The offset error, however, may be the most significant factor when measuring values at the low end of the range.

Accuracy = ±(% reading + % range)

= ±(gain error + offset error)

For example, DMM 2 V range:

Accuracy = ±(0.03% of reading + 0.01% range)

For a 0.5 V input:

Uncertainty = ±(0.03% x 0.5 V + 0.01% x 2.0 V)

= ±(0.00015 V + 0.00020 V)

= ± 350 µV

Reading = 0.49965 to 0.50035

In the next example, we have the same scenario, ie, trying to measure 0.5 V using the 2 V range, but we are now using a better quality DMM. The example has better specifications on the 2 V range, and the uncertainty is now just ±35 µV.

DMM, 6½-digit, 2 V range (2.000000)

Accuracy = ±(0.003% reading + 0.001% range)

= ±(30 ppm readings + 10 ppm range)

= ±(0.003% reading + 20 counts)

Uncertainty at 0.5 V= ±(0.000015 + 0.000020)

= ±0.000035 V

= ±35 µV

Now if we look at performing the same measurement using a data acquisition board, note that 1 LSB offset error is range/4096 = 0.024% of range. On a 2 V range, 1 LSB offset error is 0.488 millivolt.

Note that the measurement accuracy is much poorer with this data acquisition card than when using the higher quality benchtop DMM.

Analog input board, 12 bit, 2 V range

Accuracy = ±(0.01% reading + 1 LSB)

= ±(100ppm + 1 bit)

Uncertainty @ 0.5V = ±(0.000050 + 2.0/4096)

= ±(0.000050 + 0.000488)

= ±0.000538

= ±538 µV.

  • Sensitivity, the smallest observable change that can be detected by the instrument, may be limited either by noise or by the instrument’s digital resolution. The level of instrument noise is often specified as a peak-to-peak or RMS value, sometimes within a certain bandwidth.<0x000A>It is important that the sensitivity figures from the data sheet will meet your requirements, but also consider the noise figures as these will especially affect low level measurements.
  • Timing. What does the timing within a test set-up mean? Obviously, an automated PC-controlled measurement set-up allows making measurements far more quickly than manual testing. This is especially useful in manufacturing, or where many measurements are required.

However, it’s critical to ensure that measurements are taken when the equipment has ‘settled’ because there is always a trade-off between the speed with which a measurement is made and its quality.

The rise time of an analog instrument (or analog output) is generally defined as the time necessary for the output to rise from 10% to 90% of the final value when the input signal rises instantly from zero to some fixed value. Rise time affects the accuracy of the measurement when it’s of the same order of magnitude as the period of the measurement.

If the length of time allowed before taking the reading is equal to the rise time, an error of about 10% will result, because the signal will have reached only 90% of its final value. To reduce the error, more time must be allowed. To reduce the error to 1%, about two rise times must be allowed, while reducing the error to 0.1% would require roughly three rise times (or nearly seven time constants).

Step three addresses building the test system and verifying its performance, including a number of techniques that can be used to improve measurement quality.

Once a system builder has picked appropriate equipment, cables and fixtures and established that the equipment’s specifications can meet the requirements, it’s time to assemble it and verify its performance one step at a time. It is essential to check that each piece of test equipment has been calibrated within its specified calibration period, which is usually one year. If the instrument will be used for making voltage measurements, placing a short across the inputs of the meter will provide an indication of any offset errors. This can be directly compared to the specifications from the data sheet.

If the instrument will be used for current measurements, then checking to see the current level with the ammeter open circuit will give an indication of offset current. Again, this can be directly compared with the specifications from the data sheet. Next, include the system cabling and repeat the tests, followed by the test fixture, then the device under test, repeating the tests after each addition.

If the performance of the system does not meet the application’s requirements, this ‘one step at a time’ approach should help identify what is causing the problems.

Then, check the system timing to ensure there are sufficient delays to allow for settling time and reassess it to make sure it satisfies the application’s speed goals. Insufficient delay times between measurements can often create accuracy and repeatability problems.

In fact, this is among the most common sources of error in test systems, and it’s especially evident when running the test at speed produces a different result than when performing the test step by step or manually.

Although inductance can affect settling times, capacitance in the system is a more common problem. In a manual system, a delay of 0.25 to 0.5 seconds will seem to be instant. But in an automated test system, steps are typically executed in a millisecond or less and even the simplest systems may require delays of five to 10 milliseconds after a change in stimulus to get accurate results.

Large systems with lots of cabling (and therefore, lots of cable capacitance, and/or those that measure high impedances (<0x03C4> = RC) may require even longer delays or special techniques like guarding. Coaxial cable typically has capacitance in the range of 30 pF per 300 mm. The common solution is to provide sufficient delays in the measurement process to allow for settling.

Delays of several milliseconds are commonly needed, but some applications may require even longer delays. To address this need, most Keithley instruments include a programmable trigger delay.

Guarding is one technique for dealing with capacitance issues, reducing leakage errors and decreasing response time. Guarding consists of a conductor driven by a low impedance source surrounding the lead of a high impedance signal. The guard voltage is kept at or near the potential of the signal voltage.

Although all systems are unique, the following sources of error are among the most common:

  • Lead resistance. For resistance measurements, especially at lower resistances, it is important to take into account the resistance of the test leads. In the example shown in Figure 3a, the two-wire ohms method is being used to determine the resistance. A current source in the meter outputs a known and stable current, and the voltage drop is measured within the meter.

 
Figure 3a.

This method works well if the resistance to be measured is very much greater than the lead resistance. However, what if the resistance to be measured is much closer to the lead resistance or even less?

Using four-wire measurements (Figure 3b) will eliminate this problem. The voltage drop is now measured across the resistor, instead of across the resistor and leads. The input resistance of the voltmeter tends to be very high in comparison to the resistance to be measured; therefore, the lead resistances on the voltmeter path can be ignored.

If, however, the resistance to be measured is very high and approaching the resistance of the voltmeter, then an electrometer or specialised meter with extremely high input resistance may be required.

 
Figure 3b.

  • Thermoelectric EMFs in connections. In any measurement system, any connections made that are of dissimilar metals will produce a thermocouple. A thermocouple is essentially a device of two dissimilar metals that generates a voltage that varies with temperature.

These qualities can be put to good use when using thermocouples to monitor temperature, but in a standard test system, they result in the introduction of unwanted voltages. As temperatures vary, so does the magnitude of the unwanted voltages.

Table 1 lists some examples of the types of voltages that can be generated. Even when connecting copper to copper, there are typically enough differences in the composition of the two pieces of metal that voltages will be generated. If the magnitude of these errors is significant in comparison to the value to be measured, the offset-compensated ohms technique can help eliminate the effect.

 
Table 1: Thermoelectric potentials.

This offset-compensated ohms technique is built into many Keithley instruments. When this feature is enabled, the measurement cycle now consists of two parts (Figure 4): the first part is measuring the voltage with stimulus current switched on, the second part is to measure it with the stimulus current switched off.

 
Figure 4.

Subtracting the latter from the former will subtract out the errors due to the thermoelectric EMFs. Therefore, this technique will effectively eliminate accuracy issues due to temperature drift.

  • External interference. External interference introduces both AC and DC errors into signal measurements. The most common form of external noise ‘pick-up’ is 50 or 60 Hz line pick-up, depending on where in the world the measurements are being made. Picking up millivolts of noise is not uncommon, especially when measurements are made near fluorescent lights.

The signal components of noise superimposed on a DC signal being measured may result in highly inaccurate and fluctuating measurements.

As shown in Figure 5, the measured value will very much depend on where the measurement is carried out in relation to the sine wave. Many modern instruments allow users to set the integration period in relation to the number of power line cycles.

In other words, a setting of 1 NPLC will result in the measurement being integrated for 20 milliseconds (if 50 Hz) and 16.67 milliseconds (if 60 Hz), which will eliminate any mains-generated noise. The performance improvements this feature make possible are often dramatic.

 
Figure 5.

  • Theoretical measurement limits. The laws of physics provide a fundamental limit of how low a signal can be resolved because every system will generate some level of voltage and current noise.

Figure 6 identifies the levels of voltage that are impossible to measure, as well as the levels approaching the theoretical limits of voltage measurement.

 
Figure 6.

Finally, the fourth step is once the test system has been built and verified, it’s ready to begin making measurements in which users can have confidence. However, it’s important to recheck the performance of any test set-up regularly. Because of component drift, the accuracy of an instrument will vary over time, so ensure that the instrumentation is calibrated regularly.

By Derek MacLachlan, staff applications engineer, Keithley Instruments

Related Articles

Turning RFID tags into scalable passive sensors

Researchers from UC San Diego have used low-cost RFID tags to develop analog passive sensors that...

Smart buoy tech live streams data 24/7 from sea floor

A Perth-based company has developed the Nodestream protocol, a bespoke solution that uses...

Engineers discover a new way to control atomic nuclei as “qubits”

Using lasers, researchers can directly control a property of nuclei called spin, that can encode...


  • All content Copyright © 2024 Westwick-Farrow Pty Ltd