Sometimes you have to measure a voltage that has a high output impedance. To do this accurately you need to know the input impedance of your voltmeter. Ideally it should be much much larger than the impedance of what you're trying to measure. The following figure illustrates the situation.
The voltage we're trying to measure is Vx and it has an output resistance of Rx. The meter has an input resistance of Rm. The voltage measured by the meter will then be
V=VxRmRx+Rm=Vx1+Rx/Rm
So you can see that V≈Vx only when Rm>>Rx. If you happen to know the values of Rx and Rm then you can determine what Vx is even when the two resistances are of the same magnitude. How do you measure the input resistance of your voltmeter? If you apply a known Vx using a known Rx then you can solve for Rm in the above equation.
Rm=RxVxV−1
We did this for the old Scope multimeter you see in the following picture using Vx=12v, Rx=1MΩ.

The meter read 10.98v which gave us Rm=10MΩ. This is pretty typical for most handheld multimeters.
Some really cheap multimeters such as the one shown below that we got at Harbor Freight Tools showed an input resistance of only 1MΩ. You have to be careful about what you measure with a meter like that. In some cases it can give you very inaccurate readings.

© 2010-2023 Stefan Hollos and Richard Hollos
Tweet

blog comments powered by Disqus