Only specialized meters can provide virtually infinite input R at
voltages above the 10 to 20 V or so native range of conventional
amplifiers, so you have to use some kind of attenuator to cover the
higher ranges anyway. 10 megs and 1 meg (and sometimes 11) are the
traditional values used, with 10 of course providing less loading on
the signal source. It is difficult to get good resistor precision,
stability, and voltage coefficient at higher values, so 10 megs is a
good compromise.
As far as I know, no DVM uses an actual 10 or 1 gigaohm resistor for
its input termination - that's just an equivalent input R range
(sometimes just a part spec right from the datasheet) for the high
impedance opamps and JFET circuits typically used to amplify DC at
reasonably high accuracy and low noise. All it means is that it is
nearly non-loading to conventional DC circuits. If there is an actual
resistor this large in there, then it is just to get the input near
zero when it's disconnected - it will read the bias current times the
resistance, which can be quite large. If the applied input voltage
exceeds the native range, the protection circuitry will take over.
For ultra-high Z applications, the equivalent input R would need to
be in the teraohm range instead, using electrometer-class opamps,
with much lower bias current, but higher offset voltage and noise.
If you put DVMs in the low ranges below 10 or 20 V, ones without
actual termination R will tend to drift off due to input bias
current. Once it's connected, the effect is much smaller (but not
zero) since the source R is usually comparatively very small. One way
to always assure a zero reading is to have a definite and fairly low
(to not show bias current too obviously) input R, so there's the 10
megs option. It's also possible to make the actual value of the input
R (and not just the dividing ratios) very precise - or measure it -
so that its effect on measurements at known source resistances can be
figured out.
As you have already figured out, in auto-ranging, a non-terminated
DVM left disconnected and unattended will form a relaxation
oscillator and tend to wear out its front-end relays. Seeing no
signal in the higher ranges, the system will switch down to the lower
ranges and be OK until the input drifts off to a range limit, then it
will up-range until it reaches one with an attenuator, then the
signal goes back to zero, and the process repeats.
Ed
At 07:23 AM 4/10/2014, you wrote:
There is no suggestion in the specifications for the 34401A that the
accuracy suffers by selecting 10G ohm input resistance on the .1 to
10V range so why would they make 10M ohm the default? I can think of
very few cases where having the 10M ohm i/p resistor switched in is
better for accuracy than not.
On the other hand 10M is sufficiently low to produce significant
errors on a 6 1/2 digit DVM for sources with resistances as low as
10 ohms. Measuring 1V divided by a 100k/100k ohm divider for example
causes a .5% error - 502.488mV instead of 500.000mV. That might not
be a problem but I wouldn't be surprised if this catches a lot of
people out (including me) when not pausing to do the mental
arithmetic to estimate the error. It's just too easy to be seduced
by all those digits into thinking you've made an accurate
measurement even though you discarded those last three digits.
And if it's not a problem then you probably don't need an expensive
6 1/2 digit meter in the first place.
It's a small point I agree but it can get irritating to have to keep
going into the measurement menus to change it when the meter is
turned on when measuring high impedance sources (e.g. capacitor
leakage testing).
It can't be to improve i/p protection as 10M is too high to make any
significant difference to ESD and in any case there is plenty of
other over-voltage protection. OK. it provides a path for the DC
amplifier's input bias current, specified to be < 30pA at 25 degrees
C, but I imagine that varies significantly from one meter to the
next, and with temperature, so not useful for nulling out that error.
So why would they do this? Could it be psychological? By limiting
the drift caused by the i/p bias current to 300uV max when the meter
is left unconnected? A voltmeter with a rapidly drifting reading
(several mV/s) when not connected to anything is a bit disconcerting
and would probably lead to complaints that the meter is obviously
faulty to users who are used to DVMs which read 0V when open circuit
- because they have i/p resistance << 10G ohms and don't have the
resolution to show the offset voltage caused by the i/p bias current.
Personally I'd have though that the default should be the other way
round - especially given that there is no indication on the front
panel or display as to which i/p resistance is currently selected.
Any thoughts? What do other meters do?
Tony H
_______________________________________________
volt-nuts mailing list -- volt-nuts@febo.com
To unsubscribe, go to https://www.febo.com/cgi-bin/mailman/listinfo/volt-nuts
and follow the instructions there.
_______________________________________________
volt-nuts mailing list -- volt-nuts@febo.com
To unsubscribe, go to https://www.febo.com/cgi-bin/mailman/listinfo/volt-nuts
and follow the instructions there.