Hey there - I've now run into this problem twice in the last week. I want to be able to insert a constant voltage drop into a signal.
The first place where this would have been nice was that I had a comparator watching the drain and the source of an N-FET. I wanted the comparator to toggle the instant the drain went higher than the source (and the other way around, naturally). Unfortunately, the comparator that I was stuck using had fairly lousy input offset. It was OK for me to error more towards one state of the comparator, but not towards the other. So what I figured was that I'd drop one signal by as much as the worst case input offset (1.5mV). But I know of no way to do that. Instead, I just used a voltage divider that just barely dropped it by a constant amount, and since I knew the range of voltages that I'd be able to see I was able to make the minimum constant amount large enough to always be larger than the input offset.
Anyways - that worked - but dropping by a constant voltage would have better.
Now I have another application of a similar thing. I need to monitor the voltage of a signal that ranges from about 12-18V. It will be monitored by a singled ended ADC with a 3.0V max input. So I can just divide the signal by 6. However, that voltage will always range from
2-3V, so I'm losing two thirds of my ADC range. What a waste. What'd be awesome would be if I could put in a constant drop in that voltage so that I could use the full range of my ADC - either a 12V drop or a 2V drop would work, clearly. There are shunt references with voltages this large - so that'd definitely do it (but be fairly costly and large, as it seems most shunts are fairly large parts). Zeners are typically what I think about 5% or so, so that would be less than ideal.Am I missing some better way of doing this? I suspect a 10V or 2V drop is a lot easier to achieve than a 1.5mV drop, but I've been wrong before!
Thanks!
-Michael