ADC in Digital Temperature Sensor DS18B20


According to the datasheet of DS18B20 temperature sensor, it is clear that the sensor is digital so the output is in digital form.

The resolution of the temperature sensor is 9, 10, 11, or 12 bits, corresponding to increments of 0.5°C, 0.25°C, 0.125°C, and 0.0625°C, respectively, and by default it is 12-bits. Which means, for 12-bits, a change of 0.0625°C in temperature causes an increment of 1-bit.

But there is no information given on how much change in voltage will result from a change in Temperature (i.e. scale factor).
For example, there is an analog temperature sensor in which a change of 10mV in the output voltage is equal to a change of 1°C of temperature.

Is there any way to calculate the scale factor of the digital temperature sensor like the way we do for analog?

Best Answer

All you really need to know is contained here: -

enter image description here

Internally the device may have a sensor that produces 10uA per degree or 10mV per degree or it may be like a thermistor or RTD and vary its resistance to indictate temperature. This is Maxim IP and we are not allowed to know what method they use unless they tell us. If it's not in the data sheet then we cannot know - but you could assume it is one of the four methods I've mentioned and give details in your presentation of all ideas.

If you want to know the types of constant voltage or current sensors that could be used try looking up the AD590 (constant current at 10uA per degC) or the TMP36 (10mV per degC). I'm not aware of a particular name that describes the technology.