According to the datasheet of DS18B20 temperature sensor, it is clear that the sensor is digital so the output is in digital form.
The resolution of the temperature sensor is 9, 10, 11, or 12 bits, corresponding to increments of 0.5°C, 0.25°C, 0.125°C, and 0.0625°C, respectively, and by default it is 12-bits. Which means, for 12-bits, a change of 0.0625°C in temperature causes an increment of 1-bit.
But there is no information given on how much change in voltage will result from a change in Temperature (i.e. scale factor).
For example, there is an analog temperature sensor in which a change of 10mV in the output voltage is equal to a change of 1°C of temperature.
Is there any way to calculate the scale factor of the digital temperature sensor like the way we do for analog?