This function assumes that the true precision of the coordinates is equivalent to the
greatest number of decimals in either the longitude or latitude that are not trailing
zeroes. In other words:
(-130.45670, 45.53000)
is interpreted as (-130.4567, 45.5300)
(-130.20000, 45.50000)
is interpreted as (-130.2, 45.5)
If we use (-130.45670, 45.53000)
as an example, these coordinates are interpreted
as (-130.4567, 45.5300)
and the greatest possible error is inferred as two
endpoints:
(-130.45670, 45.53000)
and (-130.45679, 45.53009)
The distance between these two is then calculated and returned.