On Tuesday, 27 February 2018 at 00:04:59 UTC, H. S. Teoh wrote:
A 64-bit double can only hold about 14-15 decimal digits of precision. Anything past that, and there's a chance your "different" numbers are represented by exactly the same bits and the computer can't tell the difference.T
I really miss not having a (C# like) decimal type.