On Thursday, 23 November 2017 at 14:47:21 UTC, aberba wrote:
Some suggest working with the lowest currency denomination to avoid decimal precision handling and only convert to the highest denominations (decimal) when displaying. I've also seen some use decimal value handling libraries.

I'm thinking lowest denominations will result in extremely large values that D's type system cannot store (if such large values makes sense or can happen with money in real life).

What will be your advise on the type to use by default, the currency denominations (100p instead of 1.0 dollars), and cost of computation.

From d-money doc:

"Here the design decision is to use an integer for the internal representation. This limits the amounts you can use. For example, if you decide to use 4 digits behind the comma, the maximum number is 922,337,203,685,477.5807 or roughly 922 trillion. The US debt is currently in the trillions, so there are certainly cases where this representation is not applicable. However, we can check overflow, so if it happens, you get an exception thrown and notice it right away. The upside of using an integer is performance and a deterministic arithmetic all programmers are familiar with."

922 trillion (+4 digits after comma) should be enough for common computations.

Reply via email to