Decimal vs Double Speed

前端 未结 7 1022
暗喜
暗喜 2020-11-27 13:03

I write financial applications where I constantly battle the decision to use a double vs using a decimal.

All of my math works on numbers with no more than 5 decimal

相关标签:
7条回答
  • 2020-11-27 13:52

    Just use a long and multiply by a power of 10. After you're done, divide by the same power of 10.

    0 讨论(0)
  • 2020-11-27 13:58
    1. Floating point arithmetic will almost always be significantly faster because it is supported directly by the hardware. So far almost no widely used hardware supports decimal arithmetic (although this is changing, see comments).
    2. Financial applications should always use decimal numbers, the number of horror stories stemming from using floating point in financial applications is endless, you should be able to find many such examples with a Google search.
    3. While decimal arithmetic may be significantly slower than floating point arithmetic, unless you are spending a significant amount of time processing decimal data the impact on your program is likely to be negligible. As always, do the appropriate profiling before you start worrying about the difference.
    0 讨论(0)
  • 2020-11-27 13:59
    1. Yes; software arithmetic really is 100 times slower than hardware. Or, at least, it is a lot slower, and a factor of 100, give or take an order of magnitude, is about right. Back in the bad old days when you could not assume that every 80386 had an 80387 floating-point co-processor, then you had software simulation of binary floating point too, and that was slow.
    2. No; you are living in a fantasy land if you think that a pure binary floating point can ever exactly represent all decimal numbers. Binary numbers can combine halves, quarters, eighths, etc, but since an exact decimal of 0.01 requires two factors of one fifth and one factor of one quarter (1/100 = (1/4)*(1/5)*(1/5)) and since one fifth has no exact representation in binary, you cannot exactly represent all decimal values with binary values (because 0.01 is a counter-example which cannot be represented exactly, but is representative of a huge class of decimal numbers that cannot be represented exactly).

    So, you have to decide whether you can deal with the rounding before you call ToString() or whether you need to find some other mechanism that will deal with rounding your results as they are converted to a string. Or you can continue to use decimal arithmetic since it will remain accurate, and it will get faster once machines are released that support the new IEEE 754 decimal arithmetic in hardware.

    Obligatory cross-reference: What Every Computer Scientist Should Know About Floating-Point Arithmetic. That's one of many possible URLs.

    Information on decimal arithmetic and the new IEEE 754:2008 standard at this Speleotrove site.

    0 讨论(0)
  • 2020-11-27 14:00

    Decimals should always be used for financial calculations. The size of the numbers isn't important.

    The easiest way for me to explain is via some C# code.

    double one = 3.05;
    double two = 0.05;
    
    System.Console.WriteLine((one + two) == 3.1);
    

    That bit of code will print out False even though 3.1 is equal to 3.1...

    Same thing...but using decimal:

    decimal one = 3.05m;
    decimal two = 0.05m;
    
    System.Console.WriteLine((one + two) == 3.1m);
    

    This will now print out True!

    If you want to avoid this sort of issue, I recommend you stick with decimals.

    0 讨论(0)
  • 2020-11-27 14:02

    There are two separable issues here. One is whether the double has enough precision to hold all the bits you need, and the other is where it can represent your numbers exactly.

    As for the exact representation, you are right to be cautious, because an exact decimal fraction like 1/10 has no exact binary counterpart. However, if you know that you only need 5 decimal digits of precision, you can use scaled arithmetic in which you operate on numbers multiplied by 10^5. So for example if you want to represent 23.7205 exactly you represent it as 2372050.

    Let's see if there is enough precision: double precision gives you 53 bits of precision. This is equivalent to 15+ decimal digits of precision. So this would allow you five digits after the decimal point and 10 digits before the decimal point, which seems ample for your application.

    I would put this C code in a .h file:

    typedef double scaled_int;
    
    #define SCALE_FACTOR 1.0e5  /* number of digits needed after decimal point */
    
    static inline scaled_int adds(scaled_int x, scaled_int y) { return x + y; }
    static inline scaled_int muls(scaled_int x, scaled_int y) { return x * y / SCALE_FACTOR; }
    
    static inline scaled_int scaled_of_int(int x) { return (scaled_int) x * SCALE_FACTOR; }
    static inline int intpart_of_scaled(scaled_int x) { return floor(x / SCALE_FACTOR); }
    static inline int fraction_of_scaled(scaled_int x) { return x - SCALE_FACTOR * intpart_of_scaled(x); }
    
    void fprint_scaled(FILE *out, scaled_int x) {
      fprintf(out, "%d.%05d", intpart_of_scaled(x), fraction_of_scaled(x));
    }
    

    There are probably a few rough spots but that should be enough to get you started.

    No overhead for addition, cost of a multiply or divide doubles.

    If you have access to C99, you can also try scaled integer arithmetic using the int64_t 64-bit integer type. Which is faster will depend on your hardware platform.

    0 讨论(0)
  • 2020-11-27 14:03

    Always use Decimal for any financial calculations or you will be forever chasing 1cent rounding errors.

    0 讨论(0)
提交回复
热议问题