Edit this page on github
Frequently asked questions on comdb2
Decimal 32/64/128 datatype
Comdb2 implements the decimal format for real numbers, as described in IEEE 754-2008 standard, published in August 2008. The decimal format avoids the pitfall of floating point representation to provide error-less arithmetic using real numbers.
Comdb2 provides three datatypes,
decimal128, providing single,
double and quad precision, respectively. Each number is represented as a triplet (sign, significand, exponent),
and the number value is ((-1)^sign)significand(10^exponent).
These are the limitations of each datatype:
decimal32supports exponents between -95 and +96; significand has 7 digits (i.e. 0.000000-9.999999).
The range of numbers representable by this format is +-0.000000x10−95 to +-9.999999x10+96
decimal64supports exponents between -383 and +384; significand has 16 digits (i.e. 0.000000000000000-9.999999999999999). The range of numbers is +-0.000000000000000x10−383 to +-9.999999999999999x10+384
decimal128supports exponents between -6143 and +6144; significand has 34 digits (i.e. 0.000000000000000000000000000000000-9.999999999999999999999999999999999).
The range of numbers is +-0.000000000000000000000000000000000x10−6143 to +-9.999999999999999999999999999999999x10+6144
Note: the decimal representation allows for un-normalized significand values. This makes decimal manipulation a slower operation compared to binarty floating point. Since comdb2 allows decimal types to be used as part of the key, it was paramount that a simple binary comparison is enough to determine the ordering of two decimal values. Comdb2 chose to normalize the significand and reduce the range of the numbers by the size of the significand.