It looks like you're new here. If you want to get involved, click one of these buttons!
What is the difference between Decimal, Float and Double in C#? What are the situations that each of these should be used?
decimal is a floating decimal point type.
float and double are floating binary point types.
The main difference is the precision.
Float - 7 digits (32 bit)
Double-15-16 digits (64 bit)
Decimal -28-29 significant digits (128 bit)
Decimals have much higher precision and are usually used within financial applications that require a high degree of accuracy. Decimals are much slower (up to 20X times in some tests) than a double/float.
Decimals and Floats/Doubles cannot be compared without a cast whereas Floats and Doubles can. Decimals also allow the encoding or trailing zeros.
(Extracted answer. Please visit Here for more details.)