In general, an

operation on

floating point numbers. This term is really only used in

computing, rather than

maths, as mathematicians usually deal with

reals as

infinite precision quantities (Except when dealing with things like

sampling, or applied

numerical techniques).

Usually, one expects processors to perform floating point operations more slowly than fixed point operations, as the hardware implementation of a fixed-point operation is incredibly simple (As in UG computer science students can make an adding circuit out of nand gates if they apply themselves.)

In contrast, floating point operations have to take account of the fact that there is both a mantissa and an exponent (Either explicitly or implicitly), and so mantissa overflow has to be detected, the exponent manipulated, oh, and the exponent on the operands might not be the same. The logic is complex, making optimisations harder, and so more costly. A common way of dealing with this is to do these operations in parallel, as in a vector processor.