Click here

double precision

Refers to a type of floating-point number that has more precision (that is, more digits to the right of the decimal point) than a single-precision number. The term double precision is something of a misnomer because the precision is not really double. The word double derives from the fact that a double-precision number uses twice as many bits as a regular floating-point number. For example, if a single-precision number requires 32 bits, its double-precision counterpart will be 64 bits long.

The extra bits increase not only the precision but also the range of magnitudes that can be represented. The exact amount by which the precision and range of magnitudes are increased depends on what format the program is using to represent floating-point values. Most computers use a standard format known as the IEEE floating-point format.



Top Terms
  • 1

    private cloud project

    Companies initiate private cloud projects to enable their IT infrastructure to become more capable of quickly adapting to continually evolving...

    Read more »

  • 2

    cloud backup

    A cloud backup (cloud computer backup) refers to backing up data to a remote, cloud-based server. As a form of cloud storage, cloud backup data is...

    Read more »

  • Click Here!

Connect with Webopedia

  • Android Apps for Remote Computer Access

    Remote access is useful when you need to collaborate with co-workers, help a friend with PC problems or grab files from your desktop.

    Read More »

Did You Know? Archive »

  • Quick Reference Archive »