Jump to content

Precision (computer science)

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Minority Report (talk | contribs) at 13:13, 13 November 2004. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.
(diff) ← Previous revision | Latest revision (diff) | Newer revision → (diff)

In computer science, precision of a numerical quantity is a measure of the detail in which the quantity is expressed. This is usually measured in binary bits, but sometimes in decimal digits.

In java, one of the few programming languages with standardized precision data types, the following precisions are defined for the standard numerical types of the language. The ranges given are for signed integer values.


Precision of java data types
Type name Precision (binary bits) Range
byte 8 -128 to +127
short 16 -32,768 to 32,767
int 32 -2,147,483,648 to 2,147,483,647
long 64 -9,223,372,036,854,775,808 to -9,223,372,036,854,775,807