Jump to content

Talk:Decimal32 floating-point format

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Widefox (talk | contribs) at 19:21, 21 September 2016 (Assessment: Computing: importance=Low, hardware=y, hardware-importance=Low, software=y, software-importance=Low (assisted)). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.
WikiProject iconComputing: Software Start‑class Low‑importance
WikiProject iconThis article is within the scope of WikiProject Computing, a collaborative effort to improve the coverage of computers, computing, and information technology on Wikipedia. If you would like to participate, please visit the project page, where you can join the discussion and see a list of open tasks.
StartThis article has been rated as Start-class on Wikipedia's content assessment scale.
LowThis article has been rated as Low-importance on the project's importance scale.
Taskforce icon
This article is supported by WikiProject Software (assessed as Low-importance).
Taskforce icon
This article is supported by Computer hardware task force (assessed as Low-importance).

What implementations are there of this format?

Concerns about this page

The IEEE 754 standard goes to great lengths to separate the representation of the valid set of numbers in (say) decimal32 from the encoding of interchange formats (a particular representation in a string of binary bits), yet this article seems to muddle the concepts. Might it be better to follow the way it is described in the standard (clause 3)?

There are other problems too: 0.000000×10^−95 to 9.999999×10^96 is not the full range of the format in any sense; also (to be pedantic) the name of the format, in the standard, starts with a lower-case d.

mfc (talk) 17:41, 30 December 2009 (UTC)[reply]