Open main menu

In applied mathematics, a number is normalized when it is written in scientific notation with one non-zero decimal digit before the decimal point.[1] Thus, a real number, when written out in normalized scientific notation, is as follows:

where n is an integer, ..., are the digits of the number in base 10, and is not zero. That is, its leading digit (i.e., leftmost) is not zero and is followed by the decimal point. This is the standard form of scientific notation. An alternative style is to have the first non-zero digit after the decimal point.

Contents

ExamplesEdit

As examples, the number   in normalized form is

 ,

while the number −0.00574012 in normalized form is

 

Clearly, any non-zero real number can be normalized.

Other basesEdit

The same definition holds if the number is represented in another radix (that is, base of enumeration), rather than base 10.

In base b a normalized number will have the form

 

where again   and the digits,         ..., are integers between   and  .

In many computer systems, floating point numbers are represented internally using this normalized form for their binary representations; for details, see normal number (computing). Converting a number to base two and normalizing it are the first steps in storing a real number as a binary floating-point number in a computer, though bases of eight and sixteen are also used. Although the point is described as floating, for a normalised floating point number its position is fixed, the movement being reflected in the different values of the power.

See alsoEdit

ReferencesEdit

  1. ^ Fleisch, Daniel; Kregenow, Julia (2013), A Student's Guide to the Mathematics of Astronomy, Cambridge University Press, p. 35, ISBN 9781107292550.