Numerals in Unicode

From Wikipedia, the free encyclopedia

A numeral (often called number in Unicode) is a character that denotes a number. Decimal is used widely in various writing systems throughout the world, however the graphemes representing the decimal digits differ widely, therefore Unicode includes 22 different sets of graphemes for the decimal digits, and also various decimal points, thousands separators, negative signs, etc. Unicode also includes several non-Decimal numerals such as Aegean numerals, Roman numerals, counting rod numerals, Cuneiform numerals and ancient Greek numerals. There is also a large number of typographical variations of the Western Arabic numerals provided for specialized mathematical use and for compatibility with earlier character sets, such as ² or ②, and composite characters such as ½.

Numerals by numeric property[]

Grouped by their numerical property as used in a text, Unicode has four values for Numeric Type. First there is the "not a number" type. Then there are decimal-radix numbers, commonly used in Western style decimals (plain 0–9), there are numbers that are not part of a decimal system such as Roman numbers, and decimal numbers in typographic context, such as encircled numbers. Not noted is a numbering like "A. B. C." for chapter numbering.

Numeric Type[a][b] (Unicode character property)
Numeric type Code Has numeric value Example Remarks
Not numeric None No
  • A
  • X (Latin)
  • !
  • Д
  • μ
Numeric Value="NaN"
Decimal De Yes
  • 0
  • 1
  • 9
  •  (Devanagari 6)
  •  (Kannada 6)
WIKI