The Unicode standard is derived originally from ASCII. ASCII was an 8 bit coded character set with a limited number of individual character values. Unicode is a 16 bit coding that supports all the required characters for the major languages of the world, plus the technical symbols in common use.
For most usage, the character set will be the lower 128 characters that roughly correspond to the ASCII character table. However, internationalization and localization work in progress suggests that non-English speaking users should be able to declare identifiers in their own natural language with its particular character sets and special symbols. Internet web site domain name standards are undergoing some revision to support double byte characters. The ECMAScript edition 3 improves support for handling localized strings, numbers and dates.
The Unicode standard is due for updating to a new edition as the version 2.0 is somewhat old now and does not support Euro currency symbols among other things. A version 3.0 standard is reported to be on the way.
Unicode characters are more properly referred to as code points.
|See also:||Character set, Character-case mapping, Control character, Equal to (==), Greater than (>), Greater than or equal to (>=), Identically equal to (===), isLower(), isUpper(), Less than (<), Less than or equal to (<=), Locale-specific behavior, Multi-byte character, NOT Equal to (!=), NOT Identically equal to (!==), String, String literal|
ECMA 262 edition 2 - section - 2
ECMA 262 edition 2 - section - 6
ECMA 262 edition 3 - section - 2
ECMA 262 edition 3 - section - 6