fbpx
Wikipedia

Character encoding

Character encoding is the process of assigning numbers to graphical characters, especially the written characters of human language, allowing them to be stored, transmitted, and transformed using digital computers.[1] The numerical values that make up a character encoding are known as "code points" and collectively comprise a "code space", a "code page", or a "character map".

Punched tape with the word "Wikipedia" encoded in ASCII. Presence and absence of a hole represents 1 and 0, respectively; for example, "W" is encoded as "1010111".

Early character codes associated with the optical or electrical telegraph could only represent a subset of the characters used in written languages, sometimes restricted to upper case letters, numerals and some punctuation only. The low cost of digital representation of data in modern computer systems allows more elaborate character codes (such as Unicode) which represent most of the characters used in many written languages. Character encoding using internationally accepted standards permits worldwide interchange of text in electronic form.

History

The history of character codes illustrates the evolving need for machine-mediated character-based symbolic information over a distance, using once-novel electrical means. The earliest codes were based upon manual and hand-written encoding and cyphering systems, such as Bacon's cipher, Braille, international maritime signal flags, and the 4-digit encoding of Chinese characters for a Chinese telegraph code (Hans Schjellerup, 1869). With the adoption of electrical and electro-mechanical techniques these earliest codes were adapted to the new capabilities and limitations of the early machines. The earliest well-known electrically transmitted character code, Morse code, introduced in the 1840s, used a system of four "symbols" (short signal, long signal, short space, long space) to generate codes of variable length. Though some commercial use of Morse code was via machinery, it was often used as a manual code, generated by hand on a telegraph key and decipherable by ear, and persists in amateur radio and aeronautical use. Most codes are of fixed per-character length or variable-length sequences of fixed-length codes (e.g. Unicode).[2]

Common examples of character encoding systems include Morse code, the Baudot code, the American Standard Code for Information Interchange (ASCII) and Unicode. Unicode, a well-defined and extensible encoding system, has supplanted most earlier character encodings, but the path of code development to the present is fairly well known.

The Baudot code, a five-bit encoding, was created by Émile Baudot in 1870, patented in 1874, modified by Donald Murray in 1901, and standardized by CCITT as International Telegraph Alphabet No. 2 (ITA2) in 1930. The name baudot has been erroneously applied to ITA2 and its many variants. ITA2 suffered from many shortcomings and was often improved by many equipment manufacturers, sometimes creating compatibility issues. In 1959 the U.S. military defined its Fieldata code, a six-or seven-bit code, introduced by the U.S. Army Signal Corps. While Fieldata addressed many of the then-modern issues (e.g. letter and digit codes arranged for machine collation), it fell short of its goals and was short-lived. In 1963 the first ASCII code was released (X3.4-1963) by the ASCII committee (which contained at least one member of the Fieldata committee, W. F. Leubbert), which addressed most of the shortcomings of Fieldata, using a simpler code. Many of the changes were subtle, such as collatable character sets within certain numeric ranges. ASCII63 was a success, widely adopted by industry, and with the follow-up issue of the 1967 ASCII code (which added lower-case letters and fixed some "control code" issues) ASCII67 was adopted fairly widely. ASCII67's American-centric nature was somewhat addressed in the European ECMA-6 standard.[3]

 
Hollerith 80-column punch card with EBCDIC character set

Herman Hollerith invented punch card data encoding in the late 19th century to analyze census data. Initially, each hole position represented a different data element, but later, numeric information was encoded by numbering the lower rows 0 to 9, with a punch in a column representing its row number. Later alphabetic data was encoded by allowing more than one punch per column. Electromechanical tabulating machines represented date internally by the timing of pulses relative to the motion of the cards through the machine. When IBM went to electronic processing, starting with the IBM 603 Electronic Multiplier, it used a variety of binary encoding schemes that were tied to the punch card code.

IBM's Binary Coded Decimal (BCD) was a six-bit encoding scheme used by IBM as early as 1953 in its 702[4] and 704 computers, and in its later 7000 Series and 1400 series, as well as in associated peripherals. Since the punched card code then in use only allowed digits, upper-case English letters and a few special characters, six bits were sufficient. BCD extended existing simple four-bit numeric encoding to include alphabetic and special characters, mapping it easily to punch-card encoding which was already in widespread use. IBMs codes were used primarily with IBM equipment; other computer vendors of the era had their own character codes, often six-bit, but usually had the ability to read tapes produced on IBM equipment. BCD was the precursor of IBM's Extended Binary-Coded Decimal Interchange Code (usually abbreviated as EBCDIC), an eight-bit encoding scheme developed in 1963 for the IBM System/360 that featured a larger character set, including lower case letters.

The limitations of such sets soon became apparent,[to whom?] and a number of ad hoc methods were developed to extend them. The need to support more writing systems for different languages, including the CJK family of East Asian scripts, required support for a far larger number of characters and demanded a systematic approach to character encoding rather than the previous ad hoc approaches.[citation needed]

In trying to develop universally interchangeable character encodings, researchers in the 1980s faced the dilemma that, on the one hand, it seemed necessary to add more bits to accommodate additional characters, but on the other hand, for the users of the relatively small character set of the Latin alphabet (who still constituted the majority of computer users), those additional bits were a colossal waste of then-scarce and expensive computing resources (as they would always be zeroed out for such users). In 1985, the average personal computer user's hard disk drive could store only about 10 megabytes, and it cost approximately US$250 on the wholesale market (and much higher if purchased separately at retail),[5] so it was very important at the time to make every bit count.

The compromise solution that was eventually found and developed into Unicode[vague] was to break the assumption (dating back to telegraph codes) that each character should always directly correspond to a particular sequence of bits. Instead, characters would first be mapped to a universal intermediate representation in the form of abstract numbers called code points. Code points would then be represented in a variety of ways and with various default numbers of bits per character (code units) depending on context. To encode code points higher than the length of the code unit, such as above 256 for eight-bit units, the solution was to implement variable-length encodings where an escape sequence would signal that subsequent bits should be parsed as a higher code point.

Terminology

Informally, the terms "character encoding", "character map", "character set" and "code page" are often used interchangeably.[6] Historically, the same standard would specify a repertoire of characters and how they were to be encoded into a stream of code units — usually with a single character per code unit. However, due to the emergence of more sophisticated character encodings, the distinction between these terms has become important.

  • A character is a minimal unit of text that has semantic value.[6][7]
  • A character set is a collection of elements used to represent text.[6][7] For example, the Latin alphabet and Greek alphabet are both character sets.
  • A coded character set is a character set mapped to set of unique numbers.[7] For historical reasons, this is also often referred to as a code page.[6]
  • A character repertoire is the set of characters that can be represented by a particular coded character set.[7][8] The repertoire may be closed, meaning that no additions are allowed without creating a new standard (as is the case with ASCII and most of the ISO-8859 series); or it may be open, allowing additions (as is the case with Unicode and to a limited extent Windows code pages).[8]
  • A code point is a value or position of a character in a coded character set.[7]
  • A code space is the range of numerical values spanned by a coded character set.[7][9]
  • A code unit is the minimum bit combination that can represent a character in a character encoding (in computer science terms, it is the word size of the character encoding).[7][9] For example, common code units include 7-bit, 8-bit, 16-bit, and 32-bit. In some encodings, some characters are encoded using multiple code units; such an encoding is referred to as a variable-width encoding.

Code pages

"Code page" is a historical name for a coded character set.

Originally, a code page referred to a specific page number in the IBM standard character set manual, which would define a particular character encoding.[10] Other vendors, including Microsoft, SAP, and Oracle Corporation, also published their own sets of code pages; the most well-known code page suites are "Windows" (based on Windows-1252) and "IBM"/"DOS" (based on code page 437).

Despite no longer referring to specific page numbers in a standard, many character encodings are still referred to by their code page number; likewise, the term "code page" is often still used to refer to character encodings in general.

The term "code page" is not used in Unix or Linux, where "charmap" is preferred, usually in the larger context of locales. IBM's Character Data Representation Architecture (CDRA) designates entities with coded character set identifiers (CCSIDs), each of which is variously called a "charset", "character set", "code page", or "CHARMAP".[9]

Code units

The code unit size is equivalent to the bit measurement for the particular encoding:

Code points

A code point is represented by a sequence of code units. The mapping is defined by the encoding. Thus, the number of code units required to represent a code point depends on the encoding:

  • UTF-8: code points map to a sequence of one, two, three or four code units.
  • UTF-16: code units are twice as long as 8-bit code units. Therefore, any code point with a scalar value less than U+10000 is encoded with a single code unit. Code points with a value U+10000 or higher require two code units each. These pairs of code units have a unique term in UTF-16: "Unicode surrogate pairs".
  • UTF-32: the 32-bit code unit is large enough that every code point is represented as a single code unit.
  • GB 18030: multiple code units per code point are common, because of the small code units. Code points are mapped to one, two, or four code units.[11]

Characters

Exactly what constitutes a character varies between character encodings.

For example, for letters with diacritics, there are two distinct approaches that can be taken to encode them: they can be encoded either as a single unified character (known as a precomposed character), or as separate characters that combine into a single glyph. The former simplifies the text handling system, but the latter allows any letter/diacritic combination to be used in text. Ligatures pose similar problems.

Exactly how to handle glyph variants is a choice that must be made when constructing a particular character encoding. Some writing systems, such as Arabic and Hebrew, need to accommodate things like graphemes that are joined in different ways in different contexts, but represent the same semantic character.

Unicode encoding model

Unicode and its parallel standard, the ISO/IEC 10646 Universal Character Set, together constitute a unified standard for character encoding. Rather than mapping characters directly to bytes, Unicode separately defines a coded character set that maps characters to unique natural numbers (code points), how those code points are mapped to a series of fixed-size natural numbers (code units), and finally how those units are encoded as a stream of octets (bytes). The purpose of this decomposition is to establish a universal set of characters that can be encoded in a variety of ways. To describe this model precisely, Unicode uses its own set of terminology to describe its process:[9]

An abstract character repertoire (ACR) is the full set of abstract characters that a system supports. Unicode has an open repertoire, meaning that new characters will be added to the repertoire over time.

A coded character set (CCS) is a function that maps characters to code points (each code point represents one character). For example, in a given repertoire, the capital letter "A" in the Latin alphabet might be represented by the code point 65, the character "B" to 66, and so on. Multiple coded character sets may share the same character repertoire; for example ISO/IEC 8859-1 and IBM code pages 037 and 500 all cover the same repertoire but map them to different code points.

A character encoding form (CEF) is the mapping of code points to code units to facilitate storage in a system that represents numbers as bit sequences of fixed length (i.e. practically any computer system). For example, a system that stores numeric information in 16-bit units can only directly represent code points 0 to 65,535 in each unit, but larger code points (say, 65,536 to 1.4 million) could be represented by using multiple 16-bit units. This correspondence is defined by a CEF.

A character encoding scheme (CES) is the mapping of code units to a sequence of octets to facilitate storage on an octet-based file system or transmission over an octet-based network. Simple character encoding schemes include UTF-8, UTF-16BE, UTF-32BE, UTF-16LE, and UTF-32LE; compound character encoding schemes, such as UTF-16, UTF-32 and ISO/IEC 2022, switch between several simple schemes by using a byte order mark or escape sequences; compressing schemes try to minimize the number of bytes used per code unit (such as SCSU and BOCU).

Although UTF-32BE and UTF-32LE are simpler CESes, most systems working with Unicode use either UTF-8, which is backward compatible with fixed-length ASCII and maps Unicode code points to variable-length sequences of octets, or UTF-16BE,[citation needed] which is backward compatible with fixed-length UCS-2BE and maps Unicode code points to variable-length sequences of 16-bit words. See comparison of Unicode encodings for a detailed discussion.

Finally, there may be a higher-level protocol which supplies additional information to select the particular variant of a Unicode character, particularly where there are regional variants that have been 'unified' in Unicode as the same character. An example is the XML attribute xml:lang.

The Unicode model uses the term "character map" for other systems which directly assign a sequence of characters to a sequence of bytes, covering all of the CCS, CEF and CES layers.[9]

Unicode code points

In Unicode, a character can be referred to as 'U+' followed by its codepoint value in hexadecimal. The range of valid code points (the codespace) for the Unicode standard is U+0000 to U+10FFFF, inclusive, divided in 17 planes, identified by the numbers 0 to 16. Characters in the range U+0000 to U+FFFF are in plane 0, called the Basic Multilingual Plane (BMP). This plane contains most commonly-used characters. Characters in the range U+10000 to U+10FFFF in the other planes are called supplementary characters.

The following table shows examples of code point values:

Character Unicode code point Glyph
Latin A U+0041 Α
Latin sharp S U+00DF ß
Han for East U+6771
Ampersand U+0026 &
Inverted exclamation mark U+00A1 ¡
Section sign U+00A7 §

Example

Consider a string of the letters "ab̲c𐐀"—that is, a string containing a Unicode combining character (U+0332 ̲ ) as well a supplementary character (U+10400 𐐀 ). This string has several Unicode representations which are logically equivalent, yet while each is suited to a diverse set of circumstances or range of requirements:

  • Four composed characters:
    a, , c, 𐐀
  • Five graphemes:
    a, b, _, c, 𐐀
  • Five Unicode code points:
    U+0061, U+0062, U+0332, U+0063, U+10400
  • Five UTF-32 code units (32-bit integer values):
    0x00000061, 0x00000062, 0x00000332, 0x00000063, 0x00010400
  • Six UTF-16 code units (16-bit integers)
    0x0061, 0x0062, 0x0332, 0x0063, 0xD801, 0xDC00
  • Nine UTF-8 code units (8-bit values, or bytes)
    0x61, 0x62, 0xCC, 0xB2, 0x63, 0xF0, 0x90, 0x90, 0x80

Note in particular that 𐐀 is represented with either one 32-bit value (UTF-32), two 16-bit values (UTF-16), or four 8-bit values (UTF-8). Although each of those forms uses the same total number of bits (32) to represent the glyph, it is not obvious how the actual numeric byte values are related.

Transcoding

As a result of having many character encoding methods in use (and the need for backward compatibility with archived data), many computer programs have been developed to translate data between character encoding schemes, a process known as transcoding. Some of these are cited below.

Cross-platform:

  • Web browsers – most modern web browsers feature automatic character encoding detection. On Firefox 3, for example, see the View/Character Encoding submenu.
  • iconv – a program and standardized API to convert encodings
  • luit – a program that converts encoding of input and output to programs running interactively
  • convert_encoding.py – a Python-based utility to convert text files between arbitrary encodings and line endings[12]
  • decodeh.py – an algorithm and module to heuristically guess the encoding of a string[13]
  • International Components for Unicode – A set of C and Java libraries to perform charset conversion. uconv can be used from ICU4C.
  • chardet – This is a translation of the Mozilla automatic-encoding-detection code into the Python computer language.
  • The newer versions of the Unix file command attempt to do a basic detection of character encoding (also available on Cygwin).
  • charset – C++ template library with simple interface to convert between C++/user-defined streams. charset defined many character-sets and allows you to use Unicode formats with support of endianness.

Unix-like:

  • cmv – a simple tool for transcoding filenames.[14]
  • convmv – converts a filename from one encoding to another.[15]
  • cstocs – converts file contents from one encoding to another for the Czech and Slovak languages.
  • enca – analyzes encodings for given text files.[16]
  • recode – converts file contents from one encoding to another.[17]
  • utrac – converts file contents from one encoding to another.[18]

Windows:

  • Encoding.Convert – .NET API[19]
  • MultiByteToWideChar/WideCharToMultiByte – to convert from ANSI to Unicode & Unicode to ANSI[20][21]
  • cscvt – a character set conversion tool[22]
  • enca – analyzes encodings for given text files.[23]

See also

Common character encodings

References

  1. ^ Definition from The Tech Terms Dictionary
  2. ^ Tom Henderson (17 April 2014). "Ancient Computer Character Code Tables – and Why They're Still Relevant". Smartbear. Retrieved 29 April 2014.
  3. ^ Tom Jennings (1 March 2010). "An annotated history of some character codes". Retrieved 1 November 2018.
  4. ^ "IBM Electronic Data-Processing Machines Type 702 Preliminary Manual of Information" (PDF). 1954. p. 80. 22-6173-1. Archived (PDF) from the original on 9 October 2022.
  5. ^ Strelho, Kevin (15 April 1985). "IBM Drives Hard Disks to New Standards". InfoWorld. Popular Computing Inc. pp. 29–33. Retrieved 10 November 2020.
  6. ^ a b c d Shawn Steele (15 March 2005). "What's the difference between an Encoding, Code Page, Character Set and Unicode?". Microsoft Docs.
  7. ^ a b c d e f g "Glossary of Unicode Terms". Unicode Consortium.
  8. ^ a b "Chapter 3: Conformance". The Unicode Standard Version 15.0 – Core Specification (PDF). Unicode Consortium. September 2022. ISBN 978-1-936213-32-0.
  9. ^ a b c d e Whistler, Ken; Freytag, Asmus (11 November 2022). "UTR#17: Unicode Character Encoding Model". Unicode Consortium. Retrieved 12 August 2023.
  10. ^ "VT510 Video Terminal Programmer Information". Digital Equipment Corporation (DEC). 7.1. Character Sets - Overview. from the original on 26 January 2016. Retrieved 15 February 2017. In addition to traditional DEC and ISO character sets, which conform to the structure and rules of ISO 2022, the VT510 supports a number of IBM PC code pages (page numbers in IBM's standard character set manual) in PCTerm mode to emulate the console terminal of industry-standard PCs.
  11. ^ "Terminology (The Java Tutorials)". Oracle. Retrieved 25 March 2018.
  12. ^ convert encoding.py on GitHub
  13. ^ . Archived from the original on 8 January 2008.
  14. ^ "CharsetMove – Simple Tool for Transcoding Filenames".
  15. ^ "Convmv – converts filenames from one encoding to another".
  16. ^ . Archived from the original on 4 December 2010. Retrieved 11 March 2008.
  17. ^ recode on GitHub
  18. ^ "Utrac Homepage".
  19. ^ "Encoding.Convert Method". Microsoft .NET Framework Class Library.
  20. ^ "MultiByteToWideChar function (stringapiset.h)". Microsoft Docs. 13 October 2021.
  21. ^ "WideCharToMultiByte function (stringapiset.h)". Microsoft Docs. 9 August 2022.
  22. ^ "Kalytta's Character Set Converter".
  23. ^ . Archived from the original on 15 March 2012. Retrieved 31 March 2011.

Further reading

External links

  • Character sets registered by Internet Assigned Numbers Authority (IANA)
  • Characters and encodings, by Jukka Korpela
  • Unicode Technical Report #17: Character Encoding Model
  • Decimal, Hexadecimal Character Codes in HTML Unicode – Encoding converter
  • The Absolute Minimum Every Software Developer Absolutely, Positively Must Know About Unicode and Character Sets (No Excuses!) by Joel Spolsky (Oct 10, 2003)

character, encoding, process, assigning, numbers, graphical, characters, especially, written, characters, human, language, allowing, them, stored, transmitted, transformed, using, digital, computers, numerical, values, that, make, character, encoding, known, c. Character encoding is the process of assigning numbers to graphical characters especially the written characters of human language allowing them to be stored transmitted and transformed using digital computers 1 The numerical values that make up a character encoding are known as code points and collectively comprise a code space a code page or a character map Punched tape with the word Wikipedia encoded in ASCII Presence and absence of a hole represents 1 and 0 respectively for example W is encoded as 1010111 Early character codes associated with the optical or electrical telegraph could only represent a subset of the characters used in written languages sometimes restricted to upper case letters numerals and some punctuation only The low cost of digital representation of data in modern computer systems allows more elaborate character codes such as Unicode which represent most of the characters used in many written languages Character encoding using internationally accepted standards permits worldwide interchange of text in electronic form Contents 1 History 2 Terminology 2 1 Code pages 2 2 Code units 2 3 Code points 2 4 Characters 3 Unicode encoding model 3 1 Unicode code points 3 2 Example 4 Transcoding 5 See also 5 1 Common character encodings 6 References 7 Further reading 8 External linksHistory EditThe history of character codes illustrates the evolving need for machine mediated character based symbolic information over a distance using once novel electrical means The earliest codes were based upon manual and hand written encoding and cyphering systems such as Bacon s cipher Braille international maritime signal flags and the 4 digit encoding of Chinese characters for a Chinese telegraph code Hans Schjellerup 1869 With the adoption of electrical and electro mechanical techniques these earliest codes were adapted to the new capabilities and limitations of the early machines The earliest well known electrically transmitted character code Morse code introduced in the 1840s used a system of four symbols short signal long signal short space long space to generate codes of variable length Though some commercial use of Morse code was via machinery it was often used as a manual code generated by hand on a telegraph key and decipherable by ear and persists in amateur radio and aeronautical use Most codes are of fixed per character length or variable length sequences of fixed length codes e g Unicode 2 Common examples of character encoding systems include Morse code the Baudot code the American Standard Code for Information Interchange ASCII and Unicode Unicode a well defined and extensible encoding system has supplanted most earlier character encodings but the path of code development to the present is fairly well known The Baudot code a five bit encoding was created by Emile Baudot in 1870 patented in 1874 modified by Donald Murray in 1901 and standardized by CCITT as International Telegraph Alphabet No 2 ITA2 in 1930 The name baudot has been erroneously applied to ITA2 and its many variants ITA2 suffered from many shortcomings and was often improved by many equipment manufacturers sometimes creating compatibility issues In 1959 the U S military defined its Fieldata code a six or seven bit code introduced by the U S Army Signal Corps While Fieldata addressed many of the then modern issues e g letter and digit codes arranged for machine collation it fell short of its goals and was short lived In 1963 the first ASCII code was released X3 4 1963 by the ASCII committee which contained at least one member of the Fieldata committee W F Leubbert which addressed most of the shortcomings of Fieldata using a simpler code Many of the changes were subtle such as collatable character sets within certain numeric ranges ASCII63 was a success widely adopted by industry and with the follow up issue of the 1967 ASCII code which added lower case letters and fixed some control code issues ASCII67 was adopted fairly widely ASCII67 s American centric nature was somewhat addressed in the European ECMA 6 standard 3 Hollerith 80 column punch card with EBCDIC character setHerman Hollerith invented punch card data encoding in the late 19th century to analyze census data Initially each hole position represented a different data element but later numeric information was encoded by numbering the lower rows 0 to 9 with a punch in a column representing its row number Later alphabetic data was encoded by allowing more than one punch per column Electromechanical tabulating machines represented date internally by the timing of pulses relative to the motion of the cards through the machine When IBM went to electronic processing starting with the IBM 603 Electronic Multiplier it used a variety of binary encoding schemes that were tied to the punch card code IBM s Binary Coded Decimal BCD was a six bit encoding scheme used by IBM as early as 1953 in its 702 4 and 704 computers and in its later 7000 Series and 1400 series as well as in associated peripherals Since the punched card code then in use only allowed digits upper case English letters and a few special characters six bits were sufficient BCD extended existing simple four bit numeric encoding to include alphabetic and special characters mapping it easily to punch card encoding which was already in widespread use IBMs codes were used primarily with IBM equipment other computer vendors of the era had their own character codes often six bit but usually had the ability to read tapes produced on IBM equipment BCD was the precursor of IBM s Extended Binary Coded Decimal Interchange Code usually abbreviated as EBCDIC an eight bit encoding scheme developed in 1963 for the IBM System 360 that featured a larger character set including lower case letters The limitations of such sets soon became apparent to whom and a number of ad hoc methods were developed to extend them The need to support more writing systems for different languages including the CJK family of East Asian scripts required support for a far larger number of characters and demanded a systematic approach to character encoding rather than the previous ad hoc approaches citation needed In trying to develop universally interchangeable character encodings researchers in the 1980s faced the dilemma that on the one hand it seemed necessary to add more bits to accommodate additional characters but on the other hand for the users of the relatively small character set of the Latin alphabet who still constituted the majority of computer users those additional bits were a colossal waste of then scarce and expensive computing resources as they would always be zeroed out for such users In 1985 the average personal computer user s hard disk drive could store only about 10 megabytes and it cost approximately US 250 on the wholesale market and much higher if purchased separately at retail 5 so it was very important at the time to make every bit count The compromise solution that was eventually found and developed into Unicode vague was to break the assumption dating back to telegraph codes that each character should always directly correspond to a particular sequence of bits Instead characters would first be mapped to a universal intermediate representation in the form of abstract numbers called code points Code points would then be represented in a variety of ways and with various default numbers of bits per character code units depending on context To encode code points higher than the length of the code unit such as above 256 for eight bit units the solution was to implement variable length encodings where an escape sequence would signal that subsequent bits should be parsed as a higher code point Terminology EditInformally the terms character encoding character map character set and code page are often used interchangeably 6 Historically the same standard would specify a repertoire of characters and how they were to be encoded into a stream of code units usually with a single character per code unit However due to the emergence of more sophisticated character encodings the distinction between these terms has become important A character is a minimal unit of text that has semantic value 6 7 A character set is a collection of elements used to represent text 6 7 For example the Latin alphabet and Greek alphabet are both character sets A coded character set is a character set mapped to set of unique numbers 7 For historical reasons this is also often referred to as a code page 6 A character repertoire is the set of characters that can be represented by a particular coded character set 7 8 The repertoire may be closed meaning that no additions are allowed without creating a new standard as is the case with ASCII and most of the ISO 8859 series or it may be open allowing additions as is the case with Unicode and to a limited extent Windows code pages 8 A code point is a value or position of a character in a coded character set 7 A code space is the range of numerical values spanned by a coded character set 7 9 A code unit is the minimum bit combination that can represent a character in a character encoding in computer science terms it is the word size of the character encoding 7 9 For example common code units include 7 bit 8 bit 16 bit and 32 bit In some encodings some characters are encoded using multiple code units such an encoding is referred to as a variable width encoding Code pages Edit Main article Code page Code page is a historical name for a coded character set Originally a code page referred to a specific page number in the IBM standard character set manual which would define a particular character encoding 10 Other vendors including Microsoft SAP and Oracle Corporation also published their own sets of code pages the most well known code page suites are Windows based on Windows 1252 and IBM DOS based on code page 437 Despite no longer referring to specific page numbers in a standard many character encodings are still referred to by their code page number likewise the term code page is often still used to refer to character encodings in general The term code page is not used in Unix or Linux where charmap is preferred usually in the larger context of locales IBM s Character Data Representation Architecture CDRA designates entities with coded character set identifiers CCSIDs each of which is variously called a charset character set code page or CHARMAP 9 Code units Edit The code unit size is equivalent to the bit measurement for the particular encoding A code unit in US ASCII consists of 7 bits A code unit in UTF 8 EBCDIC and GB 18030 consists of 8 bits A code unit in UTF 16 consists of 16 bits A code unit in UTF 32 consists of 32 bits Code points Edit A code point is represented by a sequence of code units The mapping is defined by the encoding Thus the number of code units required to represent a code point depends on the encoding UTF 8 code points map to a sequence of one two three or four code units UTF 16 code units are twice as long as 8 bit code units Therefore any code point with a scalar value less than U 10000 is encoded with a single code unit Code points with a value U 10000 or higher require two code units each These pairs of code units have a unique term in UTF 16 Unicode surrogate pairs UTF 32 the 32 bit code unit is large enough that every code point is represented as a single code unit GB 18030 multiple code units per code point are common because of the small code units Code points are mapped to one two or four code units 11 Characters Edit Main article Character computing Exactly what constitutes a character varies between character encodings For example for letters with diacritics there are two distinct approaches that can be taken to encode them they can be encoded either as a single unified character known as a precomposed character or as separate characters that combine into a single glyph The former simplifies the text handling system but the latter allows any letter diacritic combination to be used in text Ligatures pose similar problems Exactly how to handle glyph variants is a choice that must be made when constructing a particular character encoding Some writing systems such as Arabic and Hebrew need to accommodate things like graphemes that are joined in different ways in different contexts but represent the same semantic character Unicode encoding model EditUnicode and its parallel standard the ISO IEC 10646 Universal Character Set together constitute a unified standard for character encoding Rather than mapping characters directly to bytes Unicode separately defines a coded character set that maps characters to unique natural numbers code points how those code points are mapped to a series of fixed size natural numbers code units and finally how those units are encoded as a stream of octets bytes The purpose of this decomposition is to establish a universal set of characters that can be encoded in a variety of ways To describe this model precisely Unicode uses its own set of terminology to describe its process 9 An abstract character repertoire ACR is the full set of abstract characters that a system supports Unicode has an open repertoire meaning that new characters will be added to the repertoire over time A coded character set CCS is a function that maps characters to code points each code point represents one character For example in a given repertoire the capital letter A in the Latin alphabet might be represented by the code point 65 the character B to 66 and so on Multiple coded character sets may share the same character repertoire for example ISO IEC 8859 1 and IBM code pages 037 and 500 all cover the same repertoire but map them to different code points A character encoding form CEF is the mapping of code points to code units to facilitate storage in a system that represents numbers as bit sequences of fixed length i e practically any computer system For example a system that stores numeric information in 16 bit units can only directly represent code points 0 to 65 535 in each unit but larger code points say 65 536 to 1 4 million could be represented by using multiple 16 bit units This correspondence is defined by a CEF A character encoding scheme CES is the mapping of code units to a sequence of octets to facilitate storage on an octet based file system or transmission over an octet based network Simple character encoding schemes include UTF 8 UTF 16BE UTF 32BE UTF 16LE and UTF 32LE compound character encoding schemes such as UTF 16 UTF 32 and ISO IEC 2022 switch between several simple schemes by using a byte order mark or escape sequences compressing schemes try to minimize the number of bytes used per code unit such as SCSU and BOCU Although UTF 32BE and UTF 32LE are simpler CESes most systems working with Unicode use either UTF 8 which is backward compatible with fixed length ASCII and maps Unicode code points to variable length sequences of octets or UTF 16BE citation needed which is backward compatible with fixed length UCS 2BE and maps Unicode code points to variable length sequences of 16 bit words See comparison of Unicode encodings for a detailed discussion Finally there may be a higher level protocol which supplies additional information to select the particular variant of a Unicode character particularly where there are regional variants that have been unified in Unicode as the same character An example is the XML attribute xml lang The Unicode model uses the term character map for other systems which directly assign a sequence of characters to a sequence of bytes covering all of the CCS CEF and CES layers 9 Unicode code points Edit In Unicode a character can be referred to as U followed by its codepoint value in hexadecimal The range of valid code points the codespace for the Unicode standard is U 0000 to U 10FFFF inclusive divided in 17 planes identified by the numbers 0 to 16 Characters in the range U 0000 to U FFFF are in plane 0 called the Basic Multilingual Plane BMP This plane contains most commonly used characters Characters in the range U 10000 to U 10FFFF in the other planes are called supplementary characters The following table shows examples of code point values Character Unicode code point GlyphLatin A U 0041 ALatin sharp S U 00DF ssHan for East U 6771 東Ampersand U 0026 amp Inverted exclamation mark U 00A1 Section sign U 00A7 Example Edit Consider a string of the letters ab c𐐀 that is a string containing a Unicode combining character U 0332 as well a supplementary character U 10400 𐐀 This string has several Unicode representations which are logically equivalent yet while each is suited to a diverse set of circumstances or range of requirements Four composed characters a b c 𐐀 Five graphemes a b c 𐐀 Five Unicode code points U 0061 U 0062 U 0332 U 0063 U 10400 Five UTF 32 code units 32 bit integer values 0x00000061 0x00000062 0x00000332 0x00000063 0x00010400 Six UTF 16 code units 16 bit integers 0x0061 0x0062 0x0332 0x0063 0xD801 0xDC00 Nine UTF 8 code units 8 bit values or bytes 0x61 0x62 0xCC 0xB2 0x63 0xF0 0x90 0x90 0x80Note in particular that 𐐀 is represented with either one 32 bit value UTF 32 two 16 bit values UTF 16 or four 8 bit values UTF 8 Although each of those forms uses the same total number of bits 32 to represent the glyph it is not obvious how the actual numeric byte values are related Transcoding EditAs a result of having many character encoding methods in use and the need for backward compatibility with archived data many computer programs have been developed to translate data between character encoding schemes a process known as transcoding Some of these are cited below Cross platform Web browsers most modern web browsers feature automatic character encoding detection On Firefox 3 for example see the View Character Encoding submenu iconv a program and standardized API to convert encodings luit a program that converts encoding of input and output to programs running interactively convert encoding py a Python based utility to convert text files between arbitrary encodings and line endings 12 decodeh py an algorithm and module to heuristically guess the encoding of a string 13 International Components for Unicode A set of C and Java libraries to perform charset conversion uconv can be used from ICU4C chardet This is a translation of the Mozilla automatic encoding detection code into the Python computer language The newer versions of the Unix file command attempt to do a basic detection of character encoding also available on Cygwin charset C template library with simple interface to convert between C user defined streams charset defined many character sets and allows you to use Unicode formats with support of endianness Unix like cmv a simple tool for transcoding filenames 14 convmv converts a filename from one encoding to another 15 cstocs converts file contents from one encoding to another for the Czech and Slovak languages enca analyzes encodings for given text files 16 recode converts file contents from one encoding to another 17 utrac converts file contents from one encoding to another 18 Windows Encoding Convert NET API 19 MultiByteToWideChar WideCharToMultiByte to convert from ANSI to Unicode amp Unicode to ANSI 20 21 cscvt a character set conversion tool 22 enca analyzes encodings for given text files 23 See also EditPercent encoding Alt code Character encodings in HTML Category Character encoding articles related to character encoding in general Category Character sets articles detailing specific character encodings Hexadecimal representations Mojibake character set mismap Mojikyō a system glyph set that includes over 100 000 Chinese character drawings modern and ancient popular and obscure Presentation layer TRON part of the TRON project is an encoding system that does not use Han Unification instead it uses control codes to switch between 16 bit planes of characters Universal Character Set characters Charset sniffing used in some applications when character encoding metadata is not availableCommon character encodings Edit ISO 646 ASCII EBCDIC ISO 8859 ISO 8859 1 Western Europe ISO 8859 2 Western and Central Europe ISO 8859 3 Western Europe and South European Turkish Maltese plus Esperanto ISO 8859 4 Western Europe and Baltic countries Lithuania Estonia Latvia and Lapp ISO 8859 5 Cyrillic alphabet ISO 8859 6 Arabic ISO 8859 7 Greek ISO 8859 8 Hebrew ISO 8859 9 Western Europe with amended Turkish character set ISO 8859 10 Western Europe with rationalised character set for Nordic languages including complete Icelandic set ISO 8859 11 Thai ISO 8859 13 Baltic languages plus Polish ISO 8859 14 Celtic languages Irish Gaelic Scottish Welsh ISO 8859 15 Added the Euro sign and other rationalisations to ISO 8859 1 ISO 8859 16 Central Eastern and Southern European languages Albanian Bosnian Croatian Hungarian Polish Romanian Serbian and Slovenian but also French German Italian and Irish Gaelic CP437 CP720 CP737 CP850 CP852 CP855 CP857 CP858 CP860 CP861 CP862 CP863 CP865 CP866 CP869 CP872 MS Windows character sets Windows 1250 for Central European languages that use Latin script Polish Czech Slovak Hungarian Slovene Serbian Croatian Bosnian Romanian and Albanian Windows 1251 for Cyrillic alphabets Windows 1252 for Western languages Windows 1253 for Greek Windows 1254 for Turkish Windows 1255 for Hebrew Windows 1256 for Arabic Windows 1257 for Baltic languages Windows 1258 for Vietnamese Mac OS Roman KOI8 R KOI8 U KOI7 MIK ISCII TSCII VISCII JIS X 0208 is a widely deployed standard for Japanese character encoding that has several encoding forms Shift JIS Microsoft Code page 932 is a dialect of Shift JIS EUC JP ISO 2022 JP JIS X 0213 is an extended version of JIS X 0208 Shift JIS 2004 EUC JIS 2004 ISO 2022 JP 2004 Chinese Guobiao GB 2312 GBK Microsoft Code page 936 GB 18030 Taiwan Big5 a more famous variant is Microsoft Code page 950 Hong Kong HKSCS Korean KS X 1001 is a Korean double byte character encoding standard EUC KR ISO 2022 KR Unicode and subsets thereof such as the 16 bit Basic Multilingual Plane UTF 8 UTF 16 UTF 32 ANSEL or ISO IEC 6937References Edit Definition from The Tech Terms Dictionary Tom Henderson 17 April 2014 Ancient Computer Character Code Tables and Why They re Still Relevant Smartbear Retrieved 29 April 2014 Tom Jennings 1 March 2010 An annotated history of some character codes Retrieved 1 November 2018 IBM Electronic Data Processing Machines Type 702 Preliminary Manual of Information PDF 1954 p 80 22 6173 1 Archived PDF from the original on 9 October 2022 Strelho Kevin 15 April 1985 IBM Drives Hard Disks to New Standards InfoWorld Popular Computing Inc pp 29 33 Retrieved 10 November 2020 a b c d Shawn Steele 15 March 2005 What s the difference between an Encoding Code Page Character Set and Unicode Microsoft Docs a b c d e f g Glossary of Unicode Terms Unicode Consortium a b Chapter 3 Conformance The Unicode Standard Version 15 0 Core Specification PDF Unicode Consortium September 2022 ISBN 978 1 936213 32 0 a b c d e Whistler Ken Freytag Asmus 11 November 2022 UTR 17 Unicode Character Encoding Model Unicode Consortium Retrieved 12 August 2023 VT510 Video Terminal Programmer Information Digital Equipment Corporation DEC 7 1 Character Sets Overview Archived from the original on 26 January 2016 Retrieved 15 February 2017 In addition to traditional DEC and ISO character sets which conform to the structure and rules of ISO 2022 the VT510 supports a number of IBM PC code pages page numbers in IBM s standard character set manual in PCTerm mode to emulate the console terminal of industry standard PCs Terminology The Java Tutorials Oracle Retrieved 25 March 2018 convert encoding py on GitHub Decodeh heuristically decode a string or text file Archived from the original on 8 January 2008 CharsetMove Simple Tool for Transcoding Filenames Convmv converts filenames from one encoding to another Extremely Naive Charset Analyser Archived from the original on 4 December 2010 Retrieved 11 March 2008 recode on GitHub Utrac Homepage Encoding Convert Method Microsoft NET Framework Class Library MultiByteToWideChar function stringapiset h Microsoft Docs 13 October 2021 WideCharToMultiByte function stringapiset h Microsoft Docs 9 August 2022 Kalytta s Character Set Converter Enca binary compiled for 32 bit Windows Archived from the original on 15 March 2012 Retrieved 31 March 2011 Further reading EditMackenzie Charles E 1980 Coded Character Sets History and Development ISBN 978 0 201 14460 4 LCCN 77 90165 a href Template Cite book html title Template Cite book cite book a work ignored help External links Edit Wikimedia Commons has media related to Encodings Wikiversity has learning resources about Character encoding Character sets registered by Internet Assigned Numbers Authority IANA Characters and encodings by Jukka Korpela Unicode Technical Report 17 Character Encoding Model Decimal Hexadecimal Character Codes in HTML Unicode Encoding converter The Absolute Minimum Every Software Developer Absolutely Positively Must Know About Unicode and Character Sets No Excuses by Joel Spolsky Oct 10 2003 Retrieved from https en wikipedia org w index php title Character encoding amp oldid 1170700523, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.