Skip to main content

What is Unicode?

Glossary image
Universal Character Set UCS

Unicode is a universal character encoding standard that assigns a unique numerical identifier, called a code point, to every character across virtually all of the world's writing systems, symbols, and scripts. Maintained by the Unicode Consortium, it serves as the foundational layer that allows computers to represent and exchange text consistently, regardless of language, platform, or software.

Before Unicode, the computing world relied on a fragmented landscape of incompatible encoding systems. ASCII, one of the earliest standards, could only represent 128 characters - enough for basic English text but wholly inadequate for languages using non-Latin scripts such as Arabic, Chinese, Japanese, or Tamil. Regional encoding schemes emerged to fill these gaps, but they were mutually incompatible and caused widespread data corruption when text crossed system boundaries. Unicode was designed to solve this problem once and for all by providing a single, unified catalog of characters.

Today, Unicode defines over 149,000 characters spanning more than 160 scripts, including historical writing systems, mathematical symbols, musical notation, and emoji. Each character is assigned a code point expressed in the format U+ followed by a hexadecimal number - for example, U+0041 represents the Latin capital letter A, while U+4E2D represents the Chinese character δΈ­.

Unicode and UTF-8

Unicode itself is an abstract standard - it defines what characters exist and their code points, but it does not dictate how those code points are stored as bytes in a file or transmitted over a network. That is the role of an encoding. UTF-8 is by far the most widely used encoding for Unicode on the web. It is a variable-width encoding, meaning it uses between one and four bytes to represent each code point. Crucially, UTF-8 is backward compatible with ASCII: the first 128 Unicode code points map directly to ASCII values and are stored as a single byte, which made adoption straightforward for systems already built around ASCII.

Other Unicode encodings include UTF-16 and UTF-32, which use fixed or semi-fixed byte widths. However, UTF-8 dominates web usage because of its efficiency with Latin-script text and its compatibility with existing infrastructure.

Unicode and Internationalization

Unicode is the technical prerequisite for internationalization (i18n) - the process of designing software so it can be adapted to different languages and regions. Without a shared character standard, building a multilingual website or application would require maintaining separate codebases for each script. Because modern browsers, operating systems, and web standards all default to Unicode (typically encoded as UTF-8), developers can write content in any language and be confident it will render correctly for users worldwide.

Declaring a UTF-8 character encoding in an HTML document's <meta> tag is a routine step in web development, but it reflects a deeper dependency: the entire global web rests on the Unicode standard as its common language for text.

Have a question?

Get in touch if you'd like to learn more about this topic.

Contact Us