SymbolFYI

Mathematical Notation in Unicode: From Clay Tablets to Code Points

History History of Symbols ม.ค. 23, 2024

The equals sign is only 467 years old.

This fact surprises many people, because the concept of equality is ancient — Babylonian mathematicians were solving quadratic equations 4,000 years ago. But the symbol for equality, the two horizontal parallel lines that every schoolchild learns, was invented in 1557 by a Welsh physician named Robert Recorde who was simply tired of writing the same phrase over and over.

Mathematical notation is not mathematics. It is a human invention — a set of visual conventions that developed over millennia through competition, accident, and deliberate design. The symbols that fill our textbooks and our Unicode block are the survivors of hundreds of competing proposals, the products of specific historical moments and specific individuals' preferences.

This is the story of how those symbols came to be, and how they found their place in the Unicode standard.

Ancient Notation: The First Mathematical Writing

The oldest mathematical documents we have are clay tablets from Mesopotamia, dating to around 2000 BCE. Babylonian mathematicians used a base-60 number system (sexagesimal — still surviving in our division of hours and degrees) and could solve problems equivalent to quadratic equations. But they had no symbols at all for mathematical operations. Their "notation" was entirely verbal: "Take the length. Add the width. Multiply by..."

Egyptian mathematical papyri (the Rhind Mathematical Papyrus, c. 1650 BCE, is the most complete) used a system of hieratic symbols that included glyphs for addition and subtraction — a pair of legs walking forward (⟶) for addition and walking backward (⟵) for subtraction. These are not ancestors of our modern +/−; they were lost to history along with the hieratic script itself.

Greek mathematics, which developed the axiomatic proof method that underlies all modern mathematics, was paradoxically poor in notation. Euclid's Elements (c. 300 BCE), the foundational text of geometry, contains essentially no symbolic notation. Everything is stated verbally: "Let a line segment AB be bisected at point C." Arithmetic quantities were referred to by the letters labeling geometric figures. There was no symbol for an unknown, no symbol for an operation, no symbol for a relationship.

This is not because Greek mathematicians were unsophisticated — it's because they conceived of mathematics primarily as geometry, and geometric reasoning was expressed in terms of figures and verbal argument. Symbolic algebra, in the sense of manipulating unknown quantities, was not a Greek invention.

Diophantus: The First Symbolic Algebra

The exception was Diophantus of Alexandria (c. 250 CE), who introduced a genuinely symbolic approach to arithmetic problems in his Arithmetica. Diophantus used abbreviations for the unknown quantity and its powers — an early step toward algebraic notation. His symbol for the unknown (roughly transliterated as s or an abbreviated Greek word) is the first true algebraic symbol we know of.

The Arithmetica was largely ignored by Latin medieval Europe but was preserved in Arabic translation, and its influence would resurface during the Renaissance through the Islamic algebraic tradition that had absorbed and extended Diophantus's work.

The Islamic Algebraic Tradition

The word "algebra" itself comes from Arabic: al-jabr, the title of a 9th-century work by Muhammad ibn Musa al-Khwarizmi (Kitab al-mukhtasar fi hisab al-jabr wal-muqabala, "The Compendious Book on Calculation by Completion and Balancing"). Al-Khwarizmi's name also gives us "algorithm."

Islamic mathematicians developed sophisticated algebraic methods but also largely avoided symbolic notation. Al-Khwarizmi's algebra, like Greek geometry, was expressed verbally. "The square plus ten times the root equals 39" — not x² + 10x = 39. This verbosity seems limiting to modern eyes, but the rhetorical tradition was deeply established and did not retard the development of the mathematical ideas themselves.

The transition to symbolic notation happened gradually in Europe during the 14th through 17th centuries, driven by practical demands — commerce, navigation, and astronomy all required rapid calculation, and symbols are more efficient than words.

The Birth of Modern Notation

Plus and Minus: Germany, Late 15th Century

The plus sign (+) and minus sign (−) in their current forms appear first in German commercial manuscripts of the late 15th century. The earliest dated occurrence of + is in a 1489 arithmetic book by Johannes Widmann, Behende und hübsche Rechenung auff allen Kauffmanschaften ("Mercantile Arithmetic"), where the signs are used to indicate surplus (+) and deficit (−) in warehouse inventory. This is commercial notation, not mathematical in the abstract sense.

The signs appeared as shorthand for the Latin words et (and) for + and minus (less) for −. The + sign is likely a ligature of et, as the ampersand is. The − sign may simply be an abbreviated m.

The mathematical use — addition and subtraction operators — followed quickly. By the early 16th century, + and − were the standard operators in German and Dutch mathematical texts, and they spread through Europe over the following century. England adopted them largely through the influence of Robert Recorde, who used them in his 1543 arithmetic textbook The Ground of Artes.

The Equals Sign: Robert Recorde, 1557

Robert Recorde, a Welsh physician and mathematician who served as Controller of the Royal Mint under Henry VIII and Edward VI, published The Whetstone of Witte in 1557. It was a textbook of algebra, and in its pages appeared, for the first time in print, the equals sign (=).

Recorde's explanation of his choice is admirably direct: "And to avoide the tediouse repetition of these woordes: is equalle to: I will sette as I doe often in woorke use, a paire of paralleles, or Gemowe [twin] lines of one lengthe, thus: =, bicause noe .2. thynges, can be moare equalle."

Two things as equal as two parallel lines of the same length. The reasoning is aesthetic and mnemonic, not mathematical. Recorde's equals sign was actually drawn longer than today's — more like ══ — and it took more than a century for the shorter = to become standard. Leibniz, in the late 17th century, still preferred the symbol ∝ (proportional) and used = only reluctantly.

Multiplication and Division: A Contested History

The multiplication sign (×) was introduced by William Oughtred in 1631 in his Clavis Mathematicae ("Key of the Mathematics"). Oughtred's × is a rotated plus sign — visually suggesting that multiplication relates to addition in the same way a diagonal relates to a vertical.

The division sign (÷) — called the obelus — has a more complex history. The obelus (from Greek obelos, a spit or pointed instrument) was used in ancient manuscripts as a critical mark indicating a passage of doubtful authenticity. In medieval manuscripts it indicated deletion. Johann Rahn appropriated it as a division operator in 1659 (Teutsche Algebra), and it spread through English-language mathematics via John Pell's translation of Rahn's work. Notably, the ÷ is largely an Anglo-American convention — most Continental European mathematics uses a colon (:) or fraction notation for division.

Leibniz, who cared deeply about notation and argued extensively for his choices, preferred × for multiplication in some contexts and a center dot (·) in others — the dot notation survives today in physics and some fields of mathematics. He famously described the dot as preferable because × "too easily becomes confused with the letter x."

The Calculus Notations: Leibniz vs. Newton

The invention of calculus in the late 17th century, independently by Isaac Newton and Gottfried Wilhelm Leibniz, produced one of the most consequential notation competitions in the history of mathematics.

Newton's notation used dots above letters: ẋ for the derivative of x with respect to time, ẍ for the second derivative. This "dot notation" is compact and useful for physics problems involving time (it's still used in mechanics), but it becomes awkward for higher derivatives and for functions of multiple variables.

Leibniz invented the notations that dominate mathematics today: - (integral sign) — a stylized elongated S, from Latin summa (sum), because an integral is a limit of sums. Leibniz introduced this notation in 1675 in a manuscript and published it in 1684. - d for differentials: dy/dx for the derivative, the notation every calculus student learns. - d²y/dx² for higher derivatives.

The elongated S for integration (∫) is Unicode U+222B. It is also available in various combinations: ∬ (double integral, U+222C), ∭ (triple integral, U+222D), ∮ (contour integral, U+222E), and others.

The Newton-Leibniz priority dispute — who invented calculus first? — became entangled with notation politics. British mathematicians, loyal to Newton, refused to adopt Leibniz's superior notation for over a century. The result was that British mathematics fell significantly behind Continental mathematics in the 18th century. It was not until the Analytical Society at Cambridge (founded 1812 by Babbage, Herschel, and Peacock) campaigned to replace Newton's dots with Leibniz's d-notation that British mathematics rejoined the European mainstream.

Sigma and Pi: Euler's Contributions

Leonhard Euler (1707–1783) was the most prolific mathematician in history and also one of the most influential architects of modern mathematical notation. Among his many contributions:

  • π (pi) for the ratio of circumference to diameter: Euler popularized this use in 1736, though William Jones had first used it in 1706. Before Euler's adoption, various other symbols and letters were used.
  • e for the base of natural logarithms: Euler introduced this notation in 1736 as well.
  • i for the imaginary unit (√−1): Euler introduced this in 1777.
  • f(x) notation for functions: Euler established this standard in 1734.
  • (capital sigma) for summation: Euler introduced this in 1755 in his Institutiones Calculi Differentialis. The sigma represents "sum" — from Greek sigma corresponding to Latin summa.

The summation notation ∑ (U+2211) is now one of the most heavily used symbols in mathematics. Its upper and lower limits (∑_{i=1}^{n}) define exactly which terms to include in the sum — a precision that verbal description requires many more words to achieve.

Similarly, (capital pi, U+220F) for products was developed in the same era and tradition. "Product" begins with P in both Latin and Greek.

Unicode Mathematical Blocks

The Unicode standard allocates significant space to mathematical notation. The major mathematical blocks are:

Block Code Points Contents
Mathematical Operators U+2200–U+22FF ∀ ∃ ∅ ∈ ∑ ∏ ∫ ∞ and 254 others
Supplemental Math Operators U+2A00–U+2AFF 256 additional operators
Misc. Mathematical Symbols-A U+27C0–U+27EF 48 symbols
Misc. Mathematical Symbols-B U+2980–U+29FF 128 symbols
Mathematical Alphanumerics U+1D400–U+1D7FF Bold, italic, script, fraktur, double-struck alphabets

The Mathematical Operators block (U+2200–U+22FF) is the core. It contains: - Logic operators: ∀ (for all), ∃ (there exists), ¬ (not), ∧ (and), ∨ (or) - Set operators: ∈ (element of), ∉ (not element of), ⊂ (subset), ∪ (union), ∩ (intersection) - Relation operators: ≤ (less than or equal), ≥ (greater than or equal), ≠ (not equal), ≈ (approximately equal) - Calculus: ∫ (integral), ∂ (partial differential), ∇ (nabla/del) - Number theory: ∑ (sum), ∏ (product), √ (square root), ∞ (infinity)

The Mathematical Alphanumerics Block

A special case in Unicode math is the Mathematical Alphanumerics Symbols block (U+1D400–U+1D7FF). This block provides distinct code points for the "styled" letter forms used extensively in mathematics:

  • Bold: 𝐀 𝐁 𝐂 (U+1D400–U+1D433)
  • Italic: 𝐴 𝐵 𝐶 (U+1D434–U+1D467)
  • Bold Italic: 𝑨 𝑩 𝑪 (U+1D468–U+1D49B)
  • Script: 𝒜 ℬ 𝒞 (U+1D49C–U+1D4CF)
  • Fraktur: 𝔄 𝔅 ℭ (U+1D504–U+1D537)
  • Double-struck: 𝔸 𝔹 ℂ (U+1D538–U+1D56B) — used for number systems (ℕ, ℤ, ℚ, ℝ, ℂ)

These styled alphabets exist as distinct code points because in mathematics, 𝐀 (bold A) and 𝐴 (italic A) and 𝒜 (script A) can denote genuinely different mathematical objects — vectors, variables, and spaces are often distinguished by the style of their letter. They cannot be treated as merely visual formatting of the same character.

This creates an interesting exception to the general Unicode principle that encoding should be semantic rather than visual. For mathematical notation, the style is the semantics.

Challenges in Encoding Mathematical Notation

The MathML Approach

Mathematical notation has a fundamental tension with Unicode's design philosophy. Unicode encodes characters — individual symbols. But mathematics is about relationships between symbols: the placement of a subscript, the position of a limit above or below a summation sign, the nesting of fractions. These spatial relationships carry mathematical meaning that Unicode cannot directly encode.

The W3C's MathML (Mathematical Markup Language) is the standard approach for representing mathematical expressions in digital contexts. MathML represents the structure of an expression as XML: a fraction is <mfrac><mi>a</mi><mi>b</mi></mfrac>, which can be rendered correctly regardless of display environment.

MathML uses Unicode for individual symbols but adds the structural layer that Unicode alone cannot provide. Modern web browsers support MathML natively (or via JavaScript libraries like MathJax and KaTeX).

LaTeX and Unicode

LaTeX, developed by Leslie Lamport (based on Donald Knuth's TeX, published 1978) is the standard typesetting system for academic mathematics and science. LaTeX uses ASCII-based commands to describe mathematical expressions: \sum_{i=1}^{n} i = \frac{n(n+1)}{2} renders as the formula for the sum of integers.

The relationship between LaTeX and Unicode is bidirectional: - Unicode includes essentially all the symbols that LaTeX's standard mathematics mode uses - LaTeX input methods increasingly support direct Unicode input - Mapping tables between LaTeX commands and Unicode code points are maintained (e.g., \sum ↔ U+2211, \int ↔ U+222B, \pi ↔ U+03C0)

For software developers working with mathematical text, understanding that most LaTeX mathematical symbols have corresponding Unicode code points is practically useful: a database storing mathematical expressions as Unicode text (rather than LaTeX commands) provides better searchability and interoperability.

A Timeline of Mathematical Notation

Year Symbol Creator Notes
c. 250 CE First algebraic unknown symbol Diophantus Abbreviation, not the modern convention
c. 825 Algebra systematized verbally Al-Khwarizmi Al-jabr — no symbolic notation
1489 +, − Johannes Widmann First print appearance as operators
1557 = Robert Recorde The Whetstone of Witte
1631 × William Oughtred Clavis Mathematicae
1637 x, y, z for unknowns; a, b, c for known René Descartes La Géométrie — still standard
1659 ÷ Johann Rahn Teutsche Algebra
1675 ∫, d notation Leibniz Calculus notation (manuscript)
1684 ∫ published Leibniz First print appearance
1706 π William Jones First use for circle ratio
1736 π, e popularized Leonhard Euler Standard notation established
1755 Leonhard Euler Institutiones Calculi Differentialis
1777 i for √−1 Leonhard Euler Published 1794
1978 TeX Donald Knuth Mathematical typesetting
1985 LaTeX Leslie Lamport Standard academic typesetting
1991 Unicode 1.0 includes math operators Unicode Consortium Core math symbols encoded
1993 Mathematical Operators block Unicode 1.1 U+2200–U+22FF
2001 Mathematical Alphanumerics Unicode 3.1 U+1D400–U+1D7FF, bold/italic/etc.
2003 MathML 2.0 W3C Web standard for math

Explore the full set of mathematical operators and symbols — including the Unicode code point, name, and properties for each — in our Symbol Table tool.


This completes the History of Symbols series. You've traced the arc from ASCII's 128 characters (1963) through Unicode's universal standard, from Shigetaka Kurita's 176 emoji to the accumulated symbolic heritage of mathematical notation. The symbols we type every day carry centuries of human ingenuity, commercial necessity, and cultural negotiation. Each code point in Unicode is the end of a history.

Explore the full character set and discover more stories behind the symbols at SymbolFYI.

สัญลักษณ์ที่เกี่ยวข้อง

อภิธานศัพท์ที่เกี่ยวข้อง

เครื่องมือที่เกี่ยวข้อง

คู่มือเพิ่มเติม