Introduction
Some people are under the misconception that Unicode is simply a 16-bit code where each character takes 16 bits and therefore there are 65,536 possible characters. This is not actually correct. It is the single most common myth about Unicode, so if you thought that, don’t feel bad.
In fact, Unicode makes you think a different way about encoding characters. Until now, we’ve assumed that a letter maps to some bits which you can store on disk or in memory:
A -> 0100 0001 Continue reading Short Discourse on Unicode