The current version of UTF-16 is capable of encoding only 1112,064 different numbers (code points); 0x0-0x10FFFF .
Unicode competition intends to force UTF-16 to end?
i.e. make code point> 0x10FFFF
If not, why would anyone write code for the utf-8 parser to be able to accept 5 or 6 byte sequences? Since this will add unnecessary instructions to their function.
Is 1,112,064 enough, do we really need more characters? I mean: how fast do we finish?
source share