We Can Build AI Systems, But Can We Handle Names With Umlauts?
My name includes three umlauts. It can be written correctly using six plus eleven unicode characters (first and last name). In the early 1990s, I never considered the idea of using my name as an identifier on a computer because most systems back then were still living in a US-ASCII world or the unfortunate ISO/IEC 8859 8-bit extensions, where the interpretation of the code points depends on the context.
Back in the 1990s, when I went to a bank to get a printout of my account status (yes, this is how banks worked back then; they had printers, and you had to visit them physically to check your account), I was always amazed at what came out of the printer. Often umlauts were replaced by curly braces and other fancy symbols. Back then, I used to be an optimist, and I believed that these problems would be overcome at the beginning of the 21st century. After all, the Unicode consortium was established in 1991, and Unicode 1.0 was released in the same year. In 1996, the initial specification of the UTF-8 transformation format was published as RFC 2044. So things looked promising. Sure, banks run mainframes, but give them 20 years and they may have made a transition to unicode.