Why does <string>.length return the number of bytes and not the number of UTF-8 characters, whereas <wstring.>length and <dstring>.length return the number of UTF-16 and UTF-32 characters? Wouldn't it be more consistent to have <string>.length return the number of UTF-8 characters as well (instead of having to use std.utf.count(<string>)?