Why Bert-chinese use do_lower_case=False?

Some Chinese Text has some English words, for example: “Apples是苹果的复数形式。”. I have questions about how to tokenize the text:

  1. why Chinese Bert Case sensitive, but I can’t find even ‘A’ in vocab.txt
  2. Because English words in Chinese vocab.txt is few, should I use wordpiece tokenizer as default, like "[‘apple’, ‘##s’, ‘是’, ‘苹’, …]"or split to char to tokenize, like “[‘a’, ‘p’, ‘p’, ‘l’, ‘e’, ‘s’, ‘是’, ‘苹’, …]”?