From my perspective it may be just fitting to the answer. We try to find symbols, without confidence, that the paper still contains information, and any text for trainig. With "rights" network you can achieve any possibile result. It is remember me a russian freak-scientist, which try to read words and texts on the detailed sun surface photos.
I wondered about that. My understanding is that the models were trained to look for letter shapes, not words. And that the models couldn't produce known words unless they were trained on the language. If it wasn't trained on a substantial text body, a model producing letter sequences that form known words means it found something and didn't hallucinate.