I am curious how I can more reliably recognize the value and image of playing cards. Here are two examples:

There may be some noise in the images, but I have a large data set that I could use for training (about 10 thousand pages, including all values ββand costumes).
I can reliably recognize images that I manually classified if I have an exact exact match using the hash method. But since I use hashing images based on their contents, then the slightest noise changes the hash and causes the image to be processed as unknown. This is what I reliably access with further automation.
I looked at the tesseract 3.05 training documentation: https://github.com/tesseract-ocr/tesseract/wiki/Training-Tesseract#automated-method
Is it possible to train tesseract only with images found in fonts? Or can I use it to recognize suits for these cards?
I was hoping I could say that all the images in this folder correspond to 4c (for example, the sample images above), and that tesseract will see the similarities in any future instances of this image (regardless of noise level), and also read that as 4c. Is it possible? Does anyone have any experience with this?
image ocr tesseract macos
Tiago
source share