Some of these images provided as the train data for Text detection and recognition are not appropriate to be used for training. For example, in recognition, we need to crop those texts and align them horizontally depending upon the angle, that gives us around ~600k images in the training data for recognition. Now there are many images, in which you can’t even see the text, and after cropping them off, they are even more unclear. And if this repo has been used to generate those data images, it will again provide such images only. On the other hand, the test dataset images look pretty good as every image is clear and the words can be made out, as they are real images and not synthetic.
Please let me know if I’m wrong. I’m attaching a sample image in which we cannot see any word but still, there are 2 words in that image. @GokulNC @Ishvinder