Text captured as part of outdoor surroundings in a photograph
Scene text is text that appears in an image captured by a camera in an outdoor environment.
The detection and recognition of scene text from camera captured images are computer vision tasks which became important after smart phones with good cameras became ubiquitous. The text in scene images varies in shape, font, colour and position. The recognition of scene text is further complicated sometimes by non-uniform illumination and focus.
To improve scene text recognition, the International Conference on Document Analysis and Recognition (ICDAR) conducts a robust reading competition once in two years. The competition was held in 2003, 2005[1][2][3] and during every ICDAR conference.[4][5][6] International association for pattern recognition (IAPR) has created a list of datasets as Reading systems.[7]
Text detection
Text detection is the process of detecting the text present in the image, followed by surrounding it with a rectangular bounding box. Text detection can be carried out using image based techniques or frequency based techniques.
In image based techniques, an image is segmented into multiple segments. Each segment is a connected component of pixels with similar characteristics. The statistical features of connected components are utilised to group them and form the text. Machine learning approaches such as support vector machine and convolutional neural networks are used to classify the components into text and non-text.
In frequency based techniques, discrete Fourier transform (DFT) or discrete wavelet transform (DWT) are used to extract the high frequency coefficients. It is assumed that the text present in an image has high frequency components and selecting only the high frequency coefficients filters the text from the non-text regions in an image.
In word recognition, the text is assumed to be already detected and located and the rectangular bounding box containing the text is available. The word present in the bounding box needs to be recognized. The methods available to perform word recognition can be broadly classified into top-down and bottom-up approaches.
In the top-down approaches, a set of words from a dictionary is used to identify which word suits the given image.[8][9][10] Images are not segmented in most of these methods. Hence, the top-down approach is sometimes referred as segmentation free recognition.
In the bottom-up approaches, the image is segmented into multiple components and the segmented image is passed through a recognition engine.[11][12][13] Either an off the shelf Optical character recognition (OCR) engine [14][15][16] or a custom-trained one is used to recognise the text.
^Novikova, Tatiana; Barinova, Olga; Kohli, Pushmeet; Lempitsky, Victor (2012). "Large-Lexicon Attribute-Consistent Text Recognition in Natural Images". Computer Vision – ECCV 2012. Lecture Notes in Computer Science. Vol. 7577. pp. 752–765. CiteSeerX10.1.1.296.4807. doi:10.1007/978-3-642-33783-3_54. ISBN978-3-642-33782-6.
^Kumar, Deepak; Ramakrishnan, A. G. (2012). "Power-law transformation for enhanced recognition of born-digital word images". D. Kumar and A. G. Ramakrishnan. Power-law transformation for enhanced recognition of born-digital word images. In Proc. 9th SPCOM, 2012. pp. 1–5. doi:10.1109/SPCOM.2012.6290009. ISBN978-1-4673-2014-6. S2CID13876092.
^D. Kumar; M. N. Anil Prasad; A. G. Ramakrishnan. "MAPS: Midline analysis and propagation of segmentation". Proc. 8th ICVGIP, 2012. doi:10.1145/2425333.2425348. S2CID13303734.
^Kumar, Deepak; Anil Prasad, M. N.; Ramakrishnan, A. G. (2013). "NESP: Nonlinear enhancement and selection of plane for optimal segmentation and recognition of scene word images". In Zanibbi, Richard; Coüasnon, Bertrand (eds.). Document Recognition and Retrieval XX. Vol. 8658. p. 865806. doi:10.1117/12.2008519. S2CID13848101.