Open Access Open Access  Restricted Access Subscription or Fee Access

Text Indicator in an Image

E. Priyanghaa, K. Pasunthamizhkumaran, J. Priyanka

Abstract


Text in an image provides vital information forinterpreting its contents, and text in a scene can aid a variety of tasks from navigation to obstacle avoidance and odometry. Despite its value, however, detecting general text in images remains a challenging research problem. Motivated by the need to consider the widely varying forms of natural text, we propose a bottom-up approach to the problem, which reflects the characterness of an image region. In this sense, our approach mirrors the move from saliency detection methods to measures of objectness. In order to measure the characterness, we develop three novel cues that are tailored for character detection and a Bayesian method for their integration. Because text is made up of sets of characters, we then design a Markov random field model so as to exploit the inherent dependencies between characters. We experimentally demonstrate the effectiveness of our characterness cues as well as the advantage of Bayesian multicue integration. The proposed text detector outperforms state-of-the-art methods on a few benchmark scene text detection data sets. We also show that our measurement of characterness is superior than state-of-the-art saliency detection models when applied to the same task.


Full Text:

PDF

References


N. D. B. Bruce and J. K. Tsotsos, “Saliency based on informa-tion maximization,” in Proc. Adv. Neural Inf. Process. Syst., 2005, 155–162.

R. Achanta, S. Hemami, F. Estrada, and S. Susstrunk, “Frequency-tuned salient region detection,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2009, pp. 1597–1604.

M.-M. Cheng, G.-X. Zhang, N. J. Mitra, X. Huang, and S.-M. Hu, “Global contrast based salient region detection,” in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2011, pp. 409–416.

[4] C. Wolf and J.-M. Jolion, “Object count/area graphs for the evaluation of object detection and segmentation algorithms,” Int. J. Document Anal.Recognit., vol. 8, no. 4, pp. 280–296, 2006.hypergraph modeling for salient object detection,” in Proc. IEEE Int.Conf. Comput. Vis., Oct. 2013, pp. 3328–3335.

Y. Zhai and M. Shah, “Visual attention detection in video sequences using spatiotemporal cues,” in Proc. 14th Annu. ACM Int. Conf. Multi-media , 2006, pp. 815–824.

R. Achanta, S. Hemami, F. Estrada, and S. Susstrunk, “Frequency-tuned salient region detection,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2009, pp. 1597–1604.

Y. Wei, F. Wen, W. Zhu, and J. Sun, “Geodesic saliency using back-ground priors,” in Proc. Eur. Conf. Comput. Vis., 2012, pp. 29–42.

J. Feng, Y. Wei, L. Tao, C. Zhang, and J. Sun, “Salient object detection by composition,” in Proc. IEEE Int. Conf. Comput. Vis., Nov. 2011, 1028–1035.

M.-M. Cheng, G.-X. Zhang, N. J. Mitra, X. Huang, and S.-M. Hu, “Global contrast based salient region detection,” in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2011, pp. 409–416.

T. Liu, Z. Yuan, J. Sun, J. Wang, N. Zheng, X. Tang, et al., “Learning to detect a salient object,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 33, no. 2, pp. 353–367, Feb. 2011.

D. A. Klein and S. Frintrop, “Center-surround divergence of feature statistics for salient object detection,” in Proc. IEEE Int. Conf. Comput.Vis., Nov. 2011, pp. 2214–2219.

B. Alexe, T. Deselaers, and V. Ferrari, “What is an object?” in Proc.IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2010, pp. 73–80.

F. Perazzi, P. Krahenbuhl, Y. Pritch, and A. Hornung, “Saliency filters: Contrast based filtering for salient region detection,” in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2012, pp. 733–740.

X. Shen and Y. Wu, “A unified approach to salient object detection via low rank matrix recovery,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2012, pp. 853–860.

H. Jiang, J. Wang, Z. Yuan, T. Liu, N. Zheng, and S. Li, “Automatic salient object segmentation based on context and shape prior,” Brit.Mach. Vis. Conf., vol. 3, no. 4, pp. 1–12, 2011.

S. Goferman, L. Zelnik-Manor, and A. Tal, “Context-aware saliency detection,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2010, pp. 2376–2383.

K.-Y. Chang, T.-L. Liu, H.-T. Chen, and S.-H. Lai, “Fusing generic objectness and visual saliency for salient object detection,” in Proc. IEEEInt. Conf. Comput. Vis., Nov. 2011, pp. 914–921.

E. Rahtu, J. Kannala, M. Salo, and J. Heikkilä, “Segmenting salient objects from images and videos,” in Proc. Eur. Conf. Comput. Vis., 2010, 366–379.

Y. Ding, J. Xiao, and J. Yu, “Importance filtering for image retarget-ing,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2011, 89–96.

J. Sun and H. Ling, “Scale and object aware image retargeting for thumbnail browsing,” in Proc. IEEE Int. Conf. Comput. Vis., Nov. 2011, 1511–1518.

G. Sharma, F. Jurie, and C. Schmid, “Discriminative spatial saliency for image classification,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2012, pp. 3506–3513.

M. Cerf, E. P. Frady, and C. Koch, “Faces and text attract gaze independent of the task: Experimental data and computer model,” J. Vis., vol. 9, no. 12, pp. 1–27, 2009.

Q. Sun, Y. Lu, and S. Sun, “A visual attention based approach to text extraction,” in Proc. IEEE 20th Int. Conf. Pattern Recognit., Aug. 2010, 3991–3995.

A. Shahab, F. Shafait, A. Dengel, and S. Uchida, “How salient is scene text?” in Proc. IEEE 10th Int. Workshop. Document Anal. Syst., Mar. 2012, pp. 317–321.

Q. Meng and Y. Song, “Text detection in natural scenes with salient region,” in Proc. IEEE 10th Int. Workshop Document Anal. Syst., Mar. 2012, pp. 384–388.

S. Uchida, Y. Shigeyoshi, Y. Kunishige, and Y. Feng, “A keypoint-based approach toward scenery character detection,” in Proc. IEEE Int. Conf.Doc. Anal. Recognit., Sep. 2011, pp. 819–823.

D. Hoiem, A. A. Efros, and M. Hebert, “Recovering occlusion bound-aries from an image,” Int. J. Comput. Vis., vol. 91, no. 3, pp. 328–346, 2011.

P. Arbelaez, B. Hariharan, C. Gu, S. Gupta, L. D. Bourdev, and J. Malik, “Semantic segmentation using regions and parts,” in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2012, pp. 3378–3385.

A. Prest, C. Leistner, J. Civera, C. Schmid, and V. Ferrari, “Learning object class detectors from weakly annotated video,” in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2012, pp. 3282–3289.

Y. Boykov, O. Veksler, and R. Zabih, “Fast approximate energy min-imization via graph cuts,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 23, no. 11, pp. 1222–1239, Nov. 2001.

J.-J. Lee, P.-H. Lee, S.-W. Lee, A. L. Yuille, and C. Koch, “Adaboost for text detection in natural scene,” in Proc. IEEE Int. Conf. DocumentAnal. Recognit., Sep. 2011, pp. 429–434.

X. Chen and A. L. Yuille, “Detecting and reading text in natural scenes,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jul. 2004, 366–373.

B. Epshtein, E. Ofek, and Y. Wexler, “Detecting text in natural scenes with stroke width transform,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2010, pp. 2963–2970.

J. Zhang and R. Kasturi, “Text detection using edge gradient and graph spectrum,” in Proc. IEEE 20th Int. Conf. Pattern Recognit., Aug. 2010, 3979–3982.

J. Zhang and R. Kasturi, “Character energy and link energy-based text extraction in scene images,” in Proc. Asian Conf. Comput. Vis., 2010, 308–320.

C. Yi and Y. Tian, “Text string detection from natural scenes by structure-based partition and grouping,” IEEE Trans. Image Process., vol. 20, no. 9, pp. 2594–2605, Apr. 2011.

H. Chen, S. S. Tsai, G. Schroth, D. M. Chen, R. Grzeszczuk, and Girod, “Robust text detection in natural images with edge-enhanced maximally stable extremal regions,” in Proc. 18th IEEE Int. Conf. ImageProcess., Sep. 2011, pp. 2609–2612.

Y. Li and H. Lu, “Scene text detection via stroke width,” in Proc. IEEEInt. Conf. Pattern Recognit., Nov. 2012, pp. 681–684.

Y. Li, C. Shen, W. Jia, and A. van den Hengel, “Leveraging surrounding context for scene text detection,” in Proc. IEEE Int. Conf. ImageProcess., Feb. 2013, pp. 2264–2268.

L. Neumann and J. Matas, “A method for text localization and recog-nition in real-world images,” in Proc. Asian Conf. Comput. Vis., 2010, pp. 770–783.

L. Neumann and J. Matas, “Real-time scene text localization and recog-nition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2012, pp. 3538–3545.

Y.-F. Pan, X. Hou, and C.-L. Liu, “A hybrid approach to detect and localize texts in natural scene images,” IEEE Trans. Image Process., vol. 20, no. 3, pp. 800–813, Mar. 2011.

C. Yao, X. Bai, W. Liu, Y. Ma, and Z. Tu, “Detecting texts of arbitrary orientations in natural images,” in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2012, pp. 1083–1090.

C. Yi and Y. Tian, “Localizing text in scene images by boundary clustering, stroke segmentation, and string fragment classification,” IEEETrans. Image Process., vol. 21, no. 9, pp. 4256–4268, Sep. 2012.

H. Koo and D. Kim, “Scene text detection via connected component clustering and non-text filtering,” IEEE Trans. Image Process., vol. 22, no. 6, pp. 2296–2305, Jun. 2013.

W. Huang, Z. Lin, J. Yang, and J. Wang, “Text localization in natural images using stroke feature transform and text covariance descriptors,” in Proc. IEEE Int. Conf. Comput. Vis., Aug. 2013, pp. 1241–1248.

J. Matas, O. Chum, M. Urban, and T. Pajdla, “Robust wide baseline stereo from maximally stable extremal regions,” in Proc. Brit. Mach.Vis. Conf., 2002, pp. 384–393.

M. Donoser and H. Bischof, “Efficient maximally stable extremal region (MSER) tracking,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2006, pp. 553–560.

P.-E. Forssén and D. G. Lowe, “Shape descriptors for maximally stable extremal regions,” in Proc. IEEE 11th Int. Conf. Comput. Vis., Oct. 2007, pp. 1–8.

L. Neumann and J. Matas, “Text localization in real-world images using efficiently pruned exhaustive search,” in Proc. IEEE Int. Conf. DocumentAnal. Recognit., Sep. 2011, pp. 687–691.

S. Tsai, V. Parameswaran, J. Berclaz, R. Vedantham, R. Grzeszczuk, and Girod, “Design of a text detection system via hypothesis generation and verification,” in Proc. Asian Conf. Comput. Vis., 2012, pp. 1–12.

K. Mikolajczyk, T. Tuytelaars, C. Schmid, A. Zisserman, J. Matas, Schaffalitzky, et al., “A comparison of affine region detectors,” Int. Comput. Vis., vol. 65, nos. 1–2, pp. 43–72, 2005


Refbacks

  • There are currently no refbacks.


Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 License.