Integrating multiple character proposals for robust scene text extraction

Cited 20 time in webofscience Cited 20 time in scopus
  • Hit : 550
  • Download : 121
DC FieldValueLanguage
dc.contributor.authorLee, SeongHunko
dc.contributor.authorKim, JinHyungko
dc.date.accessioned2014-08-29-
dc.date.available2014-08-29-
dc.date.created2014-01-02-
dc.date.created2014-01-02-
dc.date.issued2013-11-
dc.identifier.citationIMAGE AND VISION COMPUTING, v.31, no.11, pp.823 - 840-
dc.identifier.issn0262-8856-
dc.identifier.urihttp://hdl.handle.net/10203/188611-
dc.description.abstractText contained in scene images provides the semantic context of the images. For that reason, robust extraction of text regions is essential for successful scene text understanding. However, separating text pixels from scene images still remains as a challenging issue because of uncontrolled lighting conditions and complex backgrounds. In this paper, we propose a two-stage conditional random field (TCRF) approach to robustly extract text regions from the scene images. The proposed approach models the spatial and hierarchical structures of the scene text, and it finds text regions based on the scene text model. In the first stage, the system generates multiple character proposals for the given image by using multiple image segmentations and a local CRF model. In the second stage, the system selectively integrates the generated character proposals to determine proper character regions by using a holistic CRF model. Through the TCRF approach, we cast the scene text separation problem as a probabilistic labeling problem, which yields the optimal label configuration of pixels that maximizes the conditional probability of the given image. Experimental results indicate that our framework exhibits good performance in the case of the public databases.-
dc.languageEnglish-
dc.publisherELSEVIER SCIENCE BV-
dc.subjectOBJECT DETECTION-
dc.subjectIMAGES-
dc.subjectCOLOR-
dc.subjectRECOGNITION-
dc.subjectFEATURES-
dc.titleIntegrating multiple character proposals for robust scene text extraction-
dc.typeArticle-
dc.identifier.wosid000328184100001-
dc.identifier.scopusid2-s2.0-84885087482-
dc.type.rimsART-
dc.citation.volume31-
dc.citation.issue11-
dc.citation.beginningpage823-
dc.citation.endingpage840-
dc.citation.publicationnameIMAGE AND VISION COMPUTING-
dc.identifier.doi10.1016/j.imavis.2013.08.007-
dc.embargo.liftdate9999-12-31-
dc.embargo.terms9999-12-31-
dc.contributor.localauthorKim, JinHyung-
dc.contributor.nonIdAuthorLee, SeongHun-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorScene text extraction-
dc.subject.keywordAuthorTwo-stage CRF models-
dc.subject.keywordAuthorMultiple image segmentations-
dc.subject.keywordAuthorComponent-
dc.subject.keywordAuthorCharacter proposal-
dc.subject.keywordPlusOBJECT DETECTION-
dc.subject.keywordPlusIMAGES-
dc.subject.keywordPlusCOLOR-
dc.subject.keywordPlusRECOGNITION-
dc.subject.keywordPlusFEATURES-
Appears in Collection
CS-Journal Papers(저널논문)
Files in This Item
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 20 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0