Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/132177
Citations
Scopus Web of Science® Altmetric
?
?
Full metadata record
DC FieldValueLanguage
dc.contributor.authorWang, W.-
dc.contributor.authorLiu, X.-
dc.contributor.authorJi, X.-
dc.contributor.authorXie, E.-
dc.contributor.authorLiang, D.-
dc.contributor.authorYang, Z.B.-
dc.contributor.authorLu, T.-
dc.contributor.authorShen, C.-
dc.contributor.authorLuo, P.-
dc.date.issued2020-
dc.identifier.citationLecture Notes in Artificial Intelligence, 2020, vol.12359, pp.457-473-
dc.identifier.isbn3030585670-
dc.identifier.isbn9783030585679-
dc.identifier.issn0302-9743-
dc.identifier.issn1611-3349-
dc.identifier.urihttps://hdl.handle.net/2440/132177-
dc.description.abstractScene text spotting aims to detect and recognize the entire word or sentence with multiple characters in natural images. It is still challenging because ambiguity often occurs when the spacing between characters is large or the characters are evenly spread in multiple rows and columns, making many visually plausible groupings of the characters (e.g. “BERLIN” is incorrectly detected as “BERL” and “IN” in Fig. 1(c)). Unlike previous works that merely employed visual features for text detection, this work proposes a novel text spotter, named Ambiguity Eliminating Text Spotter (AE TextSpotter), which learns both visual and linguistic features to significantly reduce ambiguity in text detection. The proposed AE TextSpotter has three important benefits. 1) The linguistic representation is learned together with the visual representation in a framework. To our knowledge, it is the first time to improve text detection by using a language model. 2) A carefully designed language module is utilized to reduce the detection confidence of incorrect text lines, making them easily pruned in the detection stage. 3) Extensive experiments show that AE TextSpotter outperforms other state-of-theart methods by a large margin. For example, we carefully select a set of extremely ambiguous samples from the IC19-ReCTS dataset, where our approach surpasses other methods by more than 4%.-
dc.description.statementofresponsibilityWenhai Wang, Xuebo Liu, Xiaozhong Ji, Enze Xie, Ding Liang, ZhiBo Yang, Tong Lu, B, Chunhua Shen, and Ping Luo-
dc.language.isoen-
dc.publisherSpringer-
dc.relation.ispartofseriesLecture Notes in Computer Science; 12359-
dc.rights© Springer Nature Switzerland AG 2020-
dc.source.urihttps://link.springer.com/book/10.1007/978-3-030-58568-6-
dc.subjectText spotting; Text detection; Text recognition; Text detection ambiguity-
dc.titleAE TextSpotter: Learning visual and linguistic representation for ambiguous text spotting-
dc.typeConference paper-
dc.contributor.conferenceEuropean Conference on Computer Vision (ECCV) (23 Aug 2020 - 28 Aug 2020 : virtual online)-
dc.identifier.doi10.1007/978-3-030-58568-6_27-
dc.publisher.placeCham, Switzerland-
pubs.publication-statusPublished-
dc.identifier.orcidShen, C. [0000-0002-8648-8718]-
Appears in Collections:Computer Science publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.