method: TextFuseNet2020-07-31

Authors: Jian Ye, Zhe Chen, Juhua Liu and Bo Du

Affiliation: Wuhan University, The University of Sydney

Email: liujuhua@whu.edu.cn

Description: Arbitrary shape text detection in natural scenes is an extremely challenging task. Unlike existing text detection approaches that only perceive texts based on limited feature representations, we propose a novel framework, namely TextFuseNet, to exploit the use of richer features fused for text detection. More specifically, we propose to perceive texts from three levels of feature representations, i.e., character-, word- and global-level, and then introduce a novel text representation fusion technique to help achieve robust arbitrary text detection. The multi-level feature representation can adequately describe texts by dissecting them into individual characters while still maintaining their general semantics. TextFuseNet then collects and merges the texts’ features from different levels using a multi-path fusion architecture which can effectively align and fuse different representations. In practice, our proposed TextFuseNet can learn a more adequate description of arbitrary shapes texts, suppressing false positives and producing more accurate detection results. Our proposed framework can also be trained with weak supervision for those datasets that lack character-level annotations. Experiments on several datasets show that the proposed TextFuseNet achieves state-of-the-art performance. Specifically, we achieve an F-measure of 94.3% on ICDAR2013, 92.1% on ICDAR2015,87.1% on Total-Text and 86.6% on CTW-1500, respectively.

method: VARCO2020-12-15

Authors: Jaemyung Lee, Jusung Lee, Younghyun Lee, Joonsoo Lee

Affiliation: NCSOFT

Description: This work was supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No.1711117050, Text Localization and Recognition for Efficient Digital Contents Analysis)

method: HIT2020-05-13

Authors: Sihwan Kim and Taejang Park

Affiliation: Hana Institute of Technology

Description: we present the network architecture to maximize conditional log-likelihood by optimizing the lower bound with a proper approximate posterior that has shown impressive performance in several generative model. In addition, by extending layer of latent variables to multiple layers, the network is able to learn scale robust features with no task specific regularization or data augmentation. We provide a detailed analysis and show the results of three public benchmarks to confirm the efficiency and reliability of the proposed algorithm.

Ranking Table

Description Paper Source Code
DateMethodRecallPrecisionHmean
2020-07-31TextFuseNet90.78%95.58%93.11%
2020-12-15VARCO89.86%93.63%91.71%
2020-05-13HIT89.22%93.85%91.48%
2018-11-07CRAFT89.04%93.93%91.42%
2020-01-20VARCO90.50%92.01%91.25%
2018-01-22FOTS89.68%91.43%90.55%
2018-12-03SPCNet_TongJi & UESTC (single scale)88.68%91.86%90.24%
2019-07-12stela88.13%91.38%89.73%
2017-08-10SRC-B-MachineLearningLab-v487.49%90.81%89.12%
2017-12-15EPTN-SJTU87.31%90.62%88.93%
2020-05-19Craft++86.67%91.07%88.82%
2020-11-10Hancom Vision81.74%92.94%86.98%
2017-03-22MCLAB_TextBoxes_v283.29%89.94%86.49%
2016-12-16RRPN-483.56%89.53%86.44%
2018-01-04crpn82.28%89.65%85.81%
2018-12-08Unicamp-SRBR-v280.82%90.49%85.38%
2016-08-31MCLAB_TextBoxes82.28%87.82%84.96%
2015-03-26VGGMaxNet_cmb78.08%90.09%83.66%
2015-04-02VGGMaxNet_02579.82%87.58%83.52%
2018-12-08Unicamp-SRBR-v375.62%92.62%83.26%
2015-03-23VGGMaxNet_01376.53%90.50%82.93%
2015-03-23VGGMaxNet_1.675.89%91.32%82.89%
2019-06-26std(single-scale)76.99%80.98%78.93%
2016-11-13RRPN-370.50%88.23%78.38%
2015-01-01BUCT_YST72.15%83.60%77.45%
2016-03-16TextConv+WordGraph67.67%89.49%77.07%
2014-11-12HUST_MCLAB68.49%83.33%75.19%
2018-12-08Unicamp-SRBR-v163.20%88.04%73.58%
2015-04-03StradVision66.03%80.87%72.70%
2013-04-09I2R_NUS_FAR68.95%74.46%71.60%
2013-04-07USTB_TexStar61.46%84.76%71.25%
2017-03-16Ali-Amap-xlab-v266.03%76.35%70.81%
2016-06-23SRC-B-TextProcessingLab64.02%79.03%70.74%
2017-10-12TextFCN V274.52%66.61%70.34%
2013-04-08CASIA_NLPR66.12%74.64%70.12%
2016-12-04Ali-Amap-xlab64.93%75.96%70.01%
2013-04-05TextSpotter61.19%81.61%69.94%
2015-03-23VGGMaxNet_1054.70%96.61%69.85%
2015-07-22ZText60.00%82.95%69.63%
2015-11-04MSER_Binary_CNN63.29%76.49%69.27%
2013-04-08I2R_NUS65.30%72.08%68.52%
2018-12-29fast_ret_sh_0258.81%76.76%66.60%
2014-08-18DetectText59.82%71.58%65.17%
2013-04-08Text_detector_CASIA54.70%80.19%65.04%
2013-08-29UMD_IntegratedDisrimination52.69%81.61%64.04%
2013-04-08TH-TextLoc50.78%59.66%54.86%
2015-08-18MSER with LocalSWT38.90%60.43%47.33%
2013-04-06Text Detection34.25%60.29%43.68%
2014-06-10IWRR201432.24%56.12%40.95%
2017-03-05WeText31.23%56.16%40.14%
2016-11-08CTPN28.40%54.85%37.42%
2013-04-10Inkam28.04%29.10%28.56%

Ranking Graphic