method: Shopee MMU OCR2022-09-28

Authors: Jianqiang Liu, Hanfei Xu, Bin Zheng, Eric W, Ronnie T, Alex X

Affiliation: Shopee MMU

Description: Our method adopts a transformer-based context-aware framework. We utilize a hybrid architecture encoder and a context-aware autoregressive decoder to construct the recognition pipeline. Finally, a simple but effective multi-model fusion strategy is adopted.

method: SogouMM2019-11-07

Authors: Xu Liu, Tao Wei

Description: Our method is based on 2D-attention, we use ResNet as backbone and a tailored 2D-attention module is applied. The result is generated by single model without ensemble tricks.

method: Hancom Vision2020-10-06

Authors: Hancom Vision team

Description: Our model is featured by CNN-based, BiLSTM, and Attention.
Trained on MJSynthText + SynthText + external data (Pretrain), Focused Scene Text 2013-2015, and Incidental Scene Text 2015.

Ranking Table

Description Paper Source Code
DateMethodTotal Edit distance (case sensitive)Correctly Recognised Words (case sensitive)T.E.D. (case insensitive)C.R.W. (case insensitive)
2022-09-28Shopee MMU OCR134.811087.14%104.668289.17%
2019-11-07SogouMM144.502986.42%113.157388.11%
2020-10-06Hancom Vision160.266786.09%108.377388.93%
2019-10-31Sogou_OCR163.095484.35%129.283186.66%
2018-09-13Clova AI / Lens175.436783.00%132.422985.56%
2020-06-10test 1164.429082.91%129.243385.07%
2018-07-03Baidu VIS185.807882.85%150.852784.68%
2018-12-19SAR437.164267.40%203.144678.82%
2016-01-29SRC-B-TextProcessingLab419.741262.11%367.122264.95%
2015-11-09Megvii-Image++508.832357.82%377.652163.99%
2015-04-01MAPS1,128.007532.93%1,068.718433.90%
2015-04-01NESP1,164.496831.68%1,094.707132.98%
2015-04-02DSM1,178.614025.85%1,108.938127.97%

Ranking Graphic

Ranking Graphic