Paper ID | MLR-APPL-IVASR-4.10 | ||
Paper Title | ANALYSIS OF THE NOVEL TRANSFORMER MODULE COMBINATION FOR SCENE TEXT RECOGNITION | ||
Authors | Yeon-Gyu Kim, Hyunsu Kim, Minseok Kang, Hyug-Jae Lee, Rokkyu Lee, Gunhan Park, NHN, Republic of Korea | ||
Session | MLR-APPL-IVASR-4: Machine learning for image and video analysis, synthesis, and retrieval 4 | ||
Location | Area B | ||
Session Time: | Tuesday, 21 September, 13:30 - 15:00 | ||
Presentation Time: | Tuesday, 21 September, 13:30 - 15:00 | ||
Presentation | Poster | ||
Topic | Applications of Machine Learning: Machine learning for image & video analysis, synthesis, and retrieval | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | Various methods for scene text recognition (STR) are proposed every year. These methods dramatically increase the performance of the existing STR field; however, they have not been able to keep up with the progress of general-purpose research in image recognition, detection, speech recognition, and text analysis. In this paper, we evaluate the performance of several deep learning schemes for the encoder part of the Transformer in STR. First, we change the baseline feed forward network (FFN) module of encoder to squeeze-and- excitation (SE)-FFN or cross stage partial (CSP)-FFN. Second, the overall architecture of encoder is replaced with local dense synthesizer attention (LDSA) or Conformer structure. Conformer encoder achieves the best test accuracy in various experiments, and SE or CSP-FFN also showed competitive performance when the number of parameters is considered. Visualizing the attention maps from different encoder combinations allows for qualitative performance analysis. |