/E2E_TIT_With_MT

We will prepare our codes and datasets in this repository. Codes and Datasets are utilized to train end-to-end text image translation model with auxiliary text translation task.

Primary LanguagePython

E2E_TIT_With_MT

E2E_TIT_With_MT: End-to-end Text Image Translation with Machine Translation.

The official repository for ICPR 2022 main conference paper:

  • Cong Ma, Yaping Zhang, Mei Tu, Xu Han, Linghui Wu, Yang Zhao, Yu Zhou. Improving End-to-End Text Image Translation From the Auxiliary Text Translation Task. In Proceedings of the 26th International Conference on Pattern Recognition (ICPR 2022), Virtually, Montréal Québec, Canada. August 21-25, 2022. pp.1664-1670. arXiv_version, ieeexplore_version, GitHub.

1. Introduction

End-to-end text image translation (TIT), which aims at translating the source language embedded in images to the target language, has attracted intensive attention in recent research. However, data sparsity limits the performance of end- to-end text image translation. Multi-task learning is a nontrivial way to alleviate this problem via exploring knowledge from complementary related tasks. In this paper, we propose a novel text translation enhanced text image translation, which trains the end-to-end model with text translation as an auxiliary task. By sharing model parameters and multi-task training, our model is able to take full advantage of easily-available large-scale text parallel corpus. Extensive experimental results show our proposed method outperforms existing end-to-end methods, and the joint multi-task learning with both text translation and recognition tasks achieves better results, proving MT and OCR auxiliary tasks are complementary.

2. Usage

2.1 Requirements

  • python==3.6.2
  • pytorch == 1.3.1
  • torchvision==0.4.2
  • numpy==1.19.1
  • lmdb==0.99
  • PIL==7.2.0
  • jieba==0.42.1
  • nltk==3.5
  • six==1.15.0
  • natsort==7.0.1

2.2 Train the Model

bash ./train_model_guide.sh

2.3 Evaluate the Model

bash ./test_model_guide.sh

2.4 Datasets

We have conducted synthetic, subtitle and street-view datasets. If you want to utilize these datasets for research, please contact with cong.ma@nlpr.ia.ac.cn.

3. Acknowledgement

The reference code of the provided methods are:

We thanks for all these researchers who have made their codes publicly available.

4. Citation

If you want to cite our paper, please use this bibtex version:

  • ieeexplore offered bib citation format

    • @INPROCEEDINGS{9956695,
        author={Ma, Cong and Zhang, Yaping and Tu, Mei and Han, Xu and Wu, Linghui and Zhao, Yang and Zhou, Yu},
        booktitle={2022 26th International Conference on Pattern Recognition (ICPR)}, 
        title={Improving End-to-End Text Image Translation From the Auxiliary Text Translation Task}, 
        year={2022},
        volume={},
        number={},
        pages={1664-1670},
        doi={10.1109/ICPR56361.2022.9956695}}
  • Semantic Scholar offered bib citation format

    • @article{Ma2022ImprovingET,
        title={Improving End-to-End Text Image Translation From the Auxiliary Text Translation Task},
        author={Cong Ma and Yaping Zhang and Mei Tu and Xu Han and Linghui Wu and Yang Zhao and Yu Zhou},
        journal={2022 26th International Conference on Pattern Recognition (ICPR)},
        year={2022},
        pages={1664-1670}
      }
  • DBLP offered bib citation format

    • @inproceedings{DBLP:conf/icpr/MaZTHWZ022,
        author    = {Cong Ma and
                     Yaping Zhang and
                     Mei Tu and
                     Xu Han and
                     Linghui Wu and
                     Yang Zhao and
                     Yu Zhou},
        title     = {Improving End-to-End Text Image Translation From the Auxiliary Text
                     Translation Task},
        booktitle = {26th International Conference on Pattern Recognition, {ICPR} 2022,
                     Montreal, QC, Canada, August 21-25, 2022},
        pages     = {1664--1670},
        publisher = {{IEEE}},
        year      = {2022},
        url       = {https://doi.org/10.1109/ICPR56361.2022.9956695},
        doi       = {10.1109/ICPR56361.2022.9956695},
        timestamp = {Thu, 01 Dec 2022 15:50:19 +0100},
        biburl    = {https://dblp.org/rec/conf/icpr/MaZTHWZ022.bib},
        bibsource = {dblp computer science bibliography, https://dblp.org}
      }

If you have any issues, please contact with cong.ma@nlpr.ia.ac.cn.