TRIG: Transformer-Based Text Recognizer with Initial Embedding Guidance. Tao, Y., Jia, Z., Ma, R., & Xu, S. Electronics, 10(22):2780, January, 2021. Number: 22 Publisher: Multidisciplinary Digital Publishing Institute
TRIG: Transformer-Based Text Recognizer with Initial Embedding Guidance [link]Paper  doi  abstract   bibtex   
Scene text recognition (STR) is an important bridge between images and text, attracting abundant research attention. While convolutional neural networks (CNNS) have achieved remarkable progress in this task, most of the existing works need an extra module (context modeling module) to help CNN to capture global dependencies to solve the inductive bias and strengthen the relationship between text features. Recently, the transformer has been proposed as a promising network for global context modeling by self-attention mechanism, but one of the main short-comings, when applied to recognition, is the efficiency. We propose a 1-D split to address the challenges of complexity and replace the CNN with the transformer encoder to reduce the need for a context modeling module. Furthermore, recent methods use a frozen initial embedding to guide the decoder to decode the features to text, leading to a loss of accuracy. We propose to use a learnable initial embedding learned from the transformer encoder to make it adaptive to different input images. Above all, we introduce a novel architecture for text recognition, named TRansformer-based text recognizer with Initial embedding Guidance (TRIG), composed of three stages (transformation, feature extraction, and prediction). Extensive experiments show that our approach can achieve state-of-the-art on text recognition benchmarks.
@article{tao_trig_2021,
	title = {{TRIG}: {Transformer}-{Based} {Text} {Recognizer} with {Initial} {Embedding} {Guidance}},
	volume = {10},
	copyright = {http://creativecommons.org/licenses/by/3.0/},
	issn = {2079-9292},
	shorttitle = {{TRIG}},
	url = {https://www.mdpi.com/2079-9292/10/22/2780},
	doi = {10.3390/electronics10222780},
	abstract = {Scene text recognition (STR) is an important bridge between images and text, attracting abundant research attention. While convolutional neural networks (CNNS) have achieved remarkable progress in this task, most of the existing works need an extra module (context modeling module) to help CNN to capture global dependencies to solve the inductive bias and strengthen the relationship between text features. Recently, the transformer has been proposed as a promising network for global context modeling by self-attention mechanism, but one of the main short-comings, when applied to recognition, is the efficiency. We propose a 1-D split to address the challenges of complexity and replace the CNN with the transformer encoder to reduce the need for a context modeling module. Furthermore, recent methods use a frozen initial embedding to guide the decoder to decode the features to text, leading to a loss of accuracy. We propose to use a learnable initial embedding learned from the transformer encoder to make it adaptive to different input images. Above all, we introduce a novel architecture for text recognition, named TRansformer-based text recognizer with Initial embedding Guidance (TRIG), composed of three stages (transformation, feature extraction, and prediction). Extensive experiments show that our approach can achieve state-of-the-art on text recognition benchmarks.},
	language = {en},
	number = {22},
	urldate = {2023-09-29},
	journal = {Electronics},
	author = {Tao, Yue and Jia, Zhiwei and Ma, Runze and Xu, Shugong},
	month = jan,
	year = {2021},
	note = {Number: 22
Publisher: Multidisciplinary Digital Publishing Institute},
	keywords = {1-D split, initial embedding, scene text recognition, self-attention, transformer},
	pages = {2780},
}

Downloads: 0