TY - GEN
T1 - Improving Offline Handwritten Chinese Character Recognition by Iterative Refinement
AU - Yang, Xiao
AU - He, Dafang
AU - Zhou, Zihan
AU - Kifer, Daniel
AU - Giles, C. Lee
N1 - Publisher Copyright:
© 2017 IEEE.
PY - 2017/7/2
Y1 - 2017/7/2
N2 - We present an iterative refinement module that can be applied to the output feature maps of any existing convolutional neural networks in order to further improve classification accuracy. The proposed module, implemented by an attention-based recurrent neural network, can iteratively use its previous predictions to update attention and thereafter refine current predictions. In this way, the model is able to focus on a sub-region of input images to distinguish visually similar characters (see Figure 1 for an example). We evaluate its effectiveness on handwritten Chinese character recognition (HCCR) task and observe significant performance gain. HCCR task is challenging due to large number of classes and small differences between certain characters. To overcome these difficulties, we further propose a novel convolutional architecture that utilizes both low-level visual cues and high-level structural information. Together with the proposed iterative refinement module, our approach achieves an accuracy of 97.37%, outperforming previous methods that use raw images as input on ICDAR-2013 dataset [1].
AB - We present an iterative refinement module that can be applied to the output feature maps of any existing convolutional neural networks in order to further improve classification accuracy. The proposed module, implemented by an attention-based recurrent neural network, can iteratively use its previous predictions to update attention and thereafter refine current predictions. In this way, the model is able to focus on a sub-region of input images to distinguish visually similar characters (see Figure 1 for an example). We evaluate its effectiveness on handwritten Chinese character recognition (HCCR) task and observe significant performance gain. HCCR task is challenging due to large number of classes and small differences between certain characters. To overcome these difficulties, we further propose a novel convolutional architecture that utilizes both low-level visual cues and high-level structural information. Together with the proposed iterative refinement module, our approach achieves an accuracy of 97.37%, outperforming previous methods that use raw images as input on ICDAR-2013 dataset [1].
UR - http://www.scopus.com/inward/record.url?scp=85045202503&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85045202503&partnerID=8YFLogxK
U2 - 10.1109/ICDAR.2017.11
DO - 10.1109/ICDAR.2017.11
M3 - Conference contribution
AN - SCOPUS:85045202503
T3 - Proceedings of the International Conference on Document Analysis and Recognition, ICDAR
SP - 5
EP - 10
BT - Proceedings - 14th IAPR International Conference on Document Analysis and Recognition, ICDAR 2017
PB - IEEE Computer Society
T2 - 14th IAPR International Conference on Document Analysis and Recognition, ICDAR 2017
Y2 - 9 November 2017 through 15 November 2017
ER -