Skip to yearly menu bar Skip to main content


Poster

Connectionist Temporal Classification with Maximum Entropy Regularization

Hu Liu · Sheng Jin · Changshui Zhang

Room 517 AB #106

Keywords: [ Recurrent Networks ] [ Regularization ]


Abstract:

Connectionist Temporal Classification (CTC) is an objective function for end-to-end sequence learning, which adopts dynamic programming algorithms to directly learn the mapping between sequences. CTC has shown promising results in many sequence learning applications including speech recognition and scene text recognition. However, CTC tends to produce highly peaky and overconfident distributions, which is a symptom of overfitting. To remedy this, we propose a regularization method based on maximum conditional entropy which penalizes peaky distributions and encourages exploration. We also introduce an entropy-based pruning method to dramatically reduce the number of CTC feasible paths by ruling out unreasonable alignments. Experiments on scene text recognition show that our proposed methods consistently improve over the CTC baseline without the need to adjust training settings. Code has been made publicly available at: https://github.com/liuhu-bigeye/enctc.crnn.

Live content is unavailable. Log in and register to view live content