Journal
2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP)
Volume -, Issue -, Pages 285-289Publisher
IEEE
DOI: 10.1109/icip.2019.8802972
Keywords
sign language recognition; reinforcement learning; self-critic
Categories
Funding
- 973 Program [2015CB351803]
- NSFC [61836011, 61472379, 61632019]
- Young Elite Scientists Sponsorship Program By CAST [2016QNRC001]
Ask authors/readers for more resources
In this paper, we propose an approach to apply the Transformer with reinforcement learning (RL) for continuous sign language recognition (CSLR) task. The Transformer has an encoder-decoder structure, where the encoder network encodes the sign video into the context vector representation, while the decoder network generates the target sentence word by word based on the context vector. To avoid the intrinsic defects of supervised learning (SL) in our task, e.g., the exposure bias and non-differentiable task metrics issues, we propose to train the Transformer directly on non-differentiable metrics, i.e., word error rate (WER), through RL. Moreover, a policy gradient algorithm with baseline, which we call Self-critic REINFORCE, is employed to reduce variance while training. Experimental results on RWTH-PHOENIX-Weather benchmark verify the effectiveness of our method and demonstrate that our method achieves the comparable performance.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available