pith. machine review for the scientific record. sign in

arxiv: 1612.06138 · v2 · submitted 2016-12-19 · 💻 cs.CL

Recognition: unknown

Boosting Neural Machine Translation

Authors on Pith no claims yet
classification 💻 cs.CL
keywords trainingneuraltranslationboostingdatalearningmachinetime
0
0 comments X
read the original abstract

Training efficiency is one of the main problems for Neural Machine Translation (NMT). Deep networks need for very large data as well as many training iterations to achieve state-of-the-art performance. This results in very high computation cost, slowing down research and industrialisation. In this paper, we propose to alleviate this problem with several training methods based on data boosting and bootstrap with no modifications to the neural network. It imitates the learning process of humans, which typically spend more time when learning "difficult" concepts than easier ones. We experiment on an English-French translation task showing accuracy improvements of up to 1.63 BLEU while saving 20% of training time.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.