pith. machine review for the scientific record. sign in

arxiv: 1510.08983 · v2 · submitted 2015-10-30 · 💻 cs.NE · cs.AI· cs.CL· cs.LG· eess.AS

Recognition: unknown

Highway Long Short-Term Memory RNNs for Distant Speech Recognition

Authors on Pith no claims yet
classification 💻 cs.NE cs.AIcs.CLcs.LGeess.AS
keywords lstmshighwaymemoryconnectionsdeeperdirectdistantdlstm
0
0 comments X
read the original abstract

In this paper, we extend the deep long short-term memory (DLSTM) recurrent neural networks by introducing gated direct connections between memory cells in adjacent layers. These direct links, called highway connections, enable unimpeded information flow across different layers and thus alleviate the gradient vanishing problem when building deeper LSTMs. We further introduce the latency-controlled bidirectional LSTMs (BLSTMs) which can exploit the whole history while keeping the latency under control. Efficient algorithms are proposed to train these novel networks using both frame and sequence discriminative criteria. Experiments on the AMI distant speech recognition (DSR) task indicate that we can train deeper LSTMs and achieve better improvement from sequence training with highway LSTMs (HLSTMs). Our novel model obtains $43.9/47.7\%$ WER on AMI (SDM) dev and eval sets, outperforming all previous works. It beats the strong DNN and DLSTM baselines with $15.7\%$ and $5.3\%$ relative improvement respectively.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Training Deep Nets with Sublinear Memory Cost

    cs.LG 2016-04 accept novelty 7.0

    An algorithm trains n-layer networks with O(sqrt(n)) memory via selective recomputation of activations, at the cost of one extra forward pass.