Multi-Stream Long Short-Term Memory Neural Network Language Model

Loading...
Thumbnail Image

Date

2015

Authors

Arısoy, Ebru

Journal Title

Journal ISSN

Volume Title

Publisher

Abstract

Long Short-Term Memory (LSTM) neural networks are recurrent neural networks that contain memory units that can store contextual information from past inputs for arbitrary amounts of time. A typical LSTM neural network language model is trained by feeding an input sequence. i.e., a stream of words, to the input layer of the network and the output layer predicts the probability of the next word given the past inputs in the sequence. In this paper we introduce a multi-stream LSTM neural network language model where multiple asynchronous input sequences are fed to the network as parallel streams while predicting the output word sequence. For our experiments, we use a sub-word sequence in addition to a word sequence as the input streams, which allows joint training of the LSTM neural network language model using both information sources.

Description

Ebru Arısoy (MEF Author)

Keywords

Long short-term memory, Sub-word-based language modeling, Language modeling

Turkish CoHE Thesis Center URL

Citation

Arisoy, E., Saraclar, M., (2015). Multi-stream long short-term memory neural network language model. Conference: 16th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2015) Location: Dresden, GERMANY, vol: 1-5. p. 1413-1417.

WoS Q

N/A

Scopus Q

N/A

Source

Conference: 16th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2015) Location: Dresden, GERMANY Date: SEP 06-10, 2015

Volume

1_5

Issue

Start Page

1413

End Page

1417