Multi-Stream Long Short-Term Memory Neural Network Language Model
Loading...
Date
2015
Authors
Arısoy, Ebru
Journal Title
Journal ISSN
Volume Title
Publisher
Abstract
Long Short-Term Memory (LSTM) neural networks are recurrent neural networks that contain memory units that can store contextual information from past inputs for arbitrary amounts of time. A typical LSTM neural network language model is trained by feeding an input sequence. i.e., a stream of words, to the input layer of the network and the output layer predicts the probability of the next word given the past inputs in the sequence. In this paper we introduce a multi-stream LSTM neural network language model where multiple asynchronous input sequences are fed to the network as parallel streams while predicting the output word sequence. For our experiments, we use a sub-word sequence in addition to a word sequence as the input streams, which allows joint training of the LSTM neural network language model using both information sources.
Description
Ebru Arısoy (MEF Author)
ORCID
Keywords
Long short-term memory, Sub-word-based language modeling, Language modeling
Turkish CoHE Thesis Center URL
Citation
Arisoy, E., Saraclar, M., (2015). Multi-stream long short-term memory neural network language model. Conference: 16th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2015) Location: Dresden, GERMANY, vol: 1-5. p. 1413-1417.
WoS Q
N/A
Scopus Q
N/A
Source
Conference: 16th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2015) Location: Dresden, GERMANY Date: SEP 06-10, 2015
Volume
1_5
Issue
Start Page
1413
End Page
1417