This paper presents a new method to reduce the computational cost when using Neural Networks as Language Models, during recognition, in some particular scenarios. It is based on a Neural Network that considers input contexts of different length in order to ease the use of a fallback mechanism together with the precomputation of softmax normalization constants for these inputs. The proposed approach is empirically validated, showing their capability to emulate lower order N-grams with a single Neural Network. A machine translation task shows that the proposed model constitutes a good solution to the normalization cost of the output softmax layer of Neural Networks, for some practical cases, without a significant impact in performance while improving the system speed.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC6062053PMC
http://journals.plos.org/plosone/article?id=10.1371/journal.pone.0200884PLOS

Publication Analysis

Top Keywords

neural networks
8
neural network
8
fallback variable
4
variable history
4
history nnlms
4
nnlms efficient
4
efficient nnlms
4
nnlms precomputation
4
precomputation stochastic
4
stochastic training
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!