Continuous Space Models with Neural Networks in Natural Language Processing

Continuous Space Models with Neural Networks in Natural Language Processing
Author :
Publisher :
Total Pages : 0
Release :
ISBN-10 : OCLC:836036471
ISBN-13 :
Rating : 4/5 ( Downloads)

Book Synopsis Continuous Space Models with Neural Networks in Natural Language Processing by : Hai Son Le

Download or read book Continuous Space Models with Neural Networks in Natural Language Processing written by Hai Son Le and published by . This book was released on 2012 with total page 0 pages. Available in PDF, EPUB and Kindle. Book excerpt: The purpose of language models is in general to capture and to model regularities of language, thereby capturing morphological, syntactical and distributional properties of word sequences in a given language. They play an important role in many successful applications of Natural Language Processing, such as Automatic Speech Recognition, Machine Translation and Information Extraction. The most successful approaches to date are based on n-gram assumption and the adjustment of statistics from the training data by applying smoothing and back-off techniques, notably Kneser-Ney technique, introduced twenty years ago. In this way, language models predict a word based on its n-1 previous words. In spite of their prevalence, conventional n-gram based language models still suffer from several limitations that could be intuitively overcome by consulting human expert knowledge. One critical limitation is that, ignoring all linguistic properties, they treat each word as one discrete symbol with no relation with the others. Another point is that, even with a huge amount of data, the data sparsity issue always has an important impact, so the optimal value of n in the n-gram assumption is often 4 or 5 which is insufficient in practice. This kind of model is constructed based on the count of n-grams in training data. Therefore, the pertinence of these models is conditioned only on the characteristics of the training text (its quantity, its representation of the content in terms of theme, date). Recently, one of the most successful attempts that tries to directly learn word similarities is to use distributed word representations in language modeling, where distributionally words, which have semantic and syntactic similarities, are expected to be represented as neighbors in a continuous space. These representations and the associated objective function (the likelihood of the training data) are jointly learned using a multi-layer neural network architecture. In this way, word similarities are learned automatically. This approach has shown significant and consistent improvements when applied to automatic speech recognition and statistical machine translation tasks. A major difficulty with the continuous space neural network based approach remains the computational burden, which does not scale well to the massive corpora that are nowadays available. For this reason, the first contribution of this dissertation is the definition of a neural architecture based on a tree representation of the output vocabulary, namely Structured OUtput Layer (SOUL), which makes them well suited for large scale frameworks. The SOUL model combines the neural network approach with the class-based approach. It achieves significant improvements on both state-of-the-art large scale automatic speech recognition and statistical machine translations tasks. The second contribution is to provide several insightful analyses on their performances, their pros and cons, their induced word space representation. Finally, the third contribution is the successful adoption of the continuous space neural network into a machine translation framework. New translation models are proposed and reported to achieve significant improvements over state-of-the-art baseline systems.


Continuous Space Models with Neural Networks in Natural Language Processing Related Books

Continuous Space Models with Neural Networks in Natural Language Processing
Language: en
Pages: 0
Authors: Hai Son Le
Categories:
Type: BOOK - Published: 2012 - Publisher:

DOWNLOAD EBOOK

The purpose of language models is in general to capture and to model regularities of language, thereby capturing morphological, syntactical and distributional p
Deep Learning in Natural Language Processing
Language: en
Pages: 338
Authors: Li Deng
Categories: Computers
Type: BOOK - Published: 2018-05-23 - Publisher: Springer

DOWNLOAD EBOOK

In recent years, deep learning has fundamentally changed the landscapes of a number of areas in artificial intelligence, including speech, vision, natural langu
Neural Network Methods for Natural Language Processing
Language: en
Pages: 20
Authors: Yoav Goldberg
Categories: Computers
Type: BOOK - Published: 2022-06-01 - Publisher: Springer Nature

DOWNLOAD EBOOK

Neural networks are a family of powerful machine learning models. This book focuses on the application of neural network models to natural language data. The fi
Representation Learning for Natural Language Processing
Language: en
Pages: 319
Authors: Zhiyuan Liu
Categories: Computers
Type: BOOK - Published: 2020-07-03 - Publisher: Springer Nature

DOWNLOAD EBOOK

This open access book provides an overview of the recent advances in representation learning theory, algorithms and applications for natural language processing
Handbook of Natural Language Processing and Machine Translation
Language: en
Pages: 956
Authors: Joseph Olive
Categories: Computers
Type: BOOK - Published: 2011-03-02 - Publisher: Springer Science & Business Media

DOWNLOAD EBOOK

This comprehensive handbook, written by leading experts in the field, details the groundbreaking research conducted under the breakthrough GALE program--The Glo