elib
DLR-Header
DLR-Logo -> http://www.dlr.de
DLR Portal Home | Imprint | Privacy Policy | Contact | Deutsch
Fontsize: [-] Text [+]

Context based Text-generation using LSTM networks

Santhanam, Sivasurya (2018) Context based Text-generation using LSTM networks. Artificial Intelligence International Conference – A2IC 2018, 21- 23 Nov 2018, Barcelona, Spain. (Unpublished)

[img] PDF
45kB

Abstract

Long short-term memory(LSTM) units on sequence-based models are being used in translation, question-answering systems, classification tasks due to their capability of learning long-term dependencies. Text generation models, an application of LSTM models are recently popular due to their impressive results. LSTM models applied to natural languages are great in learning grammatically stable syntaxes. But the downside is, the system has no basic idea of the context and it generates text given a set of input words irrespective of the use-case. The proposed system trains the model to generate words given input words along with a context vector. Depending upon the use-case, the context vector is derived for a sentence or for a paragraph. A context vector could be a topic (from topic models) or the word having highest tf-idf weight in the sentence or a vector computed from word clusters. Thus, during the training phase, the same context vector is applied across the whole sentence for each window to predict successive words. Due to this structure, the model learns the relation between the context vector and the target word. During prediction, the user could provide keywords or topics to guide the system to generate words around a certain context. Apart from the syntactic structure in the current text-generation models, this proposed model will also provide semantic consistency. Based on the nature of computing context vectors, the model has been tried out with two variations (tf-idf and word clusters). The proposed system could be applied in question-answering systems to respond with a relevant topic. Also in Text-generation of stories with defined hints. The results should be evaluated manually on how semantically closer the text is generated given the context words.

Item URL in elib:https://elib.dlr.de/123892/
Document Type:Conference or Workshop Item (Speech)
Additional Information:Generation of a sequence of text based on a specific context using Long short-term memory neural networks is proposed in this work.
Title:Context based Text-generation using LSTM networks
Authors:
AuthorsInstitution or Email of AuthorsAuthors ORCID iD
Santhanam, SivasuryaSivasurya.Santhanam (at) dlr.dehttps://orcid.org/0000-0001-5117-8288
Date:November 2018
Refereed publication:No
Open Access:Yes
Gold Open Access:No
In SCOPUS:No
In ISI Web of Science:No
Status:Unpublished
Keywords:Natural language processing, Machine learning, Text generation, Neural networks
Event Title:Artificial Intelligence International Conference – A2IC 2018
Event Location:Barcelona, Spain
Event Type:international Conference
Event Dates:21- 23 Nov 2018
Organizer:PremC
HGF - Research field:Aeronautics, Space and Transport
HGF - Program:Space
HGF - Program Themes:Space Technology
DLR - Research area:Raumfahrt
DLR - Program:R SY - Technik für Raumfahrtsysteme
DLR - Research theme (Project):R - Vorhaben SISTEC
Location: Köln-Porz
Institutes and Institutions:Institut of Simulation and Software Technology > Distributed Systems and Component Software
Institut of Simulation and Software Technology
Deposited By: Santhanam, Sivasurya
Deposited On:12 Dec 2018 14:37
Last Modified:31 Jul 2019 20:21

Repository Staff Only: item control page

Browse
Search
Help & Contact
Information
electronic library is running on EPrints 3.3.12
Copyright © 2008-2017 German Aerospace Center (DLR). All rights reserved.