This paper proposes a deep learning model, StockGram, to automate financial communications via natural language generation. StockGram is a seq2seq model that generates short and coherent versions of financial news reports based on the client's point of interest from numerous pools of verified resources. The proposed model is developed to mitigate the pain points of advisors who invest numerous hours while
scanning through these news reports manually. StockGram leverages bi-directional LSTM cells that allows a recurrent system to make its prediction based on both past and future word sequences and hence predicts the next word in the sequence more precisely. The proposed model utilizes custom word-embeddings, GloVe, which incorporates global statistics to generate vector representations of news articles in an unsupervised manner and allows the model to converge faster. StockGram is evaluated based on the semantic closeness of the generated report to the provided prime words.
Sentence Validation by Statistical Language Modeling and Semantic RelationsEditor IJCATR
This paper deals with Sentence Validation - a sub-field of Natural Language Processing. It finds various applications in
different areas as it deals with understanding the natural language (English in most cases) and manipulating it. So the effort is on
understanding and extracting important information delivered to the computer and make possible efficient human computer
interaction. Sentence Validation is approached in two ways - by Statistical approach and Semantic approach. In both approaches
database is trained with the help of sample sentences of Brown corpus of NLTK. The statistical approach uses trigram technique based
on N-gram Markov Model and modified Kneser-Ney Smoothing to handle zero probabilities. As another testing on statistical basis,
tagging and chunking of the sentences having named entities is carried out using pre-defined grammar rules and semantic tree parsing,
and chunked off sentences are fed into another database, upon which testing is carried out. Finally, semantic analysis is carried out by
extracting entity relation pairs which are then tested. After the results of all three approaches is compiled, graphs are plotted and
variations are studied. Hence, a comparison of three different models is calculated and formulated. Graphs pertaining to the
probabilities of the three approaches are plotted, which clearly demarcate them and throw light on the findings of the project.
EMPLOYING PIVOT LANGUAGE TECHNIQUE THROUGH STATISTICAL AND NEURAL MACHINE TRA...ijnlc
The quality of Neural Machine Translation (NMT) systems like Statistical Machine Translation (SMT) systems, heavily depends on the size of training data set, while for some pairs of languages, high-quality parallel data are poor resources. In order to respond to this low-resourced training data bottleneck reality, we employ the pivoting approach in both neural MT and statistical MT frameworks. During our experiments on the Persian-Spanish, taken as an under-resourced translation task, we discovered that, the aforementioned method, in both frameworks, significantly improves the translation quality in comparison to the standard direct translation approach.
Our project is about guessing the correct missing
word in a given sentence. To find of guess the missing word
we have two main methods one of them statistical language
modeling, while the other is neural language models.
Statistical language modeling depend on the frequency of the
relation between words and here we use Markov chain. Since
neural language models uses artificial neural networks which
uses deep learning, here we use BERT which is the state of art
in language modeling provided by google.
TEXT ADVERTISEMENTS ANALYSIS USING CONVOLUTIONAL NEURAL NETWORKSijdms
In this paper, we describe the developed model of the Convolutional Neural Networks CNN to a
classification of advertisements. The developed method has been tested on both texts (Arabic and Slovak
texts).The advertisements are chosen on a classified advertisements websites as short texts. We evolved a
modified model of the CNN, we have implemented it and developed next modifications. We studied their
influence on the performing activity of the proposed network. The result is a functional model of the
network and its implementation in Java and Python. And analysis of model results using different
parameters for the network and input data. The results on experiments data show that the developed model
of CNN is useful in the domains of Arabic and Slovak short texts, mainly for some classification of
advertisements. This paper gives complete guidelines for authors submitting papers for the AIRCC
Journals.
GENDER AND AUTHORSHIP CATEGORISATION OF ARABIC TEXT FROM TWITTER USING PPMijcsit
In this paper we present gender and authorship categorisation using the Prediction by Partial Matching(PPM) compression scheme for text from Twitter written in Arabic. The PPMD variant of the compression scheme with different orders was used to perform the categorisation. We also applied different machine learning algorithms such as Multinational Naïve Bayes (MNB), K-Nearest Neighbours (KNN), and an
implementation of Support Vector Machine (LIBSVM), applying the same processing steps for all the algorithms. PPMD shows significantly better accuracy in comparison to all the other machine learning algorithms, with order 11 PPMD working best, achieving 90 % and 96% accuracy for gender and
authorship respectively.
presentation from my thesis defense on text summarization, discusses already existing state of art models along with efficiency of AMR or Abstract Meaning Representation for text summarization, we see how we can use AMRs with seq2seq models. We also discuss other techniques such as BPE or Byte Pair Encoding and its effectiveness for the task. Also we see how data augmentation with POS tags and AMRs effect the summarization with s2s learning.
Automatic text summarization is the process of reducing the text content and retaining the
important points of the document. Generally, there are two approaches for automatic text summarization:
Extractive and Abstractive. The process of extractive based text summarization can be divided into two
phases: pre-processing and processing. In this paper, we discuss some of the extractive based text
summarization approaches used by researchers. We also provide the features for extractive based text
summarization process. We also present the available linguistic preprocessing tools with their features,
which are used for automatic text summarization. The tools and parameters useful for evaluating the
generated summary are also discussed in this paper. Moreover, we explain our proposed lexical chain
analysis approach, with sample generated lexical chains, for extractive based automatic text summarization.
We also provide the evaluation results of our system generated summary. The proposed lexical chain
analysis approach can be used to solve different text mining problems like topic classification, sentiment
analysis, and summarization.
Sentence Validation by Statistical Language Modeling and Semantic RelationsEditor IJCATR
This paper deals with Sentence Validation - a sub-field of Natural Language Processing. It finds various applications in
different areas as it deals with understanding the natural language (English in most cases) and manipulating it. So the effort is on
understanding and extracting important information delivered to the computer and make possible efficient human computer
interaction. Sentence Validation is approached in two ways - by Statistical approach and Semantic approach. In both approaches
database is trained with the help of sample sentences of Brown corpus of NLTK. The statistical approach uses trigram technique based
on N-gram Markov Model and modified Kneser-Ney Smoothing to handle zero probabilities. As another testing on statistical basis,
tagging and chunking of the sentences having named entities is carried out using pre-defined grammar rules and semantic tree parsing,
and chunked off sentences are fed into another database, upon which testing is carried out. Finally, semantic analysis is carried out by
extracting entity relation pairs which are then tested. After the results of all three approaches is compiled, graphs are plotted and
variations are studied. Hence, a comparison of three different models is calculated and formulated. Graphs pertaining to the
probabilities of the three approaches are plotted, which clearly demarcate them and throw light on the findings of the project.
EMPLOYING PIVOT LANGUAGE TECHNIQUE THROUGH STATISTICAL AND NEURAL MACHINE TRA...ijnlc
The quality of Neural Machine Translation (NMT) systems like Statistical Machine Translation (SMT) systems, heavily depends on the size of training data set, while for some pairs of languages, high-quality parallel data are poor resources. In order to respond to this low-resourced training data bottleneck reality, we employ the pivoting approach in both neural MT and statistical MT frameworks. During our experiments on the Persian-Spanish, taken as an under-resourced translation task, we discovered that, the aforementioned method, in both frameworks, significantly improves the translation quality in comparison to the standard direct translation approach.
Our project is about guessing the correct missing
word in a given sentence. To find of guess the missing word
we have two main methods one of them statistical language
modeling, while the other is neural language models.
Statistical language modeling depend on the frequency of the
relation between words and here we use Markov chain. Since
neural language models uses artificial neural networks which
uses deep learning, here we use BERT which is the state of art
in language modeling provided by google.
TEXT ADVERTISEMENTS ANALYSIS USING CONVOLUTIONAL NEURAL NETWORKSijdms
In this paper, we describe the developed model of the Convolutional Neural Networks CNN to a
classification of advertisements. The developed method has been tested on both texts (Arabic and Slovak
texts).The advertisements are chosen on a classified advertisements websites as short texts. We evolved a
modified model of the CNN, we have implemented it and developed next modifications. We studied their
influence on the performing activity of the proposed network. The result is a functional model of the
network and its implementation in Java and Python. And analysis of model results using different
parameters for the network and input data. The results on experiments data show that the developed model
of CNN is useful in the domains of Arabic and Slovak short texts, mainly for some classification of
advertisements. This paper gives complete guidelines for authors submitting papers for the AIRCC
Journals.
GENDER AND AUTHORSHIP CATEGORISATION OF ARABIC TEXT FROM TWITTER USING PPMijcsit
In this paper we present gender and authorship categorisation using the Prediction by Partial Matching(PPM) compression scheme for text from Twitter written in Arabic. The PPMD variant of the compression scheme with different orders was used to perform the categorisation. We also applied different machine learning algorithms such as Multinational Naïve Bayes (MNB), K-Nearest Neighbours (KNN), and an
implementation of Support Vector Machine (LIBSVM), applying the same processing steps for all the algorithms. PPMD shows significantly better accuracy in comparison to all the other machine learning algorithms, with order 11 PPMD working best, achieving 90 % and 96% accuracy for gender and
authorship respectively.
presentation from my thesis defense on text summarization, discusses already existing state of art models along with efficiency of AMR or Abstract Meaning Representation for text summarization, we see how we can use AMRs with seq2seq models. We also discuss other techniques such as BPE or Byte Pair Encoding and its effectiveness for the task. Also we see how data augmentation with POS tags and AMRs effect the summarization with s2s learning.
Automatic text summarization is the process of reducing the text content and retaining the
important points of the document. Generally, there are two approaches for automatic text summarization:
Extractive and Abstractive. The process of extractive based text summarization can be divided into two
phases: pre-processing and processing. In this paper, we discuss some of the extractive based text
summarization approaches used by researchers. We also provide the features for extractive based text
summarization process. We also present the available linguistic preprocessing tools with their features,
which are used for automatic text summarization. The tools and parameters useful for evaluating the
generated summary are also discussed in this paper. Moreover, we explain our proposed lexical chain
analysis approach, with sample generated lexical chains, for extractive based automatic text summarization.
We also provide the evaluation results of our system generated summary. The proposed lexical chain
analysis approach can be used to solve different text mining problems like topic classification, sentiment
analysis, and summarization.
Increasing interpreting needs a more objective and automatic measurement. We hold a basic idea that 'translating means translating meaning' in that we can assessment interpretation quality by comparing the
meaning of the interpreting output with the source input. That is, a translation unit of a 'chunk' named
Frame which comes from frame semantics and its components named Frame Elements (FEs) which comes
from Frame Net are proposed to explore their matching rate between target and source texts. A case study in this paper verifies the usability of semi-automatic graded semantic-scoring measurement for human
simultaneous interpreting and shows how to use frame and FE matches to score. Experiments results show that the semantic-scoring metrics have a significantly correlation coefficient with human judgment.
An on-going project on Natural Language Processing (using Python and the NLTK toolkit), which focuses on the extraction of sentiment from a Question and its title on www.stackoverflow.com and determining the polarity.Based on the above findings, it is verified whether the rules and guidelines imposed by the SO community on the users are strictly followed or not.
WARRANTS GENERATIONS USING A LANGUAGE MODEL AND A MULTI-AGENT SYSTEMijnlc
Each argument begins with a conclusion, which is followed by one or more premises supporting the
conclusion. The warrant is a critical component of Toulmin's argument model; it explains why the premises
support the claim. Despite its critical role in establishing the claim's veracity, it is frequently omitted or left
implicit, leaving readers to infer. We consider the problem of producing more diverse and high-quality
warrants in response to a claim and evidence. To begin, we employ BART [1] as a conditional sequence tosequence language model to guide the output generation process. On the ARCT dataset [2], we fine-tune
the BART model. Second, we propose the Multi-Agent Network for Warrant Generation as a model for
producing more diverse and high-quality warrants by combining Reinforcement Learning (RL) and
Generative Adversarial Networks (GAN) with the mechanism of mutual awareness of agents. In terms of
warrant generation, our model generates a greater variety of warrants than other baseline models. The
experimental results validate the effectiveness of our proposed hybrid model for generating warrants.
Following are the questions which I tried to answer in this ppt
What is text summarization.
What is automatic text summarization?
How it has evolved over the time?
What are different methods?
How deep learning is used for text summarization?
business application
in first few slides extractive summarization is explained, with pro and cons in next section abstractive on is explained.
In the last section business application of each one is highlighted
Improving Neural Abstractive Text Summarization with Prior KnowledgeGaetano Rossiello, PhD
Abstractive text summarization is a complex task whose goal is to generate a concise version of a text without necessarily reusing the sentences from the original source, but still preserving the meaning and the key contents. In this position paper we address this issue by modeling the problem as a sequence to sequence learning and exploiting Recurrent Neural Networks (RNN). Moreover, we discuss the idea of combining RNNs and probabilistic models in a unified way in order to incorporate prior knowledge, such as linguistic features. We believe that this approach can obtain better performance than the state-of-the-art models for generating well-formed summaries.
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESkevig
Distributed language representation has become the most widely used technique for language representation in various natural language processing tasks. Most of the natural language processing models that are based on deep learning techniques use already pre-trained distributed word representations, commonly called word embeddings. Determining the most qualitative word embeddings is of crucial importance for such models. However, selecting the appropriate word embeddings is a perplexing task since the projected embedding space is not intuitive to humans.In this paper, we explore different approaches for creating distributed word representations. We perform an intrinsic evaluation of several state-of-the-art word embedding methods. Their performance on capturing word similarities is analysed with existing benchmark datasets for word pairs similarities. The research in this paper conducts a correlation analysis between ground truth word similarities and similarities obtained by different word embedding methods.
Prediction of Answer Keywords using Char-RNNIJECEIAES
Generating sequences of characters using a Recurrent Neural Network (RNN) is a tried and tested method for creating unique and context aware words, and is fundamental in Natural Language Processing tasks. These type of Neural Networks can also be used a question-answering system. The main drawback of most of these systems is that they work from a factoid database of information, and when queried about new and current information, the responses are usually bleak. In this paper, the author proposes a novel approach to finding answer keywords from a given body of news text or headline, based on the query that was asked, where the query would be of the nature of current affairs or recent news, with the use of Gated Recurrent Unit (GRU) variant of RNNs. Thus, this ensures that the answers provided are relevant to the content of query that was put forth.
The project is developed as a part of IRE course work @IIIT-Hyderabad.
Team members:
Aishwary Gupta (201302216)
B Prabhakar (201505618)
Sahil Swami (201302071)
Links:
https://github.com/prabhakar9885/Text-Summarization
http://prabhakar9885.github.io/Text-Summarization/
https://www.youtube.com/playlist?list=PLtBx4kn8YjxJUGsszlev52fC1Jn07HkUw
http://www.slideshare.net/prabhakar9885/text-summarization-60954970
https://www.dropbox.com/sh/uaxc2cpyy3pi97z/AADkuZ_24OHVi3PJmEAziLxha?dl=0
Semantic similarity and semantic relatedness
measure in particular is very important in the current scenario
due to the huge demand for natural language processing based
applications such as chatbots and information retrieval systems
such as knowledge base based FAQ systems. Current approaches
generally use similarity measures which does not use the context
sensitive relationships between the words. This leads to erroneous
similarity predictions and is not of much use in real life
applications. This work proposes a novel approach that gives an
accurate relatedness measure of any two words in a sentence by
taking their context into consideration. This context correction
results in a more accurate similarity prediction which results in
higher accuracy of information retrieval systems.
EXTRACTIVE SUMMARIZATION WITH VERY DEEP PRETRAINED LANGUAGE MODELijaia
Recent development of generative pretrained language models has been proven very successful on a wide
range of NLP tasks, such as text classification, question answering, textual entailment and so on. In this
work, we present a two-phase encoder decoder architecture based on Bidirectional Encoding
Representation from Transformers(BERT) for extractive summarization task. We evaluated our model by
both automatic metrics and human annotators, and demonstrated that the architecture achieves the stateof-the-art comparable result on large scale corpus – ‘CNN/Daily Mail1
As the best of our knowledge’, this
is the first work that applies BERT based architecture to a text summarization task and achieved the stateof-the-art comparable result.
An Approach To Automatic Text Summarization Using Simplified Lesk Algorithm A...ijctcm
Text Summarization is a way to produce a text, which contains the significant portion of information of the
original text(s). Different methodologies are developed till now depending upon several parameters to find
the summary as the position, format and type of the sentences in an input text, formats of different words,
frequency of a particular word in a text etc. But according to different languages and input sources, these
parameters are varied. As result the performance of the algorithm is greatly affected. The proposed
approach summarizes a text without depending upon those parameters. Here, the relevance of the
sentences within the text is derived by Simplified Lesk algorithm and WordNet, an online dictionary. This
approach is not only independent of the format of the text and position of a sentence in a text, as the
sentences are arranged at first according to their relevance before the summarization process, the
percentage of summarization can be varied according to needs. The proposed approach gives around 80%
accurate results on 50% summarization of the original text with respect to the manually summarized result,
performed on 50 different types and lengths of texts. We have achieved satisfactory results even upto 25%
summarization of the original text.
Extractive Summarization with Very Deep Pretrained Language Modelgerogepatton
Recent development of generative pretrained language models has been proven very successful on a wide range of NLP tasks, such as text classification, question answering, textual entailment and so on.In this work, we present a two-phase encoder decoder architecture based on Bidirectional Encoding Representation from Transformers(BERT) for extractive summarization task. We evaluated our model by both automatic metrics and human annotators, and demonstrated that the architecture achieves the state-of-the-art comparable result on large scale corpus - CNN/Daily Mail1. As the best of our knowledge, this is the first work that applies BERT based architecture to a text summarization task and achieved the state-of-the-art comparable result.
This paper presents machine translation based on machine learning, which learns the semantically
correct corpus. The machine learning process based on Quantum Neural Network (QNN) is used to
recognizing the corpus pattern in realistic way. It translates on the basis of knowledge gained during
learning by entering pair of sentences from source to target language. By taking help of this training data
it translates the given text. own text.The paper consist study of a machine translation system which
converts source language to target language using quantum neural network. Rather than comparing
words semantically QNN compares numerical tags which is faster and accurate. In this tagger tags the
part of sentences discretely which helps to map bilingual sentences.
BIDIRECTIONAL LONG SHORT-TERM MEMORY (BILSTM)WITH CONDITIONAL RANDOM FIELDS (...ijnlc
This study investigates the effectiveness of Knowledge Named Entity Recognition in Online Judges (OJs). OJs are lacking in the classification of topics and limited to the IDs only. Therefore a lot of time is consumed in finding programming problems more specifically in knowledge entities.A Bidirectional Long Short-Term Memory (BiLSTM) with Conditional Random Fields (CRF) model is applied for the recognition of knowledge named entities existing in the solution reports.For the test run, more than 2000 solution reports are crawled from the Online Judges and processed for the model output. The stability of the model is
also assessed with the higher F1 value. The results obtained through the proposed BiLSTM-CRF model are more effectual (F1: 98.96%) and efficient in lead-time.
Analysis of the evolution of advanced transformer-based language models: Expe...IAESIJAI
Opinion mining, also known as sentiment analysis, is a subfield of natural language processing (NLP) that focuses on identifying and extracting subjective information in textual material. This can include determining the overall sentiment of a piece of text (e.g., positive or negative), as well as identifying specific emotions or opinions expressed in the text, that involves the use of advanced machine and deep learning techniques. Recently, transformer-based language models make this task of human emotion analysis intuitive, thanks to the attention mechanism and parallel computation. These advantages make such models very powerful on linguistic tasks, unlike recurrent neural networks that spend a lot of time on sequential processing, making them prone to fail when it comes to processing long text. The scope of our paper aims to study the behaviour of the cutting-edge Transformer-based language models on opinion mining and provide a high-level comparison between them to highlight their key particularities. Additionally, our comparative study shows leads and paves the way for production engineers regarding the approach to focus on and is useful for researchers as it provides guidelines for future research subjects.
Crafting Your Customized Legal Mastery: A Guide to Building Your Private LLMChristopherTHyatt
Create a tailored legal education with a private LLM. Identify your specialization, research courses from reputable institutions, and leverage online platforms for flexibility. Craft a unique curriculum combining law with interdisciplinary studies, enhancing your expertise. Network with professionals, balance theory with practical experience, and stay updated on legal trends. Build a personalized learning journey to unlock your full potential in the legal landscape.
Identifying Hot Topic Trends in Streaming Text Data Using News Sequential Evo...Shakas Technologies
Identifying Hot Topic Trends in Streaming Text Data Using News Sequential Evolution Model Based on Distributed Representations.
Shakas Technologies ( Galaxy of Knowledge)
#11/A 2nd East Main Road,
Gandhi Nagar,
Vellore - 632006.
Mobile : +91-9500218218 / 8220150373| land line- 0416- 3552723
Shakas Training & Development | Shakas Sales & Services | Shakas Educational Trust|IEEE projects | Research & Development | Journal Publication |
Email : info@shakastech.com | shakastech@gmail.com |
website: www.shakastech.com
Facebook: https://www.facebook.com/pages/Shakas-Technologies
BIDIRECTIONAL LONG SHORT-TERM MEMORY (BILSTM)WITH CONDITIONAL RANDOM FIELDS (...kevig
This study investigates the effectiveness of Knowledge Named Entity Recognition in Online Judges (OJs). OJs are lacking in the classification of topics and limited to the IDs only. Therefore a lot of time is consumed in finding programming problems more specifically in knowledge entities.A Bidirectional Long Short-Term Memory (BiLSTM) with Conditional Random Fields (CRF) model is applied for the recognition of knowledge named entities existing in the solution reports.For the test run, more than 2000 solution reports are crawled from the Online Judges and processed for the model output. The stability of the model is also assessed with the higher F1 value. The results obtained through the proposed BiLSTM-CRF model are more effectual (F1: 98.96%) and efficient in lead-time.
Increasing interpreting needs a more objective and automatic measurement. We hold a basic idea that 'translating means translating meaning' in that we can assessment interpretation quality by comparing the
meaning of the interpreting output with the source input. That is, a translation unit of a 'chunk' named
Frame which comes from frame semantics and its components named Frame Elements (FEs) which comes
from Frame Net are proposed to explore their matching rate between target and source texts. A case study in this paper verifies the usability of semi-automatic graded semantic-scoring measurement for human
simultaneous interpreting and shows how to use frame and FE matches to score. Experiments results show that the semantic-scoring metrics have a significantly correlation coefficient with human judgment.
An on-going project on Natural Language Processing (using Python and the NLTK toolkit), which focuses on the extraction of sentiment from a Question and its title on www.stackoverflow.com and determining the polarity.Based on the above findings, it is verified whether the rules and guidelines imposed by the SO community on the users are strictly followed or not.
WARRANTS GENERATIONS USING A LANGUAGE MODEL AND A MULTI-AGENT SYSTEMijnlc
Each argument begins with a conclusion, which is followed by one or more premises supporting the
conclusion. The warrant is a critical component of Toulmin's argument model; it explains why the premises
support the claim. Despite its critical role in establishing the claim's veracity, it is frequently omitted or left
implicit, leaving readers to infer. We consider the problem of producing more diverse and high-quality
warrants in response to a claim and evidence. To begin, we employ BART [1] as a conditional sequence tosequence language model to guide the output generation process. On the ARCT dataset [2], we fine-tune
the BART model. Second, we propose the Multi-Agent Network for Warrant Generation as a model for
producing more diverse and high-quality warrants by combining Reinforcement Learning (RL) and
Generative Adversarial Networks (GAN) with the mechanism of mutual awareness of agents. In terms of
warrant generation, our model generates a greater variety of warrants than other baseline models. The
experimental results validate the effectiveness of our proposed hybrid model for generating warrants.
Following are the questions which I tried to answer in this ppt
What is text summarization.
What is automatic text summarization?
How it has evolved over the time?
What are different methods?
How deep learning is used for text summarization?
business application
in first few slides extractive summarization is explained, with pro and cons in next section abstractive on is explained.
In the last section business application of each one is highlighted
Improving Neural Abstractive Text Summarization with Prior KnowledgeGaetano Rossiello, PhD
Abstractive text summarization is a complex task whose goal is to generate a concise version of a text without necessarily reusing the sentences from the original source, but still preserving the meaning and the key contents. In this position paper we address this issue by modeling the problem as a sequence to sequence learning and exploiting Recurrent Neural Networks (RNN). Moreover, we discuss the idea of combining RNNs and probabilistic models in a unified way in order to incorporate prior knowledge, such as linguistic features. We believe that this approach can obtain better performance than the state-of-the-art models for generating well-formed summaries.
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESkevig
Distributed language representation has become the most widely used technique for language representation in various natural language processing tasks. Most of the natural language processing models that are based on deep learning techniques use already pre-trained distributed word representations, commonly called word embeddings. Determining the most qualitative word embeddings is of crucial importance for such models. However, selecting the appropriate word embeddings is a perplexing task since the projected embedding space is not intuitive to humans.In this paper, we explore different approaches for creating distributed word representations. We perform an intrinsic evaluation of several state-of-the-art word embedding methods. Their performance on capturing word similarities is analysed with existing benchmark datasets for word pairs similarities. The research in this paper conducts a correlation analysis between ground truth word similarities and similarities obtained by different word embedding methods.
Prediction of Answer Keywords using Char-RNNIJECEIAES
Generating sequences of characters using a Recurrent Neural Network (RNN) is a tried and tested method for creating unique and context aware words, and is fundamental in Natural Language Processing tasks. These type of Neural Networks can also be used a question-answering system. The main drawback of most of these systems is that they work from a factoid database of information, and when queried about new and current information, the responses are usually bleak. In this paper, the author proposes a novel approach to finding answer keywords from a given body of news text or headline, based on the query that was asked, where the query would be of the nature of current affairs or recent news, with the use of Gated Recurrent Unit (GRU) variant of RNNs. Thus, this ensures that the answers provided are relevant to the content of query that was put forth.
The project is developed as a part of IRE course work @IIIT-Hyderabad.
Team members:
Aishwary Gupta (201302216)
B Prabhakar (201505618)
Sahil Swami (201302071)
Links:
https://github.com/prabhakar9885/Text-Summarization
http://prabhakar9885.github.io/Text-Summarization/
https://www.youtube.com/playlist?list=PLtBx4kn8YjxJUGsszlev52fC1Jn07HkUw
http://www.slideshare.net/prabhakar9885/text-summarization-60954970
https://www.dropbox.com/sh/uaxc2cpyy3pi97z/AADkuZ_24OHVi3PJmEAziLxha?dl=0
Semantic similarity and semantic relatedness
measure in particular is very important in the current scenario
due to the huge demand for natural language processing based
applications such as chatbots and information retrieval systems
such as knowledge base based FAQ systems. Current approaches
generally use similarity measures which does not use the context
sensitive relationships between the words. This leads to erroneous
similarity predictions and is not of much use in real life
applications. This work proposes a novel approach that gives an
accurate relatedness measure of any two words in a sentence by
taking their context into consideration. This context correction
results in a more accurate similarity prediction which results in
higher accuracy of information retrieval systems.
EXTRACTIVE SUMMARIZATION WITH VERY DEEP PRETRAINED LANGUAGE MODELijaia
Recent development of generative pretrained language models has been proven very successful on a wide
range of NLP tasks, such as text classification, question answering, textual entailment and so on. In this
work, we present a two-phase encoder decoder architecture based on Bidirectional Encoding
Representation from Transformers(BERT) for extractive summarization task. We evaluated our model by
both automatic metrics and human annotators, and demonstrated that the architecture achieves the stateof-the-art comparable result on large scale corpus – ‘CNN/Daily Mail1
As the best of our knowledge’, this
is the first work that applies BERT based architecture to a text summarization task and achieved the stateof-the-art comparable result.
An Approach To Automatic Text Summarization Using Simplified Lesk Algorithm A...ijctcm
Text Summarization is a way to produce a text, which contains the significant portion of information of the
original text(s). Different methodologies are developed till now depending upon several parameters to find
the summary as the position, format and type of the sentences in an input text, formats of different words,
frequency of a particular word in a text etc. But according to different languages and input sources, these
parameters are varied. As result the performance of the algorithm is greatly affected. The proposed
approach summarizes a text without depending upon those parameters. Here, the relevance of the
sentences within the text is derived by Simplified Lesk algorithm and WordNet, an online dictionary. This
approach is not only independent of the format of the text and position of a sentence in a text, as the
sentences are arranged at first according to their relevance before the summarization process, the
percentage of summarization can be varied according to needs. The proposed approach gives around 80%
accurate results on 50% summarization of the original text with respect to the manually summarized result,
performed on 50 different types and lengths of texts. We have achieved satisfactory results even upto 25%
summarization of the original text.
Extractive Summarization with Very Deep Pretrained Language Modelgerogepatton
Recent development of generative pretrained language models has been proven very successful on a wide range of NLP tasks, such as text classification, question answering, textual entailment and so on.In this work, we present a two-phase encoder decoder architecture based on Bidirectional Encoding Representation from Transformers(BERT) for extractive summarization task. We evaluated our model by both automatic metrics and human annotators, and demonstrated that the architecture achieves the state-of-the-art comparable result on large scale corpus - CNN/Daily Mail1. As the best of our knowledge, this is the first work that applies BERT based architecture to a text summarization task and achieved the state-of-the-art comparable result.
This paper presents machine translation based on machine learning, which learns the semantically
correct corpus. The machine learning process based on Quantum Neural Network (QNN) is used to
recognizing the corpus pattern in realistic way. It translates on the basis of knowledge gained during
learning by entering pair of sentences from source to target language. By taking help of this training data
it translates the given text. own text.The paper consist study of a machine translation system which
converts source language to target language using quantum neural network. Rather than comparing
words semantically QNN compares numerical tags which is faster and accurate. In this tagger tags the
part of sentences discretely which helps to map bilingual sentences.
BIDIRECTIONAL LONG SHORT-TERM MEMORY (BILSTM)WITH CONDITIONAL RANDOM FIELDS (...ijnlc
This study investigates the effectiveness of Knowledge Named Entity Recognition in Online Judges (OJs). OJs are lacking in the classification of topics and limited to the IDs only. Therefore a lot of time is consumed in finding programming problems more specifically in knowledge entities.A Bidirectional Long Short-Term Memory (BiLSTM) with Conditional Random Fields (CRF) model is applied for the recognition of knowledge named entities existing in the solution reports.For the test run, more than 2000 solution reports are crawled from the Online Judges and processed for the model output. The stability of the model is
also assessed with the higher F1 value. The results obtained through the proposed BiLSTM-CRF model are more effectual (F1: 98.96%) and efficient in lead-time.
Analysis of the evolution of advanced transformer-based language models: Expe...IAESIJAI
Opinion mining, also known as sentiment analysis, is a subfield of natural language processing (NLP) that focuses on identifying and extracting subjective information in textual material. This can include determining the overall sentiment of a piece of text (e.g., positive or negative), as well as identifying specific emotions or opinions expressed in the text, that involves the use of advanced machine and deep learning techniques. Recently, transformer-based language models make this task of human emotion analysis intuitive, thanks to the attention mechanism and parallel computation. These advantages make such models very powerful on linguistic tasks, unlike recurrent neural networks that spend a lot of time on sequential processing, making them prone to fail when it comes to processing long text. The scope of our paper aims to study the behaviour of the cutting-edge Transformer-based language models on opinion mining and provide a high-level comparison between them to highlight their key particularities. Additionally, our comparative study shows leads and paves the way for production engineers regarding the approach to focus on and is useful for researchers as it provides guidelines for future research subjects.
Crafting Your Customized Legal Mastery: A Guide to Building Your Private LLMChristopherTHyatt
Create a tailored legal education with a private LLM. Identify your specialization, research courses from reputable institutions, and leverage online platforms for flexibility. Craft a unique curriculum combining law with interdisciplinary studies, enhancing your expertise. Network with professionals, balance theory with practical experience, and stay updated on legal trends. Build a personalized learning journey to unlock your full potential in the legal landscape.
Identifying Hot Topic Trends in Streaming Text Data Using News Sequential Evo...Shakas Technologies
Identifying Hot Topic Trends in Streaming Text Data Using News Sequential Evolution Model Based on Distributed Representations.
Shakas Technologies ( Galaxy of Knowledge)
#11/A 2nd East Main Road,
Gandhi Nagar,
Vellore - 632006.
Mobile : +91-9500218218 / 8220150373| land line- 0416- 3552723
Shakas Training & Development | Shakas Sales & Services | Shakas Educational Trust|IEEE projects | Research & Development | Journal Publication |
Email : info@shakastech.com | shakastech@gmail.com |
website: www.shakastech.com
Facebook: https://www.facebook.com/pages/Shakas-Technologies
BIDIRECTIONAL LONG SHORT-TERM MEMORY (BILSTM)WITH CONDITIONAL RANDOM FIELDS (...kevig
This study investigates the effectiveness of Knowledge Named Entity Recognition in Online Judges (OJs). OJs are lacking in the classification of topics and limited to the IDs only. Therefore a lot of time is consumed in finding programming problems more specifically in knowledge entities.A Bidirectional Long Short-Term Memory (BiLSTM) with Conditional Random Fields (CRF) model is applied for the recognition of knowledge named entities existing in the solution reports.For the test run, more than 2000 solution reports are crawled from the Online Judges and processed for the model output. The stability of the model is also assessed with the higher F1 value. The results obtained through the proposed BiLSTM-CRF model are more effectual (F1: 98.96%) and efficient in lead-time.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESkevig
Distributed language representation has become the most widely used technique for language representation in various natural language processing tasks. Most of the natural language processing models that are based on deep learning techniques use already pre-trained distributed word representations, commonly called word embeddings. Determining the most qualitative word embeddings is of crucial importance for such models. However, selecting the appropriate word embeddings is a perplexing task since the projected embedding space is not intuitive to humans. In this paper, we explore different approaches for creating distributed word representations. We perform an intrinsic evaluation of several state-of-the-art word embedding methods. Their performance on capturing word similarities is analysed with existing benchmark datasets for word pairs similarities. The research in this paper conducts a correlation analysis between ground truth word similarities and similarities obtained by different word embedding methods.
Chunking means splitting the sentences into tokens and then grouping them in a meaningful way. When it comes to high-performance chunking systems, transformer models have proved to be the state of the art benchmarks. To perform chunking as a task it requires a large-scale high quality annotated corpus where each token is attached with a particular tag similar as that of Named Entity Recognition Tasks. Later these tags are used in conjunction with pointer frameworks to find the final chunk. To solve this for a specific domain problem, it becomes a highly costly affair in terms of time and resources to manually annotate and produce a large-high-quality training set. When the domain is specific and diverse, then cold starting becomes even more difficult because of the expected large number of manually annotated queries to cover all aspects. To overcome the problem, we applied a grammar-based text generation mechanism where instead of annotating a sentence we annotate using grammar templates. We defined various templates corresponding to different grammar rules. To create a sentence we used these templates along with the rules where symbol or terminal values were chosen from the domain data catalog. It helped us to create a large number of annotated queries. These annotated queries were used for training the machine learning model using an ensemble transformer-based deep neural network model [24.] We found that grammar-based annotation was useful to solve domain-based chunks in input query sentences without any manual annotation where it was found to achieve a classification F1 score of 96.97% in classifying the tokens for the out of template queries.
Chunking means splitting the sentences into tokens and then grouping them in a meaningful way. When it
comes to high-performance chunking systems, transformer models have proved to be the state of the art
benchmarks. To perform chunking as a task it requires a large-scale high quality annotated corpus where
each token is attached with a particular tag similar as that of Named Entity Recognition Tasks. Later these tags are used in conjunction with pointer frameworks to find the final chunk. To solve this for a specific domain problem, it becomes a highly costly affair in terms of time and resources to manually annotate and produce a large-high-quality training set. When the domain is specific and diverse, then cold starting becomes even more difficult because of the expected large number of manually annotated queries to cover all aspects. To overcome the problem, we applied a grammar-based text generation mechanism where instead of annotating a sentence we annotate using grammar templates. We defined various templates corresponding to different grammar rules. To create a sentence we used these templates along with the rules where symbol or terminal values were chosen from the domain data catalog. It helped us to create a large number of annotated queries. These annotated queries were used for training the machine learning model using an ensemble transformer-based deep neural network model [24.] We found that grammar-based annotation was useful to solve domain-based chunks in input query sentences without any manual annotation where it was found to achieve a classification F1 score of 96.97% in classifying the tokens for the out of template queries.
Chunking means splitting the sentences into tokens and then grouping them in a meaningful way. When it comes to high-performance chunking systems, transformer models have proved to be the state of the art benchmarks. To perform chunking as a task it requires a large-scale high quality annotated corpus where each token is attached with a particular tag similar as that of Named Entity Recognition Tasks. Later these tags are used in conjunction with pointer frameworks to find the final chunk. To solve this for a specific domain problem, it becomes a highly costly affair in terms of time and resources to manually annotate and produce a large-high-quality training set. When the domain is specific and diverse, then cold starting becomes even more difficult because of the expected large number of manually annotated queries to cover all aspects. To overcome the problem, we applied a grammar-based text generation mechanism where instead of annotating a sentence we annotate using grammar templates. We defined various templates corresponding to different grammar rules. To create a sentence we used these templates along with the rules where symbol or terminal values were chosen from the domain data catalog. It helped us to create a large number of annotated queries. These annotated queries were used for training the machine learning model using an ensemble transformer-based deep neural network model [24.] We found that grammar-based annotation was useful to solve domain-based chunks in input query sentences without any manual annotation where it was found to achieve a classification F1 score of 96.97% in classifying the tokens for the out of template queries.
Sentimental analysis is a context based mining of text, which extracts and identify subjective information from a text or sentence provided. Here the main concept is extracting the sentiment of the text using machine learning techniques such as LSTM Long short term memory . This text classification method analyses the incoming text and determines whether the underlined emotion is positive or negative along with probability associated with that positive or negative statements. Probability depicts the strength of a positive or negative statement, if the probability is close to zero, it implies that the sentiment is strongly negative and if probability is close to1, it means that the statement is strongly positive. Here a web application is created to deploy this model using a Python based micro framework called flask. Many other methods, such as RNN and CNN, are inefficient when compared to LSTM. Dirash A R | Dr. S K Manju Bargavi "LSTM Based Sentiment Analysis" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-4 , June 2021, URL: https://www.ijtsrd.compapers/ijtsrd42345.pdf Paper URL: https://www.ijtsrd.comcomputer-science/data-processing/42345/lstm-based-sentiment-analysis/dirash-a-r
8 efficient multi-document summary generation using neural networkINFOGAIN PUBLICATION
From last few years online information is growing tremendously on World Wide Web or on user’s desktops and thus online information gains much more attention in the field of automatic text summarization. Text mining has become a significant research field as it produces valuable data from unstructured and large amount of texts. Summarization systems provide the possibility of searching the important keywords of the texts and so the consumer will expend less time on reading the whole document. Main objective of summarization system is to generate a new form which expresses the key meaning of the contained text. This paper study on various existing techniques with needs of novel Multi-Document summarization schemes. This paper is motivated by arising need to provide high quality summary in very short period of time. In proposed system, user can quickly and easily access correctly-developed summaries which expresses the key meaning of the contained text. The primary focus of this paper lies with thef_β-optimal merge function, a function recently presented here, that uses the weighted harmonic mean to discover a harmony in the middle of precision and recall. Proposed system utilizes Bisect K-means clustering to improve the time and Neural Networks to improve the accuracy of summary generated by NEWSUM algorithm.
Masterclass: Natural Language Processing in Trading with Terry Benzschawel & ...QuantInsti
Terry Benzschawel (Founder and Principal at Benzschawel Scientific, LLC) and Ishan Shah (AVP, Content & Research at QuantInsti) helm this Masterclass about Natural Language Processing in Trading.
In this session, Terry and Ishan discuss:
- How is Natural Language Processing applied in financial markets?
- Calculating Daily Sentiment Score on Quantra learning portal
- Compare different word embedding methods with their pros and cons
- How does Quantra learning portal provide a unique learning experience?
Summary:
This introductory webinar describes the use of Natural Language Processing (NLP) techniques in the context of building trading strategies for 1-day horizons for the corporate bond market and equities markets using news headlines.
We described various methods for converting text into digital representations and for extracting sentiment scores from those embeddings.
Looking ahead to the course, we find that approaches using the latest advances in NLP are better suited to predict future returns in credit indices, by using news headlines directly as inputs, instead of news headline sentiments
About us:
Quantra is an online education portal that specializes in Algorithmic and Quantitative trading. Quantra offers various bite-sized, self-paced and interactive courses that are perfect for busy professionals, seeking implementable knowledge in this domain.
Useful links and some bonus content:
[Blog] Step By Step Guide - http://bit.ly/StepByStepGuideNLP
[Course] Natural Language Processing in Trading - http://bit.ly/QuantraNLPT
[Courses] All Quantra courses - http://bit.ly/AllQuantraCourses
Find more info on - https://quantra.quantinsti.com/
Like us on Facebook: https://www.facebook.com/goquantra/
Follow us on Twitter: https://twitter.com/GoQuantra
A hybrid composite features based sentence level sentiment analyzerIAESIJAI
Current lexica and machine learning based sentiment analysis approaches
still suffer from a two-fold limitation. First, manual lexicon construction and
machine training is time consuming and error-prone. Second, the
prediction’s accuracy entails sentences and their corresponding training text
should fall under the same domain. In this article, we experimentally
evaluate four sentiment classifiers, namely support vector machines (SVMs),
Naive Bayes (NB), logistic regression (LR) and random forest (RF). We
quantify the quality of each of these models using three real-world datasets
that comprise 50,000 movie reviews, 10,662 sentences, and 300 generic
movie reviews. Specifically, we study the impact of a variety of natural
language processing (NLP) pipelines on the quality of the predicted
sentiment orientations. Additionally, we measure the impact of incorporating
lexical semantic knowledge captured by WordNet on expanding original
words in sentences. Findings demonstrate that the utilizing different NLP
pipelines and semantic relationships impacts the quality of the sentiment
analyzers. In particular, results indicate that coupling lemmatization and
knowledge-based n-gram features proved to produce higher accuracy results.
With this coupling, the accuracy of the SVM classifier has improved to
90.43%, while it was 86.83%, 90.11%, 86.20%, respectively using the three
other classifiers.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesChristina Lin
Traditionally, dealing with real-time data pipelines has involved significant overhead, even for straightforward tasks like data transformation or masking. However, in this talk, we’ll venture into the dynamic realm of WebAssembly (WASM) and discover how it can revolutionize the creation of stateless streaming pipelines within a Kafka (Redpanda) broker. These pipelines are adept at managing low-latency, high-data-volume scenarios.
Online aptitude test management system project report.pdfKamal Acharya
The purpose of on-line aptitude test system is to take online test in an efficient manner and no time wasting for checking the paper. The main objective of on-line aptitude test system is to efficiently evaluate the candidate thoroughly through a fully automated system that not only saves lot of time but also gives fast results. For students they give papers according to their convenience and time and there is no need of using extra thing like paper, pen etc. This can be used in educational institutions as well as in corporate world. Can be used anywhere any time as it is a web based application (user Location doesn’t matter). No restriction that examiner has to be present when the candidate takes the test.
Every time when lecturers/professors need to conduct examinations they have to sit down think about the questions and then create a whole new set of questions for each and every exam. In some cases the professor may want to give an open book online exam that is the student can take the exam any time anywhere, but the student might have to answer the questions in a limited time period. The professor may want to change the sequence of questions for every student. The problem that a student has is whenever a date for the exam is declared the student has to take it and there is no way he can take it at some other time. This project will create an interface for the examiner to create and store questions in a repository. It will also create an interface for the student to take examinations at his convenience and the questions and/or exams may be timed. Thereby creating an application which can be used by examiners and examinee’s simultaneously.
Examination System is very useful for Teachers/Professors. As in the teaching profession, you are responsible for writing question papers. In the conventional method, you write the question paper on paper, keep question papers separate from answers and all this information you have to keep in a locker to avoid unauthorized access. Using the Examination System you can create a question paper and everything will be written to a single exam file in encrypted format. You can set the General and Administrator password to avoid unauthorized access to your question paper. Every time you start the examination, the program shuffles all the questions and selects them randomly from the database, which reduces the chances of memorizing the questions.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
6th International Conference on Machine Learning & Applications (CMLA 2024)ClaraZara1
6th International Conference on Machine Learning & Applications (CMLA 2024) will provide an excellent international forum for sharing knowledge and results in theory, methodology and applications of on Machine Learning & Applications.
Water billing management system project report.pdfKamal Acharya
Our project entitled “Water Billing Management System” aims is to generate Water bill with all the charges and penalty. Manual system that is employed is extremely laborious and quite inadequate. It only makes the process more difficult and hard.
The aim of our project is to develop a system that is meant to partially computerize the work performed in the Water Board like generating monthly Water bill, record of consuming unit of water, store record of the customer and previous unpaid record.
We used HTML/PHP as front end and MYSQL as back end for developing our project. HTML is primarily a visual design environment. We can create a android application by designing the form and that make up the user interface. Adding android application code to the form and the objects such as buttons and text boxes on them and adding any required support code in additional modular.
MySQL is free open source database that facilitates the effective management of the databases by connecting them to the software. It is a stable ,reliable and the powerful solution with the advanced features and advantages which are as follows: Data Security.MySQL is free open source database that facilitates the effective management of the databases by connecting them to the software.
Using recycled concrete aggregates (RCA) for pavements is crucial to achieving sustainability. Implementing RCA for new pavement can minimize carbon footprint, conserve natural resources, reduce harmful emissions, and lower life cycle costs. Compared to natural aggregate (NA), RCA pavement has fewer comprehensive studies and sustainability assessments.
An Approach to Detecting Writing Styles Based on Clustering Techniquesambekarshweta25
An Approach to Detecting Writing Styles Based on Clustering Techniques
Authors:
-Devkinandan Jagtap
-Shweta Ambekar
-Harshit Singh
-Nakul Sharma (Assistant Professor)
Institution:
VIIT Pune, India
Abstract:
This paper proposes a system to differentiate between human-generated and AI-generated texts using stylometric analysis. The system analyzes text files and classifies writing styles by employing various clustering algorithms, such as k-means, k-means++, hierarchical, and DBSCAN. The effectiveness of these algorithms is measured using silhouette scores. The system successfully identifies distinct writing styles within documents, demonstrating its potential for plagiarism detection.
Introduction:
Stylometry, the study of linguistic and structural features in texts, is used for tasks like plagiarism detection, genre separation, and author verification. This paper leverages stylometric analysis to identify different writing styles and improve plagiarism detection methods.
Methodology:
The system includes data collection, preprocessing, feature extraction, dimensional reduction, machine learning models for clustering, and performance comparison using silhouette scores. Feature extraction focuses on lexical features, vocabulary richness, and readability scores. The study uses a small dataset of texts from various authors and employs algorithms like k-means, k-means++, hierarchical clustering, and DBSCAN for clustering.
Results:
Experiments show that the system effectively identifies writing styles, with silhouette scores indicating reasonable to strong clustering when k=2. As the number of clusters increases, the silhouette scores decrease, indicating a drop in accuracy. K-means and k-means++ perform similarly, while hierarchical clustering is less optimized.
Conclusion and Future Work:
The system works well for distinguishing writing styles with two clusters but becomes less accurate as the number of clusters increases. Future research could focus on adding more parameters and optimizing the methodology to improve accuracy with higher cluster values. This system can enhance existing plagiarism detection tools, especially in academic settings.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdf
STOCKGRAM : DEEP LEARNING MODEL FOR DIGITIZING FINANCIAL COMMUNICATIONS VIA NATURAL LANGUAGE GENERATION
1. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
DOI: 10.5121/ijnlc.2020.9401 1
STOCKGRAM : DEEP LEARNING MODEL FOR DIGITIZING
FINANCIAL COMMUNICATIONS VIA NATURAL LANGUAGE
GENERATION
Purva Singh
School of Computer Science and Engineering, VIT University, Vellore
ABSTRACT
This paper proposes a deep learning model, StockGram, to automate financial communications via natural
language generation. StockGram is a seq2seq model that generates short and coherent versions of
financial news reports based on the client's point of interest from numerous pools of verified resources.
The proposed model is developed to mitigate the pain points of advisors who invest numerous hours while
scanning through these news reports manually. StockGram leverages bi-directional LSTM cells that allows
a recurrent system to make its prediction based on both past and future word sequences and hence predicts
the next word in the sequence more precisely. The proposed model utilizes custom word-embeddings,
GloVe, which incorporates global statistics to generate vector representations of news articles in an
unsupervised manner and allows the model to converge faster. StockGram is evaluated based on the
semantic closeness of the generated report to the provided prime words.
KEYWORDS
Natural language generation, bi-directional LSTM networks, language modelling, GloVe embeddings
1. INTRODUCTION
In the past years, Artificial Intelligence has seen rapid improvements in the field of natural
language and we can see various sub-domains evolving at the intersection of AI and linguistics.
The tendency to establish a two-way communication with computers has given rise to a
distinguished sub-category of tasks that is concerned with generating (quasi)-natural speech. This
category is referred to as Natural Language Generation (NLG). Artificial Intelligence : Natural
Language Processing Fundamentals [2] defines Natural Language Generation as “process of
producing meaningful phrases and sentences in the form of natural language”. Major applications
of NLG include generating financial reports, meeting agendas, chatbots, text summarization and
many more. The goal of NLG is to predict the next word, given a sentence as input. This task of
selecting the next probable word amongst a lot of possibilities is performed by language models.
The construction of these language models can be at character level, bi-gram level, n-gram level,
sentence and even paragraph level. The usage of feed-forward networks as language models is
not encouraged as an appropriate model for natural language generation. The main reason as to
why feed-forward neural networks are not considered optimal for NLG is because they do not
take into account the order of inputs. In natural language generation, the sequence of encoding is
important [3]. This structure is analogous to time-series data and word order is of cardinal
importance. Hence, utilizing feed-forward networks for text generation would lead to information
loss and generated sentences might not make sense and could be grammatically incorrect. In
order to take into account both time and sequence, Recurrent Neural Networks (RNNs) are used
for natural language processing related tasks.
2. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
2
This paper introduces StockGram, a deep learning model to automatize financial communication
via natural language generation. StockGram is a bi-directional LSTM network that captures both
the backward and forward information about the sequence at every time-step. In order to better
capture the semantic information of words, StockGram utilizes pre-trained word-embeddings
known as GloVe for a better vectorized representation of the words. The model is trained on
latest financial tweets and news reports on stocks, which is pre-processed before feeding into the
model. StockGram takes a set of prime words as input and generates a concise report which is
syntactically inline with those prime words. In comparison to a vanilla RNN model for natural
language generation, the StockGram model converges faster and generates concise reports which
are more syntactically inline with the prime words. This difference in model’s performance is due
to the structure of bidirectional LSTMs, where the model gets to make a prediction of the next
likely word based on both past and future sequences and hence predicts the next word in the
sequence more precisely.
The main purpose of StockGram is to mitigate the pain points of advisors who scan through
multiple financial reports, decks, news articles and tweets to collate meeting points for the clients.
The proposed model takes a set of prime words which are related to the client's point of interest
and based on the pre-processed financial data fed into the model, StockGram generates a set of
consolidated financial reports which can save hours of the advisor's efforts to scan through
numerous financial resources.
This paper is divided into the following sections: Section 2 explains the reasoning behind
choosing bi-directional LSTM model with pre-trained word embeddings. Section 3 defines
related work done in the domain of natural language generation followed by Section 4 that
describes the problem statement we have addressed in the paper, followed by Section 5 that
defines the dataset and methodology adopted for text generation. Section 6 shows the
experimental results conducted for achieving the optimal model for natural language generation.
The paper concludes by discussing conclusions and mentioning future work to be done on the
proposed model.
2. MODEL DEFINITION
2.1. Bidirectional LSTM Units
Bi-directional LSTMs are an extension to traditional LSTM which tends to improvise on the
model’s performance in the text generation domain. In scenarios where, all timesteps of input
sequence are provided, bi-directional LSTMs put together two independent LSTMs instead of
one LSTM and train them on the input. Figure 6 shows the structure of a bi-directional LSTM.
This structure allows the model to have both past and future information about input sequences.
In the two networks, one runs from past to future and another one from future to past. The first
on the input sequence as-is and the second on a reversed copy of the input sequence [5]. Due to
this nature of bi-directional LSTM, the model gets additional context about the sentence and it
converges faster as compared to vanilla uni-directional LSTM. For example, if we consider a
scenario were sequence of words is as follows :
“Due to the negative impact of the corona pandemic, the oil prices have .”
Since a typical vanilla RNN retains only short-term information, it will assign the probability of
next word based on “oil prices have” and can predict words such as “risen”, “improved”,
“increasing”. But an LSTM model will retain the long-term information and take into account
3. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
3
“negative impact of coronavirus” and hence would predict the word “plunged”. It is this nature of
bidirectional LSTM that StockGram utilizes in order to generate reports that are syntactically
inline with the prime words provided as input to the model.
Figure 1: General architecture of bidirectional LSTM that uses pre-trained GloVe embeddings.
2.2. Pre-trained Word Embeddings
Pre-trained word embeddings are the embeddings that are trained on huge datasets and learned on
one task and those learnings are then used for solving another analogous task. These pre-trained
word embeddings can be thought of as a form of Transfer Learning where learnings of one task
are transferred to another task. StockGram uses pre-trained word embeddings because they
capture the semantic and syntactic meaning of a word as they are trained on large datasets. These
word embeddings boost the performance of our NLP model. One of the reasons for not choosing
StockGram’s own embeddings is because the model takes into account numerous financial
reports, news and tweets as its training data which comprises large volumes of rare corpus. The
StockGram embeddings that would be learned from this training data won’t be able to arrive at
the correct vector representation of the word. Secondly, if the model learns embeddings from
scratch, the number of training parameters would increase which would eventually lead to a
slower training process. StockGram leverages Stanford’s pre-trained word embeddings called
GloVe which relies on both local statistics (local context information of words, word-based
models) and global statistics (word co-occurrence, count-based models) in order to obtain
vectorized representation of words [7]. GloVe optimizes our model one count at a time which
leads to faster model training. The GloVe model is scalable, captures the semantics irrespective
of word frequency and performs efficiently on both large and small text corpus.
3. LITERATURE SURVEY
Dipti Pawade et. al [9] have developed a LSTM based Story Scrambler that takes into input
different stories. These stories can have different or same storyline. After pre-processing of input
text, the pre-processed words are one-hot encoded before feeding them into the model. Batches of
pre-processed data based on sequence length are then fed into the model. Next word is then
selected with the highest probability and the window is shifted by one word. This process
continues till the desired number of words required in the story is reached. They have trained the
4. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
4
model by performing various hyper-parameter tunings by varying the RNN size, number of
layers, batch size and sequence length. The accuracy of the generated text by their model w.r.t
grammar and correctness of sentences was validated by seven different people and their best
model has achieved 63% accuracy as evaluated by humans. This model can be further improved
by using bi-directional LSTM instead of uni-directional LSTM model structure and by using pre-
trained word embeddings such as GloVe or word2vec instead of one-hot encodings which are
computationally inefficient.
Gregory Luppescu et al. [10] presents a LSTM based text-generation model to predict the next
word in the sequence. They have mentioned two separate tasks for text generation purposes: per-
author and mix-of-authors text generation that determines the source of text from a single author
or different authors respectively. To minimize the errors in text generation, they have used
perplexity, which is the inverse probability of the correct word. Hence, perplexity is inversely
proportional to the correctness of generated text. They have used pre-trained GloVe embeddings
for text-generation and a dynamic learning rate for the model. The test perplexity for most of the
authors lied between 100-200, thus proving high accuracy of the model but with few outliers. The
model can be further improved by utilizing bi-directional LSTM for text generation.
Nallapati R. et. al have proposed an abstractive text summarization model using encoder and
decoder, that was originally designed for machine translation, along with attention and have
achieved state-of-the-art performance on two different corpora [14]. They have taken a baseline
model for text generation consisting of bi-directional GRU-RNN encoder and uni-directional
decoder and have proposed several models to tackle specific drawbacks of the baseline model.
They have used the Large Vocabulary Trick (LVT) to speed up the convergence by reducing the
size of decoder’s softmax layer which is a computationally inefficient operation In order to
capture key features of a document, they have added additional look-up based embedding
matrices for the vocabulary of each tag-type, similar to the embeddings for words [14]. In regards
to handling out-of-vocabulary (OOV) words, instead of emitting UNK token, their new switching
decoder/pointer architecture where if the switch is on, decoder continues to generate
words in a normal manner, otherwise, decoder generates pointer to a word present in source
which is then included in the document summary. For very long documents, where not only key
words but key sentences also play an important role in generating text summarization, their
model assigns two bi-directional RNNs, one at word level and one at sentence level, while
attention mechanism operates at both levels simultaneously.
Santhanam, S. has proposed a context based text generation model based on bi-directional LSTM
networks and uses one-hot encoding for vectorized representation of words. He has proposed two
methods for generating context vectors that capture the semantics of the sentence[13]. First using
the word importance method, where based on the frequency of the word in a sentence and
frequency of that word in the document, the importance of each word is calculated. Based on this
tf-idf method, the word with highest importance is calculated and is predicted as the next word.
Second, using the word clustering method to extract context-vectors to exploit two properties :
semantic similarity and relationships. This model states that context of a sentence is dependent on
context vector present in word vector space in such a way that semantic similarity between
context vector and accumulated sentence vector is the highest [13]. This model can be further
improved by using pre-trained word-embeddings such as GloVe or FastText.
4. PROBLEM STATEMENT
The main idea behind StockGram is to reduce the pain points of advisors who invest hours of
manual effort to gain insights from latest financial news articles, reports and tweets to present to
5. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
5
the clients. In order to reduce this manual effort, we propose a natural language generation model
that takes the latest financial articles, reports and tweets on stocks as its training data and
generates a consolidated report based on the client's sector of interest.
In order to generate reports that are syntactically and semantically aligned with client’s interest,
Stockgram leverages bidirectional LSTM which helps the model to have a complete picture of
the past and the future words around the prime word to generate a sequence of words that are
relevant to the prime word. For StockGram, we also need a better vectorized representation of
words that can greatly capture their semantic information and to increase the speed of training the
model. Keeping these properties in mind, the model proposed in this paper is a bi-directional
LSTM network that captures both the backward and forward information about the sequence at
every time-step. In order to better capture the semantic information of words, I have used pre-
trained word-embeddings called GloVe for a better vectorized representation of the words. This
language model can be applied for text generation tasks, where, given a set of prime words, the
model generates syntactically meaningful sentences.
Text generation models are usually evaluated based on whether the generated sentences are
syntactically strong and whether the model is not generating duplicate sentences [13]. One such
measure is Perplexity. In this paper, the proposed model is evaluated based on how semantically
correct sentences are and evaluated the semantic closeness of the generated text with provided
prime words.
5. DATASET
The proposed model was made keeping in mind to perform text generation on financial news
reports related to stock news. The proposed model is expected to perform neural text generation
task on an input of combined financial news on stock market from various verified sources in the
form of plain text file and based on these news reports, the model generates a concise, holistic
and latest report which are semantically related to words of interest for example IT sector, oil
prices, DJIA etc. Since manually scraping news articles or subscribing to historical news for high
quality financial news on stocks can be economically inefficient and a very tedious process,
hence, I have taken a pre-scrapped financial news article, from benzinga.com, a dataset present
in Kaggle. Name of the dataset is “Daily Financial News for 6000+ Stocks'' which contains 6
columns that define : headline (article release headline, string), URL (url of the article, string),
publisher (name of the person who wrote the article, string), date (date of article’s release,
DateTime) and stock (stock ticker , abbreviation used to uniquely identify publicly traded shares
of a particular stock on a particular stock market (NYSE /NASDAQ/ AMEX only). The only
column in which we are interested in is the “Article” column that contains news headlines related
to latest financial news on stocks. These articles are then combined together in a simple text file,
pre-processed and then fed into themodel.
6. METHODOLOGY
6.1. Pre-processing input data
The first step in a text generation task is to pre-process the input data. In the proposed framework,
we are using pre-scrapped Kaggle dataset and are only interested in the “Article” column of that
dataset which contains financial news headlines related to stocks. We parse the entries in the
“Article” column and store them in a list. Then, we open a sample text file, ‘finance_news.txt’
and join all the entries in our list and write them in the text file.
6. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
6
Now, before we proceed to feed input to the neural network, we need to pre-process the data.
Here I have followed the following steps as mentioned below:
1. Remove html tags (<href></href>, <a></a>), emoticons (:heart, :laugh), punctuations (
!”#$%&’)and stopwords ( “you’ve”, “you’ll”, “you’d” etc)
2. Convert the pre-processed text into lowercase.
3. Perform tokenization on the input text. Tokenization is the process where given a sequence
of characters tokenizing the sequence would imply chopping the sequence into pieces
known as tokens and also getting rid of punctuations [15].
After we have performed the above three steps, we are almost ready for feeding the input to the
model. After pre-processing the tokenized input data, we need to convert them into numbers. For
this, we need to create a vocabulary-to-integer and integer-to-vocabulary dictionary. There are
many ways to perform this task, but in the proposed model, I am assigning integers (indices) to
words in the inverse order of their frequencies. For example if the most common term is “the”,
then word “the” gets the index 0 and the next most frequent word gets the index 1 and so on.
6.2. Batching pre-processed data
The pre-processed input data needs to be fed in the model as small batches of data instead of
feeding the entire input at once. The figure 10 below defines how we can batch our input data.
Say for example we have a sequence [1, 2, 3, 4, 5, 6, 7, 8, 9 , 10, 11, 12] and we want a batch size
of two. In this scenario, our batching data function should split our word sequence into two
separate sequences for example batched data 1 can contain [1, 2, 3, 4, 5, 6] and batched data 2
can contain [7, 8, 9, 10, 11, 12]. Then, we specify the sequence length that tells us how big we
want our sequences to be. Figure 10 illustrated below shows the resultant data that is generated
after batching a simple sequence and how a batch size of 2 and sequence length of 3 would look
like.
Figure 2: Batching of a sample pre-processed data
6.3. Model architecture
Firstly, we provide our model with a sequence of prime words. These words will give our bi-
directional LSTM model, a clue regarding which word should be predicted next. For example,
given the following prime words [“corona”, “djia”, “oil”, “unemployment”], the output of the
proposed model would be a matrix providing the probability for each word from the dictionary to
be the next one of the given sequence [16]. Based on the news articles provided to the model, it
7. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
7
can predict that the next word in the sequence might be “plunged”. Now once the next word is
predicted, we append it in a common list and iterate the entire process.
Now in the proposed model, the batched data is first fed through the GloVe embedding layer,
where each word is converted into its vectorized representation. If the word present in an input
news article is also present as a key in GloVe matrix, then we take the corresponding GloVe
embedding of that word, otherwise, we randomly assign a numerical vector to that particular
word. The embedding matrix is of dimension V, where V is the size of our vocabulary. After the
embedding layer, the vectorized input is then passed through bi-directional LSTM with 256
hidden units and a softmax activation function for the outer layer. The learning rate for the model
is kept fixed throughout the training process. I have used Adam as the optimizer and loss
calculation is done on cross entropy. Model has been trained using the PyTorch library. Since this
model has to predict the embedding representation of the target word, that is, predicting the next
probable word in the sequence, using softmax activation function seems valid to determine
the maximum probability of the word. After training the model for ten epochs, the prediction part
requires the end-user to provide two inputs. The first input is a sequence of prime words. These
prime words are opening words that provide our model a seed for it to start the text generation
process. One can provide any number of prime words. The next input that end-use has to provide
is the number of words to generate at a time. If the number of words to be generated is n and the
number of prime words provided by the user is p, then for every prime word, n words will be
generated, i.e., a total of np words will be generated in the end. Once the next word is predicted,
the first word from the sequence is removed and the predicted word is appended. With the newly
formed input sequence, the next word is predicted. This process is iterated till the desired number
of words is reached.
7. RESULTS
Before looking at the text generated by the proposed framework, given a set of prime words, we
need to understand how the model learned over time by observing its training loss. Figure 11
shows a comparison of training loss of StockGram (bidirectional LSTM model with pre-trained
word embeddings) v/s basic RNN model (unidirectional model with 2 contiguous LSTM layers
and one-hot encoding) for natural language generation.
Figure 3: Comparison of training loss of basic RNN model v/s StockGram model
There is no point of taking validation dataset into account, since, for text generation models, there
is no particular point where the training accuracy increases and validation accuracy decreases that
the training model should be stopped. [13]. As is it clear from the training graph, the StockGram
model has converged faster than the basic RNN model for text generation for the same number of
8. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
8
iterations. For text generation models, perplexity is a fairly good evaluation metric. It tests the
syntactical integrity of the generated text. Table below shows a summarised financial report
generated after experimenting with various combinations of prime words for both the basic RNN
model and Stock Gram model.
Table 1. Financial reports generated by a basic RNN model (2 contiguous LSTM layers with one-hot
encoding) based on various combination of prime-words
Prime words Report generated
coronavirus, s&p, 500,
oil, prices,
coronavirus outbreak, which was first detected in China,
s&p 500 spasm 'new $180 peco ekspress
brief? China international group holdings posts fy net loss of hk$11. 2 mln brief?
china international group holdings posts fy net loss of hk$11. 2 mln u. s. judge says
he will not be a purge of 500 furry furry furry new ceo china says it to meet u. s.
tariffs over tariffs brief? china international group holdings posts fy net loss of
hk$11.oil prices 'new newark peco furry 'new report for first quarter earnings
release and conference call on april 24, 2018 oil prices new newark peco furry
'new report for first quarter earnings release and conference call on april 24, 2018
brief? china international group holdings posts fy net loss of hk$11.brief? china
international group holdings posts fy net loss of hk$11.brief? china international
group holdings posts fy net loss of hk$11.
Table 2. Financial reports generated by StockGram (bidirectional model with pre-trained word
embeddings) based on various combination of prime-words
Prime words Report generated
coronavirus, since coronavirus outbreak states reporting significant spikes following
pandemic, s&p, 500, attempts to ease lockdown restrictions.related: bulls beware: a
oil, prices, forecasting model model nor has been edited due to coronavirus
unemployement as the s&p 500 index rates caused the world's largest single day at a
collapse in the globe are warning that the pandemic is nowhere near
over, encouraging individuals to settle near 9. the current scenario is
different, though, from the chicago board of exchange, volatility index
rates, things don't implement to the s&p 500 triggered demand
dissipated, energy demand dissipated, and oil prices outbreak in a
shorter period of time, there is not able to work as much as 30%
as the s&p 500 index has been expanded as planned . the magnitude of
how damaged the energy industry is came into full view on april 20
when the benchmark price of us unemployment rates, it is useful to
recall that the market has had been from laying off people in bulk
By observing the reports generated by StockGram and basic RNN models, we can see that the
basic RNN model tends to repeat a lot of sentences. Also, the report generated is not in
accordance with the list of prime words provided. Whereas, some of the semantic content of the
report generated by StockGram is in line with the prime words provided. For instance, the prime
word “crude” generated phrases such as crude prices went negative for the first time in history,
prime word “covid” generated results around topic such as how covid19 is impacting oil prices
negatively, topics related to how covid19 is causing lay-offs in various companies and how
covid19 is affecting s&p 500 and DJIA. The StockGram model still needs more training as for
some of the prime words, the reports generated by StockGram were not syntactically correct.
9. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
9
8. DISCUSSION
This paper proposes a natural language generation model, StockGram, that is developed keeping
in mind to mitigate the pain points of advisors who put in hours of manual effort to scan through
financial news articles in order to generate concise reports based on client’s preference (prime
words). In order to generate reports that are semantically and syntactically inline with prime
words, StockGram utilizes bidirectional LSTM with pre-trained word embeddings. In order to
evaluate the performance of StockGram, the paper compares the proposed model with a vanilla
RNN model for natural language generation. Upon comparison, it is observed that StockGram
generates reports that are more syntactically inline with the prime words and converges faster
than basic RNN models. The proposed model can also be extended to the following areas:
1. Question-answering framework, where the question can be a sequence of prime words and
based on those prime words, the model generates the answer
2. To generate a consolidated report on stock news from various verified sources based on a
sequence of prime words. This will save a lot of time for advisors who go through a lot of
news articles online to get a complete picture of the latest trends in the financial domain.
3. Since this model has been trained on financial news related to stocks, this model can be
extended to serve as a financial assistant chatbot, where the model answers questions
related to latest trends in the stock market.
8.2. Future Work
1. The proposed model can leverage attention based models such as Transformers instead of
LSTMs.
2. More robust evaluation metrics such as BLEURT, BERTScore can be considered for
StockGram.
3. Convolutional neural networks can be leveraged instead of memory based networks such as
RNNs, to encapsulate dependencies between words.
REFERENCES
[1] Evolution of Natural Language Generation. https://medium.com/sfu-cspmp/evolution-of-natural-
language-generation-c5d7295d6517
[2] A comprehensive guide to natural language generation. https://medium.com/sciforce/a-
comprehensive-guide-to-natural-language-generation-dd63a4b6e548
[3] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–
1780, 1997.
[4] Develop bidirectional lstm sequence classification using python keras.
https://machinelearningmastery.com/develop-bidirectional-lstm-sequence-classification-python-keras/
[5] NLP 101 : Word2vec, skip gram and CBOW.
https://towardsdatascience.com/nlp-101-word2vec-skip-gram-and-cbow-93512ee24314
[6] Negative sampling and GloVe.
https://towardsdatascience.com/nlp-101-negative-sampling-and-glove-936c88f3bc68
10. International Journal on Natural Language Computing (IJNLC) Vol.9, No.4, August 2020
10
[7] GloVe embeddings by Stanford. https://nlp.stanford.edu/projects/glove/
[8] Dipti Pawade, Avani Sakhapara, Mansi Jain, Neha Jain, Krushi Gada, "Story Scrambler – Automatic
Text Generation Using Word Level RNN-LSTM", International Journal of Information Technology
and Computer Science(IJITCS), Vol.10, No.6, pp.44-53, 2018. DOI: 10.5815/ijitcs.2018.06.05
[9] Luppescu, G., & Romero, F. (2017). Comparing Deep Learning and Conventional Machine Learning
for Authorship Attribution and Text Generation.
[10] Yao, L., Peng, N., Weischedel, R., Knight, K., Zhao, D., & Yan, R. (2019, July). Plan-and-write:
Towards better automatic storytelling. In Proceedings of the AAAI Conference on Artificial
Intelligence (Vol. 33, pp. 7378-7385).
[11] Semeniuta, S., Severyn, A., & Barth, E. (2017). A hybrid convolutional variational autoencoder for
text generation. arXiv preprint arXiv:1702.02390 .
[12] Santhanam, S. (2020). Context based Text-generation using LSTM networks. arXiv preprint
arXiv:2005.00048 .
[13] Nallapati, R., Zhou, B., Gulcehre, C., & Xiang, B. (2016). Abstractive text summarization using
sequence-to-sequence rnns and beyond. arXiv preprint arXiv:1602.06023 .
[14] Tokenization.https://nlp.stanford.edu/IR-book/html/htmledition/tokenization-1.html
[15] Text generation using bidirectional LSTM and doc2vec models.
[16] PyTorch’s official documentation.
AUTHOR
Purva Singh, has completed her B.Tech in Computer Science and Engineering from
Vellore Institute of Technology, Vellore. She has been exploring the field of Artificial
Intelligence and Deep Learning for the past two years. She is passionate about Natural
Language Processing, Deep Reinforcement Learning and Big Data Analytics.