SlideShare a Scribd company logo
1 of 45
Download to read offline
Deep Temporal-Recurrent-Replicated-
Softmax for Topical Trends over Time
Intern © Siemens AG 2017
Softmax for Topical Trends over Time
Author(s): Pankaj Gupta1,2, Subbu Ram2, Bernt Andrassy2, Hinrich Schütze1
Presenter: Pankaj Gupta @NAACL/ACL New Orleans, USA. 3rd June 2018
1CIS, University of Munich (LMU) | 2Machine Intelligence, Siemens AG | June 2018
Outline
 Problem Statement
 Background Background
- Topic Modeling: Static Vs Dynamic
- Replicated Softmax Models (RSM)
 Proposal: The RNN-RSM model
 Evaluations
Intern © Siemens AG 2017
May 2017Seite 2 Corporate Technology
 Evaluations
- Generalization
- Topic Evolution and Characterization
Problem Statement: Modeling Science for Topical Overview Timelines
201420081996
Intern © Siemens AG 2017
May 2017Seite 3 Corporate Technology
Problem Statement: Modeling Science for Topical Overview Timelines
Topic Structure Detection (TSD):
 Identify the main topics in document collection
 Discover hidden structure that explains the document Discover hidden structure that explains the document
Intern © Siemens AG 2017
May 2017Seite 4 Corporate Technology
Problem Statement: Automatic Topical Overview Timelines
Topic Structure Detection (TSD):
 Identify the main topics in document collection
 Discover hidden structure that explains the document Discover hidden structure that explains the document
Topic Evolution Detection (TED):
 Recognize how the underlying topics evolves (grows or decays) over time
 Detect the emergence of a new topic
Intern © Siemens AG 2017
May 2017Seite 5 Corporate Technology
Problem Statement: Automatic Topical Overview Timelines
Topic Structure Detection (TSD):
 Identify the main topics in document collection
 Discover hidden structure that explains the document Discover hidden structure that explains the document
Topic Evolution Detection (TED):
 Recognize how the underlying topics evolves (grows or decays) over time
 Detect the emergence of a new topic
Intern © Siemens AG 2017
May 2017Seite 6 Corporate Technology
Temporal Topic Characterization (TTC):
 Characterize topics to track the words’ usage within each topic over time
 Topical/Keyword trend analysis for word evolution
Problem Statement: Automatic Topical Overview Timelines
Topic Structure Detection (TSD):
 Identify the main topics in document collection
 Discover hidden structure that explains the document Discover hidden structure that explains the document
Topic Evolution Detection (TED):
 Recognize how the underlying topics evolves (grows or decays) over time
 Detect the emergence of a new topic
Intern © Siemens AG 2017
May 2017Seite 7 Corporate Technology
Temporal Topic Characterization (TTC):
 Characterize topics to track the words’ usage within each topic over time
 Topical/Keyword trend analysis for word evolution
Background: Static Topic Model (STM)
Topic model:
 Statistical Generative modeling
 Discover abstract topics in document collection Discover abstract topics in document collection
Intuition:
 Documents exhibit multiple topics
 Each word drawn from one of the topics
Intern © Siemens AG 2017
May 2017Seite 8 Corporate Technology
Background: Static Topic Model (STM)
Topic model:
 Statistical Generative modeling
 Discover abstract topics in document collection Discover abstract topics in document collection
Intuition:
 Documents exhibit multiple topics
 Each word drawn from one of the topics
Assumption:
 Topics shared by all documents
Intern © Siemens AG 2017
May 2017Seite 9 Corporate Technology
 Document drawn from the same topics
 No temporal order of documents
 Documents are exchangeable
Background: Dynamic Topic Model (DTM)
 Model the temporal order of documents
 Capture evolution of topics in temporally organized document collections over time Capture evolution of topics in temporally organized document collections over time
 For example in scholarly articles, the research evolved over time
Intern © Siemens AG 2017
May 2017Seite 10 Corporate Technology
Background: Dynamic Topic Model
DTM (Blei and Lafferty (2006))
 Temporal stack of LDA models
Intern © Siemens AG 2017
May 2017Seite 11 Corporate Technology
Background: Dynamic Topic Model
DTM (Blei and Lafferty (2006))
 Temporal stack of LDA models
Assumption in DTMAssumption in DTM
 Assume k unique topics over time, model their evolution
 Can not model appearance of new topicsdisappearance
of old topics over time (Blei and Lafferty (2006))
Intern © Siemens AG 2017
May 2017Seite 12 Corporate Technology
Background: Dynamic Topic Model
DTM (Blei and Lafferty (2006))
 Temporal stack of LDA models
Assumption in DTMAssumption in DTM
 Assume k unique topics over time, model their evolution
 Can not model appearance of new topicsdisappearance
of old topics over time (Blei and Lafferty (2006))
Intern © Siemens AG 2017
May 2017Seite 13 Corporate Technology
Challenges:
 How new topics appear/disappear over time ?
 Explicit/long term topical dependencies ?
 Heteroscedasticity: Modeling high dimensional sequence ?
Background: Replicated Softmax Model (RSM)
 Generative Model of Word Counts
 Family of different-sized RBMs (Restricted Boltzmann Machine)
 Energy based undirected static topic model Energy based undirected static topic model
Intern © Siemens AG 2017
May 2017Seite 14 Corporate Technology
Background: Replicated Softmax Model (RSM)
 Generative Model of Word Counts
 Family of different-sized RBMs (Restricted Boltzmann Machine)
 Energy based undirected static topic model Energy based undirected static topic model
Intern © Siemens AG 2017
May 2017Seite 15 Corporate Technology
Background: Replicated Softmax Model (RSM)
 Generative Model of Word Counts
 Family of different-sized RBMs (Restricted Boltzmann Machine)
 Energy based undirected static topic model Energy based undirected static topic model
Intern © Siemens AG 2017
May 2017Seite 16 Corporate Technology
Proposed Model: Recurrent Neural Network-Replicated Softmax (RNN-RSM)
h(1)
bh
(1) h(2)
bh
(2) h(T-1)
bh
(T-1) h(T)
bh
(T)
RSM RSM RSM RSM
…
Latent Topics
V(1)
bh
bv
(1)
Wvh
u(1)
Wvu
V(2)
bh
bv
(2)
Wvh
u(2)
Wvu
V(T-1)
bh
bv
(T-1)
Wvh
u(T-1)
Wvu
u(0)
V(T)
bh
bv
(T)
Wvh
u(T)
Wvu
Wuh
Wuv Wuv Wuv
Wuh Wuh
Wuh
Wuv
RNN
Wuu Wuu Wuu Wuu
Observable
Softmax Visibles
 Neural Dynamic Topic Model
Intern © Siemens AG 2017
May 2017Seite 17 Corporate Technology
 Neural Dynamic Topic Model
 Temporal Sequence of distributional estimators (RSMs)
 Temporal Feedback connections spanning timeline to model dynamics in hidden topics
Proposed Model: RNN-RSM
h(1)
bh
(1) h(2)
bh
(2) h(T-1)
bh
(T-1) h(T)
bh
(T)
RSM RSM RSM RSM
…
Latent Topics
V(1)
bh
bv
(1)
Wvh
u(1)
Wvu
V(2)
bh
bv
(2)
Wvh
u(2)
Wvu
V(T-1)
bh
bv
(T-1)
Wvh
u(T-1)
Wvu
u(0)
V(T)
bh
bv
(T)
Wvh
u(T)
Wvu
Wuh
Wuv Wuv Wuv
Wuh Wuh
Wuh
Wuv
RNN
Wuu Wuu Wuu Wuu
Observable
Softmax Visibles
 Document collections over time
Intern © Siemens AG 2017
May 2017Seite 18 Corporate Technology
 Document collections over time
 Input to each of the RSMs
Proposed Model: RNN-RSM
h(1)
bh
(1) h(2)
bh
(2) h(T-1)
bh
(T-1) h(T)
bh
(T)
RSM RSM RSM RSM
…
Latent Topics
V(1)
bh
bv
(1)
Wvh
u(1)
Wvu
V(2)
bh
bv
(2)
Wvh
u(2)
Wvu
V(T-1)
bh
bv
(T-1)
Wvh
u(T-1)
Wvu
u(0)
V(T)
bh
bv
(T)
Wvh
u(T)
Wvu
Wuh
Wuv Wuv Wuv
Wuh Wuh
Wuh
Wuv
RNN
Wuu Wuu Wuu Wuu
Observable
Softmax Visibles
 Document collections over time
Intern © Siemens AG 2017
May 2017Seite 19 Corporate Technology
 Document collections over time
 Input to each of the RSMs
 Hidden/latent topics over time, in each of the RSMs
Training RNN-RSM: Forward Propagation Through Time
h(1)
bh
(1) h(2)
bh
(2) h(T-1)
bh
(T-1) h(T)
bh
(T)
RSM RSM RSM RSM
…
Latent Topics
V(1)
bh
bv
(1)
Wvh
u(1)
Wvu
V(2)
bh
bv
(2)
Wvh
u(2)
Wvu
V(T-1)
bh
bv
(T-1)
Wvh
u(T-1)
Wvu
u(0)
V(T)
bh
bv
(T)
Wvh
u(T)
Wvu
Wuh
Wuv Wuv Wuv
Wuh Wuh
Wuh
Wuv
RNN
Wuu Wuu Wuu Wuu
Observable
Softmax Visibles
Propagate u(t) in the RNN portion
Intern © Siemens AG 2017
May 2017Seite 20 Corporate Technology
Propagate u(t) in the RNN portion
Training RNN-RSM: Forward Propagation Through Time
h(1)
bh
(1) h(2)
bh
(2) h(T-1)
bh
(T-1) h(T)
bh
(T)
RSM RSM RSM RSM
…
Latent Topics
V(1)
bh
bv
(1)
Wvh
u(1)
Wvu
V(2)
bh
bv
(2)
Wvh
u(2)
Wvu
V(T-1)
bh
bv
(T-1)
Wvh
u(T-1)
Wvu
u(0)
V(T)
bh
bv
(T)
Wvh
u(T)
Wvu
Wuh
Wuv Wuv Wuv
Wuh Wuh
Wuh
Wuv
RNN
Wuu Wuu Wuu Wuu
Observable
Softmax Visibles
Compute RSM biases at each time step, t
Intern © Siemens AG 2017
May 2017Seite 21 Corporate Technology
Compute RSM biases at each time step, t
Training RNN-RSM: Forward Propagation Through Time
h(1)
bh
(1) h(2)
bh
(2) h(T-1)
bh
(T-1) h(T)
bh
(T)
RSM RSM RSM RSM
…
Latent Topics
V(1)
bh
bv
(1)
Wvh
u(1)
Wvu
V(2)
bh
bv
(2)
Wvh
u(2)
Wvu
V(T-1)
bh
bv
(T-1)
Wvh
u(T-1)
Wvu
u(0)
V(T)
bh
bv
(T)
Wvh
u(T)
Wvu
Wuh
Wuv Wuv Wuv
Wuh Wuh
Wuh
Wuv
RNN
Wuu Wuu Wuu Wuu
…
Observable
Softmax Visibles
Compute the Conditional distributions in each RSM
Intern © Siemens AG 2017
May 2017Seite 22 Corporate Technology
Compute the Conditional distributions in each RSM
Training RNN-RSM: Forward Propagation Through Time
h(1)
bh
(1) h(2)
bh
(2) h(T-1)
bh
(T-1) h(T)
bh
(T)
RSM RSM RSM RSM
…
Latent Topics
V(1)
bh
bv
(1)
Wvh
u(1)
Wvu
V(2)
bh
bv
(2)
Wvh
u(2)
Wvu
V(T-1)
bh
bv
(T-1)
Wvh
u(T-1)
Wvu
u(0)
V(T)
bh
bv
(T)
Wvh
u(T)
Wvu
Wuh
Wuv Wuv Wuv
Wuh Wuh
Wuh
Wuv
RNN
Wuu Wuu Wuu Wuu
…
Observable
Softmax Visibles
Neural Network Neural Network Neural NetworkNeural Network Neural Network Neural Network Topic-words
Intern © Siemens AG 2017
May 2017Seite 23 Corporate Technology
Neural Network
Language Models
Word Representation
Linear Model
Rule Set
Neural Network
Language Models
Supervised
Linear Model
Rule Set
Neural Network
Language Models
Word Embedding
Word Embeddings
Word Representation
Neural Network
Language Models
Word Representation
Linear Model
Rule Set
Neural Network
Language Models
Supervised
Linear Model
Rule Set
Neural Network
Language Models
Word Embedding
Word Embeddings
Word Representation
Topic-words
over time
for topic
‘Word Vector’
1996 1997 2014
Training RNN-RSM: Backpropagation Through Time (BPTT)
h(1)
bh
(1) h(2)
bh
(2) h(T-1)
bh
(T-1) h(T)
bh
(T)
RSM RSM RSM RSM
…
Latent Topics
V(1)
bh
bv
(1)
Wvh
u(1)
Wvu
V(2)
bh
bv
(2)
Wvh
u(2)
Wvu
V(T-1)
bh
bv
(T-1)
Wvh
u(T-1)
Wvu
u(0)
V(T)
bh
bv
(T)
Wvh
u(T)
Wvu
Wuh
Wuv Wuv Wuv
Wuh Wuh
Wuh
Wuv
RNN
Wuu Wuu Wuu Wuu
…
Observable
Softmax Visibles
Cost in RNN-RSM, the negative log-likelihood
Intern © Siemens AG 2017
May 2017Seite 24 Corporate Technology
Cost in RNN-RSM, the negative log-likelihood
Summary: RNN-RSM Offerings
 Unsupervised Neural Dynamic Topic Model
 Deterministic (RNN) + Stochastic (RSM) network Deterministic (RNN) + Stochastic (RSM) network
 Longer-term topical dependencies via RNN
 Drift/bias terms to propagate topical information
 New topics to evolve over time via sequence of RSMs
- Separate hidden vectors to capture and convey topics
- Topics over time are independent
Intern © Siemens AG 2017
May 2017Seite 25 Corporate Technology
- Topics over time are independent
Evaluation: Dataset and Experimental Setup
Data Set:
 EMNLP and ACL conference papers (Gollapalli and Li 2015) from the year 1996 -2014
 Expand Rank (Wan and Xiao 2008) to extract top 100 key-phrases for each paper (unigrams + bigrams)
 Dictionary of all unigrams and bigrams, dictionary size, K = 3390
Intern © Siemens AG 2017
May 2017Seite 26 Corporate Technology
Evaluation: Dataset and Experimental Setup
Data Set:
 EMNLP and ACL conference papers (Gollapalli and Li 2015) from the year 1996 -2014
 Expand Rank (Wan and Xiao 2008) to extract top 100 key-phrases for each paper (unigrams + bigrams)
 Dictionary of all unigrams and bigrams, dictionary size, K = 3390
Experimental Setup:
Intern © Siemens AG 2017
May 2017Seite 27 Corporate Technology
 Evaluate RNN-RSM (dynamic) with LDA-DTM (dynamic), seqRSM (static) and seqLDA (static)
 seqRSM and seqLDA: Individual 19 different RSM/LDA models trained for each year
 LDA-DTM and RNN-RSM are trained over the years with 19 time steps
 Perplexity (PPL) to choose the number of optimal topics (=30)
Evaluation (Quantitative): Dynamic Topic Models as Generative Models
 Average perplexity per word of the document collections over time
Intern © Siemens AG 2017
May 2017Seite 28 Corporate Technology
 Held-out documents: 10 documents from each time step i.e. total 190 documents
 Compute perplexity for 30 topics
Evaluation (Qualitative): Topic Detection and Evolution
 Extract the top 20 terms for every 30 topic set from 1996-2014 for the topic models,
 Resulting in |Q|_max = 19*30*20 possible topic terms
Topic count
Topic-terms count
Time steps
Intern © Siemens AG 2017
May 2017Seite 29 Corporate Technology
Evaluation (Qualitative): Topic Detection and Evolution
 Extract the top 20 terms for every 30 topic set from 1996-2014 for the topic models,
 Resulting in |Q|_max = 19*30*20 possible topic terms
 Topic Popularity (Sentiment Analysis, Word Vector and Dependency Parsing in NLP research) Topic Popularity (Sentiment Analysis, Word Vector and Dependency Parsing in NLP research)
Intern © Siemens AG 2017
May 2017Seite 30 Corporate Technology
Evaluation (Qualitative): Topic Detection and Evolution
 Extract the top 20 terms for every 30 topic set from 1996-2014 for the topic models,
 Resulting in |Q|_max = 19*30*20 possible topic terms
 Topic Popularity (Sentiment Analysis, Word Vector and Dependency Parsing in NLP research) Topic Popularity (Sentiment Analysis, Word Vector and Dependency Parsing in NLP research)
Intern © Siemens AG 2017
May 2017Seite 31 Corporate Technology
Evaluation: Topic Characterization (Trending Keywords over time)
 Analyse word evolution/usage in the complete topic sets over time (19*30*20 possible topic terms)
Keyword-trend: Appearance/disappearance of a keyword in underlying topics detected over time
Span: Length of the longest sequence of the keyword appearance in its keyword trend
Intern © Siemens AG 2017
May 2017Seite 32 Corporate Technology
Span: Length of the longest sequence of the keyword appearance in its keyword trend
Evaluation: Topic Characterization (Topic-term Usage over Time)
Intern © Siemens AG 2017
May 2017Seite 33 Corporate Technology
Evaluation: Topic Characterization (Topic-term Usage over Time)
Intern © Siemens AG 2017
May 2017Seite 34 Corporate Technology
Evaluation: Topic Characterization (Topic-term Usage over Time)
Intern © Siemens AG 2017
May 2017Seite 35 Corporate Technology
Evaluation: Topic Characterization (Topic-term Usage over Time)
Intern © Siemens AG 2017
May 2017Seite 36 Corporate Technology
Evaluation: Topic Characterization (Topic-term Usage over Time)
Intern © Siemens AG 2017
May 2017Seite 37 Corporate Technology
Evaluation: Topic Characterization (Topic-term Usage over Time)
Intern © Siemens AG 2017
May 2017Seite 38 Corporate Technology
Evaluation: Topic Characterization (Topic-term Usage over Time)
Intern © Siemens AG 2017
May 2017Seite 39 Corporate Technology
Word usage for topic ‘Word Vector’ over time
Summary
 Introduce an unsupervised neural dynamic topic model, RNN-RSM
 Model dynamics in underlying topics detected over time Model dynamics in underlying topics detected over time
 Demonstrate better generalization, topic evolution and characterization
 Analyze scholarly articles to capture evolution in NLP topics
Intern © Siemens AG 2017
May 2017Seite 40 Corporate Technology
h(1)
bh
(1) h(2)
bh
(2) h(T-1)
bh
(T-1) h(T)
bh
(T)
RSM RSM RSM RSM
…
Thanks !!!
V(1)
bh
bv
(1)
Wvh
u(1)
Wvu
V(2)
bh
bv
(2)
Wvh
u(2)
Wvu
V(T-1)
bh
bv
(T-1)
Wvh
u(T-1)
Wvu
u(0)
V(T)
bh
bv
(T)
Wvh
u(T)
Wvu
Wuh
Wuv Wuv Wuv
Wuh Wuh
Wuh
Wuv
RNN
Wuu Wuu Wuu Wuu
…
Observable
Softmax Visibles
Neural Network Neural Network Neural NetworkNeural Network Neural Network Neural Network
Intern © Siemens AG 2017
May 2017Seite 41 Corporate Technology
1996
Neural Network
Language Models
Word Representation
Linear Model
Rule Set
Neural Network
Language Models
Supervised
Linear Model
Rule Set
Neural Network
Language Models
Word Embedding
Word Embeddings
Word Representation
Neural Network
Language Models
Word Representation
Linear Model
Rule Set
Neural Network
Language Models
Supervised
Linear Model
Rule Set
Neural Network
Language Models
Word Embedding
Word Embeddings
Word Representation
Topic-words
over time
for topic
‘Word Vector’
1997 2014
References
Jean Aitchison. 2001. Language change: progress or decay?. Cambridge University Press.
Nikolaos Aletras and Mark Stevenson. 2013. Evaluating topic coherence using distributional semantics. In Proceedings of the 10th International
Conference on Computational Semantics (IWCS). Potsdam, Germany, pages 13–22.
James Allan. 2002. Introduction to topic detection and tracking. In Topic detection and tracking, Springer, pages 1–16.James Allan. 2002. Introduction to topic detection and tracking. In Topic detection and tracking, Springer, pages 1–16.
James Allan, Jaime G Carbonell, George Doddington, Jonathan Yamron, and Yiming Yang. 1998. Topic detection and tracking pilot study final
report. In Proceedings of the DARPA Broadcast News Transcription and Understanding Workshop. Virginia, US, pages 194–218.
Robert Bamler and Stephan Mandt. 2017. Dynamic word embeddings. In Proceedings of the 34th International Conference on Machine Learning.
Sydney, Australia, pages 380–389.
Pierpaolo Basile, Annalina Caputo, and Giovanni Semeraro. 2014. Analysing word meaning over time by exploiting temporal random indexing. In
Proceedings of the 1st Italian Conference on Computational Linguistics (CLiC-it). Pisa University Press, Pisa, Italy.
David M. Blei and John D. Lafferty. 2006. Dynamic topic models. In Proceedings of the 23rd International Conference on Machine Learning.
Association for Computing Machinery, Pittsburgh, Pennsylvania USA, pages 113–120.
David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent dirichlet allocation. Proceedings of Machine Learning Research 3(Jan):993–1022.
Intern © Siemens AG 2017
May 2017Seite 42 Corporate Technology
David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent dirichlet allocation. Proceedings of Machine Learning Research 3(Jan):993–1022.
Nicolas Boulanger-Lewandowski, Yoshua Bengio, and Pascal Vincent. 2012. Modeling temporal dependencies in high-dimensional sequences:
Application to polyphonic music generation and transcription. In Proceedings of the 29th International Conference on Machine Learning. Edinburgh,
Scotland UK.
Jonathan Chang, Sean Gerrish, Chong Wang, Jordan L. Boyd-Graber, and David M. Blei. 2009. Reading tea leaves: How humans interpret topic
models. In Advances in Neural Information Processing Systems. Curran Associates, Inc., Vancouver, Canada, pages 288–296.
Rajarshi Das, Manzil Zaheer, and Chris Dyer. 2015. Gaussian lda for topic models with word embeddings. In Proceedings of the 53rd Annual
Meeting of the Association for Computational Linguistics and the 7th IJCNLP. ACL, Beijing, China, volume 1, pages 795–804.
References
Peter V. Gehler, Alex D. Holub, and Max Welling. 2006. The rate adapting poisson model for information retrieval and object recognition. In
Proceedings of the 23rd International Conference on Machine Learning. Association for Computing Machinery, Pittsburgh, Pennsylvania USA, pages
337–344.
Sujatha Das Gollapalli and Xiaoli Li. 2015. Emnlp versus acl: Analyzing nlp research over time. In Proceedings of the Conference on Empirical
Methods in Natural Language Processing. Association for Computational Linguistics, Lisbon, Portugal, pages 2002–2006.Methods in Natural Language Processing. Association for Computational Linguistics, Lisbon, Portugal, pages 2002–2006.
Pankaj Gupta, Thomas Runkler, Heike Adel, Bernt Andrassy, Hans-Georg Zimmermann, and Hinrich Schutze. 2015a. Deep learning methods for the
extraction of relations in natural language text. Technical report, Technical University of Munich, Germany.
Pankaj Gupta, Thomas Runkler, and Bernt Andrassy. 2015b. Keyword learning for classifying requirements in tender documents. Technical report,
Technical University of Munich, Germany.
Pankaj Gupta, Hinrich Schutze, and Bernt Andrassy. 2016. Table filling multi-task recurrent neural network for joint entity and relation extraction. In
Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers. Osaka, Japan, pages 2537–2547.
Pankaj Gupta, Udhayaraj Sivalingam, Sebastian Polsterl, and Nassir Navab. 2015c. Identifying patients with diabetes using discriminative restricted
boltzmann machines. Technical report, Technical University of Munich, Germany.
David Hall, Daniel Jurafsky, and Christopher D. Manning. 2008. Studying the history of ideas using topic models. In Proceedings of the conference
Intern © Siemens AG 2017
May 2017Seite 43 Corporate Technology
David Hall, Daniel Jurafsky, and Christopher D. Manning. 2008. Studying the history of ideas using topic models. In Proceedings of the conference
on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Honolulu, Hawaii, pages 363–371.
Geoffrey Hinton and Ruslan Salakhutdinov. 2009. Replicated softmax: an undirected topic model. In Advances in Neural Information Processing
Systems 22. Curran Associates, Inc., Vancouver, Canada, pages 1607–1614.
Geoffrey E. Hinton. 2002. Training products of experts by minimizing contrastive divergence. Neural Computation 14(8):1771–1800.
Geoffrey E. Hinton, Simon Osindero, and Yee-Whye Teh. 2006. A fast learning algorithm for deep belief nets. Neural Computation 18(7):1527–1554.
Tomoharu Iwata, Takeshi Yamada, Yasushi Sakurai, and Naonori Ueda. 2010. Online multiscale dynamic topic models. In Proceedings of the 16th
ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, Washington DC, USA, pages 663–672.
References
David Newman, Sarvnaz Karimi, and Lawrence Cavedon. 2009. External evaluation of topic models. In Proceedings of the 14th Australasian
Document Computing Symposium. Citeseer, Sydney, Australia.
Iulian Pruteanu-Malinici, Lu Ren, John Paisley, Eric Wang, and Lawrence Carin. 2010. Hierarchical bayesian modeling of topics in time-stamped
documents. IEEE transactions on pattern analysis and machine intelligence 32(6):996–1011.
Maja Rudolph and David Blei. 2018. Dynamic bernoulli embeddings for language evolution. In Proceedings of the 27th International Conference on
World Wide Web Companion.
Lyon, France. David E. Rumelhart, Geoffrey E. Hinton, and Ronald J. Williams. 1985. Learning internal representations by error propagation.
Technical report, California Univ San Diego La Jolla Inst for Cognitive Science.
Ankan Saha and Vikas Sindhwani. 2012. Learning evolving and emerging topics in social media: a dynamic nmf approach with temporal
regularization. In Proceedings of the 5th ACM International Conference on Web Search and Data Mining. Association for Computing Machinery,
Seattle, Washington USA, pages 693–702. Aaron Schein, Hanna Wallach, and Mingyuan Zhou. 2016.
Poisson-gamma dynamical systems. In Advances in Neural Information Processing Systems 29, Curran Associates, Inc., Barcelona, Spain, pages
5005–5013. Paul Smolensky. 1986. Information processing in dynamical systems: Foundations of harmony theory. Technical report, Colorado
University at Boulder Department of Computer Science.
Intern © Siemens AG 2017
May 2017Seite 44 Corporate Technology
University at Boulder Department of Computer Science.
Ilya Sutskever, Geoffrey E. Hinton, and Graham W. Taylor. 2009. The recurrent temporal restricted boltzmann machine. In Advances in Neural
Information Processing Systems 22. Curran Associates, Inc., Vancouver, Canada, pages 1601–1608.
Graham W. Taylor, Geoffrey E. Hinton, and Sam T. Roweis. 2007. Modeling human motion using binary latent variables. In Advances in Neural
Information Processing Systems 20. Curran Associates, Inc., Vancouver, Canada, pages 1345–1352.
Ngoc Thang Vu, Heike Adel, Pankaj Gupta, and Hinrich Schutze. 2016a. Combining recurrent and convolutional neural networks for relation
classification. In Proceedings of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.
Association for Computational Linguistics, San Diego, California USA, pages 534–539.
References
Ngoc Thang Vu, Pankaj Gupta, Heike Adel, and Hinrich Schutze. 2016b. Bi-directional recurrent neural network with ranking loss for spoken
language understanding. In Proceedings of the Acoustics, Speech and Signal Processing (ICASSP). IEEE, Shanghai, China, pages 6060–6064.
Xiaojun Wan and Jianguo Xiao. 2008. Single document keyphrase extraction using neighborhood knowledge. In Proceedings of the 23rd National
Conference on Artificial Intelligence. Chicago, Illinois USA, volume 8, pages 855–860.
Xuerui Wang and Andrew McCallum. 2006. Topics over time: a non-markov continuous-time model of topical trends. In Proceedings of the 12th
ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Association for Computing Machinery, Philadelphia, Pennsylvania
USA, pages 424–433.
Eric P. Xing, Rong Yan, and Alexander G. Hauptmann. 2005. Mining associated text and images with dual-wing harmoniums. In Proceedings of the
21st Conference on Uncertainty in Artificial Intelligence. AUAI Press, Edinburgh, Scotland UK.
Zijun Yao, Yifan Sun, Weicong Ding, Nikhil Rao, and Hui Xiong. 2018. Dynamic word embeddings for evolving semantic discovery. In Proceedings of
the 11th ACM International Conference on Web Search and Data Mining (WSDM). Association for Computing Machinery, Los Angeles, California USA,
pages 673–681.
Intern © Siemens AG 2017
May 2017Seite 45 Corporate Technology

More Related Content

Similar to RNN-RSM Model for Tracking Topical Trends over Time

Building ASTM based tools for multi-language reverse engineering
Building ASTM based tools for multi-language reverse engineeringBuilding ASTM based tools for multi-language reverse engineering
Building ASTM based tools for multi-language reverse engineeringMichael Moser
 
Semantic opinion mining ontology
Semantic opinion mining ontologySemantic opinion mining ontology
Semantic opinion mining ontologyMohammed Almashraee
 
Distributed Multi-device Execution of TensorFlow – an Outlook
Distributed Multi-device Execution of TensorFlow – an OutlookDistributed Multi-device Execution of TensorFlow – an Outlook
Distributed Multi-device Execution of TensorFlow – an OutlookSebnem Rusitschka
 
Analytics with IMS Assets - 2017
Analytics with IMS Assets - 2017Analytics with IMS Assets - 2017
Analytics with IMS Assets - 2017Helene Lyon
 
Elastic SQL Database: Oxymoron or Emerging Reality? (Database Month, June 2017)
Elastic SQL Database: Oxymoron or Emerging Reality? (Database Month, June 2017)Elastic SQL Database: Oxymoron or Emerging Reality? (Database Month, June 2017)
Elastic SQL Database: Oxymoron or Emerging Reality? (Database Month, June 2017)NuoDB
 
CWIN 17 / sessions data vault modeling - f2-f - nishat gupta
CWIN 17 / sessions data vault modeling -  f2-f - nishat guptaCWIN 17 / sessions data vault modeling -  f2-f - nishat gupta
CWIN 17 / sessions data vault modeling - f2-f - nishat guptaCapgemini
 
ThyssenKrupp's Procurement Transformation: Minimizing Direct Materials Contra...
ThyssenKrupp's Procurement Transformation: Minimizing Direct Materials Contra...ThyssenKrupp's Procurement Transformation: Minimizing Direct Materials Contra...
ThyssenKrupp's Procurement Transformation: Minimizing Direct Materials Contra...SAP Ariba
 
Improving Python and Spark Performance and Interoperability with Apache Arrow...
Improving Python and Spark Performance and Interoperability with Apache Arrow...Improving Python and Spark Performance and Interoperability with Apache Arrow...
Improving Python and Spark Performance and Interoperability with Apache Arrow...Databricks
 
Ui5con virtual tour_around_a_company_site_with_ui5
Ui5con virtual tour_around_a_company_site_with_ui5Ui5con virtual tour_around_a_company_site_with_ui5
Ui5con virtual tour_around_a_company_site_with_ui5Helmut Tammen
 
LavaCon 2017 - Implementing a Customer Driven Transition to DITA - Part 1
LavaCon 2017 - Implementing a Customer Driven Transition to DITA - Part 1LavaCon 2017 - Implementing a Customer Driven Transition to DITA - Part 1
LavaCon 2017 - Implementing a Customer Driven Transition to DITA - Part 1Jack Molisani
 
Is it sensible to use Data Vault at all? Conclusions from a project.
Is it sensible to use Data Vault at all? Conclusions from a project.Is it sensible to use Data Vault at all? Conclusions from a project.
Is it sensible to use Data Vault at all? Conclusions from a project.Capgemini
 
Schema Engineering for Enterprise Knowledge Graphs
Schema Engineering for Enterprise Knowledge GraphsSchema Engineering for Enterprise Knowledge Graphs
Schema Engineering for Enterprise Knowledge GraphsVera G. Meister
 
Programming the Semantic Web
Programming the Semantic WebProgramming the Semantic Web
Programming the Semantic WebSteffen Staab
 
Kesseler - New Ways for Automation in CCMS
Kesseler - New Ways for Automation in CCMSKesseler - New Ways for Automation in CCMS
Kesseler - New Ways for Automation in CCMSLavaCon
 
Spectroscopy Ninja presents Spectragryph
Spectroscopy Ninja presents SpectragryphSpectroscopy Ninja presents Spectragryph
Spectroscopy Ninja presents SpectragryphFriedrich Menges
 
NGD Systems and Microsoft Keynote Presentation at IPDPS MPP in Vacouver
NGD Systems and Microsoft Keynote Presentation at IPDPS MPP in VacouverNGD Systems and Microsoft Keynote Presentation at IPDPS MPP in Vacouver
NGD Systems and Microsoft Keynote Presentation at IPDPS MPP in VacouverScott Shadley, MBA,PMC-III
 
Improving Python and Spark Performance and Interoperability with Apache Arrow
Improving Python and Spark Performance and Interoperability with Apache ArrowImproving Python and Spark Performance and Interoperability with Apache Arrow
Improving Python and Spark Performance and Interoperability with Apache ArrowJulien Le Dem
 

Similar to RNN-RSM Model for Tracking Topical Trends over Time (20)

NLP@DATEV: Setting up a domain specific language model, Dr. Jonas Rende & Tho...
NLP@DATEV: Setting up a domain specific language model, Dr. Jonas Rende & Tho...NLP@DATEV: Setting up a domain specific language model, Dr. Jonas Rende & Tho...
NLP@DATEV: Setting up a domain specific language model, Dr. Jonas Rende & Tho...
 
Building ASTM based tools for multi-language reverse engineering
Building ASTM based tools for multi-language reverse engineeringBuilding ASTM based tools for multi-language reverse engineering
Building ASTM based tools for multi-language reverse engineering
 
Neo4EMF eclipsecon 2013
Neo4EMF eclipsecon 2013Neo4EMF eclipsecon 2013
Neo4EMF eclipsecon 2013
 
Semantic opinion mining ontology
Semantic opinion mining ontologySemantic opinion mining ontology
Semantic opinion mining ontology
 
Distributed Multi-device Execution of TensorFlow – an Outlook
Distributed Multi-device Execution of TensorFlow – an OutlookDistributed Multi-device Execution of TensorFlow – an Outlook
Distributed Multi-device Execution of TensorFlow – an Outlook
 
Analytics with IMS Assets - 2017
Analytics with IMS Assets - 2017Analytics with IMS Assets - 2017
Analytics with IMS Assets - 2017
 
Elastic SQL Database: Oxymoron or Emerging Reality? (Database Month, June 2017)
Elastic SQL Database: Oxymoron or Emerging Reality? (Database Month, June 2017)Elastic SQL Database: Oxymoron or Emerging Reality? (Database Month, June 2017)
Elastic SQL Database: Oxymoron or Emerging Reality? (Database Month, June 2017)
 
CWIN 17 / sessions data vault modeling - f2-f - nishat gupta
CWIN 17 / sessions data vault modeling -  f2-f - nishat guptaCWIN 17 / sessions data vault modeling -  f2-f - nishat gupta
CWIN 17 / sessions data vault modeling - f2-f - nishat gupta
 
ThyssenKrupp's Procurement Transformation: Minimizing Direct Materials Contra...
ThyssenKrupp's Procurement Transformation: Minimizing Direct Materials Contra...ThyssenKrupp's Procurement Transformation: Minimizing Direct Materials Contra...
ThyssenKrupp's Procurement Transformation: Minimizing Direct Materials Contra...
 
Improving Python and Spark Performance and Interoperability with Apache Arrow...
Improving Python and Spark Performance and Interoperability with Apache Arrow...Improving Python and Spark Performance and Interoperability with Apache Arrow...
Improving Python and Spark Performance and Interoperability with Apache Arrow...
 
Salcedo BSI and ISO STS
Salcedo BSI and ISO STSSalcedo BSI and ISO STS
Salcedo BSI and ISO STS
 
Ui5con virtual tour_around_a_company_site_with_ui5
Ui5con virtual tour_around_a_company_site_with_ui5Ui5con virtual tour_around_a_company_site_with_ui5
Ui5con virtual tour_around_a_company_site_with_ui5
 
LavaCon 2017 - Implementing a Customer Driven Transition to DITA - Part 1
LavaCon 2017 - Implementing a Customer Driven Transition to DITA - Part 1LavaCon 2017 - Implementing a Customer Driven Transition to DITA - Part 1
LavaCon 2017 - Implementing a Customer Driven Transition to DITA - Part 1
 
Is it sensible to use Data Vault at all? Conclusions from a project.
Is it sensible to use Data Vault at all? Conclusions from a project.Is it sensible to use Data Vault at all? Conclusions from a project.
Is it sensible to use Data Vault at all? Conclusions from a project.
 
Schema Engineering for Enterprise Knowledge Graphs
Schema Engineering for Enterprise Knowledge GraphsSchema Engineering for Enterprise Knowledge Graphs
Schema Engineering for Enterprise Knowledge Graphs
 
Programming the Semantic Web
Programming the Semantic WebProgramming the Semantic Web
Programming the Semantic Web
 
Kesseler - New Ways for Automation in CCMS
Kesseler - New Ways for Automation in CCMSKesseler - New Ways for Automation in CCMS
Kesseler - New Ways for Automation in CCMS
 
Spectroscopy Ninja presents Spectragryph
Spectroscopy Ninja presents SpectragryphSpectroscopy Ninja presents Spectragryph
Spectroscopy Ninja presents Spectragryph
 
NGD Systems and Microsoft Keynote Presentation at IPDPS MPP in Vacouver
NGD Systems and Microsoft Keynote Presentation at IPDPS MPP in VacouverNGD Systems and Microsoft Keynote Presentation at IPDPS MPP in Vacouver
NGD Systems and Microsoft Keynote Presentation at IPDPS MPP in Vacouver
 
Improving Python and Spark Performance and Interoperability with Apache Arrow
Improving Python and Spark Performance and Interoperability with Apache ArrowImproving Python and Spark Performance and Interoperability with Apache Arrow
Improving Python and Spark Performance and Interoperability with Apache Arrow
 

More from Pankaj Gupta, PhD

textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...Pankaj Gupta, PhD
 
Neural NLP Models of Information Extraction
Neural NLP Models of Information ExtractionNeural NLP Models of Information Extraction
Neural NLP Models of Information ExtractionPankaj Gupta, PhD
 
Poster: Neural Relation ExtractionWithin and Across Sentence Boundaries
Poster: Neural Relation ExtractionWithin and Across Sentence BoundariesPoster: Neural Relation ExtractionWithin and Across Sentence Boundaries
Poster: Neural Relation ExtractionWithin and Across Sentence BoundariesPankaj Gupta, PhD
 
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...Pankaj Gupta, PhD
 
Neural Relation ExtractionWithin and Across Sentence Boundaries
Neural Relation ExtractionWithin and Across Sentence BoundariesNeural Relation ExtractionWithin and Across Sentence Boundaries
Neural Relation ExtractionWithin and Across Sentence BoundariesPankaj Gupta, PhD
 
Deep Learning for Information Extraction in Natural Language Text
Deep Learning for Information Extraction in Natural Language TextDeep Learning for Information Extraction in Natural Language Text
Deep Learning for Information Extraction in Natural Language TextPankaj Gupta, PhD
 
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...Pankaj Gupta, PhD
 
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...Pankaj Gupta, PhD
 

More from Pankaj Gupta, PhD (9)

textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
 
Neural NLP Models of Information Extraction
Neural NLP Models of Information ExtractionNeural NLP Models of Information Extraction
Neural NLP Models of Information Extraction
 
Poster: Neural Relation ExtractionWithin and Across Sentence Boundaries
Poster: Neural Relation ExtractionWithin and Across Sentence BoundariesPoster: Neural Relation ExtractionWithin and Across Sentence Boundaries
Poster: Neural Relation ExtractionWithin and Across Sentence Boundaries
 
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
 
Neural Relation ExtractionWithin and Across Sentence Boundaries
Neural Relation ExtractionWithin and Across Sentence BoundariesNeural Relation ExtractionWithin and Across Sentence Boundaries
Neural Relation ExtractionWithin and Across Sentence Boundaries
 
Deep Learning for Information Extraction in Natural Language Text
Deep Learning for Information Extraction in Natural Language TextDeep Learning for Information Extraction in Natural Language Text
Deep Learning for Information Extraction in Natural Language Text
 
Pankaj Gupta CV / Resume
Pankaj Gupta CV / ResumePankaj Gupta CV / Resume
Pankaj Gupta CV / Resume
 
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
 
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
 

Recently uploaded

VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PPRINCE C P
 
Physiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptxPhysiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptxAArockiyaNisha
 
Natural Polymer Based Nanomaterials
Natural Polymer Based NanomaterialsNatural Polymer Based Nanomaterials
Natural Polymer Based NanomaterialsAArockiyaNisha
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfSumit Kumar yadav
 
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls AgencyHire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls AgencySheetal Arora
 
Animal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxAnimal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxUmerFayaz5
 
Pests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdfPests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdfPirithiRaju
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)Areesha Ahmad
 
Botany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questionsBotany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questionsSumit Kumar yadav
 
Hubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroidsHubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroidsSérgio Sacani
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Lokesh Kothari
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPirithiRaju
 
fundamental of entomology all in one topics of entomology
fundamental of entomology all in one topics of entomologyfundamental of entomology all in one topics of entomology
fundamental of entomology all in one topics of entomologyDrAnita Sharma
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​kaibalyasahoo82800
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...RohitNehra6
 
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...ssifa0344
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfSumit Kumar yadav
 
GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)Areesha Ahmad
 
Chromatin Structure | EUCHROMATIN | HETEROCHROMATIN
Chromatin Structure | EUCHROMATIN | HETEROCHROMATINChromatin Structure | EUCHROMATIN | HETEROCHROMATIN
Chromatin Structure | EUCHROMATIN | HETEROCHROMATINsankalpkumarsahoo174
 

Recently uploaded (20)

VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C P
 
Physiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptxPhysiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptx
 
Natural Polymer Based Nanomaterials
Natural Polymer Based NanomaterialsNatural Polymer Based Nanomaterials
Natural Polymer Based Nanomaterials
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdf
 
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls AgencyHire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
 
Animal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxAnimal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptx
 
Pests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdfPests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdf
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)
 
Botany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questionsBotany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questions
 
Hubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroidsHubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroids
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
 
fundamental of entomology all in one topics of entomology
fundamental of entomology all in one topics of entomologyfundamental of entomology all in one topics of entomology
fundamental of entomology all in one topics of entomology
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...
 
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdf
 
GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)
 
Chromatin Structure | EUCHROMATIN | HETEROCHROMATIN
Chromatin Structure | EUCHROMATIN | HETEROCHROMATINChromatin Structure | EUCHROMATIN | HETEROCHROMATIN
Chromatin Structure | EUCHROMATIN | HETEROCHROMATIN
 
The Philosophy of Science
The Philosophy of ScienceThe Philosophy of Science
The Philosophy of Science
 

RNN-RSM Model for Tracking Topical Trends over Time

  • 1. Deep Temporal-Recurrent-Replicated- Softmax for Topical Trends over Time Intern © Siemens AG 2017 Softmax for Topical Trends over Time Author(s): Pankaj Gupta1,2, Subbu Ram2, Bernt Andrassy2, Hinrich Schütze1 Presenter: Pankaj Gupta @NAACL/ACL New Orleans, USA. 3rd June 2018 1CIS, University of Munich (LMU) | 2Machine Intelligence, Siemens AG | June 2018
  • 2. Outline  Problem Statement  Background Background - Topic Modeling: Static Vs Dynamic - Replicated Softmax Models (RSM)  Proposal: The RNN-RSM model  Evaluations Intern © Siemens AG 2017 May 2017Seite 2 Corporate Technology  Evaluations - Generalization - Topic Evolution and Characterization
  • 3. Problem Statement: Modeling Science for Topical Overview Timelines 201420081996 Intern © Siemens AG 2017 May 2017Seite 3 Corporate Technology
  • 4. Problem Statement: Modeling Science for Topical Overview Timelines Topic Structure Detection (TSD):  Identify the main topics in document collection  Discover hidden structure that explains the document Discover hidden structure that explains the document Intern © Siemens AG 2017 May 2017Seite 4 Corporate Technology
  • 5. Problem Statement: Automatic Topical Overview Timelines Topic Structure Detection (TSD):  Identify the main topics in document collection  Discover hidden structure that explains the document Discover hidden structure that explains the document Topic Evolution Detection (TED):  Recognize how the underlying topics evolves (grows or decays) over time  Detect the emergence of a new topic Intern © Siemens AG 2017 May 2017Seite 5 Corporate Technology
  • 6. Problem Statement: Automatic Topical Overview Timelines Topic Structure Detection (TSD):  Identify the main topics in document collection  Discover hidden structure that explains the document Discover hidden structure that explains the document Topic Evolution Detection (TED):  Recognize how the underlying topics evolves (grows or decays) over time  Detect the emergence of a new topic Intern © Siemens AG 2017 May 2017Seite 6 Corporate Technology Temporal Topic Characterization (TTC):  Characterize topics to track the words’ usage within each topic over time  Topical/Keyword trend analysis for word evolution
  • 7. Problem Statement: Automatic Topical Overview Timelines Topic Structure Detection (TSD):  Identify the main topics in document collection  Discover hidden structure that explains the document Discover hidden structure that explains the document Topic Evolution Detection (TED):  Recognize how the underlying topics evolves (grows or decays) over time  Detect the emergence of a new topic Intern © Siemens AG 2017 May 2017Seite 7 Corporate Technology Temporal Topic Characterization (TTC):  Characterize topics to track the words’ usage within each topic over time  Topical/Keyword trend analysis for word evolution
  • 8. Background: Static Topic Model (STM) Topic model:  Statistical Generative modeling  Discover abstract topics in document collection Discover abstract topics in document collection Intuition:  Documents exhibit multiple topics  Each word drawn from one of the topics Intern © Siemens AG 2017 May 2017Seite 8 Corporate Technology
  • 9. Background: Static Topic Model (STM) Topic model:  Statistical Generative modeling  Discover abstract topics in document collection Discover abstract topics in document collection Intuition:  Documents exhibit multiple topics  Each word drawn from one of the topics Assumption:  Topics shared by all documents Intern © Siemens AG 2017 May 2017Seite 9 Corporate Technology  Document drawn from the same topics  No temporal order of documents  Documents are exchangeable
  • 10. Background: Dynamic Topic Model (DTM)  Model the temporal order of documents  Capture evolution of topics in temporally organized document collections over time Capture evolution of topics in temporally organized document collections over time  For example in scholarly articles, the research evolved over time Intern © Siemens AG 2017 May 2017Seite 10 Corporate Technology
  • 11. Background: Dynamic Topic Model DTM (Blei and Lafferty (2006))  Temporal stack of LDA models Intern © Siemens AG 2017 May 2017Seite 11 Corporate Technology
  • 12. Background: Dynamic Topic Model DTM (Blei and Lafferty (2006))  Temporal stack of LDA models Assumption in DTMAssumption in DTM  Assume k unique topics over time, model their evolution  Can not model appearance of new topicsdisappearance of old topics over time (Blei and Lafferty (2006)) Intern © Siemens AG 2017 May 2017Seite 12 Corporate Technology
  • 13. Background: Dynamic Topic Model DTM (Blei and Lafferty (2006))  Temporal stack of LDA models Assumption in DTMAssumption in DTM  Assume k unique topics over time, model their evolution  Can not model appearance of new topicsdisappearance of old topics over time (Blei and Lafferty (2006)) Intern © Siemens AG 2017 May 2017Seite 13 Corporate Technology Challenges:  How new topics appear/disappear over time ?  Explicit/long term topical dependencies ?  Heteroscedasticity: Modeling high dimensional sequence ?
  • 14. Background: Replicated Softmax Model (RSM)  Generative Model of Word Counts  Family of different-sized RBMs (Restricted Boltzmann Machine)  Energy based undirected static topic model Energy based undirected static topic model Intern © Siemens AG 2017 May 2017Seite 14 Corporate Technology
  • 15. Background: Replicated Softmax Model (RSM)  Generative Model of Word Counts  Family of different-sized RBMs (Restricted Boltzmann Machine)  Energy based undirected static topic model Energy based undirected static topic model Intern © Siemens AG 2017 May 2017Seite 15 Corporate Technology
  • 16. Background: Replicated Softmax Model (RSM)  Generative Model of Word Counts  Family of different-sized RBMs (Restricted Boltzmann Machine)  Energy based undirected static topic model Energy based undirected static topic model Intern © Siemens AG 2017 May 2017Seite 16 Corporate Technology
  • 17. Proposed Model: Recurrent Neural Network-Replicated Softmax (RNN-RSM) h(1) bh (1) h(2) bh (2) h(T-1) bh (T-1) h(T) bh (T) RSM RSM RSM RSM … Latent Topics V(1) bh bv (1) Wvh u(1) Wvu V(2) bh bv (2) Wvh u(2) Wvu V(T-1) bh bv (T-1) Wvh u(T-1) Wvu u(0) V(T) bh bv (T) Wvh u(T) Wvu Wuh Wuv Wuv Wuv Wuh Wuh Wuh Wuv RNN Wuu Wuu Wuu Wuu Observable Softmax Visibles  Neural Dynamic Topic Model Intern © Siemens AG 2017 May 2017Seite 17 Corporate Technology  Neural Dynamic Topic Model  Temporal Sequence of distributional estimators (RSMs)  Temporal Feedback connections spanning timeline to model dynamics in hidden topics
  • 18. Proposed Model: RNN-RSM h(1) bh (1) h(2) bh (2) h(T-1) bh (T-1) h(T) bh (T) RSM RSM RSM RSM … Latent Topics V(1) bh bv (1) Wvh u(1) Wvu V(2) bh bv (2) Wvh u(2) Wvu V(T-1) bh bv (T-1) Wvh u(T-1) Wvu u(0) V(T) bh bv (T) Wvh u(T) Wvu Wuh Wuv Wuv Wuv Wuh Wuh Wuh Wuv RNN Wuu Wuu Wuu Wuu Observable Softmax Visibles  Document collections over time Intern © Siemens AG 2017 May 2017Seite 18 Corporate Technology  Document collections over time  Input to each of the RSMs
  • 19. Proposed Model: RNN-RSM h(1) bh (1) h(2) bh (2) h(T-1) bh (T-1) h(T) bh (T) RSM RSM RSM RSM … Latent Topics V(1) bh bv (1) Wvh u(1) Wvu V(2) bh bv (2) Wvh u(2) Wvu V(T-1) bh bv (T-1) Wvh u(T-1) Wvu u(0) V(T) bh bv (T) Wvh u(T) Wvu Wuh Wuv Wuv Wuv Wuh Wuh Wuh Wuv RNN Wuu Wuu Wuu Wuu Observable Softmax Visibles  Document collections over time Intern © Siemens AG 2017 May 2017Seite 19 Corporate Technology  Document collections over time  Input to each of the RSMs  Hidden/latent topics over time, in each of the RSMs
  • 20. Training RNN-RSM: Forward Propagation Through Time h(1) bh (1) h(2) bh (2) h(T-1) bh (T-1) h(T) bh (T) RSM RSM RSM RSM … Latent Topics V(1) bh bv (1) Wvh u(1) Wvu V(2) bh bv (2) Wvh u(2) Wvu V(T-1) bh bv (T-1) Wvh u(T-1) Wvu u(0) V(T) bh bv (T) Wvh u(T) Wvu Wuh Wuv Wuv Wuv Wuh Wuh Wuh Wuv RNN Wuu Wuu Wuu Wuu Observable Softmax Visibles Propagate u(t) in the RNN portion Intern © Siemens AG 2017 May 2017Seite 20 Corporate Technology Propagate u(t) in the RNN portion
  • 21. Training RNN-RSM: Forward Propagation Through Time h(1) bh (1) h(2) bh (2) h(T-1) bh (T-1) h(T) bh (T) RSM RSM RSM RSM … Latent Topics V(1) bh bv (1) Wvh u(1) Wvu V(2) bh bv (2) Wvh u(2) Wvu V(T-1) bh bv (T-1) Wvh u(T-1) Wvu u(0) V(T) bh bv (T) Wvh u(T) Wvu Wuh Wuv Wuv Wuv Wuh Wuh Wuh Wuv RNN Wuu Wuu Wuu Wuu Observable Softmax Visibles Compute RSM biases at each time step, t Intern © Siemens AG 2017 May 2017Seite 21 Corporate Technology Compute RSM biases at each time step, t
  • 22. Training RNN-RSM: Forward Propagation Through Time h(1) bh (1) h(2) bh (2) h(T-1) bh (T-1) h(T) bh (T) RSM RSM RSM RSM … Latent Topics V(1) bh bv (1) Wvh u(1) Wvu V(2) bh bv (2) Wvh u(2) Wvu V(T-1) bh bv (T-1) Wvh u(T-1) Wvu u(0) V(T) bh bv (T) Wvh u(T) Wvu Wuh Wuv Wuv Wuv Wuh Wuh Wuh Wuv RNN Wuu Wuu Wuu Wuu … Observable Softmax Visibles Compute the Conditional distributions in each RSM Intern © Siemens AG 2017 May 2017Seite 22 Corporate Technology Compute the Conditional distributions in each RSM
  • 23. Training RNN-RSM: Forward Propagation Through Time h(1) bh (1) h(2) bh (2) h(T-1) bh (T-1) h(T) bh (T) RSM RSM RSM RSM … Latent Topics V(1) bh bv (1) Wvh u(1) Wvu V(2) bh bv (2) Wvh u(2) Wvu V(T-1) bh bv (T-1) Wvh u(T-1) Wvu u(0) V(T) bh bv (T) Wvh u(T) Wvu Wuh Wuv Wuv Wuv Wuh Wuh Wuh Wuv RNN Wuu Wuu Wuu Wuu … Observable Softmax Visibles Neural Network Neural Network Neural NetworkNeural Network Neural Network Neural Network Topic-words Intern © Siemens AG 2017 May 2017Seite 23 Corporate Technology Neural Network Language Models Word Representation Linear Model Rule Set Neural Network Language Models Supervised Linear Model Rule Set Neural Network Language Models Word Embedding Word Embeddings Word Representation Neural Network Language Models Word Representation Linear Model Rule Set Neural Network Language Models Supervised Linear Model Rule Set Neural Network Language Models Word Embedding Word Embeddings Word Representation Topic-words over time for topic ‘Word Vector’ 1996 1997 2014
  • 24. Training RNN-RSM: Backpropagation Through Time (BPTT) h(1) bh (1) h(2) bh (2) h(T-1) bh (T-1) h(T) bh (T) RSM RSM RSM RSM … Latent Topics V(1) bh bv (1) Wvh u(1) Wvu V(2) bh bv (2) Wvh u(2) Wvu V(T-1) bh bv (T-1) Wvh u(T-1) Wvu u(0) V(T) bh bv (T) Wvh u(T) Wvu Wuh Wuv Wuv Wuv Wuh Wuh Wuh Wuv RNN Wuu Wuu Wuu Wuu … Observable Softmax Visibles Cost in RNN-RSM, the negative log-likelihood Intern © Siemens AG 2017 May 2017Seite 24 Corporate Technology Cost in RNN-RSM, the negative log-likelihood
  • 25. Summary: RNN-RSM Offerings  Unsupervised Neural Dynamic Topic Model  Deterministic (RNN) + Stochastic (RSM) network Deterministic (RNN) + Stochastic (RSM) network  Longer-term topical dependencies via RNN  Drift/bias terms to propagate topical information  New topics to evolve over time via sequence of RSMs - Separate hidden vectors to capture and convey topics - Topics over time are independent Intern © Siemens AG 2017 May 2017Seite 25 Corporate Technology - Topics over time are independent
  • 26. Evaluation: Dataset and Experimental Setup Data Set:  EMNLP and ACL conference papers (Gollapalli and Li 2015) from the year 1996 -2014  Expand Rank (Wan and Xiao 2008) to extract top 100 key-phrases for each paper (unigrams + bigrams)  Dictionary of all unigrams and bigrams, dictionary size, K = 3390 Intern © Siemens AG 2017 May 2017Seite 26 Corporate Technology
  • 27. Evaluation: Dataset and Experimental Setup Data Set:  EMNLP and ACL conference papers (Gollapalli and Li 2015) from the year 1996 -2014  Expand Rank (Wan and Xiao 2008) to extract top 100 key-phrases for each paper (unigrams + bigrams)  Dictionary of all unigrams and bigrams, dictionary size, K = 3390 Experimental Setup: Intern © Siemens AG 2017 May 2017Seite 27 Corporate Technology  Evaluate RNN-RSM (dynamic) with LDA-DTM (dynamic), seqRSM (static) and seqLDA (static)  seqRSM and seqLDA: Individual 19 different RSM/LDA models trained for each year  LDA-DTM and RNN-RSM are trained over the years with 19 time steps  Perplexity (PPL) to choose the number of optimal topics (=30)
  • 28. Evaluation (Quantitative): Dynamic Topic Models as Generative Models  Average perplexity per word of the document collections over time Intern © Siemens AG 2017 May 2017Seite 28 Corporate Technology  Held-out documents: 10 documents from each time step i.e. total 190 documents  Compute perplexity for 30 topics
  • 29. Evaluation (Qualitative): Topic Detection and Evolution  Extract the top 20 terms for every 30 topic set from 1996-2014 for the topic models,  Resulting in |Q|_max = 19*30*20 possible topic terms Topic count Topic-terms count Time steps Intern © Siemens AG 2017 May 2017Seite 29 Corporate Technology
  • 30. Evaluation (Qualitative): Topic Detection and Evolution  Extract the top 20 terms for every 30 topic set from 1996-2014 for the topic models,  Resulting in |Q|_max = 19*30*20 possible topic terms  Topic Popularity (Sentiment Analysis, Word Vector and Dependency Parsing in NLP research) Topic Popularity (Sentiment Analysis, Word Vector and Dependency Parsing in NLP research) Intern © Siemens AG 2017 May 2017Seite 30 Corporate Technology
  • 31. Evaluation (Qualitative): Topic Detection and Evolution  Extract the top 20 terms for every 30 topic set from 1996-2014 for the topic models,  Resulting in |Q|_max = 19*30*20 possible topic terms  Topic Popularity (Sentiment Analysis, Word Vector and Dependency Parsing in NLP research) Topic Popularity (Sentiment Analysis, Word Vector and Dependency Parsing in NLP research) Intern © Siemens AG 2017 May 2017Seite 31 Corporate Technology
  • 32. Evaluation: Topic Characterization (Trending Keywords over time)  Analyse word evolution/usage in the complete topic sets over time (19*30*20 possible topic terms) Keyword-trend: Appearance/disappearance of a keyword in underlying topics detected over time Span: Length of the longest sequence of the keyword appearance in its keyword trend Intern © Siemens AG 2017 May 2017Seite 32 Corporate Technology Span: Length of the longest sequence of the keyword appearance in its keyword trend
  • 33. Evaluation: Topic Characterization (Topic-term Usage over Time) Intern © Siemens AG 2017 May 2017Seite 33 Corporate Technology
  • 34. Evaluation: Topic Characterization (Topic-term Usage over Time) Intern © Siemens AG 2017 May 2017Seite 34 Corporate Technology
  • 35. Evaluation: Topic Characterization (Topic-term Usage over Time) Intern © Siemens AG 2017 May 2017Seite 35 Corporate Technology
  • 36. Evaluation: Topic Characterization (Topic-term Usage over Time) Intern © Siemens AG 2017 May 2017Seite 36 Corporate Technology
  • 37. Evaluation: Topic Characterization (Topic-term Usage over Time) Intern © Siemens AG 2017 May 2017Seite 37 Corporate Technology
  • 38. Evaluation: Topic Characterization (Topic-term Usage over Time) Intern © Siemens AG 2017 May 2017Seite 38 Corporate Technology
  • 39. Evaluation: Topic Characterization (Topic-term Usage over Time) Intern © Siemens AG 2017 May 2017Seite 39 Corporate Technology Word usage for topic ‘Word Vector’ over time
  • 40. Summary  Introduce an unsupervised neural dynamic topic model, RNN-RSM  Model dynamics in underlying topics detected over time Model dynamics in underlying topics detected over time  Demonstrate better generalization, topic evolution and characterization  Analyze scholarly articles to capture evolution in NLP topics Intern © Siemens AG 2017 May 2017Seite 40 Corporate Technology
  • 41. h(1) bh (1) h(2) bh (2) h(T-1) bh (T-1) h(T) bh (T) RSM RSM RSM RSM … Thanks !!! V(1) bh bv (1) Wvh u(1) Wvu V(2) bh bv (2) Wvh u(2) Wvu V(T-1) bh bv (T-1) Wvh u(T-1) Wvu u(0) V(T) bh bv (T) Wvh u(T) Wvu Wuh Wuv Wuv Wuv Wuh Wuh Wuh Wuv RNN Wuu Wuu Wuu Wuu … Observable Softmax Visibles Neural Network Neural Network Neural NetworkNeural Network Neural Network Neural Network Intern © Siemens AG 2017 May 2017Seite 41 Corporate Technology 1996 Neural Network Language Models Word Representation Linear Model Rule Set Neural Network Language Models Supervised Linear Model Rule Set Neural Network Language Models Word Embedding Word Embeddings Word Representation Neural Network Language Models Word Representation Linear Model Rule Set Neural Network Language Models Supervised Linear Model Rule Set Neural Network Language Models Word Embedding Word Embeddings Word Representation Topic-words over time for topic ‘Word Vector’ 1997 2014
  • 42. References Jean Aitchison. 2001. Language change: progress or decay?. Cambridge University Press. Nikolaos Aletras and Mark Stevenson. 2013. Evaluating topic coherence using distributional semantics. In Proceedings of the 10th International Conference on Computational Semantics (IWCS). Potsdam, Germany, pages 13–22. James Allan. 2002. Introduction to topic detection and tracking. In Topic detection and tracking, Springer, pages 1–16.James Allan. 2002. Introduction to topic detection and tracking. In Topic detection and tracking, Springer, pages 1–16. James Allan, Jaime G Carbonell, George Doddington, Jonathan Yamron, and Yiming Yang. 1998. Topic detection and tracking pilot study final report. In Proceedings of the DARPA Broadcast News Transcription and Understanding Workshop. Virginia, US, pages 194–218. Robert Bamler and Stephan Mandt. 2017. Dynamic word embeddings. In Proceedings of the 34th International Conference on Machine Learning. Sydney, Australia, pages 380–389. Pierpaolo Basile, Annalina Caputo, and Giovanni Semeraro. 2014. Analysing word meaning over time by exploiting temporal random indexing. In Proceedings of the 1st Italian Conference on Computational Linguistics (CLiC-it). Pisa University Press, Pisa, Italy. David M. Blei and John D. Lafferty. 2006. Dynamic topic models. In Proceedings of the 23rd International Conference on Machine Learning. Association for Computing Machinery, Pittsburgh, Pennsylvania USA, pages 113–120. David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent dirichlet allocation. Proceedings of Machine Learning Research 3(Jan):993–1022. Intern © Siemens AG 2017 May 2017Seite 42 Corporate Technology David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent dirichlet allocation. Proceedings of Machine Learning Research 3(Jan):993–1022. Nicolas Boulanger-Lewandowski, Yoshua Bengio, and Pascal Vincent. 2012. Modeling temporal dependencies in high-dimensional sequences: Application to polyphonic music generation and transcription. In Proceedings of the 29th International Conference on Machine Learning. Edinburgh, Scotland UK. Jonathan Chang, Sean Gerrish, Chong Wang, Jordan L. Boyd-Graber, and David M. Blei. 2009. Reading tea leaves: How humans interpret topic models. In Advances in Neural Information Processing Systems. Curran Associates, Inc., Vancouver, Canada, pages 288–296. Rajarshi Das, Manzil Zaheer, and Chris Dyer. 2015. Gaussian lda for topic models with word embeddings. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th IJCNLP. ACL, Beijing, China, volume 1, pages 795–804.
  • 43. References Peter V. Gehler, Alex D. Holub, and Max Welling. 2006. The rate adapting poisson model for information retrieval and object recognition. In Proceedings of the 23rd International Conference on Machine Learning. Association for Computing Machinery, Pittsburgh, Pennsylvania USA, pages 337–344. Sujatha Das Gollapalli and Xiaoli Li. 2015. Emnlp versus acl: Analyzing nlp research over time. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Lisbon, Portugal, pages 2002–2006.Methods in Natural Language Processing. Association for Computational Linguistics, Lisbon, Portugal, pages 2002–2006. Pankaj Gupta, Thomas Runkler, Heike Adel, Bernt Andrassy, Hans-Georg Zimmermann, and Hinrich Schutze. 2015a. Deep learning methods for the extraction of relations in natural language text. Technical report, Technical University of Munich, Germany. Pankaj Gupta, Thomas Runkler, and Bernt Andrassy. 2015b. Keyword learning for classifying requirements in tender documents. Technical report, Technical University of Munich, Germany. Pankaj Gupta, Hinrich Schutze, and Bernt Andrassy. 2016. Table filling multi-task recurrent neural network for joint entity and relation extraction. In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers. Osaka, Japan, pages 2537–2547. Pankaj Gupta, Udhayaraj Sivalingam, Sebastian Polsterl, and Nassir Navab. 2015c. Identifying patients with diabetes using discriminative restricted boltzmann machines. Technical report, Technical University of Munich, Germany. David Hall, Daniel Jurafsky, and Christopher D. Manning. 2008. Studying the history of ideas using topic models. In Proceedings of the conference Intern © Siemens AG 2017 May 2017Seite 43 Corporate Technology David Hall, Daniel Jurafsky, and Christopher D. Manning. 2008. Studying the history of ideas using topic models. In Proceedings of the conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Honolulu, Hawaii, pages 363–371. Geoffrey Hinton and Ruslan Salakhutdinov. 2009. Replicated softmax: an undirected topic model. In Advances in Neural Information Processing Systems 22. Curran Associates, Inc., Vancouver, Canada, pages 1607–1614. Geoffrey E. Hinton. 2002. Training products of experts by minimizing contrastive divergence. Neural Computation 14(8):1771–1800. Geoffrey E. Hinton, Simon Osindero, and Yee-Whye Teh. 2006. A fast learning algorithm for deep belief nets. Neural Computation 18(7):1527–1554. Tomoharu Iwata, Takeshi Yamada, Yasushi Sakurai, and Naonori Ueda. 2010. Online multiscale dynamic topic models. In Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, Washington DC, USA, pages 663–672.
  • 44. References David Newman, Sarvnaz Karimi, and Lawrence Cavedon. 2009. External evaluation of topic models. In Proceedings of the 14th Australasian Document Computing Symposium. Citeseer, Sydney, Australia. Iulian Pruteanu-Malinici, Lu Ren, John Paisley, Eric Wang, and Lawrence Carin. 2010. Hierarchical bayesian modeling of topics in time-stamped documents. IEEE transactions on pattern analysis and machine intelligence 32(6):996–1011. Maja Rudolph and David Blei. 2018. Dynamic bernoulli embeddings for language evolution. In Proceedings of the 27th International Conference on World Wide Web Companion. Lyon, France. David E. Rumelhart, Geoffrey E. Hinton, and Ronald J. Williams. 1985. Learning internal representations by error propagation. Technical report, California Univ San Diego La Jolla Inst for Cognitive Science. Ankan Saha and Vikas Sindhwani. 2012. Learning evolving and emerging topics in social media: a dynamic nmf approach with temporal regularization. In Proceedings of the 5th ACM International Conference on Web Search and Data Mining. Association for Computing Machinery, Seattle, Washington USA, pages 693–702. Aaron Schein, Hanna Wallach, and Mingyuan Zhou. 2016. Poisson-gamma dynamical systems. In Advances in Neural Information Processing Systems 29, Curran Associates, Inc., Barcelona, Spain, pages 5005–5013. Paul Smolensky. 1986. Information processing in dynamical systems: Foundations of harmony theory. Technical report, Colorado University at Boulder Department of Computer Science. Intern © Siemens AG 2017 May 2017Seite 44 Corporate Technology University at Boulder Department of Computer Science. Ilya Sutskever, Geoffrey E. Hinton, and Graham W. Taylor. 2009. The recurrent temporal restricted boltzmann machine. In Advances in Neural Information Processing Systems 22. Curran Associates, Inc., Vancouver, Canada, pages 1601–1608. Graham W. Taylor, Geoffrey E. Hinton, and Sam T. Roweis. 2007. Modeling human motion using binary latent variables. In Advances in Neural Information Processing Systems 20. Curran Associates, Inc., Vancouver, Canada, pages 1345–1352. Ngoc Thang Vu, Heike Adel, Pankaj Gupta, and Hinrich Schutze. 2016a. Combining recurrent and convolutional neural networks for relation classification. In Proceedings of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, San Diego, California USA, pages 534–539.
  • 45. References Ngoc Thang Vu, Pankaj Gupta, Heike Adel, and Hinrich Schutze. 2016b. Bi-directional recurrent neural network with ranking loss for spoken language understanding. In Proceedings of the Acoustics, Speech and Signal Processing (ICASSP). IEEE, Shanghai, China, pages 6060–6064. Xiaojun Wan and Jianguo Xiao. 2008. Single document keyphrase extraction using neighborhood knowledge. In Proceedings of the 23rd National Conference on Artificial Intelligence. Chicago, Illinois USA, volume 8, pages 855–860. Xuerui Wang and Andrew McCallum. 2006. Topics over time: a non-markov continuous-time model of topical trends. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Association for Computing Machinery, Philadelphia, Pennsylvania USA, pages 424–433. Eric P. Xing, Rong Yan, and Alexander G. Hauptmann. 2005. Mining associated text and images with dual-wing harmoniums. In Proceedings of the 21st Conference on Uncertainty in Artificial Intelligence. AUAI Press, Edinburgh, Scotland UK. Zijun Yao, Yifan Sun, Weicong Ding, Nikhil Rao, and Hui Xiong. 2018. Dynamic word embeddings for evolving semantic discovery. In Proceedings of the 11th ACM International Conference on Web Search and Data Mining (WSDM). Association for Computing Machinery, Los Angeles, California USA, pages 673–681. Intern © Siemens AG 2017 May 2017Seite 45 Corporate Technology