This talk covers the fundamental building blocks of neural network architectures and how they’re used to tackle problems in modern natural language processing. Topics include an overview of language vector representations, text classification, named entity recognition, and sequence-to-sequence modeling approaches. Dr. Brown emphasizes the shape of these types of problems from the perspective of deep-learning architectures, which will help attendees successfully identify the most applicable neural network techniques to new problems they encounter.
This lectures provides students with an introduction to natural language processing, with a specific focus on the basics of two applications: vector semantics and text classification.
(Lecture at the QUARTZ PhD Winter School (http://www.quartz-itn.eu/training/winter-school/ in Padua, Italy on February 12, 2018)
This lectures provides students with an introduction to natural language processing, with a specific focus on the basics of two applications: vector semantics and text classification.
(Lecture at the QUARTZ PhD Winter School (http://www.quartz-itn.eu/training/winter-school/ in Padua, Italy on February 12, 2018)
Representation Learning of Vectors of Words and PhrasesFelipe Moraes
Talk about representation learning using word vectors such as Word2Vec, Paragraph Vector. Also introduced to neural network language models. Expose some applications using NNLM such as sentiment analysis and information retrieval.
This presentation is a briefing of a paper about Networks and Natural Language Processing. It describes many graph based methods and algorithms that help in syntactic parsing, lexical semantics and other applications.
Learning to understand phrases by embedding the dictionaryRoelof Pieters
review of "Learning to Understand Phrases by Embedding the Dictionary" by Felix Hill, Kyunghyun Cho, Anna Korhonen, Yoshua Bengio
at KTH's Deep Learning reading group:
www.csc.kth.se/cvap/cvg/rg/
Is acquiring knowledge of verb subcategorization in English easier? A partial...Yu Tamura
Tamura, Y. (2016). Is acquiring knowledge of verb subcategorization in English easier? A partial replication of Jiang (2007). Paper presented at PacSLRF2016. Chuo University, Tokyo Japan. September 11, 2016
Continuous representations of words and documents, which is recently referred to as Word Embeddings, have recently demonstrated large advancements in many of the Natural language processing tasks.
In this presentation we will provide an introduction to the most common methods of learning these representations. As well as previous methods in building these representations before the recent advances in deep learning, such as dimensionality reduction on the word co-occurrence matrix.
Moreover, we will present the continuous bag of word model (CBOW), one of the most successful models for word embeddings and one of the core models in word2vec, and in brief a glance of many other models of building representations for other tasks such as knowledge base embeddings.
Finally, we will motivate the potential of using such embeddings for many tasks that could be of importance for the group, such as semantic similarity, document clustering and retrieval.
Intent Classifier with Facebook fastText
Facebook Developer Circle, Malang
22 February 2017
This is slide for Facebook Developer Circle meetup.
This is for beginner.
Presented by Ted Xiao at RobotXSpace on 4/18/2017. This workshop covers the fundamentals of Natural Language Processing, crucial NLP approaches, and an overview of NLP in industry.
The best known natural language processing tool is GPT-3, from OpenAI, which uses AI and statistics to predict the next word in a sentence based on the preceding words. NLP practitioners call tools like this “language models,” and they can be used for simple analytics tasks, such as classifying documents and analyzing the sentiment in blocks of text, as well as more advanced tasks, such as answering questions and summarizing reports. Language models are already reshaping traditional text analytics, but GPT-3 was an especially pivotal language model because, at 10x larger than any previous model upon release, it was the first large language model, which enabled it to perform even more advanced tasks like programming and solving high school–level math problems. The latest version, called InstructGPT, has been fine-tuned by humans to generate responses that are much better aligned with human values and user intentions, and Google’s latest model shows further impressive breakthroughs on language and reasoning.
For businesses, the three areas where GPT-3 has appeared most promising are writing, coding, and discipline-specific reasoning. OpenAI, the Microsoft-funded creator of GPT-3, has developed a GPT-3-based language model intended to act as an assistant for programmers by generating code from natural language input. This tool, Codex, is already powering products like Copilot for Microsoft’s subsidiary GitHub and is capable of creating a basic video game simply by typing instructions. This transformative capability was already expected to change the nature of how programmers do their jobs, but models continue to improve — the latest from Google’s DeepMind AI lab, for example, demonstrates the critical thinking and logic skills necessary to outperform most humans in programming competitions.
Models like GPT-3 are considered to be foundation models — an emerging AI research area — which also work for other types of data such as images and video. Foundation models can even be trained on multiple forms of data at the same time, like OpenAI’s DALL·E 2, which is trained on language and images to generate high-resolution renderings of imaginary scenes or objects simply from text prompts. Due to their potential to transform the nature of cognitive work, economists expect that foundation models may affect every part of the economy and could lead to increases in economic growth similar to the industrial revolution.
Representation Learning of Vectors of Words and PhrasesFelipe Moraes
Talk about representation learning using word vectors such as Word2Vec, Paragraph Vector. Also introduced to neural network language models. Expose some applications using NNLM such as sentiment analysis and information retrieval.
This presentation is a briefing of a paper about Networks and Natural Language Processing. It describes many graph based methods and algorithms that help in syntactic parsing, lexical semantics and other applications.
Learning to understand phrases by embedding the dictionaryRoelof Pieters
review of "Learning to Understand Phrases by Embedding the Dictionary" by Felix Hill, Kyunghyun Cho, Anna Korhonen, Yoshua Bengio
at KTH's Deep Learning reading group:
www.csc.kth.se/cvap/cvg/rg/
Is acquiring knowledge of verb subcategorization in English easier? A partial...Yu Tamura
Tamura, Y. (2016). Is acquiring knowledge of verb subcategorization in English easier? A partial replication of Jiang (2007). Paper presented at PacSLRF2016. Chuo University, Tokyo Japan. September 11, 2016
Continuous representations of words and documents, which is recently referred to as Word Embeddings, have recently demonstrated large advancements in many of the Natural language processing tasks.
In this presentation we will provide an introduction to the most common methods of learning these representations. As well as previous methods in building these representations before the recent advances in deep learning, such as dimensionality reduction on the word co-occurrence matrix.
Moreover, we will present the continuous bag of word model (CBOW), one of the most successful models for word embeddings and one of the core models in word2vec, and in brief a glance of many other models of building representations for other tasks such as knowledge base embeddings.
Finally, we will motivate the potential of using such embeddings for many tasks that could be of importance for the group, such as semantic similarity, document clustering and retrieval.
Intent Classifier with Facebook fastText
Facebook Developer Circle, Malang
22 February 2017
This is slide for Facebook Developer Circle meetup.
This is for beginner.
Presented by Ted Xiao at RobotXSpace on 4/18/2017. This workshop covers the fundamentals of Natural Language Processing, crucial NLP approaches, and an overview of NLP in industry.
The best known natural language processing tool is GPT-3, from OpenAI, which uses AI and statistics to predict the next word in a sentence based on the preceding words. NLP practitioners call tools like this “language models,” and they can be used for simple analytics tasks, such as classifying documents and analyzing the sentiment in blocks of text, as well as more advanced tasks, such as answering questions and summarizing reports. Language models are already reshaping traditional text analytics, but GPT-3 was an especially pivotal language model because, at 10x larger than any previous model upon release, it was the first large language model, which enabled it to perform even more advanced tasks like programming and solving high school–level math problems. The latest version, called InstructGPT, has been fine-tuned by humans to generate responses that are much better aligned with human values and user intentions, and Google’s latest model shows further impressive breakthroughs on language and reasoning.
For businesses, the three areas where GPT-3 has appeared most promising are writing, coding, and discipline-specific reasoning. OpenAI, the Microsoft-funded creator of GPT-3, has developed a GPT-3-based language model intended to act as an assistant for programmers by generating code from natural language input. This tool, Codex, is already powering products like Copilot for Microsoft’s subsidiary GitHub and is capable of creating a basic video game simply by typing instructions. This transformative capability was already expected to change the nature of how programmers do their jobs, but models continue to improve — the latest from Google’s DeepMind AI lab, for example, demonstrates the critical thinking and logic skills necessary to outperform most humans in programming competitions.
Models like GPT-3 are considered to be foundation models — an emerging AI research area — which also work for other types of data such as images and video. Foundation models can even be trained on multiple forms of data at the same time, like OpenAI’s DALL·E 2, which is trained on language and images to generate high-resolution renderings of imaginary scenes or objects simply from text prompts. Due to their potential to transform the nature of cognitive work, economists expect that foundation models may affect every part of the economy and could lead to increases in economic growth similar to the industrial revolution.
Beyond the Symbols: A 30-minute Overview of NLPMENGSAYLOEM1
This presentation delves into the world of Natural Language Processing (NLP), exploring its goal to make human language understandable to machines. The complexities of language, such as ambiguity and complex structures, are highlighted as major challenges. The talk underscores the evolution of NLP through deep learning methodologies, leading to a new era defined by large-scale language models. However, obstacles like low-resource languages and ethical issues including bias and hallucination are acknowledged as enduring challenges in the field. Overall, the presentation provides a condensed, yet comprehensive view of NLP's accomplishments and ongoing hurdles.
Building a Neural Machine Translation System From ScratchNatasha Latysheva
Human languages are complex, diverse and riddled with exceptions – translating between different languages is therefore a highly challenging technical problem. Deep learning approaches have proved powerful in modelling the intricacies of language, and have surpassed all statistics-based methods for automated translation. This session begins with an introduction to the problem of machine translation and discusses the two dominant neural architectures for solving it – recurrent neural networks and transformers. A practical overview of the workflow involved in training, optimising and adapting a competitive neural machine translation system is provided. Attendees will gain an understanding of the internal workings and capabilities of state-of-the-art systems for automatic translation, as well as an appreciation of the key challenges and open problems in the field.
Word embedding, Vector space model, language modelling, Neural language model, Word2Vec, GloVe, Fasttext, ELMo, BERT, distilBER, roBERTa, sBERT, Transformer, Attention
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingYoung Seok Kim
Review of paper
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
ArXiv link: https://arxiv.org/abs/1810.04805
YouTube Presentation: https://youtu.be/GK4IO3qOnLc
(Slides are written in English, but the presentation is done in Korean)
Word Embeddings, Application of Sequence modelling, Recurrent neural network , drawback of recurrent neural networks, gated recurrent unit, long short term memory unit, Attention Mechanism
Slides from tutorial on Deep Learning and Modern Natural Language Processing using Pytorch at PyData Miami 2019.
Note: These are the slides corresponding to the Jupiter Notebook tutorials which can be found in the following repo:
https://github.com/ZacharySBrown/deep-learning-nlp-pydata
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
JMeter webinar - integration with InfluxDB and Grafana
Deep Learning and Modern Natural Language Processing (AnacondaCon2019)
1. Deep Learning and Modern Natural
Language Processing
Zachary Brown, Lead Data Scientist, S&P Global
2. Outline
• Neural Methods for Natural Language Processing
• Shapes of Natural Language Processing Tasks
• Perceptron Text Classification
• Local vs. Global Text Representations
• Contextual Representations and Sequence Modeling
2
4. Neural Methods for Natural Language
Processing
• Natural Language Processing (NLP) has moved largely to neural
methods in recent years
4
5. Neural Methods for Natural Language
Processing
• Natural Language Processing (NLP) has moved largely to neural
methods in recent years
5
• Traditional NLP builds on years of research into
language representation
• Theoretical foundations can lead to model rigidity
• Tasks often rely on manually generated and
curated dictionaries and thesauruses
• Built upon local word representations
6. Neural Methods for Natural Language
Processing
• Natural Language Processing (NLP) has moved largely to neural
methods in recent years
6
• Few to no assumptions need to be made
• Active area of research; most open-source
• Ability to learn global and contextualized
word representations
• Purpose-built model architectures
7. Neural Methods for Natural Language
Processing
• Natural Language Processing (NLP) has moved largely to neural
methods in recent years
7
• Few to no assumptions need to be made
• Active area of research; most open-source
• Ability to learn global and contextualized
word representations
• Purpose-built model architectures
9. Shapes of Natural Language Processing
Tasks
• A general task in natural language processing often takes the form:
9
10. Shapes of Natural Language Processing
Tasks
• For binary classification processes (relevance), our target is a single
number, often interpreted as a probability
10
11. Shapes of Natural Language Processing
Tasks
• For multi-class classification processes (type of text), our target is a
set of probabilities, one for each of the output classes
11
12. Shapes of Natural Language Processing
Tasks
• For sequential classification (e.g. LM, NER, POS) the target is a
probability of each class for each element in the input
12
13. Shapes of Natural Language Processing
Tasks
• In a traditional machine learning pipeline, vectorization (feature
engineering) process is often a (very) time consuming process
13
80-90%
14. Shapes of Natural Language Processing
Tasks
• A relatively small proportion of time is spent on the actual modeling
14
10-20%
15. Shapes of Natural Language Processing
Tasks
• Neural networks allow us to develop purpose-built architectures to
solve tasks, that learn the appropriate vectorization for a task
15
100%
17. Perceptron Text Classification
• To introduce the shape of information as it flows through a neural
network, we'll first look at a network that only handles classification
17
18. Perceptron Text Classification
• For the vectorization, we'll assume that we've converted our text
into a vector using a count-based method like tf-idf
18
tf-idf
19. Perceptron Text Classification
• A perceptron is one of the simplest neural network architectures,
and is a good fit for this task
19
20. Perceptron Text Classification
• A perceptron is one of the simplest neural network architectures,
and is a good fit for this task
20
input
hidden
(linear)
activation
output
21. Perceptron Text Classification
• The hidden layer represents the weights that will be optimized by
the deep learning framework.
21
input
hidden
(linear)
activation
output
weights
22. Perceptron Text Classification
• If we want to change our task to multiclass classification, we can
simply change the size of our hidden layer (+ minor mods)
22
25. Local vs. Global Text Representations
• Let's look back to the problem of creating a vector representation for
our text
25
tf-idf
26. Local vs. Global Text Representations
• Further, let's only consider the task of how we'd represent single
words or tokens as vectors
26
dog
27. Local vs. Global Text Representations
• Traditional approaches to word representations treat each word as
a unique entity
27
28. Local vs. Global Text Representations
• Traditional approaches to word representations treat each word as
a unique entity
28
29. Local vs. Global Text Representations
• Traditional approaches to word representations treat each word as
a unique entity
29
30. Local vs. Global Text Representations
• Traditional approaches to word representations treat each word as
a unique entity
30
31. Local vs. Global Text Representations
• Modern approaches move to a fixed dimensional vector size, with
dense vectors
31
32. Local vs. Global Text Representations
• Modern approaches move to a fixed dimensional vector size, with
dense vectors
32
33. Local vs. Global Text Representations
• Modern approaches move to a fixed dimensional vector size, with
dense vectors
33
34. Local vs. Global Text Representations
• There are a variety of frameworks available that allow for computing
these vectors in an unsupervised way
34
36. Contextual Representations and Sequence
Modeling
• Global word representations are a fantastic starting point for many
problems in NLP, but consider the following sentence
36
I'm going to book our vacation then relax and read a good book
37. Contextual Representations and Sequence
Modeling
• Global word representations are a fantastic starting point for many
problems in NLP, but consider the following sentence
37
I'm going to book our vacation then relax and read a good book
38. I don't really hate horror movies, but I hate comedies
Contextual Representations and Sequence
Modeling
• Global word representations are a fantastic starting point for many
problems in NLP, but consider the following sentence
38
39. I don't really hate horror movies, but I hate comedies
Contextual Representations and Sequence
Modeling
• Global word representations are a fantastic starting point for many
problems in NLP, but consider the following sentence
39
Context
Matters
40. Contextual Representations and Sequence
Modeling
• For modeling tasks where word ordering and context matter,
sequential models are often used. These tasks often take the
following shape:
40
41. Contextual Representations and Sequence
Modeling
• Recurrent neural networks are a type of neural network architecture
that naturally handles modeling sequential data
41
42. Contextual Representations and Sequence
Modeling
• This type of network generates a new output vector for each input in
a sequence, and also feeds that same information forward
42
43. Contextual Representations and Sequence
Modeling
• By feeding the information forward, each subsequent output vector
has contextual information encoded from the preceding words
43
44. Contextual Representations and Sequence
Modeling
• This type of architecture can be used to build language models,
where the task is to predict the next word in the sequence
44
45. Contextual Representations and Sequence
Modeling
• It can also be used for problems like named entity recognition
45
animalo o o o animal
46. Contextual Representations and Sequence
Modeling
• By taking the final vector in the sequence, you can perform tasks
like sentiment classification
46
positive
47. Contextual Representations and Sequence
Modeling
• For all of these different types of tasks, a network similar to the
perceptron can be placed at the end to carry out the final
classification of each word
47
48. Contextual Representations and Sequence
Modeling
• For all of these different types of tasks, a network similar to the
perceptron can be placed at the end to carry out the final
classification of each word, or the classification of the whole
sequence
48
49. Contextual Representations and Sequence
Modeling
• In a similar manner, these individual elements can be combined in a
variety of ways to tackle very complex tasks
49
50. Contextual Representations and Sequence
Modeling
• In a similar manner, these individual elements can be combined in a
variety of ways to tackle very complex tasks
50
51. Contextual Representations and Sequence
Modeling
• In a similar manner, these individual elements can be combined in a
variety of ways to tackle very complex tasks
51
52. Contextual Representations and Sequence
Modeling
• In a similar manner, these individual elements can be combined in a
variety of ways to tackle very complex tasks
52