DeepPavlov is an open-source framework for the development of production-ready chat-bots and complex conversational systems, as well as NLP and dialog systems research.
GPT-2: Language Models are Unsupervised Multitask LearnersYoung Seok Kim
Review of paper
Language Models are Unsupervised Multitask Learners
(GPT-2)
by Alec Radford et al.
Paper link: https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf
YouTube presentation: https://youtu.be/f5zULULWUwM
(Slides are written in English, but the presentation is done in Korean)
Deep Natural Language Processing for Search and Recommender SystemsHuiji Gao
Tutorial for KDD 2019:
Search and recommender systems process rich natural language text data such as user queries and documents. Achieving high-quality search and recommendation results requires processing and understanding such information effectively and efficiently, where natural language processing (NLP) technologies are widely deployed. In recent years, the rapid development of deep learning models has been proven successful for improving various NLP tasks, indicating their great potential of promoting search and recommender systems.
In this tutorial, we summarize the current effort of deep learning for NLP in search/recommender systems. We first give an overview of search/recommender systems with NLP, then introduce basic concept of deep learning for NLP, covering state-of-the-art technologies in both language understanding and language generation. After that, we share our hands-on experience with LinkedIn applications. In the end, we highlight several important future trends.
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingYoung Seok Kim
Review of paper
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
ArXiv link: https://arxiv.org/abs/1810.04805
YouTube Presentation: https://youtu.be/GK4IO3qOnLc
(Slides are written in English, but the presentation is done in Korean)
https://imatge-upc.github.io/synthref/
Integrating computer vision with natural language processing has achieved significant progress
over the last years owing to the continuous evolution of deep learning. A novel vision and language
task, which is tackled in the present Master thesis is referring video object segmentation, in which a
language query defines which instance to segment from a video sequence. One of the biggest chal-
lenges for this task is the lack of relatively large annotated datasets since a tremendous amount of
time and human effort is required for annotation. Moreover, existing datasets suffer from poor qual-
ity annotations in the sense that approximately one out of ten language expressions fails to uniquely
describe the target object.
The purpose of the present Master thesis is to address these challenges by proposing a novel
method for generating synthetic referring expressions for an image (video frame). This method pro-
duces synthetic referring expressions by using only the ground-truth annotations of the objects as well
as their attributes, which are detected by a state-of-the-art object detection deep neural network. One
of the advantages of the proposed method is that its formulation allows its application to any object
detection or segmentation dataset.
By using the proposed method, the first large-scale dataset with synthetic referring expressions for
video object segmentation is created, based on an existing large benchmark dataset for video instance
segmentation. A statistical analysis and comparison of the created synthetic dataset with existing ones
is also provided in the present Master thesis.
The conducted experiments on three different datasets used for referring video object segmen-
tation prove the efficiency of the generated synthetic data. More specifically, the obtained results
demonstrate that by pre-training a deep neural network with the proposed synthetic dataset one can
improve the ability of the network to generalize across different datasets, without any additional annotation cost. This outcome is even more important taking into account that no additional annotation cost is involved.
GPT-2: Language Models are Unsupervised Multitask LearnersYoung Seok Kim
Review of paper
Language Models are Unsupervised Multitask Learners
(GPT-2)
by Alec Radford et al.
Paper link: https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf
YouTube presentation: https://youtu.be/f5zULULWUwM
(Slides are written in English, but the presentation is done in Korean)
Deep Natural Language Processing for Search and Recommender SystemsHuiji Gao
Tutorial for KDD 2019:
Search and recommender systems process rich natural language text data such as user queries and documents. Achieving high-quality search and recommendation results requires processing and understanding such information effectively and efficiently, where natural language processing (NLP) technologies are widely deployed. In recent years, the rapid development of deep learning models has been proven successful for improving various NLP tasks, indicating their great potential of promoting search and recommender systems.
In this tutorial, we summarize the current effort of deep learning for NLP in search/recommender systems. We first give an overview of search/recommender systems with NLP, then introduce basic concept of deep learning for NLP, covering state-of-the-art technologies in both language understanding and language generation. After that, we share our hands-on experience with LinkedIn applications. In the end, we highlight several important future trends.
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingYoung Seok Kim
Review of paper
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
ArXiv link: https://arxiv.org/abs/1810.04805
YouTube Presentation: https://youtu.be/GK4IO3qOnLc
(Slides are written in English, but the presentation is done in Korean)
https://imatge-upc.github.io/synthref/
Integrating computer vision with natural language processing has achieved significant progress
over the last years owing to the continuous evolution of deep learning. A novel vision and language
task, which is tackled in the present Master thesis is referring video object segmentation, in which a
language query defines which instance to segment from a video sequence. One of the biggest chal-
lenges for this task is the lack of relatively large annotated datasets since a tremendous amount of
time and human effort is required for annotation. Moreover, existing datasets suffer from poor qual-
ity annotations in the sense that approximately one out of ten language expressions fails to uniquely
describe the target object.
The purpose of the present Master thesis is to address these challenges by proposing a novel
method for generating synthetic referring expressions for an image (video frame). This method pro-
duces synthetic referring expressions by using only the ground-truth annotations of the objects as well
as their attributes, which are detected by a state-of-the-art object detection deep neural network. One
of the advantages of the proposed method is that its formulation allows its application to any object
detection or segmentation dataset.
By using the proposed method, the first large-scale dataset with synthetic referring expressions for
video object segmentation is created, based on an existing large benchmark dataset for video instance
segmentation. A statistical analysis and comparison of the created synthetic dataset with existing ones
is also provided in the present Master thesis.
The conducted experiments on three different datasets used for referring video object segmen-
tation prove the efficiency of the generated synthetic data. More specifically, the obtained results
demonstrate that by pre-training a deep neural network with the proposed synthetic dataset one can
improve the ability of the network to generalize across different datasets, without any additional annotation cost. This outcome is even more important taking into account that no additional annotation cost is involved.
Peter Muschick MSc thesis
Universitat Pollitecnica de Catalunya, 2020
Sign language recognition and translation has been an active research field in the recent years with most approaches using deep neural networks to extract information from sign language data. This work investigates the mostly disregarded approach of using human keypoint estimation from image and video data with OpenPose in combination with transformer network architecture. Firstly, it was shown that it is possible to recognize individual signs (4.5% word error rate (WER)). Continuous sign language recognition though was more error prone (77.3% WER) and sign language translation was not possible using the proposed methods, which might be due to low accuracy scores of human keypoint estimation by OpenPose and accompanying loss of information or insufficient capacities of the used transformer model. Results may improve with the use of datasets containing higher repetition rates of individual signs or focusing more precisely on keypoint extraction of hands.
This is a survey about Dialog System, Question and Answering, including the 03 generations: (1) Symbolic Rule/Template Based QA; (2) Data Driven, Learning; (3) Data-Driven Deep Learning. It also presents the available Frameworks and Datas for Dialog Systems.
Building DSLs: Marriage of High Essence and Groovy MetaprogrammingSkills Matter
DSLs or Domain Specific Languages focus on a domain or a particular problem. They serve as an effective human-machine interaction tool as they're highly expressive. Their scope is fairly focused and that keeps them simple and small from the user's point of view. However, designing and implementing DSLs is not easy. Typically this involves steep learning curve and difficult parsing techniques. This is where Groovy comes in. You can take advantage of the flexible syntax of Groovy and it's metaprogramming capability to create what are called internal DSLs, that is, DSLs hosted using a higher level language.
In this fast paced highly interactive presentation you will start out learning the characteristics and types of DSLs. Then you will learn about the challenges in designing DSLs and deep dive into Groovy features that can ease the pain of implementing DSLs. Then, using some live coding, Venkat will show you how to create and implement internal DSLs using Groovy. Along the way you'll learn some tricks to facilitate desirable syntax for your DSL.
DataFest 2017. Introduction to Natural Language Processing by Rudolf Eremyanrudolf eremyan
The objective of this workshop is to show how natural language processing applied in modern applications such as Google Search, Apple Siri, Bing Translator and etc. During the workshop we will go through history if natural language processing, talk about typical problems, consider classical approaches and methods, and compare them with state-of-the-art deep learning techniques.
Author: Rudolf Eremyan
Email: eremyan.rudolf@gmail.com
Phone: +995599607066
LinkedIn: https://www.linkedin.com/in/rudolferemyan/
DataFest Tbilisi 2017 website: https://datafest.ge
Deep Learning for Natural Language ProcessingJonathan Mugan
Deep Learning represents a significant advance in artificial intelligence because it enables computers to represent concepts using vectors instead of symbols. Representing concepts using vectors is particularly useful in natural language processing, and this talk will elucidate those benefits and provide an understandable introduction to the technologies that make up deep learning. The talk will outline ways to get started in deep learning, and it will conclude with a discussion of the gaps that remain between our current technologies and true computer understanding.
Getting started on your natural language processing project? First you'll need to extract some features from your corpus. Frequency, Syntax parsing, word vectors are good ones to start with.
Monthly AI Tech Talks in Toronto 2019-08-28
https://www.meetup.com/aittg-toronto
The talk will cover the end-to-end details including contextual and linguistic feature extraction, vectorization, n-grams, topic modeling, named entity resolution which are based on concepts from mathematics, information retrieval and natural language processing. We will also be diving into more advanced feature engineering strategies such as word2vec, GloVe and fastText that leverage deep learning models.
In addition, attendees will learn how to combine NLP features with numeric and categorical features and analyze the feature importance from the resulting models.
The following libraries will be used to demonstrate the aforementioned feature engineering techniques: spaCy, Gensim, fasText and Keras in Python.
https://www.meetup.com/aittg-toronto/events/261940480/
In this presentation we discuss several concepts that include Word Representation using SVD as well as neural networks based techniques. In addition we also cover core concepts such as cosine similarity, atomic and distributed representations.
Requirements Engineering: focus on Natural Language Processing, Lecture 2alessio_ferrari
In this lecture, we give a practical guide on how to detect ambiguities in natural language requirements by means of GATE and by means of Python. A brief guide to Python is also included.
The previous lecture gives an introduction to the problem of ambiguity in requirements engineering. Find it here: https://www.slideshare.net/alessio_ferrari/requirements-engineering-focus-on-natural-language-processing-lecture-1
There is an increasing demand for embedding intelligence in software systems as part of its core set of features both in the front-end (e.g. conversational user interfaces) and back-end (e.g. prediction services). This combination is usually referred to as AI-enhanced software or, simply, smart software.
The development of smart software poses new engineering challenges, as now we need to deal with the engineering of the “traditional” components, the engineering of the “AI” ones but also of the interaction between both types that need to co-exist and collaborate.
In this talk we'll see how modeling can help tame the complexity of engineering smart software by enabling software engineers specify and generate smart software systems starting from higher-level and platform-independent modeling primitives.
But, unavoidably, these models will be more diverse and complex than our usual ones. Don't despair, we'll also see how some of these same AI techniques that are making our modeling life challenging can be turned into allies and be transformed into modeling assistants to tackle the engineering of smart software with a new breed of smart modeling tools.
Peter Muschick MSc thesis
Universitat Pollitecnica de Catalunya, 2020
Sign language recognition and translation has been an active research field in the recent years with most approaches using deep neural networks to extract information from sign language data. This work investigates the mostly disregarded approach of using human keypoint estimation from image and video data with OpenPose in combination with transformer network architecture. Firstly, it was shown that it is possible to recognize individual signs (4.5% word error rate (WER)). Continuous sign language recognition though was more error prone (77.3% WER) and sign language translation was not possible using the proposed methods, which might be due to low accuracy scores of human keypoint estimation by OpenPose and accompanying loss of information or insufficient capacities of the used transformer model. Results may improve with the use of datasets containing higher repetition rates of individual signs or focusing more precisely on keypoint extraction of hands.
This is a survey about Dialog System, Question and Answering, including the 03 generations: (1) Symbolic Rule/Template Based QA; (2) Data Driven, Learning; (3) Data-Driven Deep Learning. It also presents the available Frameworks and Datas for Dialog Systems.
Building DSLs: Marriage of High Essence and Groovy MetaprogrammingSkills Matter
DSLs or Domain Specific Languages focus on a domain or a particular problem. They serve as an effective human-machine interaction tool as they're highly expressive. Their scope is fairly focused and that keeps them simple and small from the user's point of view. However, designing and implementing DSLs is not easy. Typically this involves steep learning curve and difficult parsing techniques. This is where Groovy comes in. You can take advantage of the flexible syntax of Groovy and it's metaprogramming capability to create what are called internal DSLs, that is, DSLs hosted using a higher level language.
In this fast paced highly interactive presentation you will start out learning the characteristics and types of DSLs. Then you will learn about the challenges in designing DSLs and deep dive into Groovy features that can ease the pain of implementing DSLs. Then, using some live coding, Venkat will show you how to create and implement internal DSLs using Groovy. Along the way you'll learn some tricks to facilitate desirable syntax for your DSL.
DataFest 2017. Introduction to Natural Language Processing by Rudolf Eremyanrudolf eremyan
The objective of this workshop is to show how natural language processing applied in modern applications such as Google Search, Apple Siri, Bing Translator and etc. During the workshop we will go through history if natural language processing, talk about typical problems, consider classical approaches and methods, and compare them with state-of-the-art deep learning techniques.
Author: Rudolf Eremyan
Email: eremyan.rudolf@gmail.com
Phone: +995599607066
LinkedIn: https://www.linkedin.com/in/rudolferemyan/
DataFest Tbilisi 2017 website: https://datafest.ge
Deep Learning for Natural Language ProcessingJonathan Mugan
Deep Learning represents a significant advance in artificial intelligence because it enables computers to represent concepts using vectors instead of symbols. Representing concepts using vectors is particularly useful in natural language processing, and this talk will elucidate those benefits and provide an understandable introduction to the technologies that make up deep learning. The talk will outline ways to get started in deep learning, and it will conclude with a discussion of the gaps that remain between our current technologies and true computer understanding.
Getting started on your natural language processing project? First you'll need to extract some features from your corpus. Frequency, Syntax parsing, word vectors are good ones to start with.
Monthly AI Tech Talks in Toronto 2019-08-28
https://www.meetup.com/aittg-toronto
The talk will cover the end-to-end details including contextual and linguistic feature extraction, vectorization, n-grams, topic modeling, named entity resolution which are based on concepts from mathematics, information retrieval and natural language processing. We will also be diving into more advanced feature engineering strategies such as word2vec, GloVe and fastText that leverage deep learning models.
In addition, attendees will learn how to combine NLP features with numeric and categorical features and analyze the feature importance from the resulting models.
The following libraries will be used to demonstrate the aforementioned feature engineering techniques: spaCy, Gensim, fasText and Keras in Python.
https://www.meetup.com/aittg-toronto/events/261940480/
In this presentation we discuss several concepts that include Word Representation using SVD as well as neural networks based techniques. In addition we also cover core concepts such as cosine similarity, atomic and distributed representations.
Requirements Engineering: focus on Natural Language Processing, Lecture 2alessio_ferrari
In this lecture, we give a practical guide on how to detect ambiguities in natural language requirements by means of GATE and by means of Python. A brief guide to Python is also included.
The previous lecture gives an introduction to the problem of ambiguity in requirements engineering. Find it here: https://www.slideshare.net/alessio_ferrari/requirements-engineering-focus-on-natural-language-processing-lecture-1
There is an increasing demand for embedding intelligence in software systems as part of its core set of features both in the front-end (e.g. conversational user interfaces) and back-end (e.g. prediction services). This combination is usually referred to as AI-enhanced software or, simply, smart software.
The development of smart software poses new engineering challenges, as now we need to deal with the engineering of the “traditional” components, the engineering of the “AI” ones but also of the interaction between both types that need to co-exist and collaborate.
In this talk we'll see how modeling can help tame the complexity of engineering smart software by enabling software engineers specify and generate smart software systems starting from higher-level and platform-independent modeling primitives.
But, unavoidably, these models will be more diverse and complex than our usual ones. Don't despair, we'll also see how some of these same AI techniques that are making our modeling life challenging can be turned into allies and be transformed into modeling assistants to tackle the engineering of smart software with a new breed of smart modeling tools.
Are you responsible for developing satellite on-board software? Are you the Dutch government and you have to efficiently implement the public benefits law? Are you a healthcare startup, developing companion apps that help patients through a treatment? Are you an insurance company struggling to create new, and evolve existing products quickly to keep up with the market? These are all examples of organisations who have built their own domain-specific programming language to streamline the development of applications that have a non-trivial algorithmic core. All have built their languages with Jetbrains MPS, an open source language development tool optimized for ecosystems of collaborating languages with mixed graphical, textual, tabular and mathematical notations. This talk has four parts. I start by motivating the need for DSLs based on real-world examples, including the ones above. I will then present a few high-level design practices that guide our language development work. Third, I will develop a simple language extension to give you a feel for how MPS works. And finally, I will point you to things you can read to get you started with your own language development practice.
Envisioning the Future of Language WorkbenchesMarkus Voelter
Over the last couple of years, I have used MPS successfully to build interesting (modeling and programming) languages in a wide variety of domains, targeting both business users and engineers. I’ve used MPS because it is currently the most powerful language workbench, lots of things are good about iz, in particular, its support for a multitude of notations and language modularity. But it is also obvious that MPS is not going to be viable for the medium to long term future; the most obvious reason for this statement is that it is not web/cloud-based. In this keynote, I will quickly recap why and how we have been successful with MPS, and point out how language workbenches could look like in the future; I will outline challenges, opportunities and research problems. I hope to spawn discussions for the remainder of the workshop.
Bridging the gap between AI and UI - DSI Vienna - full versionLiad Magen
This is a summary of the latest research on model interpretability, including Recurrent neural networks (RNN) for Natural Language Processing (NLP) in terms of what's in an RNN.
In addition, it contains suggestion to improve machine learning based user interface, to engage users and encourage them to contribute data to adapt the models to them.
In the rapidly evolving landscape of artificial intelligence, ChatGPT stands as a beacon of innovation, continuously pushing the boundaries of what's possible in natural language understanding. As we peer into the future, it's evident that ChatGPT is poised to become an even more integral part of our daily lives, reshaping how we communicate, learn, and interact with technology.
One of the most exciting prospects for the future of ChatGPT is its potential to become even more contextually aware and emotionally intelligent. Imagine a ChatGPT that not only comprehends the words we type but also discerns the underlying emotions and intentions behind them. This heightened level of understanding could revolutionize customer service interactions, therapy sessions, and even personal conversations, fostering deeper connections and empathy in the digital realm.In the rapidly evolving landscape of artificial intelligence, ChatGPT stands as a beacon of innovation, continuously pushing the boundaries of what's possible in natural language understanding. As we peer into the future, it's evident that ChatGPT is poised to become an even more integral part of our daily lives, reshaping how we communicate, learn, and interact with technology.
One of the most exciting prospects for the future of ChatGPT is its potential to become even more contextually aware and emotionally intelligent. Imagine a ChatGPT that not only comprehends the words we type but also discerns the underlying emotions and intentions behind them. This heightened level of understanding could revolutionize customer service interactions, therapy sessions, and even personal conversations, fostering deeper connections and empathy in the digital realm.In the rapidly evolving landscape of artificial intelligence, ChatGPT stands as a beacon of innovation, continuously pushing the boundaries of what's possible in natural language understanding. As we peer into the future, it's evident that ChatGPT is poised to become an even more integral part of our daily lives, reshaping how we communicate, learn, and interact with technology.
One of the most exciting prospects for the future of ChatGPT is its potential to become even more contextually aware and emotionally intelligent. Imagine a ChatGPT that not only comprehends the words we type but also discerns the underlying emotions and intentions behind them. This heightened level of understanding could revolutionize customer service interactions, therapy sessions, and even personal conversations, fostering deeper connections and empathy in the digital realm
.https://www.pnytrainings.com/
Generative AI in CSharp with Semantic Kernel.pptxAlon Fliess
Join Alon Fliess, Azure MVP, and Microsoft RD in an enlightening lecture where C# meets the forefront of AI. Discover how the Semantic Kernel project bridges traditional programming with advanced AI, empowering C# developers to integrate AI functionalities into their software seamlessly.
Experience a paradigm shift in diagnostics through a real-world example: a sophisticated system crafted with C#, Semantic Kernel, and Azure. Witness the synergy of C# and AI in action, optimizing system analysis and problem-solving in complex environments.
Embark on a journey where C# and AI meet.
Нейрогибридные системы: на переднем крае нейронаук и искусственного интеллектаMikhail Burtsev
В последние годы начинает зарождаться новое направление в робототехнике - разработка нейрогибридных систем. Нейрогибридные системы комбинируют сеть живых нейронов и робото-техническую платформу в единый робо-организм в надежде совместить интел-лект живого мозга и эффективность мехатроники. Какие проблемы стоят на переднем крае опытов по вживлению нейронов в тело робота? Как заставить нейроны учиться вне мозга? Что дадут исследования нейрогибридных систем для развития фундаментальной и прикладной науки?
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
"Impact of front-end architecture on development cost", Viktor TurskyiFwdays
I have heard many times that architecture is not important for the front-end. Also, many times I have seen how developers implement features on the front-end just following the standard rules for a framework and think that this is enough to successfully launch the project, and then the project fails. How to prevent this and what approach to choose? I have launched dozens of complex projects and during the talk we will analyze which approaches have worked for me and which have not.
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
6. DeepPavlov.ai
Personal AI Assistants
Voice is universal
There’s no user manual needed, and people of all ages,
across all types of devices, and in many different
geographies can use the Assistant.
Scott Huffman
VP, Engineering, Google Assistant
9. DeepPavlov.ai
Modular dialog system
Are there any comedy movies to see this
weekend?
text data
NLU
(Natural Language Understanding)
• Domain detection
• Intent detection
• Entities detection
DM
(Dialogue manager)
• Dialogue state
• Policy
intent = request_movie
entities = { genre = ‘comedy’,
date = ‘weekend ’ }
semantic frame
NLG
(Natural Language Generation)
• Generative models
• Templates
action = request_location
system action
Where are you?
text data
10. DeepPavlov.ai
Modular dialog system
• Scalability problem
Paek, Tim, and Roberto Pieraccini. "Automating spoken dialogue management design using machine learning: An industry
perspective." Speech communication 50.8 (2008): 716-729.
13. DeepPavlov.ai
DeepPavlov
• DeepPavlov is for
- development of production ready chat-bots and complex conversational systems,
- NLP and dialog systems research.
• DeepPavlov’s goal is to enable AI-application developers and researchers with:
- set of pre-trained NLP models, pre-defined dialog system components (ML/DL/Rule-
based) and conversational agents templates for a typical scenarios;
- a framework for implementing and testing their own dialog models;
- tools for application integration with adjacent infrastructure (messengers, helpdesk
software etc.);
- benchmarking environment for conversational models and uniform access to
relevant datasets.
• Distributed under Apache v2 license
17. DeepPavlov.ai
From Conversational AI to Artificial General Intelligence
• Personal assistants as platforms
for AI skills
• Evolution towards
personalization and integration
of emerging AI skills
• Minsky’s ‘Society of mind’
19. DeepPavlov.ai
GitHub Stars & Pip DownloadsGitHubStars
Total downloads
36,914
Total downloads - 30 days
4,592
Total downloads - 7 days
1,736
https://pepy.tech/project/deeppavlov
20. DeepPavlov.ai
DeepPavlov Open Source Library
Task-Oriented Factoid Chit-Chat
Named Entity Recognition √ √
Coreference resolution √ √
Intent recognition √ √
Insults detection √ √
Q&A √
Dialogue Policy √ √
Dialog history √ √
Lanaguage model √ √ √
…
Dataset DSTC-2 SQuAD reddit
Skill Skill Skill
Agent
Models/Components
Dialogue agent combines complimentary
skills to help user.
21. DeepPavlov.ai
Core concepts
Agent is a conversational agent communicating
with users in natural language (text).
Skill fulfills user’s goal in some domain. It receives
input utterance and returns response and confidence.
Skill Manager performs selection of
the Skill to generate response.
Component is a reusable
functional component of Skill.
Chainer builds an agent/component pipeline from
heterogeneous Components (rule-based/ml/dl). It allows
to train and infer models in a pipeline as a whole.
25. DeepPavlov.ai
Use cases: Human assistance
1. Request routing
1. Domain classification
2. Routing to an operator
3. Operator replies
2. Ranking of pre-defined answers
1. Semantic embedding
2. Ranking of replies
3. Best answers are presented to an operator
4. Operator replies
1
2
3
1 2
4
3
26. DeepPavlov.ai
Use cases: Question Answering
1. Semantic embedding
2. Scoring of replies
3. Automated reply if the best answer has a high
confidence
4. Routing to an operator in the case of low
confidence
5. Operator replies
1. Semantic embedding
2. Search of answer in collection of documents
3. Automated reply if the best answer has a high
confidence
4. Routing to an operator in the case of low confidence
5. Operator replies
1 2
5
4
3
1 2
5
4
3
3. Frequently asked questions 4. Knowledge base Q&A
27. DeepPavlov.ai
Use cases: Rule-based bot
1. Semantic embedding
2. Selection of the most relevant dialogue script
3. Natural language answer generation
1. Semantic embedding
2. Sentiment analysis
3. Entity recognition tagging
4. Integration with BPM system
5. Simple bot 6. Other NLP tasks
1
2
3
1 2
4
4
3
31. DeepPavlov.ai
Features
Component Description
NER Based on neural Named Entity Recognition network with Bi-LSTM+CRF architecture.
Slot filling
Based on fuzzy Levenshtein search to extract normalized slot values from text. The components either rely on NER
results or perform needle in haystack search.
Classification
Component for classification tasks (intents, sentiment, etc). Based on shallow-and-wide Convolutional Neural
Network architecture. The model allows multilabel classification of sentences.
Automatic spelling
correction component
Pipelines that use candidates search in a static dictionary and an ARPA language model to correct spelling errors.
Ranking
Based on LSTM-based deep learning models for non-factoid answer selection. The model performs ranking of
responses or contexts from some database by their relevance for the given context.
Question Answering
Based on R-NET: Machine Reading Comprehension with Self-matching Networks. The model solves the task of
looking for an answer on a question in a given context (SQuAD task format).
Morphological tagging
Based on character-based approach to morphological tagging Heigold et al., 2017. An extensive empirical
evaluation of character-based morphological tagging for 14 languages. A state-of-the-art model for Russian and
several other languages. Model assigns morphological tags in UD format to sequences of words.
Skills
Goal-oriented bot
Based on Hybrid Code Networks (HCNs) architecture. It allows to predict responses in goal-oriented dialog. The
model is customizable: embeddings, slot filler and intent classifier can switched on and off on demand.
Seq2seq goal-oriented bot
Dialogue agent predicts responses in a goal-oriented dialog and is able to handle multiple domains (pretrained bot
allows calendar scheduling, weather information retrieval, and point-of-interest navigation). The model is end-to-
end differentiable and does not need to explicitly model dialogue state or belief trackers.
ODQA
An open domain question answering skill. The skill accepts free-form questions about the world and outputs an
answer based on its Wikipedia knowledge.
Embeddings
Pre-trained embeddings
for the Russian language
Word vectors for the Russian language trained on joint Russian Wikipedia and Lenta.ru corpora.
32. DeepPavlov.ai
Automatic spelling correction
• We provide two types of pipelines for spelling correction: levenshtein_corrector uses simple Damerau-
Levenshtein distance to find correction candidates and brillmoore uses statistics based error model for
it. In both cases correction candidates are chosen based on context with the help of a kenlm language
model.
Correction method F-measure Speed
(sentences/s)
Yandex.Speller 69.59 5.
[DP] Damerau Levenstein 1 + lm 53.50 29.3
[DP] Brill Moore top 4 + lm 52.91 0.6
Hunspell + lm 44.61 2.1
JamSpell 39.64 136.2
[DP] Brill Moore top 1 39.17 2.4
Hunspell 32.06 20.3
34. DeepPavlov.ai
Sentence classification
BERT models
BERT (Bidirectional Encoder Representations from Transformers) showed state-of-the-art results on a
wide range of NLP tasks in English.
deeppavlov.models.bert.BertClassifierModel (see here) provides easy to use solution for classification
problem using pre-trained BERT.
Neural Networks on Keras
deeppavlov.models.classifiers.KerasClassificationModel (see here) contains a number of different
neural network configurations for classification task.
•dcnn_model – Deep CNN with number of layers determined by the given number of kernel sizes and
filters,
•cnn_model – Shallow-and-wide CNN 1 with max pooling after convolution,
•cnn_model_max_and_aver_pool – Shallow-and-wide CNN 1 with max and average pooling
concatenation after convolution,
•bilstm_model – Bidirectional LSTM,
•bilstm_bilstm_model – 2-layers bidirectional LSTM,
•bilstm_cnn_model – Bidirectional LSTM followed by shallow-and-wide CNN,
•cnn_bilstm_model – Shallow-and-wide CNN followed by bidirectional LSTM,
•bilstm_self_add_attention_model – Bidirectional LSTM followed by self additive attention layer,
•bilstm_self_mult_attention_model – Bidirectional LSTM followed by self multiplicative attention layer,
•bigru_model – Bidirectional GRU model.
35. DeepPavlov.ai
Intent recognition
Source of all data except DeepPavlov is https://www.slideshare.net/KonstantinSavenkov/nlu-intent-detection-benchmark-by-intento-august-2017
F1
# of training samples
36. DeepPavlov.ai
Sentence classification
• Pre-trained models
Task Dataset Lang Model Metric Valid Test Downloads
28 intents DSTC 2
En
DSTC 2 emb
Accuracy
0.7732 0.7868 800 Mb
Wiki emb 0.9602 0.9593 8.5 Gb
7 intents SNIPS-2017
DSTC 2 emb
F1
0.8685 – 800 Mb
Wiki emb 0.9811 – 8.5 Gb
Tfidf + SelectKBest + PCA + Wiki emb 0.9673 – 8.6 Gb
Wiki emb weighted by Tfidf 0.9786 – 8.5 Gb
Insult
detection
Insults Reddit emb ROC-AUC 0.9271 0.8618 6.2 Gb
5 topics AG News Wiki emb
Accuracy
0.8876 0.9011 8.5 Gb
Sentiment
Twitter
mokoron
Ru
RuWiki+Lenta emb w/o preprocessing 0.9972 0.9971 6.2 Gb
RuWiki+Lenta emb with preprocessing 0.7811 0.7749 6.2 Gb
RuSentimen
t
RuWiki+Lenta emb
F1
0.6393 0.6539 6.2 Gb
ELMo 0.7066 0.7301 700 Mb
Intent Yahoo-L31
Yahoo-L31 on ELMo pre-trained on Yahoo-
L6
ROC-AUC 0.9269 – 700 Mb
38. DeepPavlov.ai
Neural Ranking
Trained with triplet loss and hard negative sampling
Tan, Ming & Dos Santos, Cicero & Xiang, Bing & Zhou, Bowen. (2015). LSTM-based Deep Learning Models for
Non-factoid Answer Selection.
Dataset Model config Validation (Recall@1) Test1 (Recall@1) Downloads
Ubuntu V2 ranking_ubuntu_v2_interact 52.9 52.4 8913M
Ubuntu V2 ranking_ubuntu_v2_mt_interact 59.2 58.7 8906M
Dataset Model config
Val
(accuracy)
Test
(accuracy)
Val (F1) Test (F1) Val (log_loss) Test (log_loss) Downloads
paraphraser.ru
paraphrase_ident_parap
hraser
83.8 75.4 87.9 80.9 0.468 0.616 5938M
Quora
Question Pairs
paraphrase_ident_qqp 87.1 87.0 83.0 82.6 0.300 0.305 8134M
Quora
Question Pairs
paraphrase_ident_qqp 87.7 87.5 84.0 83.8 0.287 0.298 8136M
Model
Validation
(Recall@1)
Test1
(Recall@1)
Architecture II (HLQA(200) CNNQA(4000) 1-
MaxPooling Tanh)
61.8 62.8
QA-LSTM basic-model(max pooling) 64.3 63.1
ranking_insurance 72.0 72.2
39. DeepPavlov.ai
Teхt QA (SQuAD) + Open Domain QA
R-NET: Machine Reading Comprehension with Self-matching Networks. (2017)
Model
(single
model)
EM (dev) F-1 (dev)
DeepPavlov
BERT
80.88 88.49
DeepPavlov
R-Net
71.49 80.34
BiDAF + Self
Attention +
ELMo
– 85.6
R-Net 71.1 79.5
Model Lang
Ranker@5
F1 EM
DeepPavlov En 37.83 31.26
DrQA 1 - 27.1
R3 4 37.5 29.1
DeepPavlov with
RuBERT reader
Ru
42.02 29.56
DeepPavlov 28.56 18.17
Teхt QA (SQuAD) Open Domain (Wiki) QA
40. DeepPavlov.ai
Task-Oriented Dialog (DSTC-2)
Model
Test turn
textual
accuracy
basic bot 0.3809
bot with slot filler & fasttext embeddings 0.5317
bot with slot filler & intents 0.5248
bot with slot filler & intents & embeddings 0.5145
bot with slot filler & embeddings & attention 0.5551
Bordes and Weston (2016) [4] 0.411
Perez and Liu (2016) [5] 0.487
Eric and Manning (2017) [6] 0.480
Williams et al. (2017) [1] 0.556
Jason D. Williams, Kavosh Asadi, Geoffrey Zweig “Hybrid Code Networks: practical and efficient end-to-
end dialog control with supervised and reinforcement learning” – 2017
41. DeepPavlov.ai
Sequence-To-Sequence Dialogue Bot For Goal-Oriented Task
Model Test BLEU
DeepPavlov implementation of
KV Retrieval Net
13.2
KV Retrieven Net from [1] 13.2
Copy Net from [1] 11.0
Attn. Seq2Seq from [1] 10.2
Rule-Based from [1] 6.60
[1] Mihail Eric, Lakshmi Krishnan, Francois Charette, and Christopher D. Manning, “Key-Value Retrieval Networks for
Task-Oriented Dialogue – 2017
Model Test BLEU
Weather Navigation Schedules
DeepPavlov implementation of
KV Retrieval Net
14.6 12.5 11.9
Wen et al [2] 14.9 13.7 -
[2] Haoyang Wen, Yijia Liu, Wanxiang Che, Libo Qin and Ting Liu. Sequence-to-Sequence Learning for Task-
oriented Dialogue with Dialogue State Representation. COLING 2018.
42. DeepPavlov.ai
Latest release
DeepPavlov 0.3.0
* BERT-based models for ranking, NER,
classification and Text Q&A (SQuAD)
* New SMN, DAM, DAM-USE-T ranking models
* Multilingual NER for 100 languages
* New AIML wrapper component
43. DeepPavlov.ai
Future steps
• Better usability
- Improved Python API
- Tutorials, How to examples
• Support for script based skills
- Python API with script uploading from file
- DSL
- GUI tool for fast script prototyping
• Skill manager
- Implementation of baseline multi-skill manager with ranking model
- Adding of rich context to the skill manager
• Research
- Training with low data (transfer learning, language models etc.)
- Better dialogue models combining knowledge graphs and deep learning to address lack of common
sense in current solutions
44. DeepPavlov.ai
• Code
- https://github.com/deepmipt/DeepPavlov
• Documentation
- http://docs.deeppavlov.ai/
• Demo (experimental, not all models have the same performance as in the library)
- http://demo.ipavlov.ai/
• Tutorials
- Simple text classification skill of DeepPavlov
▪ https://towardsdatascience.com/simple-text-classification-skill-of-deeppavlov-54bc1b61c9ea
- Open-domain question answering with DeepPavlov
▪ https://medium.com/deeppavlov/open-domain-question-answering-with-deeppavlov-c665d2ee4d65
• References
- Burtsev M., et al. DeepPavlov: Open-Source Library for Dialogue Systems // Proceedings of ACL 2018, System Demonstrations (2018): 122-127.
- Burtsev M., et al. DeepPavlov: An Open Source Library for Conversational AI // Proceedings of NeurIPS 2018, MLOSS Workshop, 2018.
DeepPavlov.ai
45. DeepPavlov.ai
Q&A
1. What is an "ideal" framework for development of conversational
agents?
2. Do we need an "operating system" for conversational AI agents? If yes,
then how should it look like?
3. What are the most promising fields of application /verticals for the
conversational AI right now?
4. Looking into the future of ML/AI, how will conversational AI evolve
and interrelate with other research and technology directions?