Keynote talk at HILDA'2023 at SIGMOD on June 18, 2023.
Abstract: The ability to build large-scale knowledge bases that capture and extend the implicit knowledge of human experts is the foundation for many AI systems. We use an ontology-driven approach for the building, growing and serving of such knowledge bases. This approach relies on several well-known building blocks: document conversion, natural language processing, entity resolution, data transformation and fusion. In this talk, I will discuss wide range of real-world challenges related to the building of these blocks and present our work to address these challenges via better human-machine cooperation.
Learning to summarize from human feedbackharmonylab
公開URL:https://arxiv.org/abs/2009.01325
出典:Nisan Stiennon, Long Ouyang, Jeff Wu, Daniel M. Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, Paul Christiano : Learning to summarize from human feedback, arXiv:2009.01325 (2020)
概要:言語モデルが強力になるにつれて、モデルの学習と評価は特定のタスクで使用されるデータとメトリクスによってボトルネックになることが多い。要約モデルでは人間が作成した参照要約を予測するように学習され、ROUGEによって評価されることが多い。しかし、これらのメトリクスと人間が本当に気にしている要約の品質との間にはズレが存在する。本研究では、大規模で高品質な人間のフィードバックデータセットを収集し、人間が好む要約を予測するモデルを学習する。そのモデルを報酬関数として使用して要約ポリシーをfine-tuneする。TL;DRデータセットにおいて本手法を適用したところ、人間の評価において参照要約よりも上回ることがわかった。
Learning to summarize from human feedbackharmonylab
公開URL:https://arxiv.org/abs/2009.01325
出典:Nisan Stiennon, Long Ouyang, Jeff Wu, Daniel M. Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, Paul Christiano : Learning to summarize from human feedback, arXiv:2009.01325 (2020)
概要:言語モデルが強力になるにつれて、モデルの学習と評価は特定のタスクで使用されるデータとメトリクスによってボトルネックになることが多い。要約モデルでは人間が作成した参照要約を予測するように学習され、ROUGEによって評価されることが多い。しかし、これらのメトリクスと人間が本当に気にしている要約の品質との間にはズレが存在する。本研究では、大規模で高品質な人間のフィードバックデータセットを収集し、人間が好む要約を予測するモデルを学習する。そのモデルを報酬関数として使用して要約ポリシーをfine-tuneする。TL;DRデータセットにおいて本手法を適用したところ、人間の評価において参照要約よりも上回ることがわかった。
情報システム障害解析のための知識グラフ構築の試み / Constructing a knowledge graph for information sys...Shinji Takao
人工知能学会 第25回知識流通ネットワーク研究会発表 http://sigksn.html.xdomain.jp/conf25/index.html
システム障害解析に関する専門家知識の抽出、グラフ化、DB化を行った際得られた知見と、知識流通手段としての知識グラフの可能性と課題を考察した結果を報告します。
Knowledge graphs have been getting attention because of its relevance to interpretable AI. Not only that, they also can be useful as a knowledge sharing mean which enable non-experts to utilize experts’ knowledge. We aim to report findings from constructing a knowledge graph through eliciting experts’ knowledge and building a knowledge database. We also suggest the possibilities and issues of knowledge graph as a knowledge sharing mean.
Presentation for tutorial session 'Measuring scholarly impact: Methods and practice' at ISSI2015
Explains how to use linkpred: https://github.com/rafguns/linkpred
Information Technology in Industry(ITII) - November Issue 2018ITIIIndustries
IT Industry publishes original research articles, review articles, and extended versions of conference papers. Articles resulting from research of both theoretical and/or practical natures performed by academics and/or industry practitioners are welcome. IT in Industry aims to become a leading IT journal with a high impact factor.
I gave this talk at the 'Digital Twin Conference' hosted by LH Corp at COEX, Seoul on August 8th, 2019.
Abstract: 'Digital Twin' is a digital replication of real world objects, processes, phenomena that can be used for various purposes. Digital twin concept backs to manufacturing industry in early 2000s for the PLM (Product Lifecycle Management) purposes. It is based on the idea that a digital informational construct about a physical system could be created as an entity on its own. Definitions of digital twin emphasize the three important levels or characteristics. At first, there should be connection between real physical world and corresponding virtual world. To do this, Level 1 digital twin provides virtual 3D models. Secondly, this connection between real world and virtual world is established by generating (near) real time data using sensors or IoT. This is called Level 2 digital twin. Thirdly, Level 3 digital twin carries out certain analyses, predictions, and simulations using virtual 3D and (near) real time data. ‘Smart Spaces’ are interactive environments where humans and technology can openly communicate with each other in a physical or digital setting. Examples of smart spaces include smart cities, smart factories, and smart homes. ‘Smart Spaces’ is one of Garner’s Top 10 Tech Trends for 2019. As spaces are going through digital transformation with 4th industrial revolution, there are many attempts to apply digital twin technology to manage urban, spatial, and industrial issues around the world. Those attempts look set to play an increasingly important role in the creation of smart cities, smart factories, and smart homes. Bringing the virtual and real worlds together in this way can help to give better analysis, visualization, and simulation to the decision-making process. This will be a multi-way process with iterative feedback among stakeholders.
In this talk, I'll share my real experiences in carrying out digital twin and smart space projects. Also I’ll talk about what I’ve learnt from these projects.
情報システム障害解析のための知識グラフ構築の試み / Constructing a knowledge graph for information sys...Shinji Takao
人工知能学会 第25回知識流通ネットワーク研究会発表 http://sigksn.html.xdomain.jp/conf25/index.html
システム障害解析に関する専門家知識の抽出、グラフ化、DB化を行った際得られた知見と、知識流通手段としての知識グラフの可能性と課題を考察した結果を報告します。
Knowledge graphs have been getting attention because of its relevance to interpretable AI. Not only that, they also can be useful as a knowledge sharing mean which enable non-experts to utilize experts’ knowledge. We aim to report findings from constructing a knowledge graph through eliciting experts’ knowledge and building a knowledge database. We also suggest the possibilities and issues of knowledge graph as a knowledge sharing mean.
Presentation for tutorial session 'Measuring scholarly impact: Methods and practice' at ISSI2015
Explains how to use linkpred: https://github.com/rafguns/linkpred
Information Technology in Industry(ITII) - November Issue 2018ITIIIndustries
IT Industry publishes original research articles, review articles, and extended versions of conference papers. Articles resulting from research of both theoretical and/or practical natures performed by academics and/or industry practitioners are welcome. IT in Industry aims to become a leading IT journal with a high impact factor.
I gave this talk at the 'Digital Twin Conference' hosted by LH Corp at COEX, Seoul on August 8th, 2019.
Abstract: 'Digital Twin' is a digital replication of real world objects, processes, phenomena that can be used for various purposes. Digital twin concept backs to manufacturing industry in early 2000s for the PLM (Product Lifecycle Management) purposes. It is based on the idea that a digital informational construct about a physical system could be created as an entity on its own. Definitions of digital twin emphasize the three important levels or characteristics. At first, there should be connection between real physical world and corresponding virtual world. To do this, Level 1 digital twin provides virtual 3D models. Secondly, this connection between real world and virtual world is established by generating (near) real time data using sensors or IoT. This is called Level 2 digital twin. Thirdly, Level 3 digital twin carries out certain analyses, predictions, and simulations using virtual 3D and (near) real time data. ‘Smart Spaces’ are interactive environments where humans and technology can openly communicate with each other in a physical or digital setting. Examples of smart spaces include smart cities, smart factories, and smart homes. ‘Smart Spaces’ is one of Garner’s Top 10 Tech Trends for 2019. As spaces are going through digital transformation with 4th industrial revolution, there are many attempts to apply digital twin technology to manage urban, spatial, and industrial issues around the world. Those attempts look set to play an increasingly important role in the creation of smart cities, smart factories, and smart homes. Bringing the virtual and real worlds together in this way can help to give better analysis, visualization, and simulation to the decision-making process. This will be a multi-way process with iterative feedback among stakeholders.
In this talk, I'll share my real experiences in carrying out digital twin and smart space projects. Also I’ll talk about what I’ve learnt from these projects.
Abstract. Enterprise adoption of AI/ML services has significantly accelerated in the last few years. However, the majority of ML models are still developed with the goal of solving a single task, e.g., predictiction, classification. In this talk, Debmalya Biswas will present the emerging paradigm of Compositional AI, also known as, Compositional Learning. Compositional AI envisions seamless composition of existing AI/ML services, to provide a new (composite) AI/ML service, capable of addressing complex multi-domain use-cases. In an enterprise context, this enables reuse, agility, and efficiency in development and maintenance efforts.
Rethinking enterprise architecture for DevOps, Agile, and cloud native organi...Michael Coté
Current application theory says that all responsibility for software should be pushed down to the actual DevOps-style team writing, delivering, and running the software. This leaves the EA role in the dust, seemingly killing it off. In addition to this being disquieting to EAs out there who have steep mortgage payments and other expensive hobbies, it seems to drop out the original benefits of enterprise architecture, namely oversight of all IT-related activities to make sure things don’t go wrong (e.g., spending, poor tech choices, problematic integration, etc.) and that things, rather, go right.
As presented at the O'Reilly Software Architecture Conference in Berlin, November 2019.
Strata 2016 - Architecting for Change: LinkedIn's new data ecosystemShirshanka Das
Shirshanka Das and Yael Garten describe how LinkedIn redesigned its data analytics ecosystem in the face of a significant product rewrite, covering the infrastructure changes that enable LinkedIn to roll out future product innovations with minimal downstream impact. Shirshanka and Yael explore the motivations and the building blocks for this reimagined data analytics ecosystem, the technical details of LinkedIn’s new client-side tracking infrastructure, its unified reporting platform, and its data virtualization layer on top of Hadoop and share lessons learned from data producers and consumers that are participating in this governance model. Along the way, they offer some anecdotal evidence during the rollout that validated some of their decisions and are also shaping the future roadmap of these efforts.
Architecting for change: LinkedIn's new data ecosystemYael Garten
2016 StrataHadoop NYC conference talk.
http://conferences.oreilly.com/strata/hadoop-big-data-ny/public/schedule/detail/52182
Abstract:
Last year, LinkedIn embarked on an ambitious mission to completely revamp the mobile experience for its members. This would mean a completely new mobile application, reimagined user experiences, and new interaction concepts. As the team evaluated the impact of this big rewrite on the data analytics ecosystem, they observed a few problems.
Over the past few years, LinkedIn has become extremely good at incrementally changing the site one mini-feature at a time, often in conjunction with hundreds of other incremental changes. LinkedIn’s experimentation platform ensures that it is always monitoring a wide gamut of impacted metrics with every change before rolling fully forward. However, when it comes to rolling out a big change like this, different challenges crop up. You have to rollout the entire application all at once; the new experience means that you have no baseline on new metrics; and existing metrics may see double digit changes just because of the new experience or because the metric’s logic is no longer accurate—the challenge is in figuring out which is which.
Shirshanka Das and Yael Garten describe how LinkedIn redesigned its data analytics ecosystem in the face of a significant product rewrite, covering the infrastructure changes that enable LinkedIn to roll out future product innovations with minimal downstream impact. Shirshanka and Yael explore the motivations and the building blocks for this reimagined data analytics ecosystem, the technical details of LinkedIn’s new client-side tracking infrastructure, its unified reporting platform, and its data virtualization layer on top of Hadoop and share lessons learned from data producers and consumers that are participating in this governance model. Along the way, they offer some anecdotal evidence during the rollout that validated some of their decisions and are also shaping the future roadmap of these efforts.
Choosing the Right Technologies A Guide to Frameworks and Tools for Web App D...BitCot
Frameworks play a pivotal role in streamlining development workflows and ensuring scalability and maintainability of web applications. One of the most popular choices is React.js, known for its component-based architecture and virtual DOM, facilitating efficient rendering and updating of user interfaces. AngularJS is another prominent framework, offering robust features for building dynamic single-page applications. Meanwhile, Vue.js has gained traction for its simplicity and flexibility, making it ideal for projects of varying complexities.
The Role of Patterns in the Era of Large Language ModelsYunyao Li
Slides for my keynote at PAN-DL Workshop (Pattern-based Approaches to NLP in the Age of Deep Learning) at EMNLP'2023 (December. 6, 2023).
In this talk, I share our initial learnings from constructing, growing and serving large knowledge graphs
Meaning Representations for Natural Languages: Design, Models and ApplicationsYunyao Li
EMNLP'2022 Tutorial "Meaning Representations for Natural Languages: Design, Models and Applications"
Instructors: Jeffrey Flanigan, Ishan Jindal, Yunyao Li, Tim O’Gorman, Martha Palmer
Abstract:
We propose a cutting-edge tutorial that reviews the design of common meaning representations, SoTA models for predicting meaning representations, and the applications of meaning representations in a wide range of downstream NLP tasks and real-world applications. Reporting by a diverse team of NLP researchers from academia and industry with extensive experience in designing, building and using meaning representations, our tutorial has three components: (1) an introduction to common meaning representations, including basic concepts and design challenges; (2) a review of SoTA methods on building models for meaning representations; and (3) an overview of applications of meaning representations in downstream NLP tasks and real-world applications. We will also present qualitative comparisons of common meaning representations and a quantitative study on how their differences impact model performance. Finally, we will share best practices in choosing the right meaning representation for downstream tasks.
Natural language understanding is a fundamental task in artificial intelligence. English understanding has reached a mature state and successfully deployed in multiple IBM AI products and services, such as Watson Natural Language Understanding and Watson Discovery. However, scaling existing products/services to support additional languages remain an open challenge. In this talk, we will discuss the open challenges in supporting universal natural language understanding. We will share our work in the past few years in addressing these challenges. We will also showcase how universal semantic representation of natural languages can enable cross-lingual information extraction in concrete domains (e.g. compliance) and show ongoing efforts towards seamless scaling existing NLP capabilities across languages with minimal efforts.
Invited talk at Document Intelligence workshop at KDD'2021.
Harvesting information from complex documents such as in financial reports and scientific publications is critical to building AI applications for business and research. Such documents are often in PDF format with critical facts and data conveyed in table and graphs. Extracting such information is essential to extract insights from these documents. In IBM Research, we have a rich agenda in this area that we call Deep Document Understanding. In this talk, I will focus on our research on Deep Table Understanding — extracting and understanding tables from PDF documents. I will introduce key challenges in table extraction and understanding and how we address such challenges, from how to acquire data at scale to enable deep neural network models to how to build, customize and evaluate such models. I will also describe how our work enables real-world use cases in domains such as finance and life science. Finally, I will briefly present TableQA, an important downstream task enabled by Deep Table Understanding.
Explainability for Natural Language ProcessingYunyao Li
Final deck for our popular tutorial on "Explainability for Natural Language Processing" at KDD'2021. See links below for downloadable version (with higher resolution) and recording of the live tutorial.
Title: Explainability for Natural Language Processing
Presenter: Marina Danilevsky, Shipi Dhanorkar, Yunyao Li and Lucian Popa and Kun Qian and Anbang Xu
Website: http://xainlp.github.io/
Recording: https://www.youtube.com/watch?v=PvKOSYGclPk&t=2s
Downloadable version with higher resolution: https://drive.google.com/file/d/1_gt_cS9nP9rcZOn4dcmxc2CErxrHW9CU/view?usp=sharing
@article{kdd2021xaitutorial,
title={Explainability for Natural Language Processing},
author= {Marina Danilevsky, Shipi Dhanorkar and Yunyao Li and Lucian Popa and Kun Qian and Anbang Xu},
journal={KDD},
year={2021}
}
Abstract:
This lecture-style tutorial, which mixes in an interactive literature browsing component, is intended for the many researchers and practitioners working with text data and on applications of natural language processing (NLP) in data science and knowledge discovery. The focus of the tutorial is on the issues of transparency and interpretability as they relate to building models for text and their applications to knowledge discovery. As black-box models have gained popularity for a broad range of tasks in recent years, both the research and industry communities have begun developing new techniques to render them more transparent and interpretable.Reporting from an interdisciplinary team of social science, human-computer interaction (HCI), and NLP/knowledge management researchers, our tutorial has two components: an introduction to explainable AI (XAI) in the NLP domain and a review of the state-of-the-art research; and findings from a qualitative interview study of individuals working on real-world NLP projects as they are applied to various knowledge extraction and discovery at a large, multinational technology and consulting corporation. The first component will introduce core concepts related to explainability inNLP. Then, we will discuss explainability for NLP tasks and reporton a systematic literature review of the state-of-the-art literaturein AI, NLP and HCI conferences. The second component reports on our qualitative interview study, which identifies practical challenges and concerns that arise in real-world development projects that require the modeling and understanding of text data.
Explainability for Natural Language ProcessingYunyao Li
NOTE: Please check out the final version here with small but important updates and links to downloadable version and recording: https://www.slideshare.net/YunyaoLi/explainability-for-natural-language-processing-249992241
Updated version on our popular tutorial on "Explainability for Natural Language Processing" as a tutorial at KDD'2021.
Title: Explainability for Natural Language Processing
@article{kdd2021xaitutorial,
title={Explainability for Natural Language Processing},
author= {Marina Danilevsky, Dhanorkar, Shipi and Li, Yunyao and Lucian Popa and Kun Qian and Anbang Xu},
journal={KDD},
year={2021}
}
Presenter: Marina Danilevsky, Dhanorkar, Shipi and Li, Yunyao and Lucian Popa and Kun Qian and Anbang Xu
Website: http://xainlp.github.io/
Abstract:
This lecture-style tutorial, which mixes in an interactive literature browsing component, is intended for the many researchers and practitioners working with text data and on applications of natural language processing (NLP) in data science and knowledge discovery. The focus of the tutorial is on the issues of transparency and interpretability as they relate to building models for text and their applications to knowledge discovery. As black-box models have gained popularity for a broad range of tasks in recent years, both the research and industry communities have begun developing new techniques to render them more transparent and interpretable.Reporting from an interdisciplinary team of social science, human-computer interaction (HCI), and NLP/knowledge management researchers, our tutorial has two components: an introduction to explainable AI (XAI) in the NLP domain and a review of the state-of-the-art research; and findings from a qualitative interview study of individuals working on real-world NLP projects as they are applied to various knowledge extraction and discovery at a large, multinational technology and consulting corporation. The first component will introduce core concepts related to explainability inNLP. Then, we will discuss explainability for NLP tasks and reporton a systematic literature review of the state-of-the-art literaturein AI, NLP and HCI conferences. The second component reports on our qualitative interview study, which identifies practical challenges and concerns that arise in real-world development projects that require the modeling and understanding of text data.
Human in the Loop AI for Building Knowledge Bases Yunyao Li
The ability to build large-scale domain-specific knowledge bases that capture and extend the implicit knowledge of human experts is the foundation for many AI systems. We use an ontology-driven approach for the creation, representation and consumption of such domain-specific knowledge bases. This approach relies on several well-known building blocks: natural language processing, entity resolution, data transformation and fusion. I will present several human-in-the-loop work that target domain experts (rather than programmers) to extract the domain knowledge from the human expert and map it into the "right" models or algorithms. I will also share successful use cases in several domains, including Compliance, Finance, and Healthcare: by using these tools we can match the level of accuracy achieved by manual efforts, but at a significantly lower cost and much higher scale and automation.
Slides for talk given at Women in Engineering on March 20, 2021.
Abstract:
Natural language understanding is a fundamental task in artificial intelligence. English understanding has reached a mature state and successfully deployed in multiple IBM AI products and services, such as Watson Natural Language Understanding and Watson Discovery. However, scaling existing products/services to support additional languages remain an open challenge. In this talk, we will discuss the open challenges in supporting universal natural language understanding. We will share our work in the past few years in addressing these challenges. We will also showcase how universal semantic representation of natural languages can enable cross-lingual information extraction in concrete domains (e.g. compliance) and show ongoing efforts towards seamless scaling existing NLP capabilities across languages with minimal efforts.
Explainability for Natural Language ProcessingYunyao Li
Tutorial at AACL'2020 (http://www.aacl2020.org/program/tutorials/#t4-explainability-for-natural-language-processing).
More recent version: https://www.slideshare.net/YunyaoLi/explainability-for-natural-language-processing-249912819
Title: Explainability for Natural Language Processing
@article{aacl2020xaitutorial,
title={Explainability for Natural Language Processing},
author= {Dhanorkar, Shipi and Li, Yunyao and Popa, Lucian and Qian, Kun and Wolf, Christine T and Xu, Anbang},
journal={AACL-IJCNLP 2020},
year={2020}
Presenter: Shipi Dhanorkar, Christine Wolf, Kun Qian, Anbang Xu, Lucian Popa and Yunyao Li
Video: https://www.youtube.com/watch?v=3tnrGe_JA0s&feature=youtu.be
Abstract:
We propose a cutting-edge tutorial that investigates the issues of transparency and interpretability as they relate to NLP. Both the research community and industry have been developing new techniques to render black-box NLP models more transparent and interpretable. Reporting from an interdisciplinary team of social science, human-computer interaction (HCI), and NLP researchers, our tutorial has two components: an introduction to explainable AI (XAI) and a review of the state-of-the-art for explainability research in NLP; and findings from a qualitative interview study of individuals working on real-world NLP projects at a large, multinational technology and consulting corporation. The first component will introduce core concepts related to explainability in NLP. Then, we will discuss explainability for NLP tasks and report on a systematic literature review of the state-of-the-art literature in AI, NLP, and HCI conferences. The second component reports on our qualitative interview study which identifies practical challenges and concerns that arise in real-world development projects which include NLP.
Towards Universal Language Understanding (2020 version)Yunyao Li
Keynote talk given at Pacific Asia Conference on Language, Information and Computation (PACLIC 34) on Pacific Asia Conference on Language, Information and Computation (PACLIC 34) on October 24, 2020.
Title: Towards Universal Natural Language Understanding
Abstract:
Understanding the semantics of the natural language is a fundamental task in artificial intelligence. English semantic understanding has reached a mature state and successfully deployed in multiple IBM AI products and services, such as Watson Natural Language Understanding and Watson Compare and Comply. However, scaling existing products/services to support additional languages remain an open challenge. In this talk, we will discuss the open challenges in supporting universal natural language understanding. We will share our work in addressing these challenges in the past few years to provide the same unified semantic representation across languages. We will also showcase how such universal semantic understanding of natural languages can enable cross-lingual information extraction in concrete domains (e.g. insurance and compliance) and show promise towards seamless scaling existing NLP capabilities across languages with minimal efforts.
Towards Universal Semantic Understanding of Natural LanguagesYunyao Li
Keynote talk at TextXD 2019(https://www.textxd.org)
Abstract:
Understanding the semantics of the natural language is a fundamental task in artificial intelligence. English semantic understanding has reached a mature state and successfully deployed in multiple IBM AI products and services, such as Watson Natural Language Understanding and Watson Compare and Comply. However, scaling existing products/services to support additional languages remain an open challenge. In this demo, we will present Polyglot, a multilingual semantic parser capable of semantically parsing sentences in 9 different languages from 4 different language groups into the same unified semantic representation. We will also showcase how such universal semantic understanding of natural languages can enable cross-lingual information extraction in concrete domains (e.g. insurance and compliance) and show promise towards seamless scaling existing NLP capabilities across languages with minimal efforts.
An In-depth Analysis of the Effect of Text Normalization in Social MediaYunyao Li
Poster corresponding to our NAACL'2015 paper "An In-depth Analysis of the Effect of Text Normalization in Social Media"
Abstract: Recent years have seen increased interest in text normalization in social media, as the in-formal writing styles found in Twitter and other social media data often cause problems for NLP applications. Unfortunately, most current approaches narrowly regard the nor- malization task as a “one size fits all" task of replacing non-standard words with their standard counterparts. In this work we build a taxonomy of normalization edits and present a study of normalization to examine its effect on three different downstream applications (de- pendency parsing, named entity recognition, and text-to-speech synthesis). The results sug- gest that how the normalization task should be viewed is highly dependent on the targeted application. The results also show that normalization must be thought of as more than word replacement in order to produce results comparable to those seen on clean text.
Paper: https://www.aclweb.org/anthology/N15-1045
Exploiting Structure in Representation of Named Entities using Active LearningYunyao Li
Slides for our COLING'18 paper: http://aclweb.org/anthology/C18-1058
Fundamental to several knowledge-centric applications is the need to identify named entities from their textual mentions. However, entities lack a unique representation and their mentions can differ greatly. These variations arise in complex ways that cannot be captured using textual similarity metrics. However, entities have underlying structures, typically shared by entities of the same entity type, that can help reason over their name variations. Discovering, learning and manipulating these structures typically requires high manual effort in the form of large amounts of labeled training data and handwritten transformation programs. In this work, we propose an active-learning based framework that drastically reduces the labeled data required to learn the structures of entities. We show that programs for mapping entity mentions to their structures can be automatically generated using human-comprehensible labels. Our experiments show that our framework consistently outperforms both handwritten programs and supervised learning models. We also demonstrate the utility of our framework in relation extraction and entity resolution tasks.
K-SRL: Instance-based Learning for Semantic Role LabelingYunyao Li
Slides for our COLING'16 paper http://aclweb.org/anthology/C/C16/C16-1058.pdf
Abstract:
Semantic role labeling (SRL) is the task of identifying and labeling predicate-argument structures in sentences with semantic frame and role labels. A known challenge in SRL is the large number of low-frequency exceptions in training data, which are highly context-specific and difficult to generalize. To overcome this challenge, we propose the use of instance-based learning that performs no explicit generalization, but rather extrapolates predictions from the most similar instances in the training data. We present a variant of k-nearest neighbors (kNN) classification with composite features to identify nearest neighbors for SRL. We show that high-quality predictions can be derived from a very small number of similar instances. In a comparative evaluation we experimentally demonstrate that our instance-based learning approach significantly outperforms current state-of-the-art systems on both in-domain and out-of-domain data, reaching F1-scores
of 89,28% and 79.91% respectively
Natural Language Data Management and Interfaces: Recent Development and Open ...Yunyao Li
Slides deck for SIGMOD 2017 Tutorial.
ABSTRACT:
The volume of natural language text data has been rapidly increasing over the past two decades, due to factors such as the growth of the Web, the low cost associated to publishing and the progress on the digitization of printed texts. This growth combined with the proliferation of natural language systems for search and retrieving information provides tremendous opportunities for studying some of the areas where database systems and natural language processing systems overlap. This tutorial explores two more relevant areas of overlap to the database community: (1) managing
natural language text data in a relational database, and (2) developing natural language interfaces to databases. The tutorial presents state-of-the-art methods, related systems, research opportunities and challenges covering both area.
Polyglot: Multilingual Semantic Role Labeling with Unified LabelsYunyao Li
Poster for our ACL paper "Polyglot: Multilingual Semantic Role Labeling with Unified Labels".
Abstract:
We present POLYGLOT, a semantic role labeling system capable of semantically parsing sentences in 9 different languages from 4 different language groups. A core differentiator is that this system predicts English Proposition Bank labels for all supported languages. This means that
for instance a Japanese sentence will be tagged with the same labels as an English sentence with similar semantics would be. This is made possible by training the system with target language data that was automatically labeled with English PropBank labels using an annotation projection approach. We give an overview of our system, the automatically produced training data, and discuss possible applications
and limitations of this work. We present a demonstrator that accepts sentences in English, German, French, Spanish, Japanese, Chinese, Arabic, Russian and Hindi and
outputs a visualization of its shallow semantics.
Invited Talk at Modern Data Management Systems Summit on August 29-30, 2014 at Tsinghua University in Beijing, China.
http://ise.thss.tsinghua.edu.cn/MDMS/English/program.jsp
Abstract:
Modern enterprises are increasingly relying on complex analyses on large data sets to drive business decisions. Tasks such as root cause analysis from system logs and lead generation based on social media, customer retention and digital marketing are rapidly gaining importance. These applications generally consist of three major analytic phases: text analytics, semi-structured data processing (joins, group-by, aggregation), and statistical/predictive modeling. The size of the datasets in conjunction with the complexity of the analysis necessitates large-scale distributed processing of the analytical algorithms. At IBM we are building tools and technologies based on declarative languages to support each of these analytic phases. The declarative nature of the language abstracts away the need for programmer-optimization. Furthermore, the syntax of these languages is designed to appeal to the corresponding communities. As an example for statistical modeling, we expose a high-level language with syntax similar to R -- a very popular statistical processing language.
In this talk I will give an overview of some real-world big data applications we are currently working on and use that to motivate the need for declarative analytics consisting of the three major phases discussed above. I will then describe, in some detail, declarative systems for text analytics along with a discussion on speeds, feeds and comparisons.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Welocme to ViralQR, your best QR code generator.ViralQR
Welcome to ViralQR, your best QR code generator available on the market!
At ViralQR, we design static and dynamic QR codes. Our mission is to make business operations easier and customer engagement more powerful through the use of QR technology. Be it a small-scale business or a huge enterprise, our easy-to-use platform provides multiple choices that can be tailored according to your company's branding and marketing strategies.
Our Vision
We are here to make the process of creating QR codes easy and smooth, thus enhancing customer interaction and making business more fluid. We very strongly believe in the ability of QR codes to change the world for businesses in their interaction with customers and are set on making that technology accessible and usable far and wide.
Our Achievements
Ever since its inception, we have successfully served many clients by offering QR codes in their marketing, service delivery, and collection of feedback across various industries. Our platform has been recognized for its ease of use and amazing features, which helped a business to make QR codes.
Our Services
At ViralQR, here is a comprehensive suite of services that caters to your very needs:
Static QR Codes: Create free static QR codes. These QR codes are able to store significant information such as URLs, vCards, plain text, emails and SMS, Wi-Fi credentials, and Bitcoin addresses.
Dynamic QR codes: These also have all the advanced features but are subscription-based. They can directly link to PDF files, images, micro-landing pages, social accounts, review forms, business pages, and applications. In addition, they can be branded with CTAs, frames, patterns, colors, and logos to enhance your branding.
Pricing and Packages
Additionally, there is a 14-day free offer to ViralQR, which is an exceptional opportunity for new users to take a feel of this platform. One can easily subscribe from there and experience the full dynamic of using QR codes. The subscription plans are not only meant for business; they are priced very flexibly so that literally every business could afford to benefit from our service.
Why choose us?
ViralQR will provide services for marketing, advertising, catering, retail, and the like. The QR codes can be posted on fliers, packaging, merchandise, and banners, as well as to substitute for cash and cards in a restaurant or coffee shop. With QR codes integrated into your business, improve customer engagement and streamline operations.
Comprehensive Analytics
Subscribers of ViralQR receive detailed analytics and tracking tools in light of having a view of the core values of QR code performance. Our analytics dashboard shows aggregate views and unique views, as well as detailed information about each impression, including time, device, browser, and estimated location by city and country.
So, thank you for choosing ViralQR; we have an offer of nothing but the best in terms of QR code services to meet business diversity!
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
6. Human-in-the-Loop Throughout the Entire Life Cycle
of KG Construction, Growth, and Services
Data Labeling Development Deployment
Learner
raw data labeled data
1. Improve quality
2. Increase efficiency
3. Decrease skill requirements
7. Example 1: Scale Fact Collection
Missing / stale facts
Missing
Facts
Query
Synthesizer
QA System
candidate facts
Baseline
New
Facts
8. Example 1: Scale Fact Collection
Missing / stale facts
Missing
Facts
Query
Synthesizer
QA System
candidate facts
Baseline
New
Facts
Query-by-Committee
Missing
Facts
Query
Synthesizer
QA System
candidate facts
New
Facts
QA System
Q1
QA System
… …
… …
…
Qn
QbC
Selector
AnswerSet1
AnswerSetn
[EMNLP-DaSH’2022] Improving Human Annotation Effectiveness for Fact Collection by Identifying the Most Relevant Answers
Success Rate
fact collection
25%
9. Example 1: Scale Fact Collection
Missing / stale facts
Missing
Facts
Query
Synthesizer
QA System
candidate facts
Baseline
New
Facts
Open Domain Knowledge Extraction
[SIGMOD’23] Growing and Serving Large Open-domain Knowledge Graphs.
Throughput vs.
manual fact collection
>100x
Missing
Facts
Query
Synthesizer
Web Search
candidate facts w/
lower-confidence
New
Facts
Knowledge
Extractor
Fact
Corroboration
11. Example 2a: Crowd-in-the-Loop Curation
An hybrid approach
Corpus
raw data
Corpus
predicated
annotations
Annotation
Task
Corpus
curated
annotations
Task
Router
Difficult tasks are curated by experts
Easier tasks are curated by crowd
[EMNLP’17] CROWD-IN-THE-LOOP: A Hybrid Approach for Annotating Semantic Roles
13. Example 2b: Better Workflow Performs Ever Better
vs. SRL model
↑
Expert efforts
↓
10% F1
vs. SRL model
↑ 87.3%
Expert efforts
↓
Filter
unlikely options
Select
from likely options
Expert
resolve hard cases
[EMNLP’20 (Finding)] A Novel Workflow for Accurately and Efficiently
Crowdsourcing Predicate Senses and Argument Labels
14. Human-in-the-Loop Throughout the Entire Life Cycle
of KG construction, growth, and services
Data Labeling Development Deployment
Learner
Scale data labeling
raw data labeled data
IDE
Better IDE for model building
15. Different Tooling for Different Users
Full-fledged IDE
AI Engineers AI Engineers/Data Scientists
Visual IDE
[ACL’12] WizIE: A Best Practices Guided Development Environment for
Information Extraction
[CHI’13] I can do text analytics!: designing development tools for novice
developers
[VLDB’15] VINERy: A Visual IDE for Information Extraction
[KDD’19] Declarative Text Understanding with SystemT. (hands-on tutorial)
Entity Extraction in AIOps https://www.ibm.com/cloud/blog/entity-extraction-in-aiops
IBM InfoSphere BigInsights Text Analytics Eclipse Tooling
IBM Watson Knowledge Studio. Advanced Rule Editor http://ibm.biz/VineryIE
16. Human-in-the-Loop Throughout the Entire Life Cycle
of KG construction, growth, and services
Data Labeling Development Deployment
Learner
Scale data labeling
raw data labeled data
IDE
Better IDE for model building
Learner
Human-machine co-creation
17. Transparent Linguistic Models for Contract Understanding
Watson Discovery Content Intelligence
[NAACL’21] Development of an Enterprise-Grade Contract Understanding System, (Industry Track)
18. HEIDL: Human & Machine Co-Creation via Neural-Symbolic AI
[ACL’19] HEIDL: Learning Linguistic Expressions with Deep Learning and Human-in-the-Loop.
[EMNLP’20] Learning Explainable Linguistic Expressions with Neural Inductive Logic Programming for Sentence Classification
In use for major IBM customer engagements
Raises the abstraction level for domain experts to interact with
19. HEIDL: Human & Machine Co-Creation via Neural-Symbolic AI
[ACL’19] HEIDL: Learning Linguistic Expressions with Deep Learning and Human-in-the-Loop.
[EMNLP’20] Learning Explainable Linguistic Expressions with Neural Inductive Logic Programming for Sentence Classification
In use for major IBM customer engagements
Raises the abstraction level for domain experts to interact with
20. Human-in-the-Loop Throughout the Entire Life Cycle
of KG construction, growth, and services
Data Labeling Development Deployment
Learner
Scale data labeling
raw data labeled data
IDE
Better IDE for model building
Learner
Human-machine co-creation
Learner
Curb data hunger with interactive learning
21. Case 1: Example-Driven Extraction
Via pattern induction
[CHI’17] SEER: Auto-Generating Information Extraction Rules from User-Specified Examples
[SIGMOD’17] Synthesizing Extraction Rules from User Examples with SEER. SIGMOD’2017
[AAAI’22 (demo)] InteractEva: A Simulation-based Evaluation Framework for Interactive AI Systems
[AAAI’22] A Simulation-Based Evaluation Framework for Interactive AI Systems and Its Application.
IBM Watson Discovery (Beta in Plus since Oct. 2021) http://ibm.biz/SEER_IE, https://ibm.biz/WDSPressReleaseNov
22. Case 1: Example-Driven Extraction
Via pattern induction
[CHI’17] SEER: Auto-Generating Information Extraction Rules from User-Specified Examples
[SIGMOD’17] Synthesizing Extraction Rules from User Examples with SEER. SIGMOD’2017
[AAAI’22 (demo)] InteractEva: A Simulation-based Evaluation Framework for Interactive AI Systems
[AAAI’22] A Simulation-Based Evaluation Framework for Interactive AI Systems and Its Application.
IBM Watson Discovery (Beta in Plus since Oct. 2021) http://ibm.biz/SEER_IE, https://ibm.biz/WDSPressReleaseNov
23. Case 1: Example-Driven Extraction
Via pattern induction
[CHI’17] SEER: Auto-Generating Information Extraction Rules from User-Specified Examples
[SIGMOD’17] Synthesizing Extraction Rules from User Examples with SEER. SIGMOD’2017
[AAAI’22 (demo)] InteractEva: A Simulation-based Evaluation Framework for Interactive AI Systems
[AAAI’22] A Simulation-Based Evaluation Framework for Interactive AI Systems and Its Application.
IBM Watson Discovery (Beta in Plus since Oct. 2021) http://ibm.biz/SEER_IE, https://ibm.biz/WDSPressReleaseNov
24. Case 1: Example-Driven Extraction
Via pattern induction
[CHI’17] SEER: Auto-Generating Information Extraction Rules from User-Specified Examples
[SIGMOD’17] Synthesizing Extraction Rules from User Examples with SEER. SIGMOD’2017
[AAAI’22 (demo)] InteractEva: A Simulation-based Evaluation Framework for Interactive AI Systems
[AAAI’22] A Simulation-Based Evaluation Framework for Interactive AI Systems and Its Application.
IBM Watson Discovery (Beta in Plus since Oct. 2021) http://ibm.biz/SEER_IE, https://ibm.biz/WDSPressReleaseNov
25. Case 2: Entity Normalization & Variant Generation
Learning Structured Representations
Capture Entity Semantic Structure
[COLING’2018] Exploiting Structure in Representation of Named Entities using Active Learning.
[ICDE’2018] LUSTRE: An Interactive System for Entity Structured Representation and Variant
Generation.
Generated normalizers for Watson Discovery
[AAAI’2020] PARTNER: Human-in-the-Loop Entity Name Understanding with Deep
Learning.
[EMNLP’2020] Learning Structured Representations of Entity Names using Active
Learning and Weak Supervision.
“Bank of America N.A.” “Bank of America National Association”
Synthesizing Normalization and
Variant Generation Functions
26. Case 3: Deep Document Understanding
Document Ingestion
[WACV 2021] Global Table Extractor (GTE): A Framework for Joint Table Identification
and Cell Structure Recognition Using Visual Context.
[AAAI’21] KAAPA: Knowledge Aware Answers from PDF Analysis.
[ACL-CORD-19’21] CORD-19: The COVID-19 Open Research Dataset
Bringing IBM NLP capabilities to the CORD-19 Dataset. http://ibm.biz/CORD19-IBM
IBM Watson Discovery
JSON/HTML
Wide Variety in PDF Tables
Table with
graphic lines
Table with
visual clues only
Complex
table with
multi-row/column
headers
Table interleaved
with text and charts
27. Case 3: TableLab
TableLab: Easy Customization via Adaptive Deep Learning
[IUI’2021] TableLab: An Interactive Table Extraction System with Adaptive Deep Learning.
28. Case 3: Deep Document Understanding
TableLab: Easy Customization via Adaptive Deep Learning
[IUI’2021] TableLab: An Interactive Table Extraction System with Adaptive Deep Learning.
29. Case 3: Customization vis TableLab
Table Boundary Detection
Preliminary Results
Method CEDAR EDGAR Invoices Appraisals Health
Docs
GTE 0.94 0.84 0.47 0.85 0.93
GTE with
Retraining
0.96 0.91 0.92 0.96 0.98
Method CEDAR EDGAR Invoices Appraisals Health
Docs
GTE 0.88 0.62 0.42 0.71 0.55
GTE with
Retraining
0.90 0.82 0.68 0.90 0.77
Cell Adjacency Detection
Dataset
20 pages with tables per category: 10 for
retraining, 10 for testing
Evaluation Metric
F1 metric for Table Boundary and Cell Adjacency
as de
fi
ned in [1]
[1] Göbel et al. “A Methodology for Evaluating Algorithms for Table Understanding in PDF
Documents”. DocEng '12
30. Case 4: Label Sleuth
An open-source no-code system for text annotation and building text classifiers
[EMNLP’2022] Label Sleuth: From Unlabeled Text to a Classifier in a Few Hours
https://www.label-sleuth.org
1. From task definition to working
model in hours!
2. Extensible backend to integrate new
model architectures or active
learning techniques
31. Human-in-the-Loop Throughout the Entire Life Cycle
of KG construction, growth, and services
Data Labeling Development Deployment
Learner
Scale data labeling
raw data labeled data
IDE
Better IDE for model building
Learner
Human-machine co-creation
Learner
Curb data hunger with interactive learning
AutoML
Scale model building via AutoML
32. AutoAI for Text
AutoText
[AAAI’21] AutoText: An End-to-End AutoAI Framework for Text.
[NeurIPS 2022] AUTOMATA: Gradient Based Data Subset Selection for Compute-Efficient Hyper-parameter Tuning.
IBM Developer API https://developer.ibm.com/learningpaths/get-started-autoai-for-text-api
Example Use Case: Scale ML Product
Model for Text Classification
>30%
Reduction in combined
training and prediction
time
Auto weight
tuning & HPO
>10x
Speed-up in training at
comparable quality
Auto classifier
selection
33. Human-in-the-Loop Throughout the Entire Life Cycle
of KG construction, growth, and services
Data Labeling Development Deployment
Learner
Scale data labeling
raw data labeled data
IDE
Better IDE for model building
Learner
Human-machine co-creation
Learner
Curb data hunger with interactive learning
AutoML
Scale model building via AutoML
34. Human-in-the-Loop Throughout the Entire Life Cycle
of KG construction, growth, and services
Data Labeling Development Deployment
Learner
Scale data labeling
raw data labeled data
IDE
Better IDE for model building
Learner
Human-machine co-creation
Learner
Curb data hunger with interactive learning
AutoML
Scale model building via AutoML
Query
Log Tickets
User feedback influence the entire life cycle
35. Quality Evaluation
1. Measure what matters for end users
2. Identify the root cause of failures
3. Track improvements in individual
components as they evolve
The key requirements
- Who won the Paris
Paris, France Paris Masters
36. Overall Evaluation Framework
A Human-in-the-Loop Process
Annotation Quality Metrics
Dataset Collection
Tooling and annotation guidelines
for graders
Evaluation
Human in the loop to annotate/
grade queries
Logs
Synthetic Queries
Knowledge Graph Metrics
End to End Metrics
Query Understanding Metrics
37. Visual Tooling of Metrics and Loss Buckets
- Example Errors:
- Entity Prediction Error: “Who won Paris” (Paris Masters/Paris–Roubaix)
- Missing Fact: “When is the oscars in 2026”
- Fact is not present because date/location is not published yet)
- Unrecognized Entity in KG: ”Who is princess noor horse”
Facilitate Opportunity Analysis
39. ModelLens
Visual interactive tool for model improvement
[CSCW’19] ModelLens: An Interactive System to Support the Model Improvement Practices of Data Science Teams.
40. So how will EVERYTHING
change with LLMs?
Many exciting challenges and opportunities
41. Thanks!
IBM (including interns):
• Shivakumar Vaithyanathan
• Lucian Popa
• Ron Fagin
• Sriram Raghavan
• Rajasekar Krishnamurthy
• Fred Reiss
• Laura Chiticariu
• Benny Kimelfeld
• Mauricio Hernadez
• Eser Kandogan
• Huaiyu Zhu
• Kun Qian
• Dakuo Wang
• Maeda Hanafi
Many amazing collaborators and interns …
Apple (including interns):
• Ihab Ilyas
• Theodoros Rekatsinas
• Umar Farooq Minhas
• Ali Mousavi
• Jefferey Pound
• Anil Pacaci
• Shihabur R. Chowdhury
• Hongyu Ren
• Jason Mohoney
• Kun Qian
• Yiwen Sun
• Yisi Sang
• Saloni Potdar
• … …
Universities:
• Azza Abouzeid (NYU-Abu Dhabi)
• H. V. Jagadish (U. Of Michigan)
• Fei Xia (U. Of Washington)
• Kevin Chen-Chuan Chang (UIUC)
• ChengXiang Zhai (UIUC)
• Domenico Lembo(Sapienza
University of Rome)
• Dragomir R. Radev (Yale)
• Jonathan K. Kummerfeld (U. Of
Michigan)
• Walter S. Lasecki (U. Of Michigan)
• Toby Li (U. of Notre Dame)
• Rishabh Iyer (UT Dallas)
• Eduard C. Dragut (Temple Univ.)
• … ….
• Douglas Burdick’
• Alan Akbik
• Nancy Wang
• Prithiviraj Sen
• Marina Danilevsky
• Poornima Chozhiyath Raman
• Sudarshan Rangarajan
• Ramiya Venkatachalam
• Kiran Kate
• Eyal Shnarch
• Ishan Jindal
• Yiwei Yang
• Nikita Bhutani
• … ….