Abstract:
A growing number of resources are available for enriching documents with semantic annotations. While originally focused on a few standard classes of annotations, the ecosystem of annotators is now becoming increasingly diverse. Although annotators often have very different vocabularies, with both high-level and specialist
concepts, they also have many semantic interconnections. We will show that both the overlap and the diversity in annotator vocabularies motivate the need for semantic annotation integration: middleware that produces a unified annotation on top of diverse semantic annotators. On the one hand, the diversity of vocabulary allows applications
to benefit from the much richer vocabulary available in
an integrated vocabulary. On the other hand, we present evidence that the most widely-used annotators on the web suffer from serious accuracy deficiencies: the overlap in vocabularies from individual annotators allows an integrated annotator to boost accuracy by exploiting inter-annotator agreement and disagreement.
The integration of semantic annotations leads to new challenges, both compared to usual data integration scenarios and to standard aggregation of machine learning tools. We overview an approach to these challenges that performs ontology-aware aggregation. We
introduce an approach that requires no training data, making use of ideas from database repair. We experimentally compare this with a supervised approach, which adapts maximal entropy Markov models to the setting of ontology-based annotations. We further experimentally compare both these approaches with respect to ontology-unaware
supervised approaches, and to individual annotators.
This document summarizes ANNIS, a web-based search tool for multi-layer annotated corpora. ANNIS allows multiple independent annotations of the same data, supports diverse annotation types and relations between annotations. It provides search functionality for exact matches, regular expressions and complex relations. Query results can be visualized depending on annotation type, such as spans in a grid or trees in a tree view. Corpus data is imported into a relational database for performance and can be exported for data mining.
In this paper we present the SMalL Ontology for malicious software classification, SMalL Java Application for antivirus systems comparison and the SMalL knowledge based file format for malware related attacks. We believe that our ontology is able to aid the development of malware prevention software by offering a common knowledge base and a clear classification of the existing malicious software. The application is a prototype regarding how this ontology might be used in conjunction with known antivirus capabilities to offer a comprehensive comparison.
Swoogle: Showcasing the Significance of Semantic SearchIDES Editor
The World Wide Web hosts vast repositories of
information. The retrieval of required information from the
Internet is a great challenge since computer applications
understand only the structure and layout of web pages and
they do not have access to their intended meaning. Semantic
web is an effort to enhance the Internet, so that computers
can process the information presented on WWW, interpret
and communicate with it, to help humans find required
essential knowledge. Application of Ontology is the
predominant approach helping the evolution of the Semantic
web. The aim of our work is to illustrate how Swoogle, a
semantic search engine, helps make computer and WWW
interoperable and more intelligent. In this paper, we discuss
issues related to traditional and semantic web searching. We
outline how an understanding of the semantics of the search
terms can be used to provide better results. The experimental
results establish that semantic search provides more focused
results than the traditional search.
ISO 25964: Thesauri and Interoperability with Other VocabulariesMarcia Zeng
ISO 25964: Thesauri and Interoperability with Other Vocabularies + IFLA Guidelines for Multilingual Thesauri + A KOS Resource Application Profile. Presented at 2009 NKOS Workshop.
Modelling Knowledge Organization Systems and StructuresMarcia Zeng
In this paper FRSAD (as a conceptual model) is compared to SKOS and SKOS XL (as data models), with implementation examples. ISKO-UK 2011 Conference, London, July 2011.
Web classification of Digital Libraries using GATE Machine Learning sstose
This document provides an introduction to classifying digital library documents using machine learning with GATE. It discusses how text mining applies natural language processing to analyze textual content and extract knowledge. While humans can understand language intuitively, machines struggle with aspects like context and ambiguity. The document then outlines how natural language preprocessing is used for text classification, including tokenization, stopword removal, and representing documents as bags of words. The goal is to train machine learning algorithms on linguistic features to automatically categorize new documents.
Linking Folksonomies to Knowledge Organization SystemsJakob .
This document discusses linking folksonomies, which are user-generated tagging systems, to formal knowledge organization systems (KOS). It describes harvesting tags from Wikipedia and Stack Exchange to generate SKOS/RDF representations. These folksonomies are then enriched by linking tags to concepts in controlled vocabularies. The results include hierarchical links between tags, mappings to other KOS, and using these links to provide related documents and analyze document collections.
This document summarizes ANNIS, a web-based search tool for multi-layer annotated corpora. ANNIS allows multiple independent annotations of the same data, supports diverse annotation types and relations between annotations. It provides search functionality for exact matches, regular expressions and complex relations. Query results can be visualized depending on annotation type, such as spans in a grid or trees in a tree view. Corpus data is imported into a relational database for performance and can be exported for data mining.
In this paper we present the SMalL Ontology for malicious software classification, SMalL Java Application for antivirus systems comparison and the SMalL knowledge based file format for malware related attacks. We believe that our ontology is able to aid the development of malware prevention software by offering a common knowledge base and a clear classification of the existing malicious software. The application is a prototype regarding how this ontology might be used in conjunction with known antivirus capabilities to offer a comprehensive comparison.
Swoogle: Showcasing the Significance of Semantic SearchIDES Editor
The World Wide Web hosts vast repositories of
information. The retrieval of required information from the
Internet is a great challenge since computer applications
understand only the structure and layout of web pages and
they do not have access to their intended meaning. Semantic
web is an effort to enhance the Internet, so that computers
can process the information presented on WWW, interpret
and communicate with it, to help humans find required
essential knowledge. Application of Ontology is the
predominant approach helping the evolution of the Semantic
web. The aim of our work is to illustrate how Swoogle, a
semantic search engine, helps make computer and WWW
interoperable and more intelligent. In this paper, we discuss
issues related to traditional and semantic web searching. We
outline how an understanding of the semantics of the search
terms can be used to provide better results. The experimental
results establish that semantic search provides more focused
results than the traditional search.
ISO 25964: Thesauri and Interoperability with Other VocabulariesMarcia Zeng
ISO 25964: Thesauri and Interoperability with Other Vocabularies + IFLA Guidelines for Multilingual Thesauri + A KOS Resource Application Profile. Presented at 2009 NKOS Workshop.
Modelling Knowledge Organization Systems and StructuresMarcia Zeng
In this paper FRSAD (as a conceptual model) is compared to SKOS and SKOS XL (as data models), with implementation examples. ISKO-UK 2011 Conference, London, July 2011.
Web classification of Digital Libraries using GATE Machine Learning sstose
This document provides an introduction to classifying digital library documents using machine learning with GATE. It discusses how text mining applies natural language processing to analyze textual content and extract knowledge. While humans can understand language intuitively, machines struggle with aspects like context and ambiguity. The document then outlines how natural language preprocessing is used for text classification, including tokenization, stopword removal, and representing documents as bags of words. The goal is to train machine learning algorithms on linguistic features to automatically categorize new documents.
Linking Folksonomies to Knowledge Organization SystemsJakob .
This document discusses linking folksonomies, which are user-generated tagging systems, to formal knowledge organization systems (KOS). It describes harvesting tags from Wikipedia and Stack Exchange to generate SKOS/RDF representations. These folksonomies are then enriched by linking tags to concepts in controlled vocabularies. The results include hierarchical links between tags, mappings to other KOS, and using these links to provide related documents and analyze document collections.
The document discusses knowledge organization systems (KOS) and how the Simple Knowledge Organization System (SKOS) bridges KOS and the Semantic Web. It provides examples of KOS like taxonomies and thesauruses and explains how they are used differently than ontologies. SKOS is defined as an RDF vocabulary for representing KOS online in a machine-readable way and became a W3C standard in 2009.
Concept hierarchy is the backbone of ontology, and the concept hierarchy acquisition has been a hot topic in the field of ontology learning. this paper proposes a hyponymy extraction method of domain ontology concept based on cascaded conditional random field(CCRFs) and hierarchy clustering. It takes free text as extracting object, adopts CCRFs identifying the domain concepts. First the low layer of CCRFs is used to identify simple domain concept, then the results are sent to the high layer, in which the nesting concepts are recognized. Next we adopt hierarchy clustering to identify the hyponymy relation between domain ontology concepts. The experimental results demonstrate the proposed method is efficient.
The objective of this webinar is to provide a brief overview of the Knowledge Organization Systems (KOS) and the tools used for managing them. The presentation will focus on the management of the multilingual Organic.Edunet ontology as a case study. In this context it will present aspects such as the collaborative work, multilinguality needs and update of the concepts using an online KOS management tool (MoKi).
Dimensions of Media Object ComprehensibilityLawrie Hunter
This document discusses dimensions of comprehensibility in media objects. It begins by framing the topic as a pattern language approach for machine-mediated communication (MMC). It notes insights can be drawn from second language learning, where comprehension of partially acquired languages reveals aspects of text and media nature. The document then discusses various parameters that influence the difficulty of comprehending media objects, such as document purpose, content, target behaviors, and lexical items. It provides examples of how knowledge structure maps can link descriptive information about a text. The goal is to develop a pattern language to guide machines in human-like communication by understanding factors affecting media object comprehension.
Mining Opinion Features in Customer ReviewsIJCERT JOURNAL
Now days, E-commerce systems have become extremely important. Large numbers of customers are choosing online shopping because of its convenience, reliability, and cost. Client generated information and especially item reviews are significant sources of data for consumers to make informed buy choices and for makers to keep track of customer’s opinions. It is difficult for customers to make purchasing decisions based on only pictures and short product descriptions. On the other hand, mining product reviews has become a hot research topic and prior researches are mostly based on pre-specified product features to analyse the opinions. Natural Language Processing (NLP) techniques such as NLTK for Python can be applied to raw customer reviews and keywords can be extracted. This paper presents a survey on the techniques used for designing software to mine opinion features in reviews. Elven IEEE papers are selected and a comparison is made between them. These papers are representative of the significant improvements in opinion mining in the past decade.
Ontology languages are used in modelling the semantics of concepts within a particular domain and the relationships between those concepts. The Semantic Web standard provides a number of modelling languages that differ in their level of expressivity and are organized in a Semantic Web Stack in such a way that each language level builds on the expressivity of the other. There are several problems when one attempts to use independently developed ontologies. When existing ontologies are adapted for new purposes it requires that certain operations are performed on them. These operations are currently performed in a semi-automated manner. This paper seeks to model categorically the syntax and semantics of RDF ontology as a step towards the formalization of ontological operations using category theory.
This document discusses approaches to automatically structuring natural heritage literature using machine learning techniques. It compares supervised machine learning, which uses labeled examples to train models, and unsupervised machine learning, which derives structure without labeled examples. The document describes a prototype application that can convert free text to XML format in batch or online modes to support tasks like specimen identification.
KOSO Knowledge Organization Systems OntologyKatrin Weller
KOSO is a metadata ontology that aims to support knowledge exchange and reuse of existing knowledge organization systems (KOS) by providing descriptive metadata about different types of KOS, how they are classified and defined, and how they can interact. The ontology defines key concepts like KnowledgeOrganizationSystem and its subclasses like Ontology, Thesaurus, and Classification. It specifies properties to describe a KOS like its domain, language, and relations. It also models interactions between KOS through properties like has_version and is_interlinked_with. The goal is to enable discovery and understanding of existing KOS for improved reuse.
Multilingual Knowledge Organization Systems Management: Best PracticesMauro Dragoni
This presentation addresses the most well-known challenges in managing multilingual knowledge organization systems.
Such challenges are presented and it is discussed how they have been addressed with the implementation of a collaborative tool called MoKi.
Simple Knowledge Organization System (SKOS) in the Context of Semantic Web De...gardensofmeaning
The document discusses the development and use of SKOS (Simple Knowledge Organization System) for representing knowledge organization systems like thesauri and classification schemes as structured data on the semantic web. It describes how LCSH (Library of Congress Subject Headings) has been converted to RDF using SKOS and published as linked open data. It suggests further steps like linking LCSH to other metadata and developing RDF representations of additional bibliographic schemas.
The document discusses different types of information retrieval systems such as traditional query-based systems, text categorization systems, text routing systems, and text filtering systems. It also describes some common techniques used in information retrieval systems like inverted indexing, stopword removal, stemming, and vector space models. Finally, it discusses opportunities for integrating information retrieval techniques with natural language processing to develop more accurate and effective retrieval systems.
Abstract:
An increasing number of applications rely on RDF, OWL 2, and SPARQL for storing and querying data. SPARQL, however, is not targeted towards end-users, and suitable query interfaces are needed. Faceted search is a prominent approach for end-user data access, and several RDF-based faceted search systems have been developed. There is, however, a lack of rigorous theoretical underpinning for faceted search in the context of RDF and OWL 2. In this paper, we provide such solid foundations. We formalise faceted interfaces for this context, identify a fragment of first-order logic capturing the underlying queries, and study the complexity of answering such queries for RDF and OWL 2 profiles. We then study interface generation and update, and devise efficiently implementable algorithms. Finally, we have implemented and tested our faceted search algorithms for scalability, with encouraging results.
PDQ: Proof-driven Query Answering over Web=based Data
Abstract: The data needed to answer queries is often available through Web-based APIs. Indeed, for a given query there may be many Web-based sources which can be used to answer it, with the sources overlapping in their vocabularies, and differing in their access restrictions (required arguments) and cost.
We introduce PDQ (Proof-Driven Query Answering), a system for determining a query plan in the presence of web-based sources. It is: (i) constraint-aware -- exploiting relationships between sources to rewrite an expensive query into a cheaper one, (ii) access-aware -- abiding by any access restrictions known in the sources, and (iii) cost-aware -- making use of any cost information that is available about services.
PDQ takes the novel approach of generating query plans from proofs that a query is answerable. We demonstrate the use of PDQ and its effectiveness in generating low-cost plans.
RDFox is an optimized RDF triple store and parallel Datalog reasoner that supports OWL 2 RL. It features low memory consumption per triple, high scalability with CPU threads, and integration as a Java or C++ library. Evaluation shows RDFox achieves near-linear speedup on multi-core systems and outperforms competitors for query answering large datasets.
Abstract: An enhanced hybrid approach to OWL query answering that combines an RDF triple-store with an OWL reasoner in order to provide scaleable pay-as-you-go performance. The enhancements presented here include an extension to deal with arbitary OWL ontologies and optimisations that significantly improve scalability. We have implemented these techniques in a prototype system, a preliminary evaluation of which has produced very encouraging results.
Abstract: We present an enhanced hybrid approach to OWL query answering that combines an RDF triple-store with an OWL reasoner in order to provide scalable pay-as-you-go performance. The enhancements presented here include an extension to deal with arbitrary OWL ontologies, and optimisations that significantly improve scalability. We have implemented these techniques in a prototype system, a preliminary evaluation of which has produced very encouraging results.
Parallel Materialisation of Datalog Programs in Centralised, Main-Memory RDF ...DBOnto
The document describes a novel approach to parallelizing materialization (fixpoint computation) of Datalog programs in centralized, main-memory RDF systems. The approach distributes the workload evenly across cores and uses an RDF indexing data structure that supports efficient, mostly lock-free parallel updates. Evaluation showed the approach parallelizes computation very well, with up to 13.9x speedup on 16 cores and 19.3x speedup using 32 virtual cores via hyperthreading.
1. The document discusses Optique 1.0, an ontology-based data access system that allows non-expert users to access data from multiple databases through a visualized query interface.
2. Optique 1.0 uses ontologies and declarative mappings to connect data sources, allowing queries to be formulated over ontologies and then pushed to and evaluated over the databases.
3. The system was tested on a real-world dataset from the oil and gas industry containing over 70 tables and 2.3 million triples extracted from databases into an ontology and triple store.
The document discusses faceted search over ontology-enhanced RDF data. It formalizes faceted interfaces for querying RDF graphs that capture ontological information. It studies the expressivity and complexity of queries represented by faceted interfaces, and algorithms for generating and updating interfaces based on the underlying RDF and ontology information. The goal is to provide rigorous theoretical foundations for faceted search in the context of RDF and OWL 2 ontologies.
ArtForm - Dynamic analysis of JavaScript validation in web forms - PosterDBOnto
ArtForm is a tool that uses concolic analysis and symbolic execution to dynamically analyze JavaScript validation in web forms. It infers integrity constraints, models hidden data, and improves search and data extraction. It works by controlling a WebKit browser to execute code symbolically while tracking concrete and symbolic values. Challenges include event handler dependencies, implied constraints, and JavaScript semantics that are difficult to model. Future work focuses on handling more patterns and constraints and targeting exploration of interesting parts of form trees.
The document discusses knowledge organization systems (KOS) and how the Simple Knowledge Organization System (SKOS) bridges KOS and the Semantic Web. It provides examples of KOS like taxonomies and thesauruses and explains how they are used differently than ontologies. SKOS is defined as an RDF vocabulary for representing KOS online in a machine-readable way and became a W3C standard in 2009.
Concept hierarchy is the backbone of ontology, and the concept hierarchy acquisition has been a hot topic in the field of ontology learning. this paper proposes a hyponymy extraction method of domain ontology concept based on cascaded conditional random field(CCRFs) and hierarchy clustering. It takes free text as extracting object, adopts CCRFs identifying the domain concepts. First the low layer of CCRFs is used to identify simple domain concept, then the results are sent to the high layer, in which the nesting concepts are recognized. Next we adopt hierarchy clustering to identify the hyponymy relation between domain ontology concepts. The experimental results demonstrate the proposed method is efficient.
The objective of this webinar is to provide a brief overview of the Knowledge Organization Systems (KOS) and the tools used for managing them. The presentation will focus on the management of the multilingual Organic.Edunet ontology as a case study. In this context it will present aspects such as the collaborative work, multilinguality needs and update of the concepts using an online KOS management tool (MoKi).
Dimensions of Media Object ComprehensibilityLawrie Hunter
This document discusses dimensions of comprehensibility in media objects. It begins by framing the topic as a pattern language approach for machine-mediated communication (MMC). It notes insights can be drawn from second language learning, where comprehension of partially acquired languages reveals aspects of text and media nature. The document then discusses various parameters that influence the difficulty of comprehending media objects, such as document purpose, content, target behaviors, and lexical items. It provides examples of how knowledge structure maps can link descriptive information about a text. The goal is to develop a pattern language to guide machines in human-like communication by understanding factors affecting media object comprehension.
Mining Opinion Features in Customer ReviewsIJCERT JOURNAL
Now days, E-commerce systems have become extremely important. Large numbers of customers are choosing online shopping because of its convenience, reliability, and cost. Client generated information and especially item reviews are significant sources of data for consumers to make informed buy choices and for makers to keep track of customer’s opinions. It is difficult for customers to make purchasing decisions based on only pictures and short product descriptions. On the other hand, mining product reviews has become a hot research topic and prior researches are mostly based on pre-specified product features to analyse the opinions. Natural Language Processing (NLP) techniques such as NLTK for Python can be applied to raw customer reviews and keywords can be extracted. This paper presents a survey on the techniques used for designing software to mine opinion features in reviews. Elven IEEE papers are selected and a comparison is made between them. These papers are representative of the significant improvements in opinion mining in the past decade.
Ontology languages are used in modelling the semantics of concepts within a particular domain and the relationships between those concepts. The Semantic Web standard provides a number of modelling languages that differ in their level of expressivity and are organized in a Semantic Web Stack in such a way that each language level builds on the expressivity of the other. There are several problems when one attempts to use independently developed ontologies. When existing ontologies are adapted for new purposes it requires that certain operations are performed on them. These operations are currently performed in a semi-automated manner. This paper seeks to model categorically the syntax and semantics of RDF ontology as a step towards the formalization of ontological operations using category theory.
This document discusses approaches to automatically structuring natural heritage literature using machine learning techniques. It compares supervised machine learning, which uses labeled examples to train models, and unsupervised machine learning, which derives structure without labeled examples. The document describes a prototype application that can convert free text to XML format in batch or online modes to support tasks like specimen identification.
KOSO Knowledge Organization Systems OntologyKatrin Weller
KOSO is a metadata ontology that aims to support knowledge exchange and reuse of existing knowledge organization systems (KOS) by providing descriptive metadata about different types of KOS, how they are classified and defined, and how they can interact. The ontology defines key concepts like KnowledgeOrganizationSystem and its subclasses like Ontology, Thesaurus, and Classification. It specifies properties to describe a KOS like its domain, language, and relations. It also models interactions between KOS through properties like has_version and is_interlinked_with. The goal is to enable discovery and understanding of existing KOS for improved reuse.
Multilingual Knowledge Organization Systems Management: Best PracticesMauro Dragoni
This presentation addresses the most well-known challenges in managing multilingual knowledge organization systems.
Such challenges are presented and it is discussed how they have been addressed with the implementation of a collaborative tool called MoKi.
Simple Knowledge Organization System (SKOS) in the Context of Semantic Web De...gardensofmeaning
The document discusses the development and use of SKOS (Simple Knowledge Organization System) for representing knowledge organization systems like thesauri and classification schemes as structured data on the semantic web. It describes how LCSH (Library of Congress Subject Headings) has been converted to RDF using SKOS and published as linked open data. It suggests further steps like linking LCSH to other metadata and developing RDF representations of additional bibliographic schemas.
The document discusses different types of information retrieval systems such as traditional query-based systems, text categorization systems, text routing systems, and text filtering systems. It also describes some common techniques used in information retrieval systems like inverted indexing, stopword removal, stemming, and vector space models. Finally, it discusses opportunities for integrating information retrieval techniques with natural language processing to develop more accurate and effective retrieval systems.
Abstract:
An increasing number of applications rely on RDF, OWL 2, and SPARQL for storing and querying data. SPARQL, however, is not targeted towards end-users, and suitable query interfaces are needed. Faceted search is a prominent approach for end-user data access, and several RDF-based faceted search systems have been developed. There is, however, a lack of rigorous theoretical underpinning for faceted search in the context of RDF and OWL 2. In this paper, we provide such solid foundations. We formalise faceted interfaces for this context, identify a fragment of first-order logic capturing the underlying queries, and study the complexity of answering such queries for RDF and OWL 2 profiles. We then study interface generation and update, and devise efficiently implementable algorithms. Finally, we have implemented and tested our faceted search algorithms for scalability, with encouraging results.
PDQ: Proof-driven Query Answering over Web=based Data
Abstract: The data needed to answer queries is often available through Web-based APIs. Indeed, for a given query there may be many Web-based sources which can be used to answer it, with the sources overlapping in their vocabularies, and differing in their access restrictions (required arguments) and cost.
We introduce PDQ (Proof-Driven Query Answering), a system for determining a query plan in the presence of web-based sources. It is: (i) constraint-aware -- exploiting relationships between sources to rewrite an expensive query into a cheaper one, (ii) access-aware -- abiding by any access restrictions known in the sources, and (iii) cost-aware -- making use of any cost information that is available about services.
PDQ takes the novel approach of generating query plans from proofs that a query is answerable. We demonstrate the use of PDQ and its effectiveness in generating low-cost plans.
RDFox is an optimized RDF triple store and parallel Datalog reasoner that supports OWL 2 RL. It features low memory consumption per triple, high scalability with CPU threads, and integration as a Java or C++ library. Evaluation shows RDFox achieves near-linear speedup on multi-core systems and outperforms competitors for query answering large datasets.
Abstract: An enhanced hybrid approach to OWL query answering that combines an RDF triple-store with an OWL reasoner in order to provide scaleable pay-as-you-go performance. The enhancements presented here include an extension to deal with arbitary OWL ontologies and optimisations that significantly improve scalability. We have implemented these techniques in a prototype system, a preliminary evaluation of which has produced very encouraging results.
Abstract: We present an enhanced hybrid approach to OWL query answering that combines an RDF triple-store with an OWL reasoner in order to provide scalable pay-as-you-go performance. The enhancements presented here include an extension to deal with arbitrary OWL ontologies, and optimisations that significantly improve scalability. We have implemented these techniques in a prototype system, a preliminary evaluation of which has produced very encouraging results.
Parallel Materialisation of Datalog Programs in Centralised, Main-Memory RDF ...DBOnto
The document describes a novel approach to parallelizing materialization (fixpoint computation) of Datalog programs in centralized, main-memory RDF systems. The approach distributes the workload evenly across cores and uses an RDF indexing data structure that supports efficient, mostly lock-free parallel updates. Evaluation showed the approach parallelizes computation very well, with up to 13.9x speedup on 16 cores and 19.3x speedup using 32 virtual cores via hyperthreading.
1. The document discusses Optique 1.0, an ontology-based data access system that allows non-expert users to access data from multiple databases through a visualized query interface.
2. Optique 1.0 uses ontologies and declarative mappings to connect data sources, allowing queries to be formulated over ontologies and then pushed to and evaluated over the databases.
3. The system was tested on a real-world dataset from the oil and gas industry containing over 70 tables and 2.3 million triples extracted from databases into an ontology and triple store.
The document discusses faceted search over ontology-enhanced RDF data. It formalizes faceted interfaces for querying RDF graphs that capture ontological information. It studies the expressivity and complexity of queries represented by faceted interfaces, and algorithms for generating and updating interfaces based on the underlying RDF and ontology information. The goal is to provide rigorous theoretical foundations for faceted search in the context of RDF and OWL 2 ontologies.
ArtForm - Dynamic analysis of JavaScript validation in web forms - PosterDBOnto
ArtForm is a tool that uses concolic analysis and symbolic execution to dynamically analyze JavaScript validation in web forms. It infers integrity constraints, models hidden data, and improves search and data extraction. It works by controlling a WebKit browser to execute code symbolically while tracking concrete and symbolic values. Challenges include event handler dependencies, implied constraints, and JavaScript semantics that are difficult to model. Future work focuses on handling more patterns and constraints and targeting exploration of interesting parts of form trees.
Query Distributed RDF Graphs: The Effects of Partitioning PaperDBOnto
Abstract: Web-scale RDF datasets are increasingly processed using distributed RDF data stores built on top of a cluster of shared-nothing servers. Such systems critically rely on their data partitioning scheme and query answering scheme, the goal of which is to facilitate correct and ecient query processing. Existing data partitioning schemes are
commonly based on hashing or graph partitioning techniques. The latter techniques split a dataset in a way that minimises the number of connections between the resulting subsets, thus reducing the need for communication between servers; however, to facilitate ecient query answering,
considerable duplication of data at the intersection between subsets is often needed. Building upon the known graph partitioning approaches, in this paper we present a novel data partitioning scheme that employs minimal duplication and keeps track of the connections between partition elements; moreover, we propose a query answering scheme that
uses this additional information to correctly answer all queries. We show experimentally that, on certain well-known RDF benchmarks, our data partitioning scheme often allows more answers to be retrieved without distributed computation than the known schemes, and we show that our query answering scheme can eciently answer many queries.
Abstract:
The data needed to answer queries is often available through Web-based APIs. Indeed, for a given query there may be many Web-based sources which can be used to answer it, with the sources overlapping in their vocabularies, and differing in their access restrictions (required arguments) and cost.
We introduce PDQ (Proof-Driven Query Answering), a system for determining a query plan in the presence of web-based sources. It is: (i) constraint-aware -- exploiting relationships between sources to rewrite an expensive query into a cheaper one, (ii) access-aware -- abiding by any access restrictions known in the sources, and (iii) cost-aware -- making use of any cost information that is available about services.
PDQ takes the novel approach of generating query plans from proofs that a query is answerable. We demonstrate the use of PDQ and its effectiveness in generating low-cost plans.
This document describes semantic faceted search, which allows users to navigate between interconnected sets of documents through faceted filters over RDF data and OWL ontologies. It provides an example of filtering documents to find US presidents whose children graduated from Harvard or Stanford. Semantic faceted search generates faceted search interfaces automatically over structured data through translation to SPARQL queries.
Optique - to provide semantic end-to-end connection between users and data sources; enable users to rapidly formulate intuitive queries using familiar vocabularies and conceptualisations and return timely answers from large scale and heterogeneous data sources.
This document provides information on the DIADEM data extraction methodology project. It introduces the team members and lists their roles. The technology uses a domain-centric, intelligent, automated approach to extract web data at large scale from thousands of sources. The technology includes components like ROSeAnn for entity extraction, OPAL for form understanding, AMBER for record identification, and OXPath for an efficient extraction language. DIADEM aims to provide comprehensive data to help clients with tasks like price intelligence, product data, and recommendations.
PAGOdA (Pay-as-you-go OWL Query Answering Using a Triple Store) presentation by Bernardo Cuenca Grau
Abstract: We present an enhanced hybrid approach to OWL query answering that combines an RDF triple-store with an OWL reasoner in order to provide scalable pay-as-you-go performance. The enhancements presented here include an extension to deal with arbitrary OWL ontologies, and optimisations that significantly improve scalability. We have implemented these techniques in a prototype system, a preliminary evaluation of which has produced very encouraging results.
Abstract:
A growing number of resources are available for enriching documents with semantic annotations. While originally focused on a few standard classes of annotations, the ecosystem of annotators is now becoming increasingly diverse. Although annotators often have very different vocabularies, with both high-level and specialist
concepts, they also have many semantic interconnections. We will show that both the overlap and the diversity in annotator vocabularies motivate the need for semantic annotation integration: middleware that produces a unified annotation on top of diverse semantic annotators. On the one hand, the diversity of vocabulary allows applications
to benefit from the much richer vocabulary available in
an integrated vocabulary. On the other hand, we present evidence that the most widely-used annotators on the web suffer from serious accuracy deficiencies: the overlap in vocabularies from individual annotators allows an integrated annotator to boost accuracy by exploiting inter-annotator agreement and disagreement.
The integration of semantic annotations leads to new challenges, both compared to usual data integration scenarios and to standard aggregation of machine learning tools. We overview an approach to these challenges that performs ontology-aware aggregation. We
introduce an approach that requires no training data, making use of ideas from database repair. We experimentally compare this with a supervised approach, which adapts maximal entropy Markov models to the setting of ontology-based annotations. We further experimentally compare both these approaches with respect to ontology-unaware
supervised approaches, and to individual annotators.
Semantic Faceted Search with SemFacet presentationDBOnto
Abstract
An increasing number of applications rely on RDF, OWL 2, and SPARQL for storing and querying data. SPARQL, however, is not targeted towards end-users, and suitable query interfaces are needed. Faceted search is a prominent approach for end-user data access, and several RDF-based faceted search systems have been developed. There is, however, a lack of rigorous theoretical underpinning for faceted search in the context of RDF and OWL 2. In this paper, we provide such solid foundations. We formalise faceted interfaces for this context, identify a fragment of first-order logic capturing the underlying queries, and study the complexity of answering such queries for RDF and OWL 2 profiles. We then study interface generation and update, and devise efficiently implementable algorithms. Finally, we have implemented and tested our faceted search algorithms for scalability, with encouraging results.
This document describes the Oxford Database Group, including its faculty, postdocs, students, research areas, funding sources, awards, publications, invited talks, technology transfers, and the goals of the DBOnto flexible funding mechanism. The group includes over 30 faculty and students researching applications, infrastructure, and foundations of databases and ontology. Key areas of impact are numerous high-quality publications, standardization contributions, invited talks, fellowships and awards for faculty, and technology transfers.
DIADEM: domain-centric intelligent automated data extraction methodology Pres...DBOnto
This document summarizes the DIADEM data extraction methodology project. It introduces the team members and provides an overview of the technical approach and components. The technology is able to extract data from thousands of websites at large scale with high accuracy. It uses various techniques including ROSeAnn for entity extraction, OPAL for form understanding, AMBER for record identification, and OXPath for the extraction language. The system is able to adapt to new domains and outperforms other data extraction systems in precision, recall, and attribute labeling.
SWSN UNIT-3.pptx we can information about swsn professionalgowthamnaidu0986
Ontology engineering involves constructing ontologies through various methods. It begins with defining the scope and evaluating existing ontologies for reuse. Terms are enumerated and organized in a taxonomy with defined properties, facets, and instances. The ontology is checked for anomalies and refined iteratively. Popular tools for ontology development include Protege and WebOnto. Methods like Meth ontology and On-To-Knowledge methodology provide processes for building ontologies from scratch or reusing existing ones. Ontology sharing requires mapping between ontologies to allow interoperability, and libraries exist for storing and accessing ontologies.
DBpedia Spotlight is a system that automatically annotates text documents with DBpedia URIs. It identifies mentions of entities in text and links them to the appropriate DBpedia resources, addressing the challenge of ambiguity. The system is highly configurable, allowing users to specify which types of entities to annotate and the desired balance of coverage and accuracy. An evaluation found DBpedia Spotlight performed competitively compared to other annotation systems.
An Incremental Method For Meaning Elicitation Of A Domain OntologyAudrey Britton
This document describes MELIS (Meaning Elicitation and Lexical Integration System), a tool developed to support the annotation of data sources with conceptual and lexical information during the ontology generation process in MOMIS (Mediator envirOnment for Multiple Information Systems). MELIS improves upon MOMIS by enabling more automated annotation of data sources and by extracting relationships between lexical elements using lexical and domain knowledge to provide a richer semantics. The document outlines the MOMIS architecture integrated with MELIS and explains how MELIS supports key steps in the ontology creation process, including source annotation, common thesaurus generation, and relationship extraction.
An Annotation Framework For The Semantic WebAndrea Porter
This paper presents an ontology-based annotation framework called OntoAnnotate that supports the creation of semantically interlinked metadata. The framework addresses issues like avoiding duplicate annotations, managing changes to ontologies over time, and recognizing duplicate documents. OntoAnnotate allows annotators to create objects from text passages in documents and semantically relate them using properties from domain ontologies. This framework aims to enable the generation of semantically useful metadata for intelligent agents on the Semantic Web.
Cooperating Techniques for Extracting Conceptual Taxonomies from TextFulvio Rotella
The document proposes a mixed approach using existing natural language processing techniques and novel techniques to automatically construct conceptual taxonomies from text. It identifies relevant concepts from text using keyword extraction, clustering, and computing relevance weights. It then generalizes similar concepts using WordNet to group concepts and disambiguate word senses. Preliminary evaluations show promising initial results.
The document proposes a mixed approach using existing natural language processing techniques and novel techniques to automatically construct conceptual taxonomies from text. Key steps include identifying relevant concepts and attributes from text, clustering similar concepts, computing relevance weights for concepts, and generalizing concepts using WordNet. Preliminary results suggest the approach shows promise for extending and improving automatic taxonomy construction.
Semantic annotation, which is considered one of the semantic web applicative aspects, has been adopted by researchers from different communities as a paramount solution that improves searching and retrieval of information by promoting the richness of the content. However, researchers are facing challenges concerning both the quality and the relevance of the semantic annotations attached to the annotated document against its content as well as its semantics, without ignoring those regarding automation process which is supposed to ensure an optimal system for information indexing and retrieval. In this article, we will introduce the semantic annotation concept by presenting a state of the art including definitions, features and a classification of annotation systems. Systems and proposed approaches in the field will be cited, as well as a study of some existing annotation tools. This study will also pinpoint various problems and limitations related to the annotation in order to offer solutions for our future work.
SEMANTIC INTEGRATION FOR AUTOMATIC ONTOLOGY MAPPING cscpconf
In the last decade, ontologies have played a key technology role for information sharing and agents interoperability in different application domains. In semantic web domain, ontologies are efficiently used toface the great challenge of representing the semantics of data, in order to bring the actual web to its full
power and hence, achieve its objective. However, using ontologies as common and shared vocabularies requires a certain degree of interoperability between them. To confront this requirement, mapping ontologies is a solution that is not to be avoided. In deed, ontology mapping build a meta layer that allows different applications and information systems to access and share their informations, of course, after resolving the different forms of syntactic, semantic and lexical mismatches. In the contribution presented in this paper, we have integrated the semantic aspect based on an external lexical resource, wordNet, to design a new algorithm for fully automatic ontology mapping. This fully automatic character features the
main difference of our contribution with regards to the most of the existing semi-automatic algorithms of ontology mapping, such as Chimaera, Prompt, Onion, Glue, etc. To better enhance the performances of our algorithm, the mapping discovery stage is based on the combination of two sub-modules. The former
analysis the concept’s names and the later analysis their properties. Each one of these two sub-modules is
it self based on the combination of lexical and semantic similarity measures.
Knowledge Organisation Systems in Digital Libraries: A Comparative StudyBhojaraju Gunjal
The document presents a study that compares the different Knowledge Organization Systems (KOS) used in major digital libraries. It finds that while traditional libraries used standardized systems like classification schemes, digital libraries employ various KOS tools including thesauri, ontologies, and subject headings. The study analyzes the specific KOS used in different digital libraries and summarizes the current state of KOS in these libraries.
The document discusses Lin Ma's PhD research on analyzing presuppositions in natural language requirements. Presuppositions are implicit commitments in language that simplify communication but can cause misunderstanding if not made explicit. The research aims to automatically detect presuppositions triggered by definite descriptions in requirements and identify which are not explicitly stated. It will use natural language processing techniques and knowledge sources to classify definite descriptions and analyze how presuppositions project in requirements texts.
The document discusses the basics of ontologies, including their origin in philosophy, definitions, types, benefits and application areas. Some key points are:
- An ontology is a formal specification of a conceptualization used to help humans and programs share knowledge. It establishes a shared vocabulary for exchanging information.
- Ontologies describe domain knowledge and provide an agreed-upon understanding of a domain through concepts and relations. They help solve problems of ambiguity and enable knowledge sharing.
- Ontologies benefit applications like information retrieval, digital libraries, knowledge engineering and natural language processing by facilitating semantic search and integration of data.
This document summarizes ANNIS, a web-based search tool for multi-layer annotated corpora. ANNIS allows multiple independent annotations of the same data in different types and schemes. It was developed for projects analyzing information structure across languages. ANNIS supports diverse search functions, complex annotation relations, and visualization of query matches depending on annotation type. Corpus data is imported into a relational database for performance and can also be exported for other analysis tools.
14. Michael Oakes (UoW) Natural Language Processing for TranslationRIILP
This document discusses information retrieval and describes its three main phases: 1) asking a question to define an information need, 2) constructing an answer by matching queries to documents, and 3) assessing the relevance of the retrieved answers. It also covers several important information retrieval concepts like keywords, indexing documents, stemming words, calculating TF-IDF weights, and evaluating system performance using recall and precision.
The document proposes a knowledge-based approach to semi-automatic annotation of multimedia documents that takes into account the needs of producers, annotators, and consumers. It argues that current annotation approaches focus too much on the annotator's perspective and neglect the roles of other users. The proposed approach uses user modeling and intelligent interfaces to manage annotations from different types and competence levels of users, as well as relationships between multimedia annotations.
Use of ontologies in natural language processingATHMAN HAJ-HAMOU
The document discusses the limitations of classical natural language processing approaches that rely solely on text-based search and statistics. It describes how ontologies can help address these limitations by representing knowledge at the semantic level rather than just words. The document outlines how ontologies can solve problems related to vagueness, high-level concepts, semantic relations, and time dimensions. It then summarizes two research papers that explore using ontologies to index and categorize documents as well as provide multiple views and dimensions to explore a document collection.
Here are the key points about using content-based filtering techniques:
- Content-based filtering relies on analyzing the content or description of items to recommend items similar to what the user has liked in the past. It looks for patterns and regularities in item attributes/descriptions to distinguish highly rated items.
- The item content/descriptions are analyzed automatically by extracting information from sources like web pages, or entered manually from product databases.
- It focuses on objective attributes about items that can be extracted algorithmically, like text analysis of documents.
- However, personal preferences and what makes an item appealing are often subjective qualities not easily extracted algorithmically, like writing style or taste.
- So while content-based filtering can
SEMANTIC NETWORK BASED MECHANISMS FOR KNOWLEDGE ACQUISITIONcscpconf
The document describes a semantic network-based algorithm for knowledge acquisition from text. The algorithm uses the WiSENet semantic network to generate rules representing lexical relationships between concepts. It then applies these rules to text data as a finite state automaton to identify matches and acquire new concepts and relationships for expanding the semantic network. The algorithm tolerates variations in word order through its use of a "bag of concepts" approach during rule matching. Experiments showed the algorithm was effective at knowledge acquisition from text in a flexible manner.
Knowledge organization (KO) refers to activities like document description, indexing and classification performed in libraries and databases to organize documents and concepts. Knowledge organization systems (KOS) include classification schemes, subject headings, thesauri and other systems used to organize information. KOS impose structures on collections and can be used in digital libraries to provide overviews and support retrieval, though different KOS may characterize entities differently. Common types of KOS include term lists, classifications and categories, and relationship lists.
Semantic Annotation: The Mainstay of Semantic WebEditor IJCATR
Given that semantic Web realization is based on the critical mass of metadata accessibility and the representation of data with formal
knowledge, it needs to generate metadata that is specific, easy to understand and well-defined. However, semantic annotation of the
web documents is the successful way to make the Semantic Web vision a reality. This paper introduces the Semantic Web and its
vision (stack layers) with regard to some concept definitions that helps the understanding of semantic annotation. Additionally, this
paper introduces the semantic annotation categories, tools, domains and models
The document discusses the key aspects of thesauri including their purpose, structure, types of relationships displayed, and evaluation criteria. Specifically, it notes that a thesaurus provides a standardized vocabulary for information retrieval by displaying hierarchical (e.g. broader and narrower terms) and equivalence (e.g. synonyms) relationships between terms. It also discusses how terms are organized in a thesaurus and criteria for evaluating the effectiveness of a thesaurus.
Similar to Aggregating Semantic Annotators Paper (20)
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceIndexBug
Imagine a world where machines not only perform tasks but also learn, adapt, and make decisions. This is the promise of Artificial Intelligence (AI), a technology that's not just enhancing our lives but revolutionizing entire industries.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Driving Business Innovation: Latest Generative AI Advancements & Success Story
Aggregating Semantic Annotators Paper
1. Aggregating Semantic Annotators
Luying Chen, Stefano Ortona, Giorgio Orsi, and Michael Benedikt
Oxford University, UK
name.surname@cs.ox.ac.uk
ABSTRACT
A growing number of resources are available for enriching docu-ments
with semantic annotations. While originally focused on a
few standard classes of annotations, the ecosystem of annotators
is now becoming increasingly diverse. Although annotators often
have very different vocabularies, with both high-level and specialist
concepts, they also have many semantic interconnections. We will
show that both the overlap and the diversity in annotator vocabular-ies
motivate the need for semantic annotation integration: middle-ware
that produces a unified annotation on top of diverse semantic
annotators. On the one hand, the diversity of vocabulary allows ap-plications
to benefit from the much richer vocabulary available in
an integrated vocabulary. On the other hand, we present evidence
that the most widely-used annotators on the web suffer from serious
accuracy deficiencies: the overlap in vocabularies from individual
annotators allows an integrated annotator to boost accuracy by ex-ploiting
inter-annotator agreement and disagreement.
The integration of semantic annotations leads to new challenges,
both compared to usual data integration scenarios and to standard
aggregation of machine learning tools. We overview an approach
to these challenges that performs ontology-aware aggregation. We
introduce an approach that requires no training data, making use of
ideas from database repair. We experimentally compare this with a
supervised approach, which adapts maximal entropy Markov mod-els
to the setting of ontology-based annotations. We further experi-mentally
compare both these approaches with respect to ontology-unaware
supervised approaches, and to individual annotators.
1. INTRODUCTION
A growing number of resources are available for enriching doc-uments
with semantic annotations – annotations that label docu-ment
snippets as referring to either certain entities (e.g. Barack
Obama, the King of Spain) or to elements of particular semantic
categories (e.g. actors, governmental organizations). While orig-inally
focused on a few high-level standard classes of annotations
(e.g., people, places, organisations), the annotator ecosystem is be-coming
increasingly diverse. There are high-level and specialist
concepts that are supported by multiple annotators and, although
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that copies
bear this notice and the full citation on the first page. To copy otherwise, to
republish, to post on servers or to redistribute to lists, requires prior specific
permission and/or a fee. Articles from this volume were invited to present
their results at The 39th International Conference on Very Large Data Bases,
August 26th 30th
2013, Riva del Garda, Trento, Italy.
Proceedings of the VLDB Endowment, Vol. 6, No. 13
Copyright 2013 VLDB Endowment 21508097/
13/13... $ 10.00.
Location
The closure of the Subic Naval Base and Clark , a nearby
U.S. air base, put nearly 50,000 people out of jobs and starved
hundreds of bars and night spots of dollar-rich customers.
Person
An1: AlchemyAPI
An2: Extractiv
City
Facility
Facility
An3: Zemanta
An4: OpenCalais
Figure 1: Entity Annotations
their vocabularies are often quite distinct, it is possible to recognise
many semantic interconnections among them. There are annotators
that deal with very common entity classes (e.g., Person), and those
that focus on very specialized entities (e.g., Protein) to support ap-plications
domains such as biomedicine [10]. Furthermore, instead
of being custom built “white-boxes” that can be directly tuned by
developers, annotation is often available as a “black-box” service
(e.g. OpenCalais, Zemanta), where the user does not have to imple-ment
his own annotator but, on the other hand, concept classes can
be added or refined by the service host outside the control of clients.
Consider the example displayed in Figure 1, which shows a frag-ment
of text from the Reuters corpus, mentioning a US naval base
in the Philippines and classified by several Web-based annotators.
The results include some high-level concepts (i.e., Location, Person),
as well as some more specific ones (i.e., Facility). Notice that the an-notation
of type City is only returned by Extractiv: indeed, several of
the annotators simply do not have this concept in their vocabulary.
Also note that annotators give differing classifications of the same
snippet – one annotator determining that “Subic Naval Base” is a
Location and others determining that it is a Facility. A second exam-ple
is shown in Figure 2. Here we see several flavors of differences
in annotator opinions. “Nottingham Forest” is labeled as a Natu-ralFeature
by OpenCalais, and as a GeographicFeature by AlchemyAPI
– clearly these outputs are compatible, since these two concepts
represent the same meaning. On the other hand, the annotations
Facility and Sports team are clearly both incompatible with the prior
classifications, as well as with each other.
We will show that both the overlap and the diversity in anno-tator
vocabularies motivate the need for semantic annotation in-tegration:
middleware that produces a unified annotation on top
of diverse semantic annotators. On the one hand, the diversity of
vocabulary allows applications to benefit from the much richer vo-cabulary
available in an integrated vocabulary. For example, recent
work on web-scale wrapper induction (see, e.g., [14, 34]) uses se-
1486
2. Sports_team
NaturalFeature
Leeds won 1-0 at Sunderland and Coventry against Everton
and Nottingham Forest against Aston Villa ended goalless.
Facility
An1: AlchemyAPI
An3: Zemanta
GeographicFeature
An4: OpenCalais
An5: DBpedia Spotlight
SportsTeam
Figure 2: Conflicting and Re-enforcing Annotations
mantic annotations to lower the need for human supervision. On
the other hand, annotation integration can also have an impact in
terms of quality of the result. We will present experimental evi-dence
(see Section 2) indicating that the most widely-used individ-ual
annotators suffer from serious accuracy deficiencies. The over-lap
in vocabularies from individual annotators could be exploited
to allow an integrated annotator to boost accuracy by exploiting
inter-annotator agreement and disagreement.
Our approach to integration will be influenced by several distinc-tive
aspects of semantic annotation. First, semantic annotations do
not form a flat space of tags for document snippets, but rather as-sociate
snippets with a meaning, with the space of meanings being
highly structured – two annotations with different concepts may
thus still be compatible (e.g. if the concepts are in a subclass re-lation).
In many cases, Web-based annotators provide mappings
of instances and concepts to a publicly-available knowledge base
– for example DBpedia Spotlight links to the well-known DBpedia
ontology. Even in cases where such mappings are not provided,
ontological rules can often be conservatively inferred via common-sense
reasoning (e.g. a Facility is disjoint from an Attorney), and then
utilized as a precise means of capturing compatibility or incom-patibility
of the judgements given by different annotators. We will
look for integration methods that are ontology-aware – making as
much use as possible of semantic relationships that are available.
Another key factor is that many annotators are black-boxes, mak-ing
an approach based on hand-crafted “integration rules” problem-atic:
a particularWeb-based annotator may exhibit a certain pattern
of errors at one point, and later change as the underlying rule en-gine
is adapted. We will thus look for aggregation methods that
are as domain-independent as possible, and do not rely on hand-rolled
heuristics, focusing completely on the case of “black-box”
annotator aggregation.
We begin by describing an approach for on-line aggregation in
the absence of training data. We present a weighted repair ag-gregation
algorithm that assigns a score to annotations based on
the explicit opinions of individual aggregators (an annotator tags
a snippet with a class) implicit support (an annotator tags with a
subclass), and implicit opposition (e.g. an annotator tags with a
disjoint class). This method takes into account that annotators have
distinct vocabularies, thus the lack of a certain annotation may not
indicate lack of support. The weighted repair approach looks for an
aggregation which is consistent w.r.t. the ontology. We show that
for basic ontological constraints arising in the setting of concept
reasoning, an optimal solution can be found efficiently.
An obvious comparison point would be a system that can bene-fit
from off-line training in case training data is available. In this
paper, we describe an approach that uses ideas from supervised
learning. It uses a maximal entropy model, adapted to the setting
of semantic annotation. Both our supervised and unsupervised ap-proaches
support overlapping or nested spans, an occurrence that is
not uncommon in Web-based annotators.
We experimentally compare these approaches both with respect
to each other and to a number of external baselines. Our results
show clear benefits for aggregation of annotation, while also giving
new insight into the behavior of Web-based semantic annotators.
Organization. Section 2 gives background on annotators in gen-eral,
then describes the annotators we use in our examples along
with their mapping to a common ontology. It then overviews the
accuracy of the annotators, making use of a scoring model that
is tied to the ontology. Section 3 describes our aggregation algo-rithms.
Section 4 discusses the implementation of the algorithms,
and gives an experimental evaluation. Section 5 overviews related
work, while Section 6 gives conclusions.
2. BENCHMARKING ANNOTATORS
Semantic annotators take as input text documents and produce
semantic annotations in a variety of forms. The most common
ones are concept labels, e.g., for named entity recognition, entity
referencing, e.g., for named-entity disambiguation, and relation (or
role) labels, e.g., for relation extraction. Of these, the first is by far
the most mature, and we focus on it in this work. Such an anno-tation
consists of a contiguous segment of a document – namely, a
span – along with a concept (or class, or entity type: we use these
interchangeably throughout the paper), such as Person, Organization,
or Artist. The example from Figure 1 shows five concept annota-tions.
Three of them (AN1:Facility, AN4:Facility, and AN3:Location)
share the same span, with one of these (AN3:Location) disagreeing
with the other two on the concept name. Two others (AN1:Person
and AN2:City) have different (but overlapping) spans.
The state-of-the-art named entity extractors include software
frameworks such as LINGPIPE [2] and GATE [23], and web ser-vice
APIs such as OPENCALAIS, ALCHEMY, EXTRACTIV, SPOT-LIGHT
and ZEMANTA. The service-based model has particular in-terest
from the perspective of applications which want to exploit a
wide vocabulary. Each annotator may focus on a particular subset
of the concepts. They can make use of their own gazetteers and
hand-tailored concept-specific heuristics, allowing these to remain
proprietary, at the same time freeing applications from having to
constantly upload or update whenever the annotator is tuned. On
the other hand, the service-based model poses specific challenges
to integrators, since the performance of the annotators, and even
the vocabulary, may be highly dynamic.
Thus in our work we will target web service-based annotators,
focusing on the task of named entity recognition, which is the
most widely-supported function. OPENCALAIS, EXTRACTIV and
SPOTLIGHT return entities, including their position and entity type,
while ALCHEMY and ZEMANTA simply return text anchors with
their types. More importantly, the taxonomies of named entity
types they exploit are quite distinct. OPENCALAIS deals with 39
types, no pair of which are in a subclass relationship. Most of these
are quite general concepts such as Person and Organization. SPOT-LIGHT
and ZEMANTA adopt existing ontology resources (DBpedia
and Freebase respectively) as the labeling vocabulary. ALCHEMY
and EXTRACTIV have taxonomies in which the deepest subclass
hierarchy is quite short (usually no more than 2), including both
abstract and specialised concepts.
Vocabulary Alignment. As shown in the examples above, the
annotators do not have a flat vocabulary – annotations that are syn-tactically
distinct may have the same meaning, like NaturalFeature
and GeographicFeature in Figure 2. Some annotators, such as OPEN-CALAIS,
have mostly high-level concepts in their vocabulary – e.g.
Organization; others such as EXTRACTIV, have many much finer
classifications – e.g. FinancialOrg, CommercialOrg. Clearly, these
1487
3. <Class rdf:about="global:GeographicFeature ">
<owl:equivalentClass rdf:resource="alchemyAPI:GeographicFeature"/>
<owl:equivalentClass rdf:resource="OpenCalais:NaturalFeature"/>
<rdfs:subClassOf rdf:resource="extractiv:LOCATION"/>
<owl:disjointWith rdf:resource="OpenCalais:Facility"/>
…
</Class>
Figure 3: Example Snippet of Merged Ontology
should not be considered as conflicting within aggregation. To cap-ture
the relationships in the vocabulary, we will make use of an
ontology W which states the relationships between types of enti-ties.
In this work, an ontology consists of a finite set of concepts
(classes, types), including the special type ? (the empty concept)
and > (the universal concept), along with constraints restricting or
enforcing the co-occurrence of concepts. The most common con-straints
supported for concepts are:
Subclass (subsumption) constraintsC vD, stating that an en-tity
of type C is also an entity of type D.
Disjointness constraintsCuDv?, stating thatC and D have
an empty intersection.
The above constraints are supported by the major languages for
describing Web semantics, such as the OWL2-QL fragment of
OWL. As mentioned above, SPOTLIGHT and ZEMANTA directly
use an external ontology. The other annotators do not come with
any mapping, and so the meaning of the concepts had to be re-covered
from the documentation. In some cases the concepts of
the other annotators could be mapped into Freebase or DBpedia
concepts, but in other cases new concepts needed to be added.
We believe that as these annotators mature, the use of standard-ized
vocabularies will increase, making the use of rules as “hard
facts”, as we do throughout this work, appropriate. An effort in
this direction is the Schema.org initiative. We are focusing in this
paper on reconciling annotator outputs using schema information,
not on the very different issue of mapping/aligning concepts and
schemas (in other terms, “integrating the ABox”, not “integrating
the TBox”). We have thus created a merged ontology manually,
aligning the relevant fragments of Freebase and DBpedia within
it, focusing on rules that are clear from the documentation and by
validating them against Schema.org knowledge base when applica-ble.
The two external ontologies focus on subsumption relations,
while disjointness relationships are crucial for gauging whether two
annotations are compatible. We thus manually added disjointness
relationships that were obvious from the documentation. Figure
3 gives a snippet of our merged ontology that maps the concepts
GeographicFeature from ALCHEMY and NaturalFeature from OPEN-CALAIS
as equivalent classes of a global concept GeographicFeature,
while the concept Location from EXTRACTIV is mapped as a super-class,
and the concept Facility from OPENCALAIS as disjoint class.
The merged ontology used in this paper is accessible from http:
//diadem.cs.ox.ac.uk/roseann/MergedOntology.owl and can
be queried through a SPARQL endpoint available at URL: http:
//163.1.88.38:8081/openrdf-sesame.
Dataset. In order to assess the performance of those individual
annotators, we conducted experiments over four benchmark cor-pora:
(i) the MUC7 dataset [3], which contains 300 newswire arti-cles
for evaluating Named Entity recognition, annotated with stan-dard
high-level concepts such as Person, Organization, Location and
Money; (ii) the Reuters corpus (RCV1) [5], a general information
retrieval corpus of English newswire documents, subdivided into
topics; (iii) the corpus used by the FOX [1] entity extractor and
consisting of 100 snippets of text from news headlines annotated
with three concept types, namely Person, Location, and Organisation;
(iv) the corpus used by the NETAGGER [31] entity extractor and
consisting of text sourced from 20 web pages mostly about aca-demics
and annotated with the same three concepts above plus a
MISC concept that we excluded from the evaluation due to the ambi-guity
in the annotated instances. E.g., the string Computer Science
is annotated as MISC, but other similar topics e.g., Biomolecular
Computation are not. This corpus also allowed us to see the per-formance
of our own annotators in a very different domain from
Reuters and MUC7. Note that for our supervised approach, MEMM
(described in Section 3) we performed training/testing on represen-tative
samples of the gold-standards provided by the two datasets.
Since the original Reuters corpus is very large (810,000 articles),
we looked at five of the most common Reuters topics – Entertain-ment
Sports, FinancialEconomics, HealthcareSocial, Prod-ucts
and TourismTravel – and randomly sampled 250 documents,
distributing evenly over the topics. For this sub-corpus of Reuters,
we manually tagged the texts by using the most specific concepts
from the global vocabulary of the aggregated ontology discussed
above. Details on the coverage of the testset are shown in Table 1.
Our gold standard annotation of the Reuters corpus is available on-line
from [6]. Note that for the other corpora, we directly use the
original gold standard provided by the benchmarks.
Table 1: Description of Testing Corpora
Test Corpus Docs Covered Types Entities ()
MUC7 300 7 18,700
Reuters 250 215 51,100
Fox 100 3 395
NETagger 20 3 624
Precision and Recall. We measure PRECISION and RECALL in
a way that is ontology-aware: for example, given an ontology W,
if an annotator declares a given span to be a Person while our gold
standard indicates that it is an Artist, then this annotator will be even-tually
penalized in recall for Artist (since it had a miss), but not in
precision for Person (since it can be inferred via Artist). There is no
standard way to do this. Euzenat [18] has defined semantic preci-sion
and recall measures in the context of ontology alignment – but
the goal there is to assign static measures to an alignment, which
approximate the precision and recall one would get from an “ideal”
extension of the concepts via sets of instances. In our case, we can
use concrete cardinality estimates since we are interested only in
the precision and recall for an algorithm on a particular dataset.
More precisely, we define the precision of an annotator AN for a
concept C as:
PRECISIONW(C) =
jInstAN(C+)InstGS(C+)j
jInstAN(C+)j
where InstAN(C+) denotes all instances annotated as (a subclass
of) C by AN, and InstGS(C+) denotes all instances determined to
be (a subclass of) C in the gold standard. In computing the in-tersection,
we use a “relaxed” span matching, which requires only
that the spans overlap. Here and throughout the remainder of the
paper, when we say that the concept A is a subclass of a concept B
we mean that the rules of the ontology derive that B is a subclass,
possibly improper (that is, A may be equivalent to B). Similarly,
when we talk about superclasses of a concept A, we include those
derivable via transitivity, and include A as a superclass of itself.
We define the recall of an annotator AN for a concept C in an
analogous way, again using the “relaxed” notion of matching for
the intersection:
RECALLW(C) =
jInstAN(C+)InstGS(C+)j
jInstGS(C+)j
Based on the extended definitions of precision and recall, the F-SCORE
for concept C is defined in the standard way:
1488
4. F-SCOREW(C) =
2PRECISIONW(C)RECALLW(C)
PRECISIONW(C)+RECALLW(C)
Evaluation Result of Individual Annotators. We tested indi-vidual
annotators over the test datasets using the entire merged on-tology
as a reference schema. We implemented ontology-aware
scorers in order to calculate the precision and recall defined above.
The full list of types with score for each individual annotator, as
well as the source code of the scores are available online from [6].
We present selected results in Table 2, where highlighted val-ues
represent the best performance. For each annotator AN, we
display results for all of the MUC7, FOX, and NETAGGER con-cepts
that are in the vocabulary of AN. For each corpus, we first
mapped the response of individual annotators into the ones with
the entity types in the corpus and then compared the matching. For
the Reuters dataset we give a selection of concepts in the following
manner: (i) filter down to concepts with instances occurring in all
of the testing folders and with at least 10 occurrences; (ii) organize
the concepts by their F-score for AN, restricting to a subset includ-ing
roughly half of the concepts from the top quartile, and half from
the lowest quartile (iii) within the subsets above, select subsets so
as to avoid having more than two subclasses of any given concept.
We thus selected concepts with sufficient support, diversity in per-formance,
and diversity within the ontology.
We see from the table that each annotator varied greatly in per-formance,
ranging from below 10% in F-score to above 90%. If
we focus on concepts that occur in the list for multiple annotators,
we also see tremendous variation in performance, with no single
annotator having a clear advantage. OpenCalais has a higher perfor-mance
on most of the concepts that are within its vocabulary, but
it has a deficit on a few shared concepts, and on many important
classes, such as Date, it can identify only small subclasses. The
performance differences within an annotator are quite dramatic,
and surprisingly many of the concepts which exhibit the worst per-formance
are not obviously the most semantically complex. For
example, Movie would seem like something straightforward to han-dle
by referencing an online database such as IMDB, but the recall
numbers for ALCHEMY, SPOTLIGHT, and ZEMANTA were 20%,
28%, and 17% respectively. Also, the performance of annotators
on very high-level concepts, such as those in MUC7, was generally
higher than the performance on more specific concepts. However
the example of ZEMANTA shows that this is not universally true.
Another interesting observation is that all annotators contribute
to recall, i.e., each annotator contributes some annotations that are
not produced by any of the others.
Conflicts of Individual Annotators. Given that annotators can
vary radically in accuracy and that we have to consider all of them,
a question is to what extent the semantics of the annotations can
be useful to determine errors in the annotations. We analyzed one
coarse measure of this, based on the amount of annotator conflicts.
We measure two kinds of annotation conflicts: a basic conflict
occurs when one annotator annotates the span with a concept C,
and another annotator which has (a superclass of) C in its vocabu-lary
fails to annotate the same span with it. Clearly, this represents
a situation where the annotators have “differing opinions”. For an-notators
with low recall, a basic conflict may be a weak indicator of
a problematic annotation. Thus we also track strong conflicts, de-noting
situations when two annotators annotate the same span with
disjoint concepts C and C0. Table 3 shows the number of basic and
strong conflicts in each datasets – both the number of annotated
spans in which a given type of conflict (basic and strong) occurred,
and the total number of conflicts of each type. An annotated span
is a span which contained at least one annotation.
Table 3: Conflicts Statistic of Individual Annotators
Test Corpus Annot.
Span
Basic
Confl.
Basic
Span
Strong
Confl.
Strong
Span
MUC7 30,689 36,756 26,174 3,501 2,483
Reuters 29,162 21,639 15,654 2,937 1,981
Fox 798 943 605 185 68
NETagger 1,493 1,486 1,195 179 83
We have also tracked conflicts in spans, e.g. annotators agree-ing
on a concept, but disagreeing on the span. Table 4 reports, for
each corpus, on the number and type of span conflicts, in partic-ular:
(i) the total number of annotations, (ii) the number of anno-tations
with same span, (iii) the number of annotations having one
span strictly contained into the other, (iv) the number of annotations
having overlapping spans, and (v) the total number of annotations
having conflicting spans (i.e., either contained or overlapping).
Table 4: Span conflicts.
Test Corpus Tot. Annot. Same Span Containment Overlap Conflicts
MUC7 96,111 205,289 43,262 958 44,220
Reuters 87,666 198,745 37,743 898 38,641
Fox 2,730 8,107 2,012 38 2,050
NETagger 5,430 7,029 4,719 263 4,982
Although one might expect that conflicts are rare, the results
show that they are extremely frequent. A chief source of conflict is
that annotators have very limited capability of using context to dis-tinguish
ambiguity in meaning. Bloomberg sometimes refers to a
company, and sometimes to a person; Jaguar can refer to an animal
or to a car; Chelsea can refer to a place, a person, or a sports team;
Notting Hill can refer to a place or a movie. Different annotators
tend to favor particular solutions to these ambiguities. Note that
the number of conflicts is restricted by the limited overlap in the
vocabularies and limited recall of the annotators. For example, it is
very rare for three annotators to be mutually strongly conflicting on
the same span, since it is unlikely that all three will simultaneously
annotate the span.
Overall, the results show both the need for annotator integration
and the possibility of using conflict and agreement signals in con-structing
superior aggregate annotations. The results are not meant
to indicate an intrinsic pattern of who is correct on which concepts.
Results can vary as different datasets are used; further, these an-notators
are frequently modified, with the modifications impacting
both their vocabularies and their scope. Thus our goal in the rest
of the paper will not be to tune an aggregation algorithm to this
particular dataset (e.g. by hard-coded rules capturing which anno-tator
is seen to be reliable), but to come up with a general domain-independent
methodology for aggregating annotator values, not re-lying
on hand-tuned rules or weights.
3. AGGREGATION ALGORITHMS
Unsupervised aggregation via weighted repair. We introduce
an aggregation approach that requires no supervision and no infor-mation
on the reliability of aggregators.
Aggregation of conflicting data in the presence of integrity con-straints
is a long-standing topic in AI, semantic web, and database
research. In the case where the input consists of a set of opinions
from multiple “experts”, there are a number of judgement aggrega-tion
methods that have been studied – for example, based on vari-ations
of voting [24]. These algorithms have been examined not
1489
5. Table 2: Performance of individual annotators.
MUC7 NETagger Fox Reuters
Date
Organization
Person
Location
Money
Percent
Time
Person
Location
Organization
Person
Location
Organization
Person
Location
Organization
Date
Movie
ALCHEMY
Prec. .50 .88 .80 .90 - - - .81 .63 .74 1 .89 .79 .75 .92 .80 .68 .42
Rec. - .49 .55 .69 - - - .75 .52 .45 .92 .89 .62 .55 .70 .47 - .20
F1 - .49 .65 .78 - - - .78 .57 .56 .96 .89 .70 .64 .64 .60 .02 .27
SPOTLIGHT
Prec. - .63 .70 .76 - - - .58 .30 .66 1 .75 .65 .71 .81 .58 - .12
Rec. - .40 .22 .46 - - - .14 .40 .14 .40 .44 .33 .31 .57 .38 - .29
F1 - .49 .34 .57 - - - .23 .34 .23 .57 .56 .43 .42 .67 .46 - .17
EXTRACTIV
Prec. .91 .89 .87 .80 .98 1 .95 .82 .46 .50 .80 .72 .69 .81 .75 .76 .85 .54
Rec. .59 .72 .71 .78 .68 .89 .11 .44 .48 .34 .73 .95 .65 .70 .76 .60 .84 .42
F1 .71 .80 .78 .79 .80 .94 .21 .57 .47 .41 .76 .82 .67 .75 .76 .67 .85 .47
OPENCALAIS
Prec. .60 .85 .97 .94 .93 - - .83 .64 .69 .80 .87 .69 .95 .92 .85 .68 .51
Rec. - .73 .88 .71 .92 - - .52 .33 .69 .87 .90 .67 .87 .76 .69 .01 .53
F1 .01 .79 .92 .81 .93 - - .64 .44 .69 .83 .89 .68 .91 .83 .76 .02 .52
ZEMANTA
Prec. - .70 .89 .72 - - - .92 .36 .52 .62 .76 .67 .64 .83 .70 - .45
Rec. - .28 .07 .17 - - - .10 .38 .07 .58 .75 .54 .16 .40 .23 - .17
F1 - .40 .13 .28 - - - .17 .37 .13 .60 .76 .60 .26 .54 .35 - .25
only from the point of view of guaranteeing consistency, but from
the perspective of social choice theory – e.g. can they be considered
to be fair to the parties involved ([22]). When, on the contrary, the
input is seen as a single inconsistent dataset, a common approach
in the database and semantic web community is based on repair:
finding a “minimal” (in some metric) set of update operations to
the initial data so that the updated data is consistent [33].
Our solution in the unsupervised setting will use ideas from both
voting and repair. Assume for the moment a single span ˆ s that is
tagged by several annotators. In this case, concepts can be iden-tified
with atomic propositions, and the ontology relationships can
be considered as propositional constraints – e.g. if concepts C and
D are disjoint, our logical theory includes the constraint C !:D.
Thus we can translate the ontology W to a propositional theory TW.
One component of our solution is a voting-based AtomicScore func-tion
AtomicScore(C), assigning an integer to each concept C. This
represents the degree of support for or opposition to C by annota-tors.
We will always consider an annotator to support a concept C
if it tags the span ˆ s with a subclass of C – each such supporter will
add to AtomicScore(C). AtomicScore(C) will drop for each annota-tor
that tags ˆ s with a class disjoint from C. We allow in addition
the possibility of “opposition by omission” – AtomicScore(C) drops
when an annotator fails to tag ˆ s with a superclass of C that is in its
vocabulary. Such a penalty is justified under the assumption that
annotators have high recall. The general form of our scoring func-tion
allows the possibility of tuning the assumptions about both
precision and recall, as well as support for more general concept
constraints than simply subsumption and disjointness:
AtomicScore(C) = SA2Anns( SDvC2WSupportWeightA;D Support(A;D)
SDuC=?2WSupportWeightA;D Support(A;D)
SCvD2WOmitWeightA;D Omit(A;D) )
Above, Anns denotes the set of annotations, D v C 2 W indicates
that from the rules of ontology W one can prove D is a subclass
of C, and DuC = ? 2 W indicates that W implies disjointness of
D and C. Support(A;D) is 1 if annotator A tags the span s with
D, and is 0 otherwise. Omit(A;D) is 1 iff A has D in its vocabu-lary,
but failed to tag span s with D. Lastly SupportWeightA;D and
OmitWeightA;D are non-negative [0;1]-valued weights that indicate
how much weight the tagging of A with D or the omission of D
by A should have. Note that a term in the sum above will only be
present if the corresponding type of axiom is asserted (or inferred)
in the ontology. The above sum considers two ways for an anno-tator
can “oppose” a concept C, by tagging with a concept disjoint
from C or by omitting to tag with (a superclass of) C in its vocabu-lary.
If our ontology language was rich enough to have (e.g.) rules
saying that C and D partition all entities, we would also have an
additional term representing “support by omission”: an annotator
omitting D is supporting C. Since we have no such rules in our
example ontology, we do not examine this setting.
We will discuss ways of setting the weights further below. A
straightforward approach is to assume annotators have low recall,
and thus annotator A failing to tag with a concept D gives no sig-nal
that the entity is not in class D – this corresponds to setting
each OmitWeightA;D to 0, eliminating two of the terms above. In
our experiments with wide-spectrum web-based annotators we find
that recall is indeed low, and hence we utilize this approach in our
implementation (see Section 4 for details). However, for domain-specific
annotators, other approaches, such as those described be-low
based on bootstrapping, may be appropriate.
Another assumption would be that the precision does not vary ei-ther
across annotators or across concepts – hence SupportWeightA;D
can be set uniformly to 1. We will refer to this as simple atomic
scoring, and under this policy AtomicScore(C) simplifies to:
SDvC2WfA j A annotates with Dg
SDuC=?2WfA j A annotates with Dg
EXAMPLE 1. Assume the simple atomic scoring model, and
consider a span annotated by annotator A1 with City, annotator A2
with Country, and annotator A3 with NaturalFeature. Since these con-cepts
are disjoint in the ontology, under the simple scoring model
each of City, Country, and NaturalFeature will have AtomicScore 1,
since each is supported by one annotator and opposed by 2. The
least common ancestor of these three concepts, namely Location,
will have AtomicScore 3.
Given the atomic scores, a score combination function is respon-sible
for calculating a boolean combination of concepts that is con-sistent
with the ontology, based on the scores of constituent literals.
For the combination function we propose a declarative approach,
denoted as weighted repair (WR). We begin with the union of all
concepts returned by any annotator, which can be considered as a
1490
6. conjunction of concepts sInit. A repair operation Op is either a dele-tion
of a concept occurring as a conjunct within sInit or an insertion
of a concept that is absent from sInit. The application of Op to sInit
is a new formula. For a deletion of class C, it is formed by remov-ing
every conjunct corresponding to a subclass of C while adding
the negation of C, while for an insertion it is formed by conjoin-ing
with a proposition corresponding to C (the superclasses could
also be inserted, but this will have no impact on the calculation be-low).
A set of repairs is internally-consistent if no two operations
conflict: we do not delete a class C and also insert a subclass of
C. For an internally-consistent set of repairs S = fOp1 : : :Opng, the
application on sInit, denoted S(sInit) is defined to be the result of
applying the Opi in any order – one can easily check that this gives
a unique result. A repair set is non-redundant if we do not delete
or insert two concepts in a subclass relation. We call an internally-consistent,
non-redundant repair set S such that S(sInit) is consistent
with TW a solution.
EXAMPLE 2. Let us return to Example 1. sInit is City^Country^
NaturalFeature. The repair set fDel(City); Del(Country)g is internally-consistent
and non-redundant, and its application to sInit gives
the formula NaturalFeature ^ :Country ^ :City. Since this is con-sistent
with the ontology, the repair set is a solution. The repair
set fDel(City); Del(Country); Del(NaturalFeature)g is another solution,
yielding the formula :NaturalFeature^:Country^:City.
Our goal is to find a solution S such that its aggregate score is
maximal among all solutions, where the aggregate score of S is:
SIns(C)2SAtomicScore(C)SDel(C)2SAtomicScore(C)
That is, an operation that deletes a concept C incurs the penalty
AtomicScore(C), while an insertion of a concept C incurs the nega-tive
of AtomicScore(C) as a penalty.
There can be many repairs that achieve the maximal score, and
it is natural to impose other criteria to prune them: 1. Given two
solutions with the same score and different numbers of repairs, we
prefer the smaller one. 2. Given solution S1 = S0 [fIns(C1)g S2 =
S0 [fIns(C2)g, with C2 a subclass of C1, we prefer S2 – that is we
prefer those that insert deeper (that is, more specific) classes. Note
that, unlike in other database repair contexts, repair operations can
have positive score (net benefit), negative score (net penalty), or
zero score (cost neutral, but perhaps needed to gain consistency).
We can solve this optimization problem by a reduction to integer
linear programming (ILP). We have variables XDel;C for each con-cept
C that can be deleted, with a value of 1 denoting deletion and
0 denoting no deletion. Similarly we have variables XIns;C for each
concept that can be inserted. Each disjointness or subsumption re-lation
in the ontology corresponds to an integer constraint; e.g. for
concepts B;C 2 W with B 2 sInit and C62 sInit, a disjointness con-straint
BuC =? implies an integer inequality XIns;C SB0BXDel;B0
– that is, if we insert C we must delete a subclass of B. The scoring
policy described above is easily translation to maximization of a
corresponding linear objective function.
EXAMPLE 3. Consider an annotated span with 5 differ-ent
concepts C = fCommercial Org; Financial Org; Educational Org;
Government Org; Organization Otherg, where all the concepts are mu-tually
disjoint and in subclass relationship with the concept
Organization. Moreover Organization is the disjoint union of all
the concepts in C. The set of concepts involved is L = C [
fOrganizationg. For each concept l involved we have two variables,
XDel;l and XIns;l to determine whether a concept must be deleted or
inserted in the final solution. Since each concept in C is already
in the initial solution and cannot be inserted, we have XIns;l = 0
8l 2 C. On the other hand, Organization can only be inserted, there-fore
XDel;organization = 0. Note that Organization is the only concept in
the example eligible to be inserted in the final solution. For each
pair of disjoint concepts in the initial solution we add an integer
constraint to ensure that at least one of them will be deleted, that is
XDel;l +XDel;h 1 8l;h 2C j l uh v?. Then we add a constraint to
ensure that Organization is the disjoint union of classes in C, that is
jCjSl2CXDel;l = XIns;organization. Eventually we try to maximize
k (Sl2LAtomicScore(l) XIns;lSl2LAtomicScore(l) XDel;l)
(Sl2LXDel;l+XIns;l)
where k is the total number of concepts involved + 1.
This optimization problem requires (in particular) checking con-sistency
of various boolean combinations with respect to the propo-sitional
theory TW. For rich enough ontological rules (enough to
give an arbitrary propositional theory), is NP-hard, and the use of
a robust ILP solver is needed. In our implementation of the ILP-based
approach, we used IBM’s ILOG CPLEX as a solver, coupled
with OWLIM-Lite to query the ontology. For each span a collection
of a few hundred constraints is produced. We can prune these by
considering only super-classes of concepts that are returned by one
of the annotators on the span; given that the depth of the concept
hierarchy is in single digits, this results in a drastic reduction in the
number of variables. We also generate constraints corresponding to
disjointness axioms from the top of the ontology hierarchy down-ward,
allowing us to suppress redundant disjointness constraints,
further reducing the constraint set. Note that the ILP reduction also
gives an NP upper bound for the decision problem corresponding
to finding an optimal repair (checking the existence of a repair with
a given score).
However, in the special case of disjointness and subsumption
constraints only, we can find at least one optimal solution tractably,
without appeal to ILP:
THEOREM 1. Assuming only disjointness and subsumption
constraints, an annotation set corresponding to some optimal re-pair
set can be achieved by inserting every non-annotated con-cept
of positive AtomicScore, deleting every concept of negative
AtomicScore and deleting some concepts with score 0.
The proof relies on the following simple observation about the
AtomicScore function, which can be proven simply by unwinding
the definitions:
CLAIM 2. Given two disjoint concepts C1 and C2 at most one
can have a positive AtomicScore, and if one does, the other must
have negative AtomicScore.
The proof of Theorem 1 now follows easily: it suffices to con-sider
any repair set S0 that yields a boolean combination containing
all concepts with positive AtomicScore, and which deletes all con-cepts
with AtomicScore 0. Using the claim above we see that S0
leaves the database consistent. Now consider S an arbitrary non-redundant
repair set of maximal score, whose application yields a
consistent formula. First, suppose the application of S deletes a
concept C of positive AtomicScore. Thus S contains Del(C0) for C0 a
superclass of C. Consider the repair S0 formed by removing Del(C0)
from S. It is easy to see, using the definition of AtomicScore, that
AtomicScore(C0) must also be positive. If S0 were inconsistent, this
means that C0 was disjoint from some class D either inserted or left
intact by S0. But using the claim we see that any such D that was
inserted must have had negative AtomicScore, as would any D left
intact. Thus a modification to S by deleting or not inserting any
1491
7. such D, as appropriate, would result in a repair set that yields con-sistency
and has higher score. Now suppose that the application
of S fails to insert a C of positive AtomicScore. We can argue, as
above, that the application S[fIns(C)g gives a consistent formula.
Finally, we consider a concept C of negative AtomicScore. Clearly
adding Del(C) to S does not lose consistency, and S[fDel(C)g will
have higher score than S; thus if S has optimal score its application
must already delete C.
Theorem 1 gives a simple algorithm for finding an annotation
that results from an optimal repair set: start with inserting every
non-annotated concept of positive AtomicScore, deleting all concepts
that are disjoint from such concepts, and deleting all concepts with
score 0. Applying this to the initial formula, one gets a for-mula
that is consistent, since it cannot contain disjoint concepts
with AtomicScore 0, and also cannot contain disjoint concepts where
one has positive AtomicScore. By the theorem above, this gives an
optimal result.
EXAMPLE 4. Returning to Example 1, an optimal score can be
obtained by a repair set that deletes all 3 base annotations and in-serts
Location. This matches the intuition that the preponderance
of the evidence points against each specific annotation, while sup-porting
their common superclass Location.
We return to the question of the settings for SupportWeightA;D
and OmitWeightA;D, beyond that given in the simple atomic scoring
model. We have also experimented with a bootstrapping approach
setting all the weights to be initially uniform at 1, and then iterating:
SupportWeightk+1
A;D = AVGSupport(A;D;s) NormAtomicScorek(D)
OmitWeightk+1
A;D = AVGOmit(A;D;s) (1NormAtomicScorek(D))
where Support(A;D; s) indicates that s is a span which A tags with
D, Omit(A;D; s) indicates that A has D in its vocabulary but omitted
it on s, and NormAtomicScorek(D) is a version of the AtomicScore(D)
calculated based on SupportWeightk and OmitWeightk, but normalized
to be in [0;1] (e.g. by shifting the scores and dividing by their max-imum).
The intuition is that if A has high precision for D, spans
that it annotates with D should get high support in round k, and
hence NormAtomicScorek(D) should be be close to 1, making its
SupportWeight high in round k+1. Similarly, if A has high recall
for D, spans for which it omits D should get low support in round
k, which will make OmitWeightk+1 high.
The set of documents over which the spans are calculated obvi-ously
will have a large impact on the results. We have tested the
bootstrapping approach using the entire test corpus from Section 2.
We have found that it works well for very general concepts, such
as Person, Place, and Organization, where examples are plentiful, but
hurts on a more specific (and less frequent) classes – increasing the
Micro-average F-score but decreasing the Macro-average F-score
(see Section 4 for the definitions of these).
Dealing with overlapping spans. The procedure described
above assumes that annotations are related only if they have the
same span, i.e., they insist on the same tokens of text. The repair
is therefore computed one span at a time. On the other hand, anno-tations
often have nested or overlapping spans, as in Figure 1. We
therefore consider a more careful definition of annotation span that
takes into account overlaps.
Overlaps: Let T be a document represented as a sequence of
tokens T =ht1; : : : ; tni and a an annotation of a sequence hti; : : : ; t ji,
1 i j n, of tokens of T provided by an annotator ANi using
a label lwi from its local vocabulary Lwi . We denote by type(a) the
label lW from the vocabulary LW of the global ontology to which
lwi is mapped.
We also denote by intv(a) the interval [i, j) of tokens covered by
a in D. Given two annotations a1, a2 we say that they are disjoint
iff intv(a1)intv(a2) = / 0, they overlap iff intv(a1)intv(a2)6= / 0, in
particular they coincide iff intv(a1) = intv(a2). Moreover, we say
that a1 is contained in a2 iff intv(a1) intv(a2).
Orthogonal spans. A way of dealing with overlapping spans
is to consider distinct spans as mutually independent, i.e., anno-tations
conflict or support one another only if their spans coincide.
This obviously leads to a loss of precision in the aggregation, since
overlapping spans do often refer to the same entity.
Coalesced spans. A completely opposite strategy is to take the
connected component of spans under the overlap relation, consider-ing
each component as a span having all of the annotations associ-ated
with its component. We found that in practice these connected
components are quite small – at most four tokens. But consider-ing
a disjoint pair of annotations connected in this way as being in
conflict can lower the recall: e.g. in the case of nested annotations
such as “Sweden Central Bank” where the whole string is anno-tated
as Organization, e.g., by OpenCalais, and the string “Sweden”
is annotated as Country, e.g., by AlchemyAPI.
Hybrid spans. To circumvent the above problem, we introduce
a hybrid annotation span defined as follows. Given an ontology W,
we define two annotations a1, a2 (with a16= a2) as belonging to the
same annotation span if either:
1. a1 is contained into a2 (resp. a2 is contained into a1) and it
is not the case that W j= type(a1)utype(a2) v ?.
2. a1 overlaps with a2 and neither a1 nor a2 is contained into
the other.
3. a1 coincides with a2.
Condition 1 relates overlapping annotations that are possibly re-dundant.
This is the case, e.g., when one annotation is contained in
another with types in superclass relationship. Condition 2 captures
the case of annotations with ambiguous spans, i.e., they overlap on
one or more tokens but neither contains the other. Note that we
purposely excluded those cases of (strictly) contained annotations
with logically conflicting types.
Consider again the annotations in the example of Figure 1.
According to the definition of hybrid spans, AN1:Facility and
AN4:Facility belong to the same annotation span, due to Condition 1.
The same applies to AN2:City and AN3:Location, since City v Loca-tion.
On the other hand, AN1:Person and AN3:Location do not belong
to the same annotation span, because W j= Person u City v ?. The
same applies to AN1:City and AN2:Facility. Therefore, for this exam-ple,
the repair procedure will operate on AN1:Facility, AN4:Facility,
AN3:Location, and AN2:City, leaving out AN1:Person. Notice that,
for the example of Figure 2, the repair procedure will operate on all
annotations due to Condition 3.
Our experiments have shown that the use of hybrid spans in-creases
the recall but lowers the precision w.r.t. coalesced spans.
However, the change in performance is below 1% and the over-all
F-score remains practically unchanged (0.1%) on all datasets.
In the following we will stick to hybrid spans since containment
among annotations is very frequent in practice.
Comparison with supervised aggregation. In Section 4 we
benchmark our repair-based approach against several baselines, in-cluding
all freely-available annotation aggregators. But a second
natural comparison point is against what can be accomplished with
supervision. Currently we know of only two approaches (i.e., FOX
and NERD) to supervised aggregation of Web-based annotators. In
FOX a single web annotator is considered together with standalone
annotators and aggregated via neural networks, while in NERD the
1492
8. details of the machine learning technique used for aggregation is
not explicitly stated. We have thus implemented our own super-vised
approach to see to what extent one can leverage the existence
of training data to find patterns that indicate (in)correctness of an-notations.
Can supervision be used not just to look for patterns in
reliability of annotators (e.g. from tracking the recall or precision
of annotators on concepts), but to detect and exploit more complex
patterns in the behaviour of annotators?
We build on maximal entropy models, which have gained wide
popularity for a variety of individual extraction tasks in recent
years [8]. In our case, we wish to use a sequential model, which
will allow us to reconcile span boundaries jointly with concept
names, which also allows span boundary differences to serve as
clues. Also, this model is general enough to detect a wide class
of behavioural patterns within annotators. When non-sequential
approaches have been applied to aggregation of annotations (e.g.
the neural networks used in [1]) they have ignored span boundary
issues. We thus use a Maximal Entropy Markov Model (MEMM)
[29], which combines features of maximal entropy models with
the sequential approach of a Markov model. A MEMM has a set
of states, an input and output alphabet, and a probabilistic transi-tion
and emission functions that capture the system dynamics when
a new input item is consumed. The inputs consist of sets of features
token by token. As with other maximal entropy models, they allow
for feature functions that overlap, without requiring strong inde-pendence
assumptions.
Applied to this setting, we let AN = fAN1;AN2; g be the
collection of base annotators, Classes(W) be the concept classes in
ontology W, and D be a document represented as a sequence of
tokens D = ht1; : : : ; tni. By adopting the standard begin/in/out (B-I-O)
encoding scheme, let L(W)=fB ; I gCon(Classes(W)), where
Con(Classes(W)) is the set of collections of concepts that are con-sistent
with ontology W. A symbol (B;S), where S is a consistent
collection of concepts, indicates a position that is the beginning of
a snippet for an entity that is a member of each element of S, while
(I;S) indicates that the position is inside of such a snippet. We let
States(W) be the power set of L(W). Elements of States(W) can be
mapped to a boolean combination in the obvious way: conjoining
all the concepts in the list and dropping the set of decorations. We
identify the empty subset as “Other” meaning that a token is out of
any of the classes in W.
The input alphabet for observations about tokens, denoted O,
consists of the evidence vectors, that correspond to a token t –
the labelings of the different base annotators. We denote an ev-idence
vector w.r.t. token t as ~Evt = fltA
NjAN 2 ANg, where
ltA
N 2 States(W). Note that we do not assume that the individual
annotator opinions are consistent. In our prototype, the output al-phabet
was identical to the state space, and we denote both by S
below. We can easily extend to finer-grained state spaces (e.g. dis-tinguishing
between sections of an annotation).
As in any Markov model, the key problem is to define the transi-tion
probabilities: given two states l; l0 along with an input symbol
o, define the probability of transiting from l to l0 when traversing
o. In the maximal entropy approach, this conditional probability is
set to be the one which maximizes the entropy among those distri-butions
in which the expectation of a distinguished set of feature
functions agree with their expectation over a sample distribution.
Such a conditional probability must have the form:
Pl(l0jo) =
1
Z(l;o)
exp(åll
i fi(l0;o))
Above F = f f1; f2; ; fmg is a finite set of feature functions, that
take as input the target states along with the input symbol. ll =
fl1;l2; ;lmg are the corresponding weight parameters learned
in a training phase of source state l, and Z(l;o) is a normalization
factor for the weights, which is calculated using the formula:
Z(l;o) = å
l02S
exp(åll
i fi(l0;o))
The weight parameters are set so as to make the expectation of
the feature functions match their empirical expectation. There are
standard methods to calculate the weights by training that achieve
this. In particular, we can use the generalized iterative scaling
(GIS) [29] algorithm that iteratively updates the parameters based
on expectations defined over their current values.
Thus the transition probability calculation is straightforward
once the feature functions are defined and the corresponding
weights are trained. Our most fundamental features are based on
annotator opinions. In order to deal with data sparsity issues,
we construct features based on decomposed evidence on token t,
which focuses on the individual opinions of the annotators ANi, as
pANi ( ~Evt
) and the co-occurrence of pairwise opinions between dif-ferent
annotators ANi and ANj, as pANi ;ANj ( ~Evt
). Here pA is the
projection on a set of annotators A. So that we can capture very
fine-grained patterns, we allow these features to be specific to a
particular piece of evidence and target state. For example, the fol-lowing
feature function corresponds to the event where one anno-tator
annotates a token as the beginning of a Facility and the token
is actually the beginning of a Person:
f1(l0;o)=
8
:
1; if lOpenCalais = B Facility, lAlchemyAPI = B Person
and l0= B Person ;
0; otherwise.
We train separate Maximum Entropy models via the GIS algo-rithm
for each source state l with its corresponding training set T l .
T l consists of token events, of which the previous token label is
l, as T l = fhpANi ( ~Evtk );pANi ;ANj ( ~Evtk ); labeltk ijlabeltk1 = l and
ANi;ANj 2 AN and ANi6= ANjg.
We then process the input evidence sequence by constructing a
Markov sequence, after which we decode using the standard Viterbi
algorithm, calculating the most likely annotation sequence accord-ing
to the model and returning the output annotations. More details
on the implementation of MEMM are deferred to [6].
Note that the Markov model resulting from this training pro-cesses
the document per-token, not per-span, and makes no as-sumptions
about the way annotator spans intersect. Thus nested
or overlapping spans of base annotators do not require any special
pre-processing. The model will enforce that an output annotation
is always well-formed (e.g. that is does not begin a new span be-fore
closing an old one), and thus that the token-string produced by
MEMM will correspond to an annotation of the document, which can
be returned as the final output.
We utilize the library OpenNLP-MAXENT-3.0.0 as the core solution
for the Maximum Entropy framework. This is a mature JAVA pack-age
for training and running maximum entropy models. The toolkit
OpenNLP Tools [4] is used to deal with common NLP tasks such
as tokenization and sentence segmentation.
As the performance of predicting the most likely annotation se-quence
is highly sensitive to the state space of the MEMM, we per-form
some post-processing on the Markov sequence resulting from
the product of the Maximum Entropy model with the input se-quence,
removing transitions that are (heuristically) unlikely to be
relevant. For example, if there is an outgoing transition from a state
having probability over a given threshold (currently set to 0:6), we
will remove all other outgoing transitions from the representation.
1493
9. We use a map to index only the transitions with nonzero probabil-ity,
taking advantage of the sparsity of the transition matrix. We
further exploit the fact that our Markov model does not consider
interactions across sentences; thus we can run Viterbi in parallel
over each sentence.
4. EXPERIMENTS
We carried out an extensive experimental evaluation of our ag-gregation
methods. A first set of experiments has been devoted to
demonstrate and quantify the benefit of aggregation over individual
annotators. A second set of experiments compares the various ag-gregation
techniques, including those provided by state-of-the art
aggregators such as FOX [1] and NERD [32].
Individual vs aggregated. We evaluated our aggregation meth-ods
using the same datasets and experimental setting which have
been described in Section 2. We compared the performance of both
WR and MEMM against individual annotators and against a na¨ıve
Baseline aggregation that simply collects all annotations from indi-vidual
annotators and, after mapping the returned concepts to the
global ontology, returns the union of all annotations regardless of
consistency. Table 5 reports the 10-fold cross-validation perfor-mance
of each aggregation approach against each of the individ-ual
annotators, where highlighted values represent the best perfor-mance.
Since the focus is on accuracy (and not, e.g., vocabulary
coverage), the comparison for each annotator is made only w.r.t.
the concepts that are in the annotator’s vocabulary – thus a differ-ent
set of key concepts for each annotator. We give averages of the
precision, recall, and F-score, measured in the ontology-aware way
given in Section 2. We provide: a macro-average that first averages
over each concept, and then over concepts, and a micro-average
that averages over each annotation (thus giving more weight to the
concepts that are more highly represented in the testset).
Our first observation is that all three aggregation methods con-sistently
outperform individual annotators in F-score except for
OPENCALAIS and ALCHEMY. OPENCALAIS and ALCHEMY have
a slightly higher F-score (i.e., 3% better on average) than any of
the aggregators on the NETagger and Reuters corpus respectively.
However, it is worth noting that their vocabularies represent only
18% of all concepts in the gold standard. On average, aggrega-tion
via WR results in a 15% increase in performance, while MEMM
produces an average of 22% increase with a peak of 66% improve-ment
over ZEMANTA. As expected, the Baseline aggregation deliv-ers
worse overall performance than our aggregation methods, with
a relatively high recall but a drastically lower precision, worse than
any of the individual annotators.
Supervised vs unsupervised. MEMM delivers the best perfor-mance
among the aggregators. We see from the figures that for
MUC7, with rich training examples over a small number of target
concepts, MEMM delivers high micro-average and macro-average
F-scores (mostly above 90%). There is a degradation in macro-average
scores in the case of concepts recognised by OPENCALAIS
and EXTRACTIV. The decrease is caused by some key concepts
which have limited training data support in our corpus. For ex-ample,
PhoneNumber has very few instances in the news-oriented
Reuters corpus (a total of 36 occurrences). Since the examples are
extremely sparse in a training set, MEMM has only 5% in F-score
over PhoneNumber, while the other annotators are able to achieve
around 80%. It is unsurprising that individual annotators can do
well on phone numbers, given that they can be mostly recognised
via regular expressions.
In general, MEMM learns which annotator to trust at a concept-level
through global training. Consider, e.g., the key concept Per-son:
the performance of individual annotators varies within a wide
range in terms of F-score – that is from 14% to 93% on MUC7
(see Table 2); 26% for ZEMANTA, 43% for SPOTLIGHT, 63% for
ALCHEMY, 75% for EXTRACTIV and 91% for OPENCALAIS on
Reuters. MEMM learns to trust the one with more reliable behaviour
over Person – namely, OPENCALAIS– and thus achieves a compar-atively
high accuracy (94% and 89% on two dataset respectively).
There are extreme cases where MEMM does better than all annota-tors
put together: for example, MEMM, on CommercialOrg improves
the accuracy by more than 20% against individual annotators and
more than 10% against the other two aggregators that are aware of
the concept. This is possible because the annotators often correctly
annotate Organisations, and the probability of an Organization being
a CommercialOrg is high; by detecting this pattern, MEMM can notice
many CommercialOrg instances. Obviously, this kind of pattern is
highly data-dependent, and in scaling the dataset out this particu-lar
inference may not longer hold (and hence would not be applied
by MEMM given sufficient training). But the result does show that
when patterns within the data or the annotators do exist, they can
be detected by our aggregator.
WR has precision competitive with MEMM. This is because it be-lieves
in a concept C only if the majority of the “judges” who know
about C labelled the span with C or one of its subclasses – indeed,
such a signal turns out to be strongly correlated with correctness.
When all the annotators are competent over a certain concept, the
recall is also comparable with individual annotators, and in some
cases is superior. For example, the majority of the annotators are
competent on concept Country, with F-score above 75%, with only
ZEMANTA being significantly lower. For this concept WR achieves
a precision of 90% while still having recall comparable with the
best annotator, OPENCALAIS. However a more common situation
is when the recall of individual annotator is really low. In this sce-nario
WR has much lower recall, since most of the instances are
considered to lack support. For the concept Person discussed above,
only OPENCALAIS and EXTRACTIV have good recall, which low-ers
the recall, and hence the F-score, of WR.
The key advantage of WR is its resilience to sparsity in the
dataset. We see from the example of PhoneNumber discussed above
that MEMM performs quite poorly when training data is sparse,
while WR has score comparable with individual annotators.
Comparative evaluation. We compared the performance of our
aggregation methods against the FOX [1] and NERD [32] NER ag-gregators
(discussed in Section 5) on all datasets.
Although our focus is onWeb-based annotators, the comparative
evaluation required the use of non web annotators used by FOX,
namely the STANFORDNER and Illinois NETAGGER, as well as
other web-based annotators such as SAPLO, WIKIMETA, LUPE-DIA,
and YQL. Individual results about these annotators have not
been included in this paper due to space reasons but can be ex-amined
online at [6]. Figure 4 summarizes the results and clearly
shows that our aggregation techniques are at worst comparable with
FOX and NERD and in average outperform the competitor aggrega-tors
of more than 15% in F-score on the MUC7, NETagger, and
Reuters datasets. An interesting case is the NETagger corpus that
suffers from high data sparsity. In fact, not only does NERD show
a better behavior, but also MEMM is often outperformed by WR on
this dataset. On the other hand, the difference in performance is
less evident on the Fox corpus where FOX and MEMM behave very
similarly, but they are often outperformed by WR due again to data-sparsity
issues in the dataset.
The full suite of test results, together with all the text resources
and annotated gold standards is available online at [6].
Performance and Scalability. The last set of experiments stud-ies
the amount of resources required to compute a solution for the
1494
10. Table 5: Individual annotators vs aggregators.
MUC7 NETagger Web Reuters
micro macro micro macro micro macro micro macro
Prec.
Rec.
F1
Prec.
Rec.
F1
Prec.
Rec.
F1
Prec.
Rec.
F1
Prec.
Rec.
F1
Prec.
Rec.
F1
Prec.
Rec.
F1
Prec.
Rec.
F1
ALCHEMY .86 .45 .59 .76 .44 .55 .78 .60 .67 .73 .57 .64 .90 .81 .85 .89 .81 .85 .80 .59 .68 .62 .44 .51
Baseline .44 .84 .57 .55 .84 .66 .40 .79 .53 .41 .75 .53 .35 .90 .50 .36 .91 .51 .43 .89 .58 .43 .75 .54
WR .73 .84 .78 .77 .83 .80 .51 .78 .61 .53 .72 .60 .71 .88 .79 .72 .89 .79 .68 .82 .75 .61 .64 .63
MEMM .95 .88 .92 .95 .87 .91 .77 .55 .64 .74 .52 .60 .89 .88 .88 .89 .88 .88 .83 .79 .81 .70 .60 .64
SPOTLIGHT .69 .37 .49 .70 .36 .48 .30 .15 .19 .30 .17 .20 .77 .38 .50 .77 .38 .50 .66 .45 .53 .45 .32 .37
Baseline .40 .91 .56 .43 .91 .59 .40 .79 .53 .41 .75 .53 .35 .90 .50 .36 .91 .51 .36 .88 .52 .38 .66 .48
WR .70 .91 .79 .73 .91 .81 .51 .78 .61 .53 .72 .60 .71 .88 .79 .72 .89 .79 .64 .83 .72 .54 .56 .55
MEMM .96 .94 .95 .96 .94 .95 .77 .55 .64 .74 .52 .60 .89 .88 .88 .89 .88 .88 .84 .78 .81 .61 .52 .56
EXTRACTIV .87 .68 .76 .92 .64 .75 .56 .35 .40 .50 .35 .41 .73 .79 .76 .74 .78 .76 .71 .69 .70 .65 .62 .63
Baseline .45 .82 .58 .73 .75 .74 .40 .79 .53 .41 .75 .53 .35 .90 .50 .36 .92 .51 .52 .81 .63 .62 .72 .67
WR .73 .82 .77 .86 .75 .80 .51 .78 .61 .53 .72 .60 .71 .88 .79 .72 .89 .79 .72 .76 .74 .69 .65 .67
MEMM .95 .85 .90 .94 .76 .85 .77 .55 .64 .74 .52 .60 .89 .88 .88 .89 .88 .88 .83 .79 .81 .70 .63 .66
OPENCALAIS .91 .61 .73 .85 .65 .73 .73 .51 .60 .60 .45 .51 .80 .82 .81 .79 .82 .80 .80 .72 .76 .69 .59 .64
Baseline .45 .85 .58 .62 .85 .72 .40 .79 .53 .41 .75 .53 .35 .90 .50 .36 .91 .51 .49 .86 .62 .48 .77 .59
WR .73 .84 .78 .81 .85 .83 .51 .78 .61 .53 .72 .60 .71 .88 .79 72 .89 .80 .66 .79 .72 .60 .67 .63
MEMM .95 .88 .92 .95 .88 .92 .77 .55 .64 .74 .52 .60 .89 .88 .88 .89 .88 .88 .80 .72 .76 .69 .55 .61
ZEMANTA .72 .20 .31 .76 .18 .29 .47 .14 .20 .37 .15 .19 .71 .64 .67 .69 .63 .65 .69 .29 .41 .51 .21 .29
Baseline .41 .91 .56 .43 .91 .59 .40 .79 .53 .41 .75 .53 .35 .90 .50 .36 .91 .51 .42 .87 .56 .40 .60 .48
WR .70 .91 .79 .73 .91 .81 .51 .78 .61 .53 .72 .60 .71 .88 .78 .72 .89 .79 .67 .80 .73 .53 .53 .49
MEMM .96 .93 .95 .96 .94 .95 .77 .55 .64 .74 .51 .60 .89 .88 .88 .89 .88 .88 .83 .76 .79 .59 .49 .54
1
0.9
0.8
0.7
0.6
0.5
0.4
m Fscore
M Fscore
m Fscore
MEMM WR Fox
M Fscore
m Fscore
M Fscore
m Fscore
M Fscore
Reuters MUC7 NETagger Fox
(a) vs FOX
0.95
0.85
0.75
0.65
0.55
0.45
0.35
m Fscore
M Fscore
M Fscore
MEMM WR NERD
M Fscore
m Fscore
M Fscore
m Fscore
M Fscore
Reuters MUC7 NETagger Fox
(b) vs NERD
Figure 4: Comparative Evaluation.
512
256
128
64
32
16
Reuters MUC7 NETTagger FOX
2 5 8 11
msec
# Annotators
(a) WR
10000
1000
100
10
1
Reuters MUC7 NETTagger FOX
2 5 8 11
msec
# Annotators
(b) MEMM
Figure 5: Aggregation Performance.
integration. Figure 5 plots the (per-document) average computation
time for both WR and MEMM w.r.t. an increasing number of anno-tators
(2 to 11) and for different corpora. WR is substantially more
performant than MEMM. An optimal solution for a single document
can be produced on the order of 300 milliseconds in the worst-case.
In addition, WR’s performance is directly correlated with the num-ber
of annotated spans to be processed, and that is considerably
higher in the MUC7, NETagger and Reuters corpora than in the
Fox corpus. This is different from MEMM, where the time to com-pute
an optimal solution is affected by the number of concepts to
be taken into account and on which MEMM has been trained. On the
Reuters corpus, solutions can be computed on the order of 1-2 sec-onds
while on the other corpora, when fewer concepts are involved,
is much lower and MEMM’s performance is comparable with WR.
Training MEMM takes between 68 msec (2 annotators and 3 con-cepts
on the Fox corpus) and 2 mins (11 annotators and 215 con-cepts
on the Reuters corpus). Overall, the time required to produce
an optimal solution with both MEMM (resp. WR) is dominated by
at least one (resp. two) orders of magnitude by the time required
to send the various requests to the online annotation services and
collect the results. As an example, EXTRACTIV takes about 60 sec-onds
to annotate a medium-sized (i.e., 3-4 Kb) document. Also,
both MEMM and WR have negligible memory consumption when
compared with the size of the raw annotations.
Summary. The finding of our experimental evaluations can be
summarized as follows:
1. Aggregating opinions of individual annotators is always ben-eficial
in terms of recall. In terms of overall accuracy, aggre-gation
shows consistent benefit except few cases where the
majority of annotators contribute wrong annotations.
2. Although supervised aggregation delivers overall higher ac-curacy,
unsupervised aggregation is to be preferred when no
or sparse training data is available.
3. Both our aggregation techniques consistently outperform ex-isting
annotation aggregation approaches.
4. Aggregation does not impact the scalability of the perfor-mance
of annotation tasks, since it is always dominated by
the annotation time of the individual annotators.
5. RELATED WORK
Our work is rooted in information extraction, but uses techniques
from data integration, inconsistency-tolerant reasoning, and truth-finding:
we review a few of the most-related works below.
Named Entity Recognition. Research on Named Entity Recog-nition
(NER) systems can be currently grouped into three strands,
according to the sources of information that are leveraged in order
to produce textual annotations. Language-driven approaches (e.g.
KnowItAll [17]) represent the tradition in NER and exploit knowledge
about the natural language to derive linguistic patterns – typically in
the form of rules – that, once applied to the text, extract instances
of the entities of interest. Knowledge-driven systems (e.g. Learn-ingPinocchio
[13], PANKOW [12], KIM [27]) use unstructured, i.e.,
1495
11. gazetted, or structured, i.e., ontological, background knowledge to
locate the entities of interest in a text document.
Combining annotators. Many previous approaches addressing
the integration of multiple NER systems have been based on varia-tions
of voting mechanisms [28]. As also noticed in [35], the reli-ability
of voting-based aggregation is strictly connected to the type
and the number of the annotators. In particular, annotators have of-ten
a very limited coverage of concepts that are claimed to be recog-nised
and this affects the performance of voting mechanisms that do
not assume prior-knowledge about single annotators. An obvious
way of overcoming this problem is biasing the voting mechanism
by assigning different weights to the annotators that somehow re-flect
the confidence in its annotations. In [35] the weights are de-termined
using an exponential model borrowed from meta infor-mation
retrieval systems [7]. A major problem with this model is
that weights are determined on a per-annotator basis without con-sidering
that an annotator might have very different performance
on different concepts, as also proven by our experimental evalua-tion.
Our unsupervised approach makes use of the ontology in vot-ing,
thus distinguishing the annotator-concept pairs. Our bootstrap-ping
procedure described in Section 3 also works at a fine granu-larity,
producing weights for each annotator-concept pair, ensuring
that these peculiarities are considered in the computation. Biasing
is also often complemented with thresholding [26], that considers
only annotations with a minimum of support. As shown by our ex-perimental
evaluation, there are entities that can be correctly typed
by a single annotator and thresholding can lead to a noticeable drop
in recall. An example of this is the concept CommercialOrg that can
only be recognised by Extractiv. Our repair-based approach, in con-trast,
does not need to consider any arbitrary threshold.
A different way of combining annotators is to use machine learn-ing
techniques to determine the optimal combination of annotators.
The most popular techniques are neural networks [1], support vec-tor
machines [16], classifier stacking [38, 39, 20], and conditional
random fields (CRFs) [35]. The underlying idea of all these ap-proaches
is to combine classifiers in a meaningful way to obtain
a composite model. Being generic, these techniques often fail to
take into account specific information about the semantics of the
text. None of the approaches above uses ontological knowledge
to determine logical conflicts caused by the aggregation. A recent
approach, part of the NERD project, proved that background knowl-edge
[32] – that already proved beneficial for generating the anno-tations
– can also help in the integration of multiple extractors by
locating conflicting and logically incoherent annotations. However,
the NERD ontology does not contain disjointness constraints and
therefore these conflicts cannot be automatically detected. More-over,
we are aware that NERD uses machine learning techniques
for annotation integration as reported in [37], but we do not know
which technique is applied when invoking the NERD web service.
Our approach explicitly uses ontological knowledge to locate and
discard logically inconsistent opinions from the individual annota-tors.
To the best of our knowledge, our approach is also the first
one considering MEMM for semantic annotator aggregation. In fu-ture
work we are considering how other machine learning meth-ods,
such as CRFs, can be adapted to take into account ontological
rules. There is a wide range of variants for both CRFs and Max-imal
Entropy Markov models (e.g. in the objective function used
in training [25]), and so the trade-off between them, even in the
absence of semantic considerations, is quite complex. In Section
4, we compared our approach experimentally to the main learning-based
approach which was freely available, FOX [1]. We could not
find an implementation of [20], which won a CoNLL competition,
but we note that it performed experiments on a Reuters corpus sim-ilar
to the one we tested on, reporting F-scores that are very similar
to the macro- and micro- F-scores for both of our approaches.
In either the unsupervised or supervised setting, it is possible to
adopt more sophisticated techniques when deep knowledge about
the annotators is available. On the learning side, Michelakis et.
al. [30] look at the combination of the annotation rules adopted
by each annotator. As in MEMM, it makes use of a maximal en-tropy
classifier, but with “evidence vectors” being the output of
rule-based annotators. Unlike either of our approaches, semantic
relationships within the vocabulary of annotations are not consid-ered.
Our unsupervised method is close in spirit to classic judge-ment
aggregation techniques [24], where the final solution captures
judgements that are at a minimum aggregate-distance from those of
individual experts. However, the traditional judgement aggregation
setting assumes a “white-box” model where the methods used by
individual experts are transparent to the aggregator.
Relation to other areas. In a different context – that of web-pages
asserting facts about entities – work has been done on trying
to judge the trust and dependence of sources [15, 21] and produce
a truth-finding model. For example, iterative algorithms within the
PageRank/Hits family are applied in Galland et. al. [21] to de-termine
trustworthiness. Here trust/reliability/dependence is natu-rally
attributed globally to a web source, since the assertions being
judged are not assumed to belong to a hierarchy of types. We see
all these techniques as complementary to our aggregation methods
since they exploit redundancy across different web sources to de-termine
true assertions. It is possible, in principle, to apply these
techniques on top of our aggregation algorithm to further validate
the annotations across documents deemed to refer to the same facts,
e.g., news from different websites describing the same event.
Another interesting setting is the one of SOFIE [36] where
common-knowledge rules are applied on top of automatically-generated
ontological assertions to determine, discard or repair log-ically
inconsistent facts in the ontology. Again, these techniques
are complementary to aggregation when additional knowledge is
available from the annotation sources. In particular, common-knowledge
rules can be used to improve the recognition of nested
annotations, like those in Figure 1, by restricting the valid combi-nations
of concepts that can appear in a containment relationship,
e.g., the name of a university usually contains also the name of the
city in which it is located, while illness names can embed the name
of the scientist discovering them.
Our repair-based aggregation is close in spirit to consistent rea-soning
for inconsistent ontologies [33]. However, such techniques
cannot be straightforwardly applied to our setting where inconsis-tent
assertions can be stated multiple times and the logical consis-tency
also relates to the way annotations are contained into each
other. We consider scoring of repairs due to the support of the facts
they remove. Scoring of repairs also arises for numerical integrity
constraints [9, 19]. Unlike in our case repairs can only have non-negative
scores, and this impacts the technical development of the
solution dramatically.
6. CONCLUSIONS FUTURE WORK
We have presented a set of approaches for knowledge-aware in-tegration
of entity extractors that consider each entity extractor as
a black box and do not assume any prior knowledge about their
performance or competence w.r.t. a certain set of entities.
These approaches have been proposed as alternatives, depend-ing
on the application scenario. But of course it will often occur
that one approach or the other is more appropriate on a per-concept
basis. Thus we are currently investigating how to combine these
methods. We currently produce an aggregate annotation without
1496
12. distinguishing whether an annotation has strong or weak support
for correctness, but each of our aggregators can be adapted to give
a measure of confidence. For example, for MEMM we can return
the probability of the given annotation at each span. This is con-ceptually
straightforward, since it is a particular kind of query on
the probabilistic annotation generated by the Markov sequence. We
are currently investigating a general approach to efficiently query-ing
Markov sequences, tuned to the case of annotations. We are
also preparing a study of the adaptation of other supervised learn-ing
methods, such as conditional random fields, for the annota-tion
reconciliation setting, along with a comparison of the resulting
adapted method to MEMM.
Another interesting direction is to progressively reduce the de-pendency
on the annotator’s documentation when creating the map-ping
between the recognized concept classes and the merged ontol-ogy.
This can be addressed either by schema/ontology matching
techniques or by ontology learning approaches. For now, the refer-ence
implementation of WR and MEMM called ROSeAnn [11] simply
reports newly discovered classes, leaving to the programmer the
task of updating the mapping to the merged ontology.
We have focused on concept annotation here, since it is the most
widely-supported function. Our methods can be extended to do
entity reconciliation jointly with reconciliation of entity instance
disambiguation – e.g. by considering instances as bottom-level el-ements
of an ontology. Again, ROSeAnn simply reports the disam-biguation
information provided by the annotators. A more difficult
task is relation extraction, which is also emerging in service-based
annotators: it currently has some support from OpenCalais and Ex-tractiv.
This is important from the point of view of information ex-traction,
since it identifies part-whole relationships that are needed
to piece together a tuple or object. Our focus on concept annotation
allowed us to use a very simple kind of reasoning – based on sub-sumption
and disjointness. More costly reasoning would be needed
for relations, since the ontology may have richer constraints (e.g.
foreign keys, uniqueness constraints) available at the relationship
level.
7. REFERENCES
[1] FOX. http://ontowiki.net/Projects/FOX?v=4e5.
[2] LingPipe. http://alias-i.com/lingpipe/.
[3] MUC7. http://www.ldc.upenn.edu/Catalog/catalogEntry.
jsp?catalogId=LDC2001T02.
[4] OpenNLP Tools. http://opennlp.apache.org/index.html.
[5] Reuters. http://about.reuters.com/researchandstandards/
corpus/index.asp.
[6] ROSeAnn. http://diadem.cs.ox.ac.uk/roseann.
[7] J. A. Aslam and M. Montague. Models for metasearch. In SIGIR,
pages 276–284, 2001.
[8] O. Bender, F. J. Och, and H. Ney. Maximum entropy models for
named entity recognition. In CoNLL, pages 148–151, 2003.
[9] L. Bertossi, L. Bravo, E. Franconi, and A. Lopatenko. The
complexity and approximation of fixing numerical attributes in
databases under integrity constraints. Inf. Sys., 33:407–434, 2008.
[10] R. Carreira, S. Carneiro, R. Pereira, M. Rocha, I. Rocha, E. Ferreira,
and A. Lourenc¸o. Semantic annotation of biological concepts
interplaying microbial cellular responses. BMC Bioinformatics,
12:1–10, 2011.
[11] L. Chen, S. Ortona, G. Orsi, and M. Benedikt. ROSeAnn:
Reconciling opinions of semantic annotators. PVLDB, To appear
2013.
[12] P. Cimiano, S. Handschuh, and S. Staab. Towards the self-annotating
web. In WWW, pages 462–471, 2004.
[13] F. Ciravegna and A. Lavelli. LearningPinocchio: Adaptive
information extraction for real world applications. Nat. Lang. Eng.,
10(2):145–165, 2004.
[14] N. Dalvi, R. Kumar, and M. Soliman. Automatic wrappers for large
scale web extraction. PVLDB, 4(4):219–230, 2011.
[15] X. L. Dong, L. Berti-Equille, and D. Srivastava. Integrating
conflicting data: the role of source dependence. PVLDB,
2(1):550–561, 2009.
[16] D. Duong, J. Venuto, B. Goertzel, R. Richardson, S. Bohner, and
E. Fox. Support vector machines to weight voters in a voting system
of entity extractors. In IJCNN, pages 1226–1230, 2006.
[17] O. Etzioni, M. Cafarella, D. Downey, S. Kok, A. M. Popescu,
T. Shaked, S. Soderland, D. S. Weld, and A. Yates. Web-scale
information extraction in KnowItAll. In WWW, pages 100–110, 2004.
[18] J. Euzenat. Semantic precision and recall for ontology alignment
evaluation. In IJCAI, pages 348–353, 2007.
[19] S. Flesca, F. Furfaro, and F. Parisi. Querying and repairing
inconsistent numerical databases. ACM Trans. Database Syst.,
35(2):14:1–14:50, May 2010.
[20] R. Florian, A. Ittycheriah, H. Jing, and T. Zhang. Named entity
recognition through classifier combination. In CoNLL, pages
168–171, 2003.
[21] A. Galland, S. Abiteboul, A. Marian, and P. Senellart. Corroborating
information from disagreeing views. In WSDM, pages 131–140,
2010.
[22] D. Grossi and G. Pigozzi. Introduction to judgment aggregation. In
Lectures on Logic and Computation, pages 160–209. 2012.
[23] H. Cunningham et. al. Text Processing with GATE (Version 6). U.
Sheffield Dept. of CS, 2011.
[24] S. Hartmann, G. Pigozzi, and J. Sprenger. Reliable methods of
judgement aggregation. J. Log. Comp., 20(2):603–617, 2010.
[25] S. Kakade, Y. W. Teh, and S. T. Roweis. An alternate objective
function for markovian fields. In ICML, pages 275–282, 2002.
[26] N. Kambhatla. Minority vote: at-least-n voting improves recall for
extracting relations. In COLING, pages 460–466, 2006.
[27] A. Kiryakov, B. Popov, I. Terziev, D. Manov, and D. Ognyanoff.
Semantic annotation, indexing, and retrieval. Web Semantics:
Science, Services and Agents on the World Wide Web, 2(1):49–79,
2004.
[28] Z. Kozareva, O. Ferr´andez, A. Montoyo, R. Mu˜noz, A. Su´arez, and
J. G´omez. Combining data-driven systems for improving named
entity recognition. Data Knowl. Eng., 61(3):449–466, 2007.
[29] A. McCallum, D. Freitag, and F. C. Pereira. Maximum entropy
markov models for information extraction and segmentation. In
ICML, pages 591–598, 2000.
[30] E. Michelakis, R. Krishnamurthy, P. J. Haas, and S. Vaithyanathan.
Uncertainty management in rule-based information extraction
systems. In SIGMOD, pages 101–114, 2009.
[31] L. Ratinov and D. Roth. Design challenges and misconceptions in
named entity recognition. In CoNLL, pages 147–155, 2009.
[32] G. Rizzo and R. Troncy. Nerd: A framework for unifying named
entity recognition and disambiguation extraction tools. In EACL,
pages 73–76, 2012.
[33] R. Rosati. On the complexity of dealing with inconsistency in
description logic ontologies. In IJCAI, pages 1057–1062, 2011.
[34] P. Senellart, A. Mittal, D. Muschick, R. Gilleron, and M. Tommasi.
Automatic wrapper induction from hidden-web sources with domain
knowledge. In WIDM, pages 9–16, 2008.
[35] L. Si, T. Kanungo, and X. Huang. Boosting performance of bio-entity
recognition by combining results from multiple systems. In
BIOKDD, pages 76–83, 2005.
[36] F. M. Suchanek, M. Sozio, and G. Weikum. SOFIE: A
self-organizing framework for information extraction. In WWW,
pages 631–640, 2009.
[37] M. van Erp, G. Rizzo, and R. Troncy. Learning with the Web:
Spotting named entities on the intersection of NERD and machine
learning. In MSM, 2013.
[38] H. Wang and T. Zhao. Identifying named entities in biomedical text
based on stacked generalization. In WCICA, pages 160–164, 2008.
[39] D. Wu, G. Ngai, and M. Carpuat. A stacked, voted, stacked model for
named entity recognition. In CoNLL, pages 200–203, 2003.
1497