KeyNote talk Given at the DanTermBank workshop on Januaray,9th 2015.
http://dantermbank.cbs.dk/dtb_uk/the_dantermbank_project_launches_a_new_website/dantermbank_workshop_revealing_hidden_knowledge_9_january_2015
The document introduces ontology and describes what it is from both philosophical and computer science perspectives. An ontology in computers consists of a vocabulary to describe a domain, specifications of the meaning of terms, and constraints capturing additional knowledge about the domain. It then provides an example ontology and discusses applications of ontologies such as for the semantic web. It also discusses important considerations for building ontologies such as collaboration, versioning, and ease of use.
SPARQL is a semantic query language used to retrieve and manipulate data stored in RDF format. An ontology represents concepts within a domain and provides specific meanings of terms within that domain, such as modeling playing cards within a poker ontology. While ontologies are similar to object-oriented class hierarchies, ontologies are meant to evolve constantly to represent diverse internet data, whereas class hierarchies evolve slowly from structured corporate databases. The Protege tool can be used to create domain-specific ontologies and publish them on the web with the .owl extension to then run SPARQL queries to retrieve information.
This document discusses using databases and SQL to store and organize text data. It explains that arrays in PHP can be used to represent text as data structures like tables and trees, but databases provide more efficient storage and retrieval. Specifically, relational databases use SQL, which allows defining schemas to represent ontologies and then querying the data through logical operations. The document introduces MySQL as an open source relational database and phpMyAdmin as a PHP interface for managing MySQL databases.
Ontologies provide a shared understanding of a domain by formally defining concepts, properties, and relationships. An ontology introduces vocabulary relevant to a domain and specifies the meaning of terms. Ontologies are machine-readable and enable overcoming differences in terminology across complex, distributed applications. Examples include gene ontologies, pharmaceutical drug ontologies, and customer profile ontologies. Semantic technologies use ontologies to provide semantic search, integration, reasoning, and analysis capabilities.
This document summarizes an OKFN Korea hackathon event focused on open data. It discusses modeling Seoul open government data using ontologies, linking it to external datasets like cultural heritage data, and publishing the enriched data in RDF format. It covers topics like linked data, modeling with RDF/RDFS/OWL, reusing existing vocabularies, ontology development best practices, and triple store storage solutions.
Ontology development in protégé-آنتولوژی در پروتوغهsadegh salehi
This document describes an agenda for an ontology development presentation in Protégé. It discusses the syntactic web and its limitations, as well as the promise of the semantic web to address these issues by adding meaning to web content that is understandable to machines. It outlines two sessions on ontology and OWL basics, Protégé, and developing a pizza ontology in Protégé.
The document introduces ontology and describes what it is from both philosophical and computer science perspectives. An ontology in computers consists of a vocabulary to describe a domain, specifications of the meaning of terms, and constraints capturing additional knowledge about the domain. It then provides an example ontology and discusses applications of ontologies such as for the semantic web. It also discusses important considerations for building ontologies such as collaboration, versioning, and ease of use.
SPARQL is a semantic query language used to retrieve and manipulate data stored in RDF format. An ontology represents concepts within a domain and provides specific meanings of terms within that domain, such as modeling playing cards within a poker ontology. While ontologies are similar to object-oriented class hierarchies, ontologies are meant to evolve constantly to represent diverse internet data, whereas class hierarchies evolve slowly from structured corporate databases. The Protege tool can be used to create domain-specific ontologies and publish them on the web with the .owl extension to then run SPARQL queries to retrieve information.
This document discusses using databases and SQL to store and organize text data. It explains that arrays in PHP can be used to represent text as data structures like tables and trees, but databases provide more efficient storage and retrieval. Specifically, relational databases use SQL, which allows defining schemas to represent ontologies and then querying the data through logical operations. The document introduces MySQL as an open source relational database and phpMyAdmin as a PHP interface for managing MySQL databases.
Ontologies provide a shared understanding of a domain by formally defining concepts, properties, and relationships. An ontology introduces vocabulary relevant to a domain and specifies the meaning of terms. Ontologies are machine-readable and enable overcoming differences in terminology across complex, distributed applications. Examples include gene ontologies, pharmaceutical drug ontologies, and customer profile ontologies. Semantic technologies use ontologies to provide semantic search, integration, reasoning, and analysis capabilities.
This document summarizes an OKFN Korea hackathon event focused on open data. It discusses modeling Seoul open government data using ontologies, linking it to external datasets like cultural heritage data, and publishing the enriched data in RDF format. It covers topics like linked data, modeling with RDF/RDFS/OWL, reusing existing vocabularies, ontology development best practices, and triple store storage solutions.
Ontology development in protégé-آنتولوژی در پروتوغهsadegh salehi
This document describes an agenda for an ontology development presentation in Protégé. It discusses the syntactic web and its limitations, as well as the promise of the semantic web to address these issues by adding meaning to web content that is understandable to machines. It outlines two sessions on ontology and OWL basics, Protégé, and developing a pizza ontology in Protégé.
a system called natural language interface which transforms user's natural language question into SPARQL query
find related papers here https://sites.google.com/site/fadhlinams81/publication
The document describes a project to publish mathematics lecture notes as linked data. Key points:
1) Lecture notes containing 2,000 slides and 1,000 homework problems were semantically annotated and converted to RDF to create structured data.
2) The RDF is stored in a triplestore and can be queried with an OMDoc-aware SPARQL endpoint or full-text search.
3) Annotations in the human-readable XHTML documents link to services for interactivity. The goal is to scale this to 300,000 annotated publications and link to external datasets.
Neno/Fhat: Semantic Network Programming Language and Virtual Machine Specific...Marko Rodriguez
• The Semantic Web is a distributed, flexible modeling framework.
• The Semantic Web is primarily descriptive in nature. The Semantic Web is used to describe web-pages, services, systems, etc.
• Neno is an object-oriented language that was designed specifically for the Semantic Web.
• Fhat is a virtual machine represented in the Semantic Web.
• With Neno/Fhat the Semantic Web now has a procedural component. The Semantic Web now includes object methods, algorithms, and computing machines.
• The Semantic Web can be made to behave like a distributed, general-purpose computer. Not just an information repository.
This document provides an overview of taxonomy, ontology, folksonomies, and SKOS (Simple Knowledge Organization Systems). It defines each concept and provides examples. Taxonomy is described as a subject-based classification system. Ontology is defined as a formal specification of concepts and relationships. Folksonomies allow user-generated tagging. SKOS provides a standard for sharing and linking knowledge organization systems on the web. Bibliographies with relevant references are also included for each topic.
The document discusses ontologies, vocabularies, and semantic web technologies. It provides an overview of RDF, RDF Schema, and OWL, including their semantics and capabilities. It describes how ontologies can constrain models and enable reasoning to derive inferences from class definitions and axioms. The document also addresses some common misconceptions regarding ontology modeling concepts.
The document discusses ontology engineering and provides details about:
1. Ontology engineering is the process of developing ontologies for a particular domain by defining concepts, arranging them hierarchically, and defining their properties and relationships.
2. Ontology engineering is analogous to object-oriented database design but ontologies reflect the structure of the world using open world assumptions.
3. Popular ontology engineering tools include Protégé, which supports ontology development and knowledge modeling.
The document discusses approaches to representing terminology in the Semantic Web. It proposes a semiotic or sign-based view where terms are treated as first-class citizens along with concepts and real-world referents. Current models like SKOS are described as either too concept-centric or lacking context. The document suggests introducing "meaning" resources to explicitly capture the context and possible senses of a term, addressing limitations of existing approaches.
Modelling Knowledge Organization Systems and StructuresMarcia Zeng
In this paper FRSAD (as a conceptual model) is compared to SKOS and SKOS XL (as data models), with implementation examples. ISKO-UK 2011 Conference, London, July 2011.
This document discusses modelling and representing social network data ontologically. It covers representing social individuals and relationships ontologically, as well as aggregating and reasoning with social network data. It discusses ontology languages like RDF, OWL, and FOAF that can be used to represent social network data and individuals semantically. It also talks about state-of-the-art approaches for representing network structure and attribute data, and the need for representations that can integrate different data sources and maintain identity.
Tutorial at OAI5 (cern.ch/oai5). Abstract: This tutorial will provide a practical overview of current practices in modelling complex or compound digital objects. It will examine some of the key scenarios around creating complex objects and will explore a number of approaches to packaging and transport. Taking research papers, or scholarly works, as an example, the tutorial will explore the different ways in which these, and their descriptive metadata, can be treated as complex objects. Relevant application profiles and metadata formats will be introduced and compared, such as Dublin Core, in particular the DCMI Abstract Model, and MODS, alongside content packaging standards, such as METS MPEG 21 DIDL and IMS CP. Finally, we will consider some future issues and activities that are seeking to address these. The tutorial will be of interest to librarians and technical staff with an interest in metadata or complex objects, their creation, management and re-use.
Bernhard Haslhofer is a postdoc researcher at Cornell University studying linked data, user-contributed data, and data interoperability. He discusses Linked (Open) Data, which uses URIs and RDF to publish and link structured data on the web. The key principles are using URIs to identify things, providing useful information about those URIs when dereferenced, and including links to other URIs. Enabling technologies include URIs, RDF, RDFS/OWL for vocabularies, SPARQL for querying, and best practices for publishing vocabularies and data. Useful tools are also presented.
Mining and Supporting Community Structures in Sensor Network ResearchMarko Rodriguez
The document discusses mining and supporting community structures in sensor network research. It summarizes a study that analyzed the co-authorship network of researchers at the Center for Embedded Network Sensing (CENS) to determine if structural communities detected in the network are independent of socio-academic communities like academic department or affiliation. The study found that structural communities correspond more to department and affiliation, while academic position and country of origin are independent of structural communities.
Gathering Lexical Linked Data and Knowledge Patterns from FrameNetAndrea Nuzzolese
The document discusses transforming FrameNet, a lexical knowledge base, into Linked Open Data (LOD) and knowledge patterns. It presents several semantic issues with representing linguistic resources and proposes a two-step method using Semion to address these issues. The method first syntactically transforms FrameNet data into RDF triples, then applies a rule-based refactoring to add semantics. Ongoing work includes linking FrameNet to other LOD resources like WordNet and VerbNet. The transformation aims to publish FrameNet as a LOD dataset and convert its data into reusable knowledge patterns.
Franz Et Al - Concepts and Tools Needed to Increase Bottom-Up Taxonomic Exper...taxonbytes
We discuss the perceived requirements – conceptual, technical, and social – for the creation of a “Taxonomic Clearing House” (TCH) that will enfranchise and enhance contributions by individual taxonomic experts and collaboratives in a global, names-based infrastructure. In terms of scale, such an infrastructure must be suited to assemble, retrieve, and editing contemporary taxonomic and phylogenetic classifications that involve some 22 million name strings representing 2.3 million living and extinct species; and serve diverse contributor and user communities including 6-40 thousand experts, 400,000 biologists, and more than 100 million citizen scientists. Existing classification synthesis platforms fall short of this grand challenge because they (1) may be limited to living or fossil taxa, (2) fail to show alternative points of view or (3) integrate molecularly-defined entities (“dark taxa”), (4) do not automatically monitor new data, (5) lack scalable solutions for providing feedback and credit, (6) have slow revisionary processes, (7) lack effective machine-to-machine services, or (8) cannot represent finer-grained insights such as evolving taxonomic concepts. Jointly these factors can produce a disconnect of the expert community that leads the global, piece-meal process of advancing classifications from large-scale platforms that purport to represent and unify their individual contributions. A suitable TCH should counteract this by acting as an open communal environment allowing expert contributors to jointly assemble and edit evolving taxonomic and phylogenetic content leading to large-scale classifications. In particular, it must (1) engage major collaborating taxonomic ad phylogenetic initiatives and facilitate diverse information flow; (2) expand information acquisition capabilities to harvest names and classifications from diverse sources; (3) create a powerful interface for taxonomic editing, including a topology assembly and visualization layer, nomenclatural and taxonomic editing layers, a Filtered Push-based service (http://wiki.filteredpush.org/wiki/) for submitting, tracking and accrediting edits to expert contributors, and taxonomically intelligent alerts; and (4) leverage these efforts towards a “Union” reference classification holding two million taxa and multiple alternative perspectives as indicated. To promote the engagement and acceptance, a TCH should target existing expert communities such as contributor to the Symbiota collections or TimeTree phylogenetics platforms. The presentation will both introduce the elements of this TCH vision and assess their merits and current progress and challenges towards realization.
Franz. 2014. Explaining taxonomy's legacy to computers – how and why?taxonbytes
Slides presented on the Euler/X projected (http://taxonbytes.org/prior-work-on-concept-taxonomy-2013/ & https://bitbucket.org/eulerx/euler-project) - for the conference "The Meaning of Names: Naming Diversity in the 21st Century", CU Natural History Museum, September 30, 2014.
Extending models for controlled vocabularies to classification systems: model...Marcia Zeng
Mitchell, Joan S., Marcia Lei Zeng, and Maja Zumer. Presented at the International UDC Seminar 2011, Classification & Ontology, The Hague, The Netherlands, Sept. 19-20, 2011.
Franz Et Al. Using ASP to Simulate the Interplay of Taxonomic and Nomenclatur...taxonbytes
Answer Set Programming (ASP) is a declarative, stable model approach to logic programming with an under-realized potential for representing and reasoning over biological information. ASP is particularly suited to address reasoning challenges with complex starting conditions and rule sets. One such challenge is the interplay of taxonomic and nomenclatural change in biological taxonomy that often results when a taxonomy is revised based on a previously published perspective. Depending on the nature of the taxonomic changes to be undertaken, one or more Code-mandated principles will apply to regulate specific and concomitant name changes. In the case of the International Code of Zoological Nomenclature, two principles of significance include the Principles of Priority and Typification. Although the relationship between the number of taxonomic and nomenclatural adjustments under a given transition scenario is not linear, the application of the name-changing rules is usually unambiguous and therefore amenable to logic representation. Here we explore the modeling of the taxonomy/nomenclature interplay in ASP with a simple, abstract nine-taxon use case that contains four terminal species of which two are type-bearers for their respective genera. Four distinct one-taxon transfer scenarios are simulated through a transition system approach, requiring 1-7 concomitant nomenclatural changes depending (1) on the priority relationships among the terminal taxa being repositioned and (2) the type-bearing name dependencies of their higher-level parents. ASP can simulate these rules faithfully and thus reason over situations that range from a one-to-one match of taxonomic and nomenclatural changes to situations where they two kinds of change become increasingly disconnected (e.g., transfer of non-type genera among tribes without name change, or "transfer" [in reverse direction] of a single priority-carrying name/taxon into a larger yet junior entity with numerous required name changes). Our results, though very preliminary, illustrate how ASP logic approach may be utilized to perform optimizations at the taxonomy/nomenclature intersection, and generally represent a novel step towards translating Code-mandated naming rules into logic, with potential benefits for virtual taxonomic domains.
The document discusses the relationship between research and teaching in library and information science (LIS) curricula. It argues that research is essential to LIS teaching and that teaching should focus on asking questions rather than just transmitting established knowledge. It also notes the need for LIS terminology to change as the field evolves from focusing on catalogs to linked open data and graphs.
folksonomy, social tagging, tag clouds, automatic folksonomy construction, word clouds, wordle,context-preserving word cloud visualisation, CPEWCV, seam carving, inflate and push, star forest, cycle cover, quantitative metrics, realized adjacencies, distortion, area utilization, compactness, aspect ratio, running time, semantics in language technology
Tutorial: Building and using ontologies - E.Simperl - ESWC SS 2014eswcsummerschool
This document discusses building and using ontologies. It defines an ontology as defining a domain of interest in terms of things, attributes, and relationships. Ontologies are used to share a common understanding of a domain among people and machines. The document then discusses ontology engineering processes, examples of ontologies like DBpedia, and semantic technologies for creating intelligent applications.
This document discusses building and using ontologies. It defines an ontology as defining a domain of interest in terms of things, attributes, and relationships. Ontologies are used to share a common understanding of a domain among people and machines. The document then discusses ontology engineering processes, examples of ontologies like DBpedia, and semantic technologies used to create intelligent applications.
a system called natural language interface which transforms user's natural language question into SPARQL query
find related papers here https://sites.google.com/site/fadhlinams81/publication
The document describes a project to publish mathematics lecture notes as linked data. Key points:
1) Lecture notes containing 2,000 slides and 1,000 homework problems were semantically annotated and converted to RDF to create structured data.
2) The RDF is stored in a triplestore and can be queried with an OMDoc-aware SPARQL endpoint or full-text search.
3) Annotations in the human-readable XHTML documents link to services for interactivity. The goal is to scale this to 300,000 annotated publications and link to external datasets.
Neno/Fhat: Semantic Network Programming Language and Virtual Machine Specific...Marko Rodriguez
• The Semantic Web is a distributed, flexible modeling framework.
• The Semantic Web is primarily descriptive in nature. The Semantic Web is used to describe web-pages, services, systems, etc.
• Neno is an object-oriented language that was designed specifically for the Semantic Web.
• Fhat is a virtual machine represented in the Semantic Web.
• With Neno/Fhat the Semantic Web now has a procedural component. The Semantic Web now includes object methods, algorithms, and computing machines.
• The Semantic Web can be made to behave like a distributed, general-purpose computer. Not just an information repository.
This document provides an overview of taxonomy, ontology, folksonomies, and SKOS (Simple Knowledge Organization Systems). It defines each concept and provides examples. Taxonomy is described as a subject-based classification system. Ontology is defined as a formal specification of concepts and relationships. Folksonomies allow user-generated tagging. SKOS provides a standard for sharing and linking knowledge organization systems on the web. Bibliographies with relevant references are also included for each topic.
The document discusses ontologies, vocabularies, and semantic web technologies. It provides an overview of RDF, RDF Schema, and OWL, including their semantics and capabilities. It describes how ontologies can constrain models and enable reasoning to derive inferences from class definitions and axioms. The document also addresses some common misconceptions regarding ontology modeling concepts.
The document discusses ontology engineering and provides details about:
1. Ontology engineering is the process of developing ontologies for a particular domain by defining concepts, arranging them hierarchically, and defining their properties and relationships.
2. Ontology engineering is analogous to object-oriented database design but ontologies reflect the structure of the world using open world assumptions.
3. Popular ontology engineering tools include Protégé, which supports ontology development and knowledge modeling.
The document discusses approaches to representing terminology in the Semantic Web. It proposes a semiotic or sign-based view where terms are treated as first-class citizens along with concepts and real-world referents. Current models like SKOS are described as either too concept-centric or lacking context. The document suggests introducing "meaning" resources to explicitly capture the context and possible senses of a term, addressing limitations of existing approaches.
Modelling Knowledge Organization Systems and StructuresMarcia Zeng
In this paper FRSAD (as a conceptual model) is compared to SKOS and SKOS XL (as data models), with implementation examples. ISKO-UK 2011 Conference, London, July 2011.
This document discusses modelling and representing social network data ontologically. It covers representing social individuals and relationships ontologically, as well as aggregating and reasoning with social network data. It discusses ontology languages like RDF, OWL, and FOAF that can be used to represent social network data and individuals semantically. It also talks about state-of-the-art approaches for representing network structure and attribute data, and the need for representations that can integrate different data sources and maintain identity.
Tutorial at OAI5 (cern.ch/oai5). Abstract: This tutorial will provide a practical overview of current practices in modelling complex or compound digital objects. It will examine some of the key scenarios around creating complex objects and will explore a number of approaches to packaging and transport. Taking research papers, or scholarly works, as an example, the tutorial will explore the different ways in which these, and their descriptive metadata, can be treated as complex objects. Relevant application profiles and metadata formats will be introduced and compared, such as Dublin Core, in particular the DCMI Abstract Model, and MODS, alongside content packaging standards, such as METS MPEG 21 DIDL and IMS CP. Finally, we will consider some future issues and activities that are seeking to address these. The tutorial will be of interest to librarians and technical staff with an interest in metadata or complex objects, their creation, management and re-use.
Bernhard Haslhofer is a postdoc researcher at Cornell University studying linked data, user-contributed data, and data interoperability. He discusses Linked (Open) Data, which uses URIs and RDF to publish and link structured data on the web. The key principles are using URIs to identify things, providing useful information about those URIs when dereferenced, and including links to other URIs. Enabling technologies include URIs, RDF, RDFS/OWL for vocabularies, SPARQL for querying, and best practices for publishing vocabularies and data. Useful tools are also presented.
Mining and Supporting Community Structures in Sensor Network ResearchMarko Rodriguez
The document discusses mining and supporting community structures in sensor network research. It summarizes a study that analyzed the co-authorship network of researchers at the Center for Embedded Network Sensing (CENS) to determine if structural communities detected in the network are independent of socio-academic communities like academic department or affiliation. The study found that structural communities correspond more to department and affiliation, while academic position and country of origin are independent of structural communities.
Gathering Lexical Linked Data and Knowledge Patterns from FrameNetAndrea Nuzzolese
The document discusses transforming FrameNet, a lexical knowledge base, into Linked Open Data (LOD) and knowledge patterns. It presents several semantic issues with representing linguistic resources and proposes a two-step method using Semion to address these issues. The method first syntactically transforms FrameNet data into RDF triples, then applies a rule-based refactoring to add semantics. Ongoing work includes linking FrameNet to other LOD resources like WordNet and VerbNet. The transformation aims to publish FrameNet as a LOD dataset and convert its data into reusable knowledge patterns.
Franz Et Al - Concepts and Tools Needed to Increase Bottom-Up Taxonomic Exper...taxonbytes
We discuss the perceived requirements – conceptual, technical, and social – for the creation of a “Taxonomic Clearing House” (TCH) that will enfranchise and enhance contributions by individual taxonomic experts and collaboratives in a global, names-based infrastructure. In terms of scale, such an infrastructure must be suited to assemble, retrieve, and editing contemporary taxonomic and phylogenetic classifications that involve some 22 million name strings representing 2.3 million living and extinct species; and serve diverse contributor and user communities including 6-40 thousand experts, 400,000 biologists, and more than 100 million citizen scientists. Existing classification synthesis platforms fall short of this grand challenge because they (1) may be limited to living or fossil taxa, (2) fail to show alternative points of view or (3) integrate molecularly-defined entities (“dark taxa”), (4) do not automatically monitor new data, (5) lack scalable solutions for providing feedback and credit, (6) have slow revisionary processes, (7) lack effective machine-to-machine services, or (8) cannot represent finer-grained insights such as evolving taxonomic concepts. Jointly these factors can produce a disconnect of the expert community that leads the global, piece-meal process of advancing classifications from large-scale platforms that purport to represent and unify their individual contributions. A suitable TCH should counteract this by acting as an open communal environment allowing expert contributors to jointly assemble and edit evolving taxonomic and phylogenetic content leading to large-scale classifications. In particular, it must (1) engage major collaborating taxonomic ad phylogenetic initiatives and facilitate diverse information flow; (2) expand information acquisition capabilities to harvest names and classifications from diverse sources; (3) create a powerful interface for taxonomic editing, including a topology assembly and visualization layer, nomenclatural and taxonomic editing layers, a Filtered Push-based service (http://wiki.filteredpush.org/wiki/) for submitting, tracking and accrediting edits to expert contributors, and taxonomically intelligent alerts; and (4) leverage these efforts towards a “Union” reference classification holding two million taxa and multiple alternative perspectives as indicated. To promote the engagement and acceptance, a TCH should target existing expert communities such as contributor to the Symbiota collections or TimeTree phylogenetics platforms. The presentation will both introduce the elements of this TCH vision and assess their merits and current progress and challenges towards realization.
Franz. 2014. Explaining taxonomy's legacy to computers – how and why?taxonbytes
Slides presented on the Euler/X projected (http://taxonbytes.org/prior-work-on-concept-taxonomy-2013/ & https://bitbucket.org/eulerx/euler-project) - for the conference "The Meaning of Names: Naming Diversity in the 21st Century", CU Natural History Museum, September 30, 2014.
Extending models for controlled vocabularies to classification systems: model...Marcia Zeng
Mitchell, Joan S., Marcia Lei Zeng, and Maja Zumer. Presented at the International UDC Seminar 2011, Classification & Ontology, The Hague, The Netherlands, Sept. 19-20, 2011.
Franz Et Al. Using ASP to Simulate the Interplay of Taxonomic and Nomenclatur...taxonbytes
Answer Set Programming (ASP) is a declarative, stable model approach to logic programming with an under-realized potential for representing and reasoning over biological information. ASP is particularly suited to address reasoning challenges with complex starting conditions and rule sets. One such challenge is the interplay of taxonomic and nomenclatural change in biological taxonomy that often results when a taxonomy is revised based on a previously published perspective. Depending on the nature of the taxonomic changes to be undertaken, one or more Code-mandated principles will apply to regulate specific and concomitant name changes. In the case of the International Code of Zoological Nomenclature, two principles of significance include the Principles of Priority and Typification. Although the relationship between the number of taxonomic and nomenclatural adjustments under a given transition scenario is not linear, the application of the name-changing rules is usually unambiguous and therefore amenable to logic representation. Here we explore the modeling of the taxonomy/nomenclature interplay in ASP with a simple, abstract nine-taxon use case that contains four terminal species of which two are type-bearers for their respective genera. Four distinct one-taxon transfer scenarios are simulated through a transition system approach, requiring 1-7 concomitant nomenclatural changes depending (1) on the priority relationships among the terminal taxa being repositioned and (2) the type-bearing name dependencies of their higher-level parents. ASP can simulate these rules faithfully and thus reason over situations that range from a one-to-one match of taxonomic and nomenclatural changes to situations where they two kinds of change become increasingly disconnected (e.g., transfer of non-type genera among tribes without name change, or "transfer" [in reverse direction] of a single priority-carrying name/taxon into a larger yet junior entity with numerous required name changes). Our results, though very preliminary, illustrate how ASP logic approach may be utilized to perform optimizations at the taxonomy/nomenclature intersection, and generally represent a novel step towards translating Code-mandated naming rules into logic, with potential benefits for virtual taxonomic domains.
The document discusses the relationship between research and teaching in library and information science (LIS) curricula. It argues that research is essential to LIS teaching and that teaching should focus on asking questions rather than just transmitting established knowledge. It also notes the need for LIS terminology to change as the field evolves from focusing on catalogs to linked open data and graphs.
folksonomy, social tagging, tag clouds, automatic folksonomy construction, word clouds, wordle,context-preserving word cloud visualisation, CPEWCV, seam carving, inflate and push, star forest, cycle cover, quantitative metrics, realized adjacencies, distortion, area utilization, compactness, aspect ratio, running time, semantics in language technology
Tutorial: Building and using ontologies - E.Simperl - ESWC SS 2014eswcsummerschool
This document discusses building and using ontologies. It defines an ontology as defining a domain of interest in terms of things, attributes, and relationships. Ontologies are used to share a common understanding of a domain among people and machines. The document then discusses ontology engineering processes, examples of ontologies like DBpedia, and semantic technologies for creating intelligent applications.
This document discusses building and using ontologies. It defines an ontology as defining a domain of interest in terms of things, attributes, and relationships. Ontologies are used to share a common understanding of a domain among people and machines. The document then discusses ontology engineering processes, examples of ontologies like DBpedia, and semantic technologies used to create intelligent applications.
The document provides an overview of a tutorial on semantic digital libraries. It introduces the speakers and schedule, which includes an introduction to semantic digital libraries and existing solutions, followed by discussions on conclusions and future directions. It also briefly covers the semantic web, ontologies, RDF, and how these technologies can help digital libraries by making metadata machine-understandable.
Knowledge Organisation Systems in Digital Libraries: A Comparative StudyBhojaraju Gunjal
The document presents a study that compares the different Knowledge Organization Systems (KOS) used in major digital libraries. It finds that while traditional libraries used standardized systems like classification schemes, digital libraries employ various KOS tools including thesauri, ontologies, and subject headings. The study analyzes the specific KOS used in different digital libraries and summarizes the current state of KOS in these libraries.
UMBEL: Subject Concepts Layer for the WebMike Bergman
UMBEL is a lightweight ontology and subject concept framework comprised of around 20,000 concepts and their relationships that aims to provide context for web content and datasets. It serves as a reference structure for placing information into context with other data by defining common subject concepts and mapping entities and datasets to these concepts. UMBEL is freely available under an open source license and relies on existing vocabularies and ontologies like SKOS, RDFS, and OWL to provide interoperability.
This document discusses ontologies, which organize and describe related items to represent semantics. An ontology has several components: classes (collections, concepts), individuals (instances, objects), attributes (aspects, properties), values/properties (specific data for attributes), and relations (how individuals/classes relate). Good ontologies have well-defined syntax, structure, semantics, and pragmatics. They are useful for categorizing large amounts of data to improve integration and allow machine interpretation.
This document discusses ontologies, which organize and describe related items to represent semantics. An ontology has several components: classes (collections, concepts), individuals (instances, objects), attributes (aspects, properties), values/properties (specific data for attributes), and relations (how individuals/classes relate). Good ontologies have well-defined syntax, structure, semantics, and pragmatics. They are useful for categorizing large amounts of data to improve integration and allow machine interpretation.
This document discusses text encoding and markup. It introduces XML and the Text Encoding Initiative (TEI), which uses XML to encode scholarly documents. Key points include:
- XML allows users to define their own semantic markup languages and impose interpretive models on texts through schemas like TEI.
- TEI is the dominant language for encoding scholarly texts and primary sources. It allows scholars to select elements to match their areas of interest.
- XML and TEI view texts as ordered hierarchies of content objects (OHCO), representing them as trees. This has advantages like easy processing but also limitations regarding overlaps in logical and physical structure.
- Different representational tools like tables and trees can be used to reconcile textual
The document provides an overview of semantic technologies and discusses their increasing mainstream adoption. It notes that Microsoft purchased Powerset in 2008, Apple purchased Siri in 2010, and Google bought Metaweb and released semantic search in 2013. It discusses how semantic technologies allow for interoperability through shared representations and reasoning. Examples are given of early semantic search applications from 1999-2002 and an operational semantic electronic medical record application deployed in 2006.
Using construction grammar in conversational systemsCJ Jenkins
This thesis explored using construction grammar and ontologies in conversational systems. The author built two early experimental systems using these techniques. Construction grammar represents language as constructions pairing form and meaning. Ontologies allow for more explicit semantics compared to databases. The author developed a stemmer called UEA-Lite and a system called KIA that incorporated construction grammar, ontologies, and machine learning to understand and respond to natural language.
study or concern about what kinds of things exist
what entities there are in the universe.
the ontology derives from the Greek onto (being) and logia (written or spoken). It is a branch of metaphysics , the study of first principles or the root of things.
The document discusses knowledge representation on the Semantic Web. It introduces the need to formally represent information on the web using languages that allow computers to process and reason with the information. It describes the approach of using ontology languages like RDF and OWL to develop domain models and conceptualizations that provide shared interpretations of information across sources. It explains some of the basic constructs in ontology-based knowledge representation using these languages, including classes, properties, subclasses and restrictions.
The document discusses ontologies, including their definition, purpose, and typical engineering process. It provides examples of existing ontologies such as DBpedia, Wikidata, and WordNet. It also outlines some key activities for developing ontologies, such as finding relevant existing ontologies, selecting which to use or extend, and adjusting or expanding them as needed. Some basics of ontology conceptualization are also introduced, such as modeling classes, instances, attributes, and relationships between classes.
Relations for Reusing (R4R) in A Shared Context: An Exploration on Research P...andrea huang
Will the rich domain knowledge from research publications and the implicit cross-domain metadata of cultural objects be compliant with each other? A contextual framework is proposed as dynamic and relational in supporting three different contexts: Reusing, Publication and Curation, which are individually constructed but overlapped with major conceptual elements. A Relations for Reusing (R4R) ontology has been devised for modeling these overlapping
conceptual components (Article, Data, Code, Provence, and License) for interlinking research outputs and cultural heritage data. In particular, packaging and citation relations are key to build up interpretations for dynamic contexts. Examples are provided for illustrating how the linking mechanism can be constructed and represented as a result to reveal the data linked in different contexts.
ESWC SS 2013 - Wednesday Tutorial Elena Simperl: Creating and Using Ontologie...eswcsummerschool
1. The document discusses building and using ontologies in computer science. An ontology defines a domain of interest by describing the things in the domain, their attributes, and relationships. Ontologies are used to share a common understanding of a domain among people and machines.
2. The document then outlines the typical process for ontology engineering including requirements analysis, conceptualization, implementation, knowledge acquisition, testing and documentation. It provides examples of ontologies like Project Halo and how they are used.
3. The core activities in the given scenario are finding existing relevant ontologies, selecting which to use and extend, rather than creating one from scratch. It describes several repositories and vocabularies to find existing ontologies and gives tips
This document describes the BLOOMS+ approach for performing contextual ontology alignment of Linked Open Data datasets with an upper ontology. It presents the challenges of existing ontology matching approaches when applied to large, diverse LOD datasets. BLOOMS+ leverages structured knowledge from Wikipedia categories to generate "BLOOMS trees" representing different senses of concept names, and calculates similarity between trees to identify subclass, equivalence and other relationships between concepts in different datasets. It outperforms existing approaches on aligning three real-world LOD ontologies to the PROTON upper ontology. Future work aims to improve the weighting and identify additional contextual sources to assist with schema alignment across datasets.
(1) The document discusses the Semantic Web, ontologies, and ontology learning. It defines the Semantic Web as an extension of the current web that gives information well-defined meaning. (2) Ontologies are formal specifications of concepts and relations that provide shared meanings between machines and humans. (3) Ontology learning is the automatic or semi-automatic process of extracting ontological concepts and relations from text to build or enrich ontologies. The document outlines methods for ontology learning and its applications.
Similar to Semantic relations: new (terminological) challenges in a world of Linked Data (20)
KeyNote @SEMANTICS 2017 (Amsterdam, sept 2017) about convergences between NLP and KE at the era of the semantic web, with a focus on semantic relation extraction from text.
Caractériser le contenu des exigences pour aider à leur rédaction : un défi p...Nathalie Aussenac-Gilles
Journée "Linguistique et Traitement Automatique des Langues pour l'Aéronautique et l'Espace :Dimensions langagières du risque" à l'université Toulouse Jean Jaurès, 2 juillet 2013
The Ipsos - AI - Monitor 2024 Report.pdfSocial Samosa
According to Ipsos AI Monitor's 2024 report, 65% Indians said that products and services using AI have profoundly changed their daily life in the past 3-5 years.
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Aggregage
This webinar will explore cutting-edge, less familiar but powerful experimentation methodologies which address well-known limitations of standard A/B Testing. Designed for data and product leaders, this session aims to inspire the embrace of innovative approaches and provide insights into the frontiers of experimentation!
Global Situational Awareness of A.I. and where its headedvikram sood
You can see the future first in San Francisco.
Over the past year, the talk of the town has shifted from $10 billion compute clusters to $100 billion clusters to trillion-dollar clusters. Every six months another zero is added to the boardroom plans. Behind the scenes, there’s a fierce scramble to secure every power contract still available for the rest of the decade, every voltage transformer that can possibly be procured. American big business is gearing up to pour trillions of dollars into a long-unseen mobilization of American industrial might. By the end of the decade, American electricity production will have grown tens of percent; from the shale fields of Pennsylvania to the solar farms of Nevada, hundreds of millions of GPUs will hum.
The AGI race has begun. We are building machines that can think and reason. By 2025/26, these machines will outpace college graduates. By the end of the decade, they will be smarter than you or I; we will have superintelligence, in the true sense of the word. Along the way, national security forces not seen in half a century will be un-leashed, and before long, The Project will be on. If we’re lucky, we’ll be in an all-out race with the CCP; if we’re unlucky, an all-out war.
Everyone is now talking about AI, but few have the faintest glimmer of what is about to hit them. Nvidia analysts still think 2024 might be close to the peak. Mainstream pundits are stuck on the wilful blindness of “it’s just predicting the next word”. They see only hype and business-as-usual; at most they entertain another internet-scale technological change.
Before long, the world will wake up. But right now, there are perhaps a few hundred people, most of them in San Francisco and the AI labs, that have situational awareness. Through whatever peculiar forces of fate, I have found myself amongst them. A few years ago, these people were derided as crazy—but they trusted the trendlines, which allowed them to correctly predict the AI advances of the past few years. Whether these people are also right about the next few years remains to be seen. But these are very smart people—the smartest people I have ever met—and they are the ones building this technology. Perhaps they will be an odd footnote in history, or perhaps they will go down in history like Szilard and Oppenheimer and Teller. If they are seeing the future even close to correctly, we are in for a wild ride.
Let me tell you what we see.
Orchestrating the Future: Navigating Today's Data Workflow Challenges with Ai...Kaxil Naik
Navigating today's data landscape isn't just about managing workflows; it's about strategically propelling your business forward. Apache Airflow has stood out as the benchmark in this arena, driving data orchestration forward since its early days. As we dive into the complexities of our current data-rich environment, where the sheer volume of information and its timely, accurate processing are crucial for AI and ML applications, the role of Airflow has never been more critical.
In my journey as the Senior Engineering Director and a pivotal member of Apache Airflow's Project Management Committee (PMC), I've witnessed Airflow transform data handling, making agility and insight the norm in an ever-evolving digital space. At Astronomer, our collaboration with leading AI & ML teams worldwide has not only tested but also proven Airflow's mettle in delivering data reliably and efficiently—data that now powers not just insights but core business functions.
This session is a deep dive into the essence of Airflow's success. We'll trace its evolution from a budding project to the backbone of data orchestration it is today, constantly adapting to meet the next wave of data challenges, including those brought on by Generative AI. It's this forward-thinking adaptability that keeps Airflow at the forefront of innovation, ready for whatever comes next.
The ever-growing demands of AI and ML applications have ushered in an era where sophisticated data management isn't a luxury—it's a necessity. Airflow's innate flexibility and scalability are what makes it indispensable in managing the intricate workflows of today, especially those involving Large Language Models (LLMs).
This talk isn't just a rundown of Airflow's features; it's about harnessing these capabilities to turn your data workflows into a strategic asset. Together, we'll explore how Airflow remains at the cutting edge of data orchestration, ensuring your organization is not just keeping pace but setting the pace in a data-driven future.
Session in https://budapestdata.hu/2024/04/kaxil-naik-astronomer-io/ | https://dataml24.sessionize.com/session/667627
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data LakeWalaa Eldin Moustafa
Dynamic policy enforcement is becoming an increasingly important topic in today’s world where data privacy and compliance is a top priority for companies, individuals, and regulators alike. In these slides, we discuss how LinkedIn implements a powerful dynamic policy enforcement engine, called ViewShift, and integrates it within its data lake. We show the query engine architecture and how catalog implementations can automatically route table resolutions to compliance-enforcing SQL views. Such views have a set of very interesting properties: (1) They are auto-generated from declarative data annotations. (2) They respect user-level consent and preferences (3) They are context-aware, encoding a different set of transformations for different use cases (4) They are portable; while the SQL logic is only implemented in one SQL dialect, it is accessible in all engines.
#SQL #Views #Privacy #Compliance #DataLake
Analysis insight about a Flyball dog competition team's performanceroli9797
Insight of my analysis about a Flyball dog competition team's last year performance. Find more: https://github.com/rolandnagy-ds/flyball_race_analysis/tree/main
Predictably Improve Your B2B Tech Company's Performance by Leveraging DataKiwi Creative
Harness the power of AI-backed reports, benchmarking and data analysis to predict trends and detect anomalies in your marketing efforts.
Peter Caputa, CEO at Databox, reveals how you can discover the strategies and tools to increase your growth rate (and margins!).
From metrics to track to data habits to pick up, enhance your reporting for powerful insights to improve your B2B tech company's marketing.
- - -
This is the webinar recording from the June 2024 HubSpot User Group (HUG) for B2B Technology USA.
Watch the video recording at https://youtu.be/5vjwGfPN9lw
Sign up for future HUG events at https://events.hubspot.com/b2b-technology-usa/
2. Outline of the talk
• Semantic relations
– what do we mean?
– why do we need them for?
• Finding semantic relations
– what are the issues?
– what can large corpora and machine learning do
for you?
– how can terminology contribute to Linked Data?
2Semantic relations - Aussenac09/01/2015
3. Semantic relations,
what do we mean?
Research field
• Linguistics: semantic
relations, semantic roles,
discourse relations
• Terminology
– Weak structure
– Stored in DB or SKOS
models
• Information extraction
What is a relation
A tree comprises at least a trunk,
roots and branches.
A tree [Plants] comprises [meronymy] at
least a trunk, roots and branches.
tree has_parts trunk, roots, branches
(tree, has parts, trunk) …
in a gardening terminology
looks for relations between instances
09/01/2015 Semantic relations - Aussenac 3
tree Plantation year Species Branches
Tree1 1990 Oak > 20
Tree2 1995 Oak 15
whole
parts
4. Semantic relations,
what do we mean?
Research field
• Domain Ontology engineering
– Formal (logic, RDF, OWL …) and
may lead to infer new
knowledge
– The relation is part of a network
– May be shared or not
• Semantic web
– Independent triples
– Publically available in data
repositories with W3C Standard
format
– Connect triples with existing
ones, with web ontologies
What is a relation
bot:Tree bot:has_part bot:Branch
09/01/2015 Semantic relations - Aussenac 4
Trunk
Has-part
Root
Plant
Fonguscereals
Has-
part
Root
is_a
Tree
Has-
part
Branch
bot:myTr
ee
bot:has-
part
bot:MyTre
eRoots
bot:Tree
bot:has-
part
bot:Branch
rdf:Type
5. Example: tree in DBPedia
5
dbpedia-
owl:tree
dbpedia-
owl:Species
dbpedia-
owl:Place
Semantic relations - Aussenac09/01/2015
6. dbpedia-
owl:PhysicalEntity
rdfs:subClassOf
dbpedia-
owl:Organism
Example: Plants in DbPedia
6
owl:SameAs
yago:WordNet_Plant_
100017222
dbpedia-
owl:Plant
Semantic web specificities
- New types of relations: mappings
- Focus 1: resources, either classes or instances
- Focus 2: relations between resources
- Focus 3: give a type to resources
dbpedia-
owl:Acer_Stone
bergae
dbpedia-
owl:Alopecurus_ca
rolinianus
dbpedia-
owl:Alsmithia_long
ipes
dbpedia-owl:…
rdf:typerdf:type
rdf:typerdf:type
Semantic relations - Aussenac09/01/2015
7. dbpedia-
owl:PhysicalEntity
rdfs:subClassOf
dbpedia-
owl:Organism
Example: Plants in DbPedia
7
owl:SameAs
yago:WordNet_Plant_
100017222
dbpedia-
owl:Plant
Semantic web specificities
- New types of relations: mappings
- Focus 1: resources, either classes or instances
- Focus 2: relations between resources
- Focus 3: give a type to resources
dbpedia-
owl:Acer_Stone
bergae
dbpedia-
owl:Alopecurus_ca
rolinianus
dbpedia-
owl:Alsmithia_long
ipes
dbpedia-owl:…
rdf:typerdf:type
rdf:typerdf:type
Semantic relations - Aussenac09/01/2015
8. dbpedia-
owl:PhysicalEntity
rdfs:subClassOf
dbpedia-
owl:Organism
Example: Plants in DbPedia
8
owl:SameAs
yago:WordNet_Plant_
100017222
dbpedia-
owl:Plant
Semantic web specificities
- New types of relations: mappings
- Focus 1: resources, either classes or instances
- Focus 2: relations between resources
- Focus 3: give a type to resources
dbpedia-
owl:Acer_Stone
bergae
dbpedia-
owl:Alopecurus_ca
rolinianus
dbpedia-
owl:Alsmithia_long
ipes
dbpedia-owl:…
rdf:typerdf:type
rdf:typerdf:type
Semantic relations - Aussenac09/01/2015
9. Semantic relations:
why do we need them for?
Research field
• Linguistics: to understand how
language produces meaning
• Terminology: to capture domain
specific terms and meanings
• Information extraction: to
collect structured data – the
schema (classes and relations) is
known
Relations contribute to
• semantics and language
interpretation
• formal semantics and
discourse semantics
• Structure terminologies and
make browsing easier
• Connect terms
• Give meaning to “concepts”
• Find related entities (values) to
build DB and then mine these
data (statistics)
09/01/2015 Semantic relations - Aussenac 9
10. Semantic relations:
why do we need them for?
Research field
• Formal ontologies: to define
axioms and inferences
associated with relations
• Domain ontologies: human
and machine “understanding”
• Linked data: interoperability,
data connection from various
sites or applications
Relations contribute to
• Inferences and reasoning
• Ex: sub-classes inherit of some class
properties
• Ex : transitivity (cf. some parthood rel.),
symmetry
• Ex: cardinality …
• Relations define concepts by
differences and similarities
• Relations have labels and are human
readable ; each one can be processed in
a specific way
• Relations connect resources (data)
• Their semantics is defined in ontologies
or produce behaviors in inference
engines (rdf:subClassOf)
09/01/2015 Semantic relations - Aussenac 10
11. 09/01/2015 Semantic relations - Aussenac 11
Erarht Rahm http://dbs.uni-leipzig.de/file/paris-Octob2014.pdf
Semantic web specificities
- Relations connect web data called resources
- Relations connect data with ontology classes: importance of hypernymy
- Relations may map ontology classes
12. Finding semantic relations,
what are the issues?
• Knowledge sources:
– where can we find relations?
• Extraction techniques
– How can we identify them?
• Representation
– Which way do I represent this information?
• Validation
– What makes a relation representation valild? Relevant?
09/01/2015 Semantic relations - Aussenac 12
13. Finding semantic relations,
what are the issues?
• Knowledge sources
– text, human experts, existing “semantic” resources (lexicon,
terminologies, ontologies, Linked Data vocabularies)
– Domain specific vs general knowledge
• Extraction techniques
– “obvious” language regularities, known relations and classes (or
entities) -> Patterns
• Issues : domain dependence, domain coverage, variation and
flexibility, rigidity (need to be regularly updated)
• Research issues: automatic building by machine learning
– “more implicit” language regularities, medium size corpora,
open list of classes/entities -> supervised learning
– Very large corpora, unexpected relations -> unsupervised
learning
13Semantic relations - Aussenac09/01/2015
14. Pattern based relation extraction,
an issue: variation
• A tree comprises at least a trunk,
roots and branches.
• With branches reaching the ground,
the willow is an ornamental tree.
• The tree of the neighbor has been
delimed.
• He climbs on the branches of the tree.
• This tree is wonderful. Its branches
reach the ground.
• Contains: very systematic pattern; the
parts may be difficult to spot;
enumeration > various parts
• With: meronymy pattern only in some
genres (such as catalogs, biology
documents)
• Delimed : Term and pattern are in the
same word; requires background
knowledge: delimed -> has_part
branches (and branches are cut)
• Of : Very ambiguous pattern; polysemy
reduced in [verb N1 of N2]
• Its : very ambiguous pattern; necessity
to take into account two sentences
14Semantic relations - Aussenac09/01/2015
15. Pattern based relation extraction,
learning patterns (1)
• Patterns are seen (and stored) as lexicalizations of ontology properties
• Patterns are “extracted” from syntactic dependencies between related
entities (in triples)
• Assumes that patterns are structured around ONE lexical entry
• Lemon format for lexical ontologies
• Entries can be frames
09/01/2015 Semantic relations - Aussenac 15
ATOLL—A framework for the automatic induction of ontology lexica
S. Walter, C. Unger, P. Cimiano, DKE (94), 148-162 (2014)
16. Pattern based relation extraction,
learning patterns (1)
• Patterns are seen (and stored) as lexicalizations of ontology properties
• Patterns are “extracted” from syntactic dependencies between related
entities (in triples)
• Assumes that patterns are structured around ONE lexical entry
• Lemon format for lexical ontologies
• Entries can be frames
09/01/2015 Semantic relations - Aussenac 16
ATOLL—A framework for the automatic induction of ontology lexica
S. Walter, C. Unger, P. Cimiano, DKE (94), 148-162 (2014)
17. Pattern based relation extraction,
learning patterns (2)
09/01/2015 Semantic relations - Aussenac 17
Michelle Obama is the wife of Barack Obama, the current president.
Michelle Obama allegedly told her husband, Barack Obama, to ..
Michelle Obama, the 44th first lady and wife of President Barack
Dbpedia:spouse
Find all lexicalizations of the entities: Michelle Obama, Mrs. Obama, Michelle
Robinson …
18. Pattern based relation extraction,
learning patterns (3)
09/01/2015 Semantic relations - Aussenac 18
• Pattern = shortest path btw the 2 entities in the dependency graph
[MichelleObama (subject), wife (root), of (preposition), BarackObama (object)]
• Lexical entry in the ontology
19. Relation extraction:
learning relations from enumerative structures
09/01/2015 Semantic relations - Aussenac 19
IS_A
IS_A
Learning relations from an parallel enumerative structure =
- classification task to identify the relation (IS_A, part_Of, other)
- Term extraction to identify the primer and the items
20. Relation extraction:
learning relations from enumerative structures
• Corpus
– 745 enumerative structures from
Wikipedia pages
– 3 relation types: taxonomic,
ontological_non_taxonomic,
non_ontological
• Classification task
– Feature definition
– Automatic evaluation of features
– 3 algorithms are compared : SVM,
MaxEntropy and baseline (majority)
– Training of the 2 algorithms
• Results
– 82% f-measure for SVM
– Best result with a 2 step process
(ontological yes/no -> feature and
then taxonomic yes/no)
09/01/2015 Semantic relations - Aussenac 20
21. From intepretation to representation
• A tree comprises at least a trunk,
roots and branches.
• With branches reaching the
ground, the willow is an
ornamental tree.
• The tree of the neighbor has been
delimed.
Tree
Trunk
Branches
Has-part Roots
Ornamental
Tree
Willow Tree Has-part Branches
Has-part Branches
Has-part Branches
Semantic relations - Aussenac 2109/01/2015
22. From intepretation to representation
• A tree comprises at least a trunk,
roots and branches.
• With branches reaching the
ground, the willow is an
ornamental tree.
• The tree of the neighbor has been
delimed.
• He’s climbing on the branches of
the tree.
• This tree is wonderful. Its
branches reach the ground.
Tree
Trunk
Branches
Has-part Roots
Has-part Branches
Has-part Branches
Semantic relations - Aussenac 2209/01/2015
Neighbor
Tree
Instance _of
23. Finding semantic relations:
what can large corpora and machine
learning do for you ?
• Learning patterns
– Poor results
– Requires very large data sets
– Reasonable for general
knowledge
• Learning relations
– Much more relevant
– Large variety of approaches in
the state of the art
– Key step = select feature
– more features, better the results
09/01/2015 Semantic relations - Aussenac 23
24. Finding semantic relations,
what can terminology do for the semantic web?
• Terminology can play a key role
– change its practices
– contribute to enrich the LOD with QUALITY data
• Contribute to define /evaluate tools
– evaluate machine learning tools
– Test and experiment them on various textual genre, in
various domains
– Propose more features
• Propose terminology tools
– … to improve these approaches with the studies carried
out during the last 20 years to build terminologies and
terminological knowledge bases.
– term extractors, relation extractors, patterns
24Semantic relations - Aussenac09/01/2015
25. Finding semantic relations,
what can terminology do for the semantic web?
• Contribute to enrich the LOD with terminologies
– Terminologies as structuring networked vocabularies
– Publish your ontologies as public web resources
– Conform to linked data requirements
– Use standards format like SKOS (Simple Knowledge
Organization System) the W3C’s OWL ontology for creating
thesauruses, taxonomies, and controlled vocabularies
• Contribute to the multilingual LOD
– Add lexicon to ontologies
– Multilingual efforts: LEMON, MONNET projects DID NOT
include terminologists
09/01/2015 Semantic relations - Aussenac 25
26. The LOD and the
linguistic LOD
need you!
http://linguistics.okfn.org/resources/llod/
09/01/2015 Semantic relations - Aussenac 26
27. Some landmark KOS LD
implementations
• Many Libraries
– Swedish National Library’s Libris catalogue and thesaurus http://libris.kb.se/
– Library of Congress’ vocabularies, including LCSH http://id.loc.gov/
– DNB’s Gemeinsame Normdatei (incl. SWD subject headings) http://dnb.info/gnd/
• Documentation at https://wiki.dnb.de/display/LDS
– BnF’s RAMEAU subject headings http://stitch.cs.vu.nl/
– OCLC’s DDC classification http://dewey.info/ and VIAF http://viaf.org/
– STW economy thesaurus http://zbw.eu/stw
– National Library of Hungary’s catalogue and thesauri
– http://oszkdk.oszk.hu/resource/DRJ/404(example)
• Other fields
– Wikipedia categories through Dbpedia http://dbpedia.org/
– New York Times subject headings http://data.nytimes.com/
– IVOA astronomy vocabularies http://www.ivoa.net/Documents/latest/Vocabularies.html
– GEMET environmental thesaurus http://eionet.europa.eu/gemet
– Agrovoc http://aims.fao.org/
– Linked Life Data http://linkedlifedata.com/
– Taxonconcept http://www.taxonconcept.org/
– UK Public sector vocabularies http://standards.esd.org.uk/ (e.g., http://id.esd.org.uk/lifeEvent/7)
09/01/2015 Semantic relations - Aussenac 27