SlideShare a Scribd company logo
1 of 19
Download to read offline
Ontology Evaluation:
Methods and Metrics
             MITRE Research Interest


                Dr. Joanne S. Luciano




                    In collaboration with

               Dr. Leo Obrst, PhD
               Suzette Stoutenburg
                  Kevin Cohen
                 Jean Stanford
Approved for Public Release. Distribution Unlimited. Case Number: 08
                                                                  08-0849   1
Ontology: A Key Technology
 for Knowledge Management
Used to describe terms in a vocabulary
and the relationships among them.
Ontology languages vary in their
semantic expressiveness.

                                         Based on work by Leo Obrst of MITRE as interpreted by
                                         Dan McCreary. This can be viewed as a trade-off of
                                         semantic clarity v. the time and money it takes to construct
                                         http://www.mkbergman.com/?m=20070516.




Ontologies have become the most
widespread form of knowledge
representation for multiple purposes       Ontology Summit 2007 (NIST, Gaithersburg, MD,
                                           April 23-24, 2007)                                           2
The Problem
   Ontology Elephants
There is no single real elephant                                                   An elephant is abstract
                                   There must be a purpose for
                                   an elephant: use cases?




                                                                                An elephant is very abstract

 There must be an
 upper elephant


                                                              An elephant
                                                              is really very                 An elephant is the
                                                              simple                         result of consensus
                                     There are only
Open vs.                             distributed elephants
Closed                               & their mappings
Elephant
                             Prospects and Possibilities for Ontology Evaluation: The View from NCOR
                                                                                                NCOR,
                             Obrst, Hughes, Ray, WWW 2006, May 22   22–26, 2006, Edinburgh, UK.
                                                                                                               3
The Problem

Users need to be able to build sound ontologies and to reuse ontologies for different
purposes. There is no standard no way to do that now.

At the Ontology Summit 2008 two competing “state of the art” evaluation proposals for the
Open Ontology Repository were presented. Both treat ontologies as “black boxes” and both
are subjective evaluations.
 – Peer Review – self appointed editorial review board decides, non
                                                                non-overlapping domain so first one
   gets preference, ‘best practices’
 – User Ratings – users report their experience and rate them on a website

    Language was inserted into the communiqué to provide mechanisms that enable ontology evaluation
    by other metrics

Prior work on evaluating ontologies is limited, no consensus
                      - Ontology Workshop Methods and Metrics October 2007
                                Workshop materials at: http://sites.google.com/a/cme.nist.gov/workshop
                                                           ://sites.google.com/a/cme.nist.gov/workshop-on-ontology-evaluation/


 – Formal logic based applications can use software reasoners to address logical consistency and
   classification; many don’t take advantage of these tools
 – Natural Language Processing (NLP) application use NLP evaluation methods, but address only NLP
                                                 ons
   applications (text mark-up – aka “annotation”), information retrieval and extraction
 – Alignment (mapping of ontologies) for data mining, integration, fusion
    Ontology Summit 2007 (NIST, Gaithersburg, MD, April 23-24, 2007 See slide at end with notes.
                                                           24, 2007)                                                             4
Objective: Evolve toward Science &
Engineering Discipline for Ontology

• Create procedures, processes, methods to help define,
  adjudicate, & ensure quality of knowledge
  capture/representation

• Facilitate the education of communities on ontology
  development & promote best practices for ontology
  development

• Enable the best standards related to ontologies, &
  promote linkages, liaisons among standards
  organizations


                                                          5
Approach:
• Two stages:
   – Recast use case into its components:              Novel
      • Functional objective                          Approach
      • Design objective & requirements specification
      • Semantic components required to achieve above
   – Evaluate components using objective metrics

• Place existing evaluation methods in context by utility

• Engage and rally the community / stakeholders
   – Participate at appropriate meetings: present work and facilitate
     community focus on objective metrics for evaluation
   – Introduce users with complementary skills, joint vision, shared
     needs to develop needed metrics, content, tests, tools
   – Involve multiple government agencies, industry and academia
     to support initiative
                                                                        6
Research Plan: (1) Identify use cases
For each, recast the use cases into their components:

         a.   Specify functional objective (what it is, what it does)
                 e.g. enable investigation of data collected on influenza strain mutations
                 that cause death in birds

         b.   Specify design objective (how good it has to be e.g., what
              specifications have to be met? Is it a prototype, for commercial
              use, or must it meet military specifications?)
                  e.g. Must meet: Minimum Information about an Influenza Genotype
                      and a Nomenclature Standard (MIIGNS)
         c.   Identify (or specify) the semantic components required to achieve
              the functional objective to the level specified by the design
              objective (Authoritative Sources such as engineering tolerances,
              physical constants, legal jurisdictions, company policies)
                  e.g. To meet MIIGNS, the following semantic components must
                      be included:
                                  biomaterial transformations
                                  assays
                                  data transformations

                                                                                             7
Research Plan (2) Develop Metrics

Develop metrics for 3 criteria for evaluation:
  1.   Correctness: how well the f
                               e functional components express the
       design objectives
         a) Language expressiveness
         b) Fluency / competency
  2.   Completeness: combines use case with design criteria
       (requirement specification)
         a) To what extent can requirements be met?
         b) Which semantic components (authoritative sources) are needed/missing?
                                     nts
  3.   Utility: Is it useful? Does it work?
          Combine 1 and 2 (correctness and completeness) and evaluate
          against the use case (by competency questions, or other challenge
          tests).


                                                                                    8
Examples:


• BioPAX (prior work)

• Habitat-Lite (subset of Environmental
          Lite
  Ontology to support of NSF funded Mining
  Metadata for Metagenomics)

• Influenza Infectious Disease Ontology (for
  Genomics for Bioforensics MSR)
                                               9
Example (1) BioPAX lack of fluency
 chemical structure & pathway steps incorrectly modeled
        - misunderstanding of the language (language has capability)
        - modeled disjoint from the biology & chemistry
        - leads to logical inconsistency




OWL has a steep learning curve it’s easy to get things wrong.
                           rve,
                                                                       10
Example (2) Habitat-
                   -Lite:
   correctness & completeness
Objective: facilitate capture of habitat and
environmental metadata on genomic sequences
Approach: select subset of terms with highest frequency and
evaluate usefulness by correctness and completeness metrics

 – Evaluated correctness
     • 64% agreement (84 of 132 terms) of automated and expert mapping of
       terms

 – Evaluated coverage of terms
     • 84% exact matches (“host,” “aquatic,” and “soil” covered 75%)


Hirschman, Clark, Cohen, Mardis, Luciano, Kottmann, Cole, Markowitz, Kyprpides, Field
Habitat-Lite: a GSC Case Study Based on Free Text Terms for Environmental Metadata
        Lite:
OMICS A Journal of Integrative Biology Volume 12, Number 2, 2008 (in press)
                                                                                        11
Example (3) Enable Influenza Research
        (proposed construction and subsequent evaluation)

Function: enable investigation of data collected on
influenza strain mutations that cause death in birds
Design objective: Minimum Information about an Influenza
                :
     Genotype and a Nomenclature Standard (MIIGNS)
Semantic components:
   1.   biomaterial transformations
            a. recombinant plasmid biomaterial transformation
            b. site-directed mutagenesis biomaterial transformation
                    directed
            c. reverse genetic virus production biomaterial transformation
            d. Mouse infection biomaterial transformation
   2.   assays
           a. weight assay
           b. virus replication / mouse lung assay
           c. Cytokine quantification assay
   3.   data transformations
           a. statistical difference evaluation

                                                                             12
Example (3) Enable Influenza Research
    (proposed construction and subsequent evaluation)

Correctness:
  Language expressivity: validate definitions against
    OBO Foundry relations
  Fluency: inter-developer agreement (3 developers, 2
                  developer
      code same, 3rd validates)
Completeness:
  Calculate % coverage of minimum terms (18 terms)
  Calculate % coverage of full terms (196 terms)
Utility: Challenge Questions
  To be developed (by our collaborator BioHealthBase)

                                                        13
Impact

Communities of Practice areas need objective methods and
Metrics to facilitate the development, interoperability and reuse
of their ontologies

Some examples:

   – Message Based Data Exchange
   – BioSecurity
   – Health Care and Biomedicine
   – Life Sciences
   – Disease
   – Metagenomics
   – Agile systems for rapid enterprise integration of
     heterogeneous data
   – Intelligence Community

                                                                    14
Why MITRE?



 MITRE is uniquely positioned to act as an
 impartial experimental designer and
 arbitrator in the development of an
 evaluation methodology for ontologies

 – MITRE has acted in the past for natural language
  technologies such as text summarization in the Text
  REtrieval Conferences (TREC) and Information
  Extraction in the BioCreAtIvE challenge

                                                        15
Additional Notes on Specific Slides
Slide 1: Lower right gaphic: Ontology Summit 2007 (NIST, Gaithersburg, MD, April 23 23-24, 2007) effort. See the following:
Ontology Summit 2007 - Ontology, Taxonomy, Folksonomy: Understanding the Distinctions. http://ontolog.cim3.net/cgi-
                                                            :
bin/wiki.pl?OntologySummit2007.
Ontology Summit 2007 Communique. http://ontolog.cim3.net/cgi
                                     http://ontolog.cim3.net/cgi-bin/wiki.pl?OntologySummit2007_Communique.
Ontology Summit 2007 Ontology Dimensions Map. http://ontolog.cim3.net/cgi
                                                    http://ontolog.cim3.net/cgi-
bin/wiki.pl?OntologySummit2007_FrameworksForConsideration/DimensionsMap.
bin/wiki.pl?OntologySummit2007_FrameworksForConsideration/DimensionsMap
Gruninger, Michael; Olivier Bodenreider; Frank Olken; Leo Obrst; Peter Yim. 2008. The 2007 Ontology Summit Joint
          ,
Communiqué. Ontology, Taxonomy, Folksonomy: Understanding the Distinctions. Journal of Applied Ontology, forthcoming.
                                                :

Slide 2: Ontology Summit 2008 (NIST, Gaithersburg, MD, April 28
                                                              28-29, 2008).
Ontology Summit 2008: Toward an Open Ontology Repository. http://ontolog.cim3.net/cgi
                                                            http://ontolog.cim3.net/cgi-bin/wiki.pl?OntologySummit2008.
Ontology Summit 2008 Communique. http://ontolog.cim3.net/cgi
                                   http://ontolog.cim3.net/cgi-bin/wiki.pl?OntologySummit2008_Communique.

Slide 4:
Concerning:
“At the Ontology Summit 2008 two competing “state of the art” evaluation proposals for the Open Ontology Repository were
presented. Both treat ontologies as “black boxes” and both are subjective evaluations
                                                                          evaluations.

–Peer Review –self appointed editorial review board decides, non
               self                                          non-overlapping domain so first one gets preference, „best
practices‟
–User Ratings –users report their experience and rate them on a website
               users

These points were made during the summit and discussed more fully during the Quality and Gatekeeping session of the
Ontology Summit 2008 (NIST, Gaithersburg, MD, April 28-29, 2008
                                                        29, 2008.


                                                                                                                              16
Background References
The BioCreAtIvE (Critical Assessment of Information Extraction systems in Biology) challenge evaluation consists of a communi
                                                                                                                            community-wide effort for evaluating text mining
      and information extraction systems applied to the biological domain. BioCreative (http://biocreative.sourceforge.net/)
                                                                              BioCreative.
The Text REtrieval Conference (TREC), co-sponsored by the National Institute of Standards and Technology (NIST) and U.S. Department of Defense, wa started in
                                                sponsored                                                                                                     was
      1992 as part of the TIPSTER Text program. Its purpose was to support research within the information retrieval community by p        providing the infrastructure
      necessary for large-scale evaluation of text retrieval methodologies. (http://trec.nist.gov/overview.html
                                                                               http://trec.nist.gov/overview.html)
The Message Understanding Conferences (MUC) were initiated and financed by DARPA to encourage the development of new and bett methods of information
                                                                                                                                           better
      extraction. The character of this competition -- many concurrent research teams competing against one another -- necessitated the development of standards for
      evaluation, e.g. the adoption of recall and precision. (http://en.wikipedia.org/wiki/Message_Understanding_Conference
                                                              http://en.wikipedia.org/wiki/Message_Understanding_Conference)
Lenat, Douglas B. “CYC: a large-scale investment in knowledge infrastructure,” Communications of the ACM Volume 38 , Issue 11 (November 1995) Pages: 33 – 38.
                                                                                                             ACM,
Project Halo (http://www.projecthalo.com/), is a project funded by Paul Allen's Vulcan Ventures. The project was initially led by Oliver Roup and Noah Friedland but is
                                             ),
      currently led by Mark Greaves, a former DARPA Program Manager. Project Halo is an attempt to apply Artificial Intelligence te      techniques to the problem of producing
      a "digital Aristotle" that might serve as a mentor, providing comprehensive access to the world's knowledge (    (http://en.wikipedia.org/wiki/Project_Halo).
Ontoprise is a commerial software provider of ontology-based solutions. (http://www.ontoprise.de/content/index_eng.html
                                                                             http://www.ontoprise.de/content/index_eng.html)
Maedche, Alexander and Staab, Steffen. Ontology learning for the semantic web. Special Issue on Semantic Web. IEEE Intelligent Systems, 16(2):72    16(2):72-79, MAR 2001.
López, M. F.; Gómez-Pérez, A.; Sierra, J. P. & Sierra, A. P. Building a chemical ontology using Methontology and the OntologyDesign Environment IEEE Intelligent
                               ,
      Systems and Their Applications, 1999, 14, 37-46
Oltramari, A.; Gangemi, A.; Guarino, N. & Masolo, C. Restructuring WordNet's Top-        -Level: The OntoClean approach Proceedings of the Workshop OntoLex'2, Ontologies
      and Lexical Knowledge Bases, 2002
Guarino, N. & Welty, C. Evaluating ontological decisions with OntoClean Commun. ACM, ACM Press, 2002, 45, 61
                                                                                        .                             61-65
Smith, B.; Williams, J. & Schulze-Kremer, S. The ontology of the gene ontology. AMIA Annu Symp Proc, Institute for Formal Ontology and Medical Information Science,
                                    Kremer,
      University of Leipzig., 2003, 609-613.
Smith, B. From concepts to clinical reality: an essay on the benchmarking of biomedical terminologies. J Biomed Inform, Depar  Department of Philosophy and National Center
      for Biomedical Ontology, University at Buffalo, Buffalo, NY 14260, USA. phismith@buffalo.edu, 2006, 39, 288   288-298
Obrst, Leo; Todd Hughes; Steve Ray. 2006. Prospects and Possibilities for Ontology Evaluation: The View from NCOR. Workshop o Evaluation of Ontologies for the
                                                                                                                                          on
      Web (EON2006), Edinburgh, UK, May 22, 2006.
Obrst, Leo; Werner Ceusters; Inderjeet Mani; Steve Ray; Barry Smith. 2007 The Evaluation of Ontologies: Toward Improved Semantic Interoperability. Chapt in:
                                 ;                                                                                                                            Chapter
      Semantic Web: Revolutionizing Knowledge Discovery in the Life Sciences, Christopher J. O. Baker and Kei      Kei-Hoi Cheung, Eds., Springer, 2007.
Gangemi, Aldo; Carola Catenacci1; Massimiliano Ciaramita1; Jos Lehmann; contributions by: Rosa Gil (in section 2.2). Francesco Bolici and Onofrio Strignano (in
      section 2.4). 2004. Ontology evaluation and validation: An integrated formal model for the quality diagnostic task. ). OntoEval 2004.
Gangemi, A.; Catenacci, C.; Ciaramita, M.; Lehmann, J. 2005. A Theoretical Framework for Ontology Evaluation and Validation. In Proceedings of SWAP2005.
      http://www.loa-cnr.it/Papers/swap_final_v2.pdf
Gangemi, Aldo; Carola Catenacci; Massimiliano Ciaramita; and Jos Lehmann. 2006. Modelling ontology evaluation and validation. To appear in Proceedings of
      ESWC2006, Springer.
Lawrence Hunter, Mike Bada, K. Bretonnel Cohen, Helen Johnson, William Baumgartner, Jr. and Philip V. Ogren. "Ontology Quality Metrics," Center for Computational
      Pharmacology University of Colorado School of Medicine, October 8, 2007.
Lynette Hirschman , Jong C. Park , Junichi Tsujii , Limsoon Wong , and Cathy H. Wu. Accomplishments and challenges in literature data mining for biology.
      Bioinformatics 18: 1553-1561.
Proceedings of NIST 2007 Automatic Content Extraction Workshop (ACE), 2007. http://www.nist.gov/speech/tests/ace/ace07/
                                                                                       http://www.nist.gov/speech/tests/ace/ace07/.
      Methods and Metrics for Ontology Evaluation Workshop (Sponsors: NIST and NIH), National Institute of Standards and Technology Gaithersburg, MD, October 25-
                                                                                                                                    Technology,
      26.
Mani, I., B. Sondheim, D. House, L. Obrst. 1998. TIPSTER Text Summarization Evaluation: Final Report, MITRE technical report, Reston, VA, September, 1998.
      Proceedings of NIST 2007 Automatic Content Extraction Workshop (ACE), 2007. http://www.nist.gov/speech/tests/ace/ace07/.
                                                                                                                                                                             17
Background: Prior Technical
Approaches

• Evaluation in use - Navigli et al. 2003, Porzel and Malaka
  2005
   – Best case: Halo Project - Friedland et al. 2004
• Data-driven evaluation - essentially the fit between the
  ontology and a knowledge source e.g. a corpus - Brewster
  et al. 2004
• Gold Standard approaches, very common, used by for
  example, Cimiano et al. 2005
   – Major problem is the arbitrary choice of an ontology
   – Dellschaft and Staab 2006, proposed a method to derive
     IR/NLP type Precision/Recall/F
                  Precision/Recall/F-Measure



                                                               18
Background : Philosophical and Methodological
Approaches

Methontology approach of Gomez
                         Gomez-Perez:
  • Focus on evaluating procedural or formative aspects ontology
    construction
  • Criteria included: Consistency, Completeness, Conciseness,
    Expandability
  • Some tools developed reflecting these approaches:
     o Lam et al. 2004, Knublauch et al. 2004, Alani 2005, 2006
OntoClean approach of Guarino and Welty:
  • Philosophical approach based on theoretical principles:
  • Proposed a set of "metaproperties":
     o Rigidity
     o Identity
     o Unity
  • Much focus on "cleaning up" existing "ontologies" such as
    WordNet so as to make them more rigororous
                                                                   19

More Related Content

Similar to Ontology Evaluation Methods: Assessing Correctness, Completeness, and Utility

ISWC2023-McGuinnessTWC16x9FinalShort.pdf
ISWC2023-McGuinnessTWC16x9FinalShort.pdfISWC2023-McGuinnessTWC16x9FinalShort.pdf
ISWC2023-McGuinnessTWC16x9FinalShort.pdfDeborah McGuinness
 
SWSN UNIT-3.pptx we can information about swsn professional
SWSN UNIT-3.pptx we can information about swsn professionalSWSN UNIT-3.pptx we can information about swsn professional
SWSN UNIT-3.pptx we can information about swsn professionalgowthamnaidu0986
 
Knowledge Organization System (KOS) for biodiversity information resources, G...
Knowledge Organization System (KOS) for biodiversity information resources, G...Knowledge Organization System (KOS) for biodiversity information resources, G...
Knowledge Organization System (KOS) for biodiversity information resources, G...Dag Endresen
 
A Comparative Study of Ontology building Tools in Semantic Web Applications
A Comparative Study of Ontology building Tools in Semantic Web Applications A Comparative Study of Ontology building Tools in Semantic Web Applications
A Comparative Study of Ontology building Tools in Semantic Web Applications dannyijwest
 
A Comparative Study Ontology Building Tools for Semantic Web Applications
A Comparative Study Ontology Building Tools for Semantic Web Applications A Comparative Study Ontology Building Tools for Semantic Web Applications
A Comparative Study Ontology Building Tools for Semantic Web Applications IJwest
 
A Comparative Study Ontology Building Tools for Semantic Web Applications
A Comparative Study Ontology Building Tools for Semantic Web Applications A Comparative Study Ontology Building Tools for Semantic Web Applications
A Comparative Study Ontology Building Tools for Semantic Web Applications dannyijwest
 
0810ijdms02
0810ijdms020810ijdms02
0810ijdms02ayu dewi
 
Iot ontologies state of art$$$
Iot ontologies state of art$$$Iot ontologies state of art$$$
Iot ontologies state of art$$$Sof Ouni
 
A Comparative Study of Recent Ontology Visualization Tools with a Case of Dia...
A Comparative Study of Recent Ontology Visualization Tools with a Case of Dia...A Comparative Study of Recent Ontology Visualization Tools with a Case of Dia...
A Comparative Study of Recent Ontology Visualization Tools with a Case of Dia...IJORCS
 
Knowledge Organization System (KOS) for biodiversity information resources, G...
Knowledge Organization System (KOS) for biodiversity information resources, G...Knowledge Organization System (KOS) for biodiversity information resources, G...
Knowledge Organization System (KOS) for biodiversity information resources, G...Dag Endresen
 
A Framework for Ontology Usage Analysis
A Framework for Ontology Usage AnalysisA Framework for Ontology Usage Analysis
A Framework for Ontology Usage AnalysisJamshaid Ashraf
 
Tools for Ontology Building from Texts: Analysis and Improvement of the Resul...
Tools for Ontology Building from Texts: Analysis and Improvement of the Resul...Tools for Ontology Building from Texts: Analysis and Improvement of the Resul...
Tools for Ontology Building from Texts: Analysis and Improvement of the Resul...IOSR Journals
 
Big Data Standards - Workshop, ExpBio, Boston, 2015
Big Data Standards - Workshop, ExpBio, Boston, 2015Big Data Standards - Workshop, ExpBio, Boston, 2015
Big Data Standards - Workshop, ExpBio, Boston, 2015Susanna-Assunta Sansone
 
Ontology Mapping
Ontology MappingOntology Mapping
Ontology Mappingsamhati27
 
Metadata Analyser: measuring metadata quality
Metadata Analyser: measuring metadata qualityMetadata Analyser: measuring metadata quality
Metadata Analyser: measuring metadata qualityFrancisco Couto
 
Keystone Summer School 2015: Mauro Dragoni, Ontologies For Information Retrieval
Keystone Summer School 2015: Mauro Dragoni, Ontologies For Information RetrievalKeystone Summer School 2015: Mauro Dragoni, Ontologies For Information Retrieval
Keystone Summer School 2015: Mauro Dragoni, Ontologies For Information RetrievalMauro Dragoni
 
A Survey of Ontology-based Information Extraction for Social Media Content An...
A Survey of Ontology-based Information Extraction for Social Media Content An...A Survey of Ontology-based Information Extraction for Social Media Content An...
A Survey of Ontology-based Information Extraction for Social Media Content An...ijcnes
 
OOR--Open-Ontology-Repository--jun2010
OOR--Open-Ontology-Repository--jun2010OOR--Open-Ontology-Repository--jun2010
OOR--Open-Ontology-Repository--jun2010Peter Yim
 

Similar to Ontology Evaluation Methods: Assessing Correctness, Completeness, and Utility (20)

ISWC2023-McGuinnessTWC16x9FinalShort.pdf
ISWC2023-McGuinnessTWC16x9FinalShort.pdfISWC2023-McGuinnessTWC16x9FinalShort.pdf
ISWC2023-McGuinnessTWC16x9FinalShort.pdf
 
SWSN UNIT-3.pptx we can information about swsn professional
SWSN UNIT-3.pptx we can information about swsn professionalSWSN UNIT-3.pptx we can information about swsn professional
SWSN UNIT-3.pptx we can information about swsn professional
 
Knowledge Organization System (KOS) for biodiversity information resources, G...
Knowledge Organization System (KOS) for biodiversity information resources, G...Knowledge Organization System (KOS) for biodiversity information resources, G...
Knowledge Organization System (KOS) for biodiversity information resources, G...
 
A Comparative Study of Ontology building Tools in Semantic Web Applications
A Comparative Study of Ontology building Tools in Semantic Web Applications A Comparative Study of Ontology building Tools in Semantic Web Applications
A Comparative Study of Ontology building Tools in Semantic Web Applications
 
A Comparative Study Ontology Building Tools for Semantic Web Applications
A Comparative Study Ontology Building Tools for Semantic Web Applications A Comparative Study Ontology Building Tools for Semantic Web Applications
A Comparative Study Ontology Building Tools for Semantic Web Applications
 
A Comparative Study Ontology Building Tools for Semantic Web Applications
A Comparative Study Ontology Building Tools for Semantic Web Applications A Comparative Study Ontology Building Tools for Semantic Web Applications
A Comparative Study Ontology Building Tools for Semantic Web Applications
 
0810ijdms02
0810ijdms020810ijdms02
0810ijdms02
 
Iot ontologies state of art$$$
Iot ontologies state of art$$$Iot ontologies state of art$$$
Iot ontologies state of art$$$
 
A Comparative Study of Recent Ontology Visualization Tools with a Case of Dia...
A Comparative Study of Recent Ontology Visualization Tools with a Case of Dia...A Comparative Study of Recent Ontology Visualization Tools with a Case of Dia...
A Comparative Study of Recent Ontology Visualization Tools with a Case of Dia...
 
Knowledge Organization System (KOS) for biodiversity information resources, G...
Knowledge Organization System (KOS) for biodiversity information resources, G...Knowledge Organization System (KOS) for biodiversity information resources, G...
Knowledge Organization System (KOS) for biodiversity information resources, G...
 
A Framework for Ontology Usage Analysis
A Framework for Ontology Usage AnalysisA Framework for Ontology Usage Analysis
A Framework for Ontology Usage Analysis
 
Tools for Ontology Building from Texts: Analysis and Improvement of the Resul...
Tools for Ontology Building from Texts: Analysis and Improvement of the Resul...Tools for Ontology Building from Texts: Analysis and Improvement of the Resul...
Tools for Ontology Building from Texts: Analysis and Improvement of the Resul...
 
Big Data Standards - Workshop, ExpBio, Boston, 2015
Big Data Standards - Workshop, ExpBio, Boston, 2015Big Data Standards - Workshop, ExpBio, Boston, 2015
Big Data Standards - Workshop, ExpBio, Boston, 2015
 
Ontology Mapping
Ontology MappingOntology Mapping
Ontology Mapping
 
Metadata Analyser: measuring metadata quality
Metadata Analyser: measuring metadata qualityMetadata Analyser: measuring metadata quality
Metadata Analyser: measuring metadata quality
 
Seronto Process
Seronto ProcessSeronto Process
Seronto Process
 
Keystone Summer School 2015: Mauro Dragoni, Ontologies For Information Retrieval
Keystone Summer School 2015: Mauro Dragoni, Ontologies For Information RetrievalKeystone Summer School 2015: Mauro Dragoni, Ontologies For Information Retrieval
Keystone Summer School 2015: Mauro Dragoni, Ontologies For Information Retrieval
 
DR KL CV v5
DR KL CV v5DR KL CV v5
DR KL CV v5
 
A Survey of Ontology-based Information Extraction for Social Media Content An...
A Survey of Ontology-based Information Extraction for Social Media Content An...A Survey of Ontology-based Information Extraction for Social Media Content An...
A Survey of Ontology-based Information Extraction for Social Media Content An...
 
OOR--Open-Ontology-Repository--jun2010
OOR--Open-Ontology-Repository--jun2010OOR--Open-Ontology-Repository--jun2010
OOR--Open-Ontology-Repository--jun2010
 

More from Joanne Luciano

Luciano uvi hackfest.28.10.2020
Luciano uvi hackfest.28.10.2020Luciano uvi hackfest.28.10.2020
Luciano uvi hackfest.28.10.2020Joanne Luciano
 
Indiana University 2018 SICE summer camp slides
Indiana University 2018 SICE summer camp slidesIndiana University 2018 SICE summer camp slides
Indiana University 2018 SICE summer camp slidesJoanne Luciano
 
Why are some websites successful (at behavioral change) Informs International...
Why are some websites successful (at behavioral change) Informs International...Why are some websites successful (at behavioral change) Informs International...
Why are some websites successful (at behavioral change) Informs International...Joanne Luciano
 
The General Ontology Evaluation Framework (GOEF) & the I-Choose Use Case A ...
The General Ontology Evaluation Framework (GOEF) & the I-Choose Use CaseA ...The General Ontology Evaluation Framework (GOEF) & the I-Choose Use CaseA ...
The General Ontology Evaluation Framework (GOEF) & the I-Choose Use Case A ...Joanne Luciano
 
Luciano informs healthcare_2015 Nashville, TN USA July 30 2015
Luciano informs healthcare_2015 Nashville, TN USA July 30 2015Luciano informs healthcare_2015 Nashville, TN USA July 30 2015
Luciano informs healthcare_2015 Nashville, TN USA July 30 2015Joanne Luciano
 
Ontology Support for Influenza and Surveillance
Ontology Support for Influenza and Surveillance Ontology Support for Influenza and Surveillance
Ontology Support for Influenza and Surveillance Joanne Luciano
 
2013 dec 26_bgu_israel_seminar_l_luciano
2013 dec 26_bgu_israel_seminar_l_luciano2013 dec 26_bgu_israel_seminar_l_luciano
2013 dec 26_bgu_israel_seminar_l_lucianoJoanne Luciano
 
2013 dec bgu_israel_luciano_dec_22
2013 dec bgu_israel_luciano_dec_222013 dec bgu_israel_luciano_dec_22
2013 dec bgu_israel_luciano_dec_22Joanne Luciano
 
2013 dec bgu_israel_luciano_day_1_dec_22
2013 dec bgu_israel_luciano_day_1_dec_222013 dec bgu_israel_luciano_day_1_dec_22
2013 dec bgu_israel_luciano_day_1_dec_22Joanne Luciano
 
2013 dec bgu_israel_luciano_day_3_dec_25
2013 dec bgu_israel_luciano_day_3_dec_252013 dec bgu_israel_luciano_day_3_dec_25
2013 dec bgu_israel_luciano_day_3_dec_25Joanne Luciano
 
Translational Medicine: Patterns of Response to Antidepressant Treatment and ...
Translational Medicine: Patterns of Response to Antidepressant Treatment and ...Translational Medicine: Patterns of Response to Antidepressant Treatment and ...
Translational Medicine: Patterns of Response to Antidepressant Treatment and ...Joanne Luciano
 
Amia tbi 2010_pmi_luciano.ppt
Amia tbi 2010_pmi_luciano.pptAmia tbi 2010_pmi_luciano.ppt
Amia tbi 2010_pmi_luciano.pptJoanne Luciano
 
Luciano pr 08-849_ontology_evaluation_methods_metrics
Luciano pr 08-849_ontology_evaluation_methods_metricsLuciano pr 08-849_ontology_evaluation_methods_metrics
Luciano pr 08-849_ontology_evaluation_methods_metricsJoanne Luciano
 
Bio onttalk 30minutes-june2003[1]
Bio onttalk 30minutes-june2003[1]Bio onttalk 30minutes-june2003[1]
Bio onttalk 30minutes-june2003[1]Joanne Luciano
 
06317731 Patent page 1
06317731 Patent page 106317731 Patent page 1
06317731 Patent page 1Joanne Luciano
 
Bio it 2005_rdf_workshop05
Bio it 2005_rdf_workshop05Bio it 2005_rdf_workshop05
Bio it 2005_rdf_workshop05Joanne Luciano
 
The Translational Medicine
The Translational MedicineThe Translational Medicine
The Translational MedicineJoanne Luciano
 

More from Joanne Luciano (19)

Luciano uvi hackfest.28.10.2020
Luciano uvi hackfest.28.10.2020Luciano uvi hackfest.28.10.2020
Luciano uvi hackfest.28.10.2020
 
Indiana University 2018 SICE summer camp slides
Indiana University 2018 SICE summer camp slidesIndiana University 2018 SICE summer camp slides
Indiana University 2018 SICE summer camp slides
 
Why are some websites successful (at behavioral change) Informs International...
Why are some websites successful (at behavioral change) Informs International...Why are some websites successful (at behavioral change) Informs International...
Why are some websites successful (at behavioral change) Informs International...
 
The General Ontology Evaluation Framework (GOEF) & the I-Choose Use Case A ...
The General Ontology Evaluation Framework (GOEF) & the I-Choose Use CaseA ...The General Ontology Evaluation Framework (GOEF) & the I-Choose Use CaseA ...
The General Ontology Evaluation Framework (GOEF) & the I-Choose Use Case A ...
 
Luciano informs healthcare_2015 Nashville, TN USA July 30 2015
Luciano informs healthcare_2015 Nashville, TN USA July 30 2015Luciano informs healthcare_2015 Nashville, TN USA July 30 2015
Luciano informs healthcare_2015 Nashville, TN USA July 30 2015
 
Ontology Support for Influenza and Surveillance
Ontology Support for Influenza and Surveillance Ontology Support for Influenza and Surveillance
Ontology Support for Influenza and Surveillance
 
2013 dec 26_bgu_israel_seminar_l_luciano
2013 dec 26_bgu_israel_seminar_l_luciano2013 dec 26_bgu_israel_seminar_l_luciano
2013 dec 26_bgu_israel_seminar_l_luciano
 
2013 dec bgu_israel_luciano_dec_22
2013 dec bgu_israel_luciano_dec_222013 dec bgu_israel_luciano_dec_22
2013 dec bgu_israel_luciano_dec_22
 
2013 dec bgu_israel_luciano_day_1_dec_22
2013 dec bgu_israel_luciano_day_1_dec_222013 dec bgu_israel_luciano_day_1_dec_22
2013 dec bgu_israel_luciano_day_1_dec_22
 
2013 dec bgu_israel_luciano_day_3_dec_25
2013 dec bgu_israel_luciano_day_3_dec_252013 dec bgu_israel_luciano_day_3_dec_25
2013 dec bgu_israel_luciano_day_3_dec_25
 
Translational Medicine: Patterns of Response to Antidepressant Treatment and ...
Translational Medicine: Patterns of Response to Antidepressant Treatment and ...Translational Medicine: Patterns of Response to Antidepressant Treatment and ...
Translational Medicine: Patterns of Response to Antidepressant Treatment and ...
 
Amia tbi 2010_pmi_luciano.ppt
Amia tbi 2010_pmi_luciano.pptAmia tbi 2010_pmi_luciano.ppt
Amia tbi 2010_pmi_luciano.ppt
 
Luciano pr 08-849_ontology_evaluation_methods_metrics
Luciano pr 08-849_ontology_evaluation_methods_metricsLuciano pr 08-849_ontology_evaluation_methods_metrics
Luciano pr 08-849_ontology_evaluation_methods_metrics
 
Bio onttalk 30minutes-june2003[1]
Bio onttalk 30minutes-june2003[1]Bio onttalk 30minutes-june2003[1]
Bio onttalk 30minutes-june2003[1]
 
06063028 face page
06063028 face page06063028 face page
06063028 face page
 
06317731 Patent page 1
06317731 Patent page 106317731 Patent page 1
06317731 Patent page 1
 
Bio it 2005_rdf_workshop05
Bio it 2005_rdf_workshop05Bio it 2005_rdf_workshop05
Bio it 2005_rdf_workshop05
 
Luciano phddefense
Luciano phddefenseLuciano phddefense
Luciano phddefense
 
The Translational Medicine
The Translational MedicineThe Translational Medicine
The Translational Medicine
 

Ontology Evaluation Methods: Assessing Correctness, Completeness, and Utility

  • 1. Ontology Evaluation: Methods and Metrics MITRE Research Interest Dr. Joanne S. Luciano In collaboration with Dr. Leo Obrst, PhD Suzette Stoutenburg Kevin Cohen Jean Stanford Approved for Public Release. Distribution Unlimited. Case Number: 08 08-0849 1
  • 2. Ontology: A Key Technology for Knowledge Management Used to describe terms in a vocabulary and the relationships among them. Ontology languages vary in their semantic expressiveness. Based on work by Leo Obrst of MITRE as interpreted by Dan McCreary. This can be viewed as a trade-off of semantic clarity v. the time and money it takes to construct http://www.mkbergman.com/?m=20070516. Ontologies have become the most widespread form of knowledge representation for multiple purposes Ontology Summit 2007 (NIST, Gaithersburg, MD, April 23-24, 2007) 2
  • 3. The Problem Ontology Elephants There is no single real elephant An elephant is abstract There must be a purpose for an elephant: use cases? An elephant is very abstract There must be an upper elephant An elephant is really very An elephant is the simple result of consensus There are only Open vs. distributed elephants Closed & their mappings Elephant Prospects and Possibilities for Ontology Evaluation: The View from NCOR NCOR, Obrst, Hughes, Ray, WWW 2006, May 22 22–26, 2006, Edinburgh, UK. 3
  • 4. The Problem Users need to be able to build sound ontologies and to reuse ontologies for different purposes. There is no standard no way to do that now. At the Ontology Summit 2008 two competing “state of the art” evaluation proposals for the Open Ontology Repository were presented. Both treat ontologies as “black boxes” and both are subjective evaluations. – Peer Review – self appointed editorial review board decides, non non-overlapping domain so first one gets preference, ‘best practices’ – User Ratings – users report their experience and rate them on a website Language was inserted into the communiqué to provide mechanisms that enable ontology evaluation by other metrics Prior work on evaluating ontologies is limited, no consensus - Ontology Workshop Methods and Metrics October 2007 Workshop materials at: http://sites.google.com/a/cme.nist.gov/workshop ://sites.google.com/a/cme.nist.gov/workshop-on-ontology-evaluation/ – Formal logic based applications can use software reasoners to address logical consistency and classification; many don’t take advantage of these tools – Natural Language Processing (NLP) application use NLP evaluation methods, but address only NLP ons applications (text mark-up – aka “annotation”), information retrieval and extraction – Alignment (mapping of ontologies) for data mining, integration, fusion Ontology Summit 2007 (NIST, Gaithersburg, MD, April 23-24, 2007 See slide at end with notes. 24, 2007) 4
  • 5. Objective: Evolve toward Science & Engineering Discipline for Ontology • Create procedures, processes, methods to help define, adjudicate, & ensure quality of knowledge capture/representation • Facilitate the education of communities on ontology development & promote best practices for ontology development • Enable the best standards related to ontologies, & promote linkages, liaisons among standards organizations 5
  • 6. Approach: • Two stages: – Recast use case into its components: Novel • Functional objective Approach • Design objective & requirements specification • Semantic components required to achieve above – Evaluate components using objective metrics • Place existing evaluation methods in context by utility • Engage and rally the community / stakeholders – Participate at appropriate meetings: present work and facilitate community focus on objective metrics for evaluation – Introduce users with complementary skills, joint vision, shared needs to develop needed metrics, content, tests, tools – Involve multiple government agencies, industry and academia to support initiative 6
  • 7. Research Plan: (1) Identify use cases For each, recast the use cases into their components: a. Specify functional objective (what it is, what it does) e.g. enable investigation of data collected on influenza strain mutations that cause death in birds b. Specify design objective (how good it has to be e.g., what specifications have to be met? Is it a prototype, for commercial use, or must it meet military specifications?) e.g. Must meet: Minimum Information about an Influenza Genotype and a Nomenclature Standard (MIIGNS) c. Identify (or specify) the semantic components required to achieve the functional objective to the level specified by the design objective (Authoritative Sources such as engineering tolerances, physical constants, legal jurisdictions, company policies) e.g. To meet MIIGNS, the following semantic components must be included: biomaterial transformations assays data transformations 7
  • 8. Research Plan (2) Develop Metrics Develop metrics for 3 criteria for evaluation: 1. Correctness: how well the f e functional components express the design objectives a) Language expressiveness b) Fluency / competency 2. Completeness: combines use case with design criteria (requirement specification) a) To what extent can requirements be met? b) Which semantic components (authoritative sources) are needed/missing? nts 3. Utility: Is it useful? Does it work? Combine 1 and 2 (correctness and completeness) and evaluate against the use case (by competency questions, or other challenge tests). 8
  • 9. Examples: • BioPAX (prior work) • Habitat-Lite (subset of Environmental Lite Ontology to support of NSF funded Mining Metadata for Metagenomics) • Influenza Infectious Disease Ontology (for Genomics for Bioforensics MSR) 9
  • 10. Example (1) BioPAX lack of fluency chemical structure & pathway steps incorrectly modeled - misunderstanding of the language (language has capability) - modeled disjoint from the biology & chemistry - leads to logical inconsistency OWL has a steep learning curve it’s easy to get things wrong. rve, 10
  • 11. Example (2) Habitat- -Lite: correctness & completeness Objective: facilitate capture of habitat and environmental metadata on genomic sequences Approach: select subset of terms with highest frequency and evaluate usefulness by correctness and completeness metrics – Evaluated correctness • 64% agreement (84 of 132 terms) of automated and expert mapping of terms – Evaluated coverage of terms • 84% exact matches (“host,” “aquatic,” and “soil” covered 75%) Hirschman, Clark, Cohen, Mardis, Luciano, Kottmann, Cole, Markowitz, Kyprpides, Field Habitat-Lite: a GSC Case Study Based on Free Text Terms for Environmental Metadata Lite: OMICS A Journal of Integrative Biology Volume 12, Number 2, 2008 (in press) 11
  • 12. Example (3) Enable Influenza Research (proposed construction and subsequent evaluation) Function: enable investigation of data collected on influenza strain mutations that cause death in birds Design objective: Minimum Information about an Influenza : Genotype and a Nomenclature Standard (MIIGNS) Semantic components: 1. biomaterial transformations a. recombinant plasmid biomaterial transformation b. site-directed mutagenesis biomaterial transformation directed c. reverse genetic virus production biomaterial transformation d. Mouse infection biomaterial transformation 2. assays a. weight assay b. virus replication / mouse lung assay c. Cytokine quantification assay 3. data transformations a. statistical difference evaluation 12
  • 13. Example (3) Enable Influenza Research (proposed construction and subsequent evaluation) Correctness: Language expressivity: validate definitions against OBO Foundry relations Fluency: inter-developer agreement (3 developers, 2 developer code same, 3rd validates) Completeness: Calculate % coverage of minimum terms (18 terms) Calculate % coverage of full terms (196 terms) Utility: Challenge Questions To be developed (by our collaborator BioHealthBase) 13
  • 14. Impact Communities of Practice areas need objective methods and Metrics to facilitate the development, interoperability and reuse of their ontologies Some examples: – Message Based Data Exchange – BioSecurity – Health Care and Biomedicine – Life Sciences – Disease – Metagenomics – Agile systems for rapid enterprise integration of heterogeneous data – Intelligence Community 14
  • 15. Why MITRE? MITRE is uniquely positioned to act as an impartial experimental designer and arbitrator in the development of an evaluation methodology for ontologies – MITRE has acted in the past for natural language technologies such as text summarization in the Text REtrieval Conferences (TREC) and Information Extraction in the BioCreAtIvE challenge 15
  • 16. Additional Notes on Specific Slides Slide 1: Lower right gaphic: Ontology Summit 2007 (NIST, Gaithersburg, MD, April 23 23-24, 2007) effort. See the following: Ontology Summit 2007 - Ontology, Taxonomy, Folksonomy: Understanding the Distinctions. http://ontolog.cim3.net/cgi- : bin/wiki.pl?OntologySummit2007. Ontology Summit 2007 Communique. http://ontolog.cim3.net/cgi http://ontolog.cim3.net/cgi-bin/wiki.pl?OntologySummit2007_Communique. Ontology Summit 2007 Ontology Dimensions Map. http://ontolog.cim3.net/cgi http://ontolog.cim3.net/cgi- bin/wiki.pl?OntologySummit2007_FrameworksForConsideration/DimensionsMap. bin/wiki.pl?OntologySummit2007_FrameworksForConsideration/DimensionsMap Gruninger, Michael; Olivier Bodenreider; Frank Olken; Leo Obrst; Peter Yim. 2008. The 2007 Ontology Summit Joint , Communiqué. Ontology, Taxonomy, Folksonomy: Understanding the Distinctions. Journal of Applied Ontology, forthcoming. : Slide 2: Ontology Summit 2008 (NIST, Gaithersburg, MD, April 28 28-29, 2008). Ontology Summit 2008: Toward an Open Ontology Repository. http://ontolog.cim3.net/cgi http://ontolog.cim3.net/cgi-bin/wiki.pl?OntologySummit2008. Ontology Summit 2008 Communique. http://ontolog.cim3.net/cgi http://ontolog.cim3.net/cgi-bin/wiki.pl?OntologySummit2008_Communique. Slide 4: Concerning: “At the Ontology Summit 2008 two competing “state of the art” evaluation proposals for the Open Ontology Repository were presented. Both treat ontologies as “black boxes” and both are subjective evaluations evaluations. –Peer Review –self appointed editorial review board decides, non self non-overlapping domain so first one gets preference, „best practices‟ –User Ratings –users report their experience and rate them on a website users These points were made during the summit and discussed more fully during the Quality and Gatekeeping session of the Ontology Summit 2008 (NIST, Gaithersburg, MD, April 28-29, 2008 29, 2008. 16
  • 17. Background References The BioCreAtIvE (Critical Assessment of Information Extraction systems in Biology) challenge evaluation consists of a communi community-wide effort for evaluating text mining and information extraction systems applied to the biological domain. BioCreative (http://biocreative.sourceforge.net/) BioCreative. The Text REtrieval Conference (TREC), co-sponsored by the National Institute of Standards and Technology (NIST) and U.S. Department of Defense, wa started in sponsored was 1992 as part of the TIPSTER Text program. Its purpose was to support research within the information retrieval community by p providing the infrastructure necessary for large-scale evaluation of text retrieval methodologies. (http://trec.nist.gov/overview.html http://trec.nist.gov/overview.html) The Message Understanding Conferences (MUC) were initiated and financed by DARPA to encourage the development of new and bett methods of information better extraction. The character of this competition -- many concurrent research teams competing against one another -- necessitated the development of standards for evaluation, e.g. the adoption of recall and precision. (http://en.wikipedia.org/wiki/Message_Understanding_Conference http://en.wikipedia.org/wiki/Message_Understanding_Conference) Lenat, Douglas B. “CYC: a large-scale investment in knowledge infrastructure,” Communications of the ACM Volume 38 , Issue 11 (November 1995) Pages: 33 – 38. ACM, Project Halo (http://www.projecthalo.com/), is a project funded by Paul Allen's Vulcan Ventures. The project was initially led by Oliver Roup and Noah Friedland but is ), currently led by Mark Greaves, a former DARPA Program Manager. Project Halo is an attempt to apply Artificial Intelligence te techniques to the problem of producing a "digital Aristotle" that might serve as a mentor, providing comprehensive access to the world's knowledge ( (http://en.wikipedia.org/wiki/Project_Halo). Ontoprise is a commerial software provider of ontology-based solutions. (http://www.ontoprise.de/content/index_eng.html http://www.ontoprise.de/content/index_eng.html) Maedche, Alexander and Staab, Steffen. Ontology learning for the semantic web. Special Issue on Semantic Web. IEEE Intelligent Systems, 16(2):72 16(2):72-79, MAR 2001. López, M. F.; Gómez-Pérez, A.; Sierra, J. P. & Sierra, A. P. Building a chemical ontology using Methontology and the OntologyDesign Environment IEEE Intelligent , Systems and Their Applications, 1999, 14, 37-46 Oltramari, A.; Gangemi, A.; Guarino, N. & Masolo, C. Restructuring WordNet's Top- -Level: The OntoClean approach Proceedings of the Workshop OntoLex'2, Ontologies and Lexical Knowledge Bases, 2002 Guarino, N. & Welty, C. Evaluating ontological decisions with OntoClean Commun. ACM, ACM Press, 2002, 45, 61 . 61-65 Smith, B.; Williams, J. & Schulze-Kremer, S. The ontology of the gene ontology. AMIA Annu Symp Proc, Institute for Formal Ontology and Medical Information Science, Kremer, University of Leipzig., 2003, 609-613. Smith, B. From concepts to clinical reality: an essay on the benchmarking of biomedical terminologies. J Biomed Inform, Depar Department of Philosophy and National Center for Biomedical Ontology, University at Buffalo, Buffalo, NY 14260, USA. phismith@buffalo.edu, 2006, 39, 288 288-298 Obrst, Leo; Todd Hughes; Steve Ray. 2006. Prospects and Possibilities for Ontology Evaluation: The View from NCOR. Workshop o Evaluation of Ontologies for the on Web (EON2006), Edinburgh, UK, May 22, 2006. Obrst, Leo; Werner Ceusters; Inderjeet Mani; Steve Ray; Barry Smith. 2007 The Evaluation of Ontologies: Toward Improved Semantic Interoperability. Chapt in: ; Chapter Semantic Web: Revolutionizing Knowledge Discovery in the Life Sciences, Christopher J. O. Baker and Kei Kei-Hoi Cheung, Eds., Springer, 2007. Gangemi, Aldo; Carola Catenacci1; Massimiliano Ciaramita1; Jos Lehmann; contributions by: Rosa Gil (in section 2.2). Francesco Bolici and Onofrio Strignano (in section 2.4). 2004. Ontology evaluation and validation: An integrated formal model for the quality diagnostic task. ). OntoEval 2004. Gangemi, A.; Catenacci, C.; Ciaramita, M.; Lehmann, J. 2005. A Theoretical Framework for Ontology Evaluation and Validation. In Proceedings of SWAP2005. http://www.loa-cnr.it/Papers/swap_final_v2.pdf Gangemi, Aldo; Carola Catenacci; Massimiliano Ciaramita; and Jos Lehmann. 2006. Modelling ontology evaluation and validation. To appear in Proceedings of ESWC2006, Springer. Lawrence Hunter, Mike Bada, K. Bretonnel Cohen, Helen Johnson, William Baumgartner, Jr. and Philip V. Ogren. "Ontology Quality Metrics," Center for Computational Pharmacology University of Colorado School of Medicine, October 8, 2007. Lynette Hirschman , Jong C. Park , Junichi Tsujii , Limsoon Wong , and Cathy H. Wu. Accomplishments and challenges in literature data mining for biology. Bioinformatics 18: 1553-1561. Proceedings of NIST 2007 Automatic Content Extraction Workshop (ACE), 2007. http://www.nist.gov/speech/tests/ace/ace07/ http://www.nist.gov/speech/tests/ace/ace07/. Methods and Metrics for Ontology Evaluation Workshop (Sponsors: NIST and NIH), National Institute of Standards and Technology Gaithersburg, MD, October 25- Technology, 26. Mani, I., B. Sondheim, D. House, L. Obrst. 1998. TIPSTER Text Summarization Evaluation: Final Report, MITRE technical report, Reston, VA, September, 1998. Proceedings of NIST 2007 Automatic Content Extraction Workshop (ACE), 2007. http://www.nist.gov/speech/tests/ace/ace07/. 17
  • 18. Background: Prior Technical Approaches • Evaluation in use - Navigli et al. 2003, Porzel and Malaka 2005 – Best case: Halo Project - Friedland et al. 2004 • Data-driven evaluation - essentially the fit between the ontology and a knowledge source e.g. a corpus - Brewster et al. 2004 • Gold Standard approaches, very common, used by for example, Cimiano et al. 2005 – Major problem is the arbitrary choice of an ontology – Dellschaft and Staab 2006, proposed a method to derive IR/NLP type Precision/Recall/F Precision/Recall/F-Measure 18
  • 19. Background : Philosophical and Methodological Approaches Methontology approach of Gomez Gomez-Perez: • Focus on evaluating procedural or formative aspects ontology construction • Criteria included: Consistency, Completeness, Conciseness, Expandability • Some tools developed reflecting these approaches: o Lam et al. 2004, Knublauch et al. 2004, Alani 2005, 2006 OntoClean approach of Guarino and Welty: • Philosophical approach based on theoretical principles: • Proposed a set of "metaproperties": o Rigidity o Identity o Unity • Much focus on "cleaning up" existing "ontologies" such as WordNet so as to make them more rigororous 19