This document provides an overview of a tutorial on data integration and open information systems. It discusses the goals of the semantic web and linked data, which aim to create a universal medium for data exchange by publishing and connecting structured data on the web. Currently, web APIs allow access to data but use different data models and formats. Linked data uses common RDF standards and links entities to enable querying across diverse domains and data sources, forming a global data web.
This document contains information about current and past projects and members of the Sina Institute. For current projects, it lists the names, descriptions, and URLs of 7 active projects. For past projects, it lists the name and URL of 1 past project. It also provides details on current members, including names, titles, email addresses, and profiles for 4 members.
The document discusses Oracle Semantic Technologies for storing and querying RDF data. It provides an overview of how RDF data is stored and organized in Oracle databases using ID triples and URI mapping tables. It describes how the SEM_MATCH SQL function allows querying RDF data using a SPARQL-like syntax. Optimization techniques for SEM_MATCH queries include indexes and materialized views. The core entities in the Oracle Semantic Store include semantic networks, models, rulebases, and entailments. Functionality includes bulk loading, incremental loading, SPARQL querying, and built-in or user-defined inference rules.
This document provides an overview of the Web Ontology Language (OWL). OWL is built on top of RDF and is used to process information on the web by computers. It allows for stronger constraints and rules than RDF. There are three sublanguages of OWL with varying expressiveness. OWL is written in XML and is a W3C standard, making it suitable for exchanging and processing web information across different systems.
This tutorial discusses the Web Ontology Language (OWL). OWL is built on top of RDF and is used to process information on the web by computers. It allows for stronger constraints and rules than RDF. There are three sublanguages of OWL with varying expressiveness. OWL is written in XML and is a W3C standard for representing ontologies on the semantic web.
This document provides an overview of storing Resource Description Framework (RDF) graphs in relational database management systems. Specifically:
- RDF represents data as subject-predicate-object triples that form a directed graph. This triples-based data model allows for easy data integration.
- RDF graphs are typically stored as a single subject-predicate-object table in a relational database for persistent storage.
- Queries to retrieve and manipulate data in the RDF graph can then be performed using SQL on this table.
1) The document discusses RDFa, which bridges the gap between the Web of Documents and the Web of Data (Web 3.0) by embedding RDF triples inside XHTML pages.
2) An example shows how to annotate HTML with RDFa by adding attributes to designate namespaces, types of entities (people), and properties (name, nickname, affiliation).
3) Nesting is used to include one entity (address) within another (person), with the relationship indicated by the "rel" attribute. A graph is drawn to represent the RDF data.
Pal gov.tutorial2.session13 3.data integration and fusion using rdfMustafa Jarrar
This document discusses data integration and fusion using RDF. It provides an example of integrating data from three governmental agencies about companies by transforming each database into RDF and then concatenating the RDF graphs. Entities are linked across datasets using URIs and properties like owl:sameAs. Integrating RDF data in this way provides an integrated view where applications can query the datasets as a single database. References are provided on linked data and the emerging web of linked data.
This document provides information about a practical session on building an ontology for a national student registry in Palestine. The session aims to have students specify a shared ontology in OWL that universities can use to exchange student profile data in RDF format. Each student will work individually to extend their existing RDF model with OWL constructs to better define and constrain the semantics of the exchanged data. Students will then present their OWL model and build example RDF data files from two universities to validate the model.
This document contains information about current and past projects and members of the Sina Institute. For current projects, it lists the names, descriptions, and URLs of 7 active projects. For past projects, it lists the name and URL of 1 past project. It also provides details on current members, including names, titles, email addresses, and profiles for 4 members.
The document discusses Oracle Semantic Technologies for storing and querying RDF data. It provides an overview of how RDF data is stored and organized in Oracle databases using ID triples and URI mapping tables. It describes how the SEM_MATCH SQL function allows querying RDF data using a SPARQL-like syntax. Optimization techniques for SEM_MATCH queries include indexes and materialized views. The core entities in the Oracle Semantic Store include semantic networks, models, rulebases, and entailments. Functionality includes bulk loading, incremental loading, SPARQL querying, and built-in or user-defined inference rules.
This document provides an overview of the Web Ontology Language (OWL). OWL is built on top of RDF and is used to process information on the web by computers. It allows for stronger constraints and rules than RDF. There are three sublanguages of OWL with varying expressiveness. OWL is written in XML and is a W3C standard, making it suitable for exchanging and processing web information across different systems.
This tutorial discusses the Web Ontology Language (OWL). OWL is built on top of RDF and is used to process information on the web by computers. It allows for stronger constraints and rules than RDF. There are three sublanguages of OWL with varying expressiveness. OWL is written in XML and is a W3C standard for representing ontologies on the semantic web.
This document provides an overview of storing Resource Description Framework (RDF) graphs in relational database management systems. Specifically:
- RDF represents data as subject-predicate-object triples that form a directed graph. This triples-based data model allows for easy data integration.
- RDF graphs are typically stored as a single subject-predicate-object table in a relational database for persistent storage.
- Queries to retrieve and manipulate data in the RDF graph can then be performed using SQL on this table.
1) The document discusses RDFa, which bridges the gap between the Web of Documents and the Web of Data (Web 3.0) by embedding RDF triples inside XHTML pages.
2) An example shows how to annotate HTML with RDFa by adding attributes to designate namespaces, types of entities (people), and properties (name, nickname, affiliation).
3) Nesting is used to include one entity (address) within another (person), with the relationship indicated by the "rel" attribute. A graph is drawn to represent the RDF data.
Pal gov.tutorial2.session13 3.data integration and fusion using rdfMustafa Jarrar
This document discusses data integration and fusion using RDF. It provides an example of integrating data from three governmental agencies about companies by transforming each database into RDF and then concatenating the RDF graphs. Entities are linked across datasets using URIs and properties like owl:sameAs. Integrating RDF data in this way provides an integrated view where applications can query the datasets as a single database. References are provided on linked data and the emerging web of linked data.
This document provides information about a practical session on building an ontology for a national student registry in Palestine. The session aims to have students specify a shared ontology in OWL that universities can use to exchange student profile data in RDF format. Each student will work individually to extend their existing RDF model with OWL constructs to better define and constrain the semantics of the exchanged data. Students will then present their OWL model and build example RDF data files from two universities to validate the model.
Pal gov.tutorial2.session12 1.the problem of data integrationMustafa Jarrar
This document discusses the problem of data integration. It provides examples of data integration challenges from the government domain, including different agencies registering different but related information about the same business. The main challenges of data integration discussed are heterogeneities in database schemas, including name and meaning heterogeneities, differences in structure and type, different rules and constraints, and differences in data models used. Addressing these heterogeneities is a key challenge in achieving effective data integration.
The document provides an overview of XML schemas, which are used to define and validate the structure of XML documents. Some key points:
- XML schemas allow defining data types and content models for XML elements and attributes. They provide better features than DTDs for describing documents.
- An XML schema is defined in its own XML document with the .xsd file extension. It can then be referenced by XML documents to validate their structure.
- The root <schema> element defines the schema. It contains definitions of elements, attributes, data types etc. Namespaces are used to identify element types.
- Benefits of schemas over DTDs include support for namespaces, built-in and custom data types, and
Pal gov.tutorial2.session14.lab rdf-dataintegrationMustafa Jarrar
This document provides information about a practical session on integrating and fusing heterogeneous data using RDF. Students will work in groups of two, with each group containing students from different universities. Each group will construct three hypothetical student record databases based on different university data schemes. They will populate the databases with sample data and then integrate and fuse all the data into a single RDF dataset. Students are expected to use existing ontologies and write SPARQL queries on the integrated RDF data. The final deliverable will include snapshots of the original databases, the RDF mappings, the integrated RDF dataset, executed SPARQL queries and results.
Pal gov.tutorial2.session13 1.data schema integrationMustafa Jarrar
This document discusses data schema integration, which involves identifying correspondences between different data schemas and resolving conflicts between them to create an integrated schema. It describes challenges in schema integration including identifying corresponding concepts and analyzing conflicts. It then presents a generic framework for schema integration involving schema transformation, schema matching to identify correspondences, and integration and mapping generation to create the integrated schema and mappings. Finally, it provides examples of different types of conflicts and integration methods.
The document discusses XML document type definitions (DTDs). It provides an overview of DTDs, explaining that they define the legal structure and elements of an XML document. It describes how DTDs can be declared internally or externally. The document also explains how to write internal and external DTD declarations, and how external DTD declarations can reference external files using system identifiers.
Pal gov.tutorial2.session13 2.gav and lav integrationMustafa Jarrar
This document discusses Global-As-View (GAV) and Local-As-View (LAV) integration approaches. GAV defines the global schema in terms of the local schemas by writing views over the local schemas. LAV defines the local schemas in terms of the global schema by writing views from the global schema to the local schemas. The document provides an example of each approach and discusses how queries are executed differently under GAV versus LAV.
Pal gov.tutorial2.session4.lab xml document and schemasMustafa Jarrar
1) The Palestinian Ministry of Education is building a National Student Registry to centrally manage student profiles and academic records as students transfer between universities.
2) Each semester, universities must send academic data for every student to the ministry to update and integrate into the registry.
3) The practical session involves building the necessary XML documents and schemas to collect student data from university databases for the registry as required by the ministry.
This document provides an overview of RDF Schema (RDFS), which extends RDF to enable the description of classes and properties. RDFS defines classes using rdfs:Class and subclasses using rdfs:subClassOf. Properties are instances of rdf:Property, and their domains and ranges can be specified using rdfs:domain and rdfs:range. RDFS provides a basic vocabulary for defining other application-specific vocabularies through classes, subclasses, and properties.
Pal gov.tutorial2.session12 2.architectural solutions for the integration issuesMustafa Jarrar
This document discusses two main architectural solutions for data integration issues: application-driven integration and data-driven integration. Application-driven integration uses middleware like web services or publish/subscribe architectures. Data-driven integration uses techniques like data consolidation, data warehousing, or virtual data integration to reconcile data schemas and queries. The document provides examples of architectures for each approach.
This document provides an introduction to the Resource Description Framework (RDF). RDF is a framework for describing resources on the web using Uniform Resource Identifiers (URIs) and properties. It represents data as a directed labeled graph consisting of triples with a subject, predicate, and object. Examples are provided to demonstrate how RDF can be used to describe resources and their properties. Key concepts explained include URIs, triples, and representing RDF data as a graph.
Pal gov.tutorial2.session1.xml basics and namespacesMustafa Jarrar
The document discusses XML (Extensible Markup Language) basics and namespaces. It provides an overview of XML, describing it as a protocol for containing and managing information by allowing users to create their own markup languages. The document also discusses the need for namespaces to avoid conflicts between element names and introduces the syntax for using namespaces, which involves associating namespaces with prefixes.
This document provides an outline for a tutorial on data integration and open information systems. The tutorial consists of 16 sessions over a total of approximately 40 hours. It will cover topics such as XML, RDF, OWL, data integration, linked data, and the semantic web. The intended learning outcomes include understanding data models, semantic web languages, integrating and querying heterogeneous data using techniques such as SPARQL and RDF. Students will gain practical skills in tools like Oracle Semantic Technology and Virtuoso for storing and querying RDF data. Attendance is mandatory for all sessions.
Pal gov.tutorial4.session1 1.needforsharedsemanticsMustafa Jarrar
This document provides an overview of the need for shared semantics and ontologies. It discusses how a lack of shared understanding limits communication between people and systems. XML provides syntax but not semantics, while standard vocabularies are often ambiguous. Ontologies aim to formally specify meaning and resolve discrepancies in semantics through shared conceptualizations. They allow computers and people to communicate meaningfully.
Producing, Publishing and Consuming Linked Data Three lessons from the Bio2RD...François Belleau
The document discusses three lessons learned from the Bio2RDF project about producing, publishing, and consuming linked data. Lesson 1 is that data transformation to RDF is an ETL task best done using frameworks like Talend. Lesson 2 is to publish semantic data using triplestores like Virtuoso and make SPARQL endpoints publicly available. Lesson 3 is that semantic data sources can be consumed in various ways, including SPARQL queries, HTTP requests, and SOAP services.
Pal gov.tutorial3.session2.xml ns and schemaMustafa Jarrar
Here are some examples of information item types in an XML document:
- The <book> element is an element information item that contains child elements and attributes.
- The "title" attribute of the <book> element is an attribute information item.
- The text "Learning XML" between the <title> tags is a character data information item.
- A comment like <!-- This is a book details --> is a comment information item.
- The XML declaration at the top <?xml version="1.0"?> provides version and encoding details and is also an information item.
- The document type declaration <!DOCTYPE book> is a document type information item.
- The entire XML document
Project Places, Periods and Properties for Linked Science discusses linking scientific data from the University of Münster (WWU) using linked open data principles. Currently, WWU's knowledge is siloed across organizations, terminology, formats and literature. The project aims to open these silos by allowing researchers to easily publish, find and reuse data and models. It will interconnect data about people, publications, projects, locations and time periods. This will make WWU's knowledge more transparent, accessible and foster transdisciplinary research. Several use cases are discussed, including archaeology, bioinformatics and social science interviews.
A Web 2.0 Personal Learning Environment for Classical Chinese PoetryRalf Klamma
A Web 2.0 Personal Learning Environmentfor Classical Chinese Poetry
Yiwei Cao 曹怡蔚, Ralf Klamma, Yan Gao 高岩, Rynson W.H. Lau 劉永雄, and Matthias Jarke
Informatik 5, RWTH Aachen University
Department of Computer Science, City University of Hong Kong
Aachen, Germany
ICWL 2009
20.08.2009
Learning Analytics for the Lifelong Long Tail LearnerRalf Klamma
Learning Analytics for the Lifelong Long Tail Learner
Ralf Klamma
RWTH Aachen University
Informatik 5 (DBIS)
CELSTEC, Heerlen, The Netherlands
February 24, 2011
Multimedia Processing on Multimedia Semantics and Multimedia ContextRalf Klamma
The 10thWorkshop on Multimedia Metadata (SeMuDaTe‘09)
Yiwei Cao, Ralf Klamma, and Dejan KovachevI
Informatik 5 (Information Systems), RWTH Aachen University
2.12.2009
Graz, Austria
Easysoft Limited do data access, we love data, nothing scares us.
Whether you have a data integration, data migration or data synchronisation project you need help with, get in touch with us sales@easysoft.com
The document discusses different approaches to integrating information from multiple systems, including:
1. Providing a uniform logical view of distributed data through approaches like mediated query systems, portals, federated database systems, and web services.
2. Realizing a common data storage through data warehouses and operational data stores that load and aggregate data from multiple sources.
3. Achieving integration through applications like workflow management systems that coordinate interactions between different systems and users.
Pal gov.tutorial2.session12 1.the problem of data integrationMustafa Jarrar
This document discusses the problem of data integration. It provides examples of data integration challenges from the government domain, including different agencies registering different but related information about the same business. The main challenges of data integration discussed are heterogeneities in database schemas, including name and meaning heterogeneities, differences in structure and type, different rules and constraints, and differences in data models used. Addressing these heterogeneities is a key challenge in achieving effective data integration.
The document provides an overview of XML schemas, which are used to define and validate the structure of XML documents. Some key points:
- XML schemas allow defining data types and content models for XML elements and attributes. They provide better features than DTDs for describing documents.
- An XML schema is defined in its own XML document with the .xsd file extension. It can then be referenced by XML documents to validate their structure.
- The root <schema> element defines the schema. It contains definitions of elements, attributes, data types etc. Namespaces are used to identify element types.
- Benefits of schemas over DTDs include support for namespaces, built-in and custom data types, and
Pal gov.tutorial2.session14.lab rdf-dataintegrationMustafa Jarrar
This document provides information about a practical session on integrating and fusing heterogeneous data using RDF. Students will work in groups of two, with each group containing students from different universities. Each group will construct three hypothetical student record databases based on different university data schemes. They will populate the databases with sample data and then integrate and fuse all the data into a single RDF dataset. Students are expected to use existing ontologies and write SPARQL queries on the integrated RDF data. The final deliverable will include snapshots of the original databases, the RDF mappings, the integrated RDF dataset, executed SPARQL queries and results.
Pal gov.tutorial2.session13 1.data schema integrationMustafa Jarrar
This document discusses data schema integration, which involves identifying correspondences between different data schemas and resolving conflicts between them to create an integrated schema. It describes challenges in schema integration including identifying corresponding concepts and analyzing conflicts. It then presents a generic framework for schema integration involving schema transformation, schema matching to identify correspondences, and integration and mapping generation to create the integrated schema and mappings. Finally, it provides examples of different types of conflicts and integration methods.
The document discusses XML document type definitions (DTDs). It provides an overview of DTDs, explaining that they define the legal structure and elements of an XML document. It describes how DTDs can be declared internally or externally. The document also explains how to write internal and external DTD declarations, and how external DTD declarations can reference external files using system identifiers.
Pal gov.tutorial2.session13 2.gav and lav integrationMustafa Jarrar
This document discusses Global-As-View (GAV) and Local-As-View (LAV) integration approaches. GAV defines the global schema in terms of the local schemas by writing views over the local schemas. LAV defines the local schemas in terms of the global schema by writing views from the global schema to the local schemas. The document provides an example of each approach and discusses how queries are executed differently under GAV versus LAV.
Pal gov.tutorial2.session4.lab xml document and schemasMustafa Jarrar
1) The Palestinian Ministry of Education is building a National Student Registry to centrally manage student profiles and academic records as students transfer between universities.
2) Each semester, universities must send academic data for every student to the ministry to update and integrate into the registry.
3) The practical session involves building the necessary XML documents and schemas to collect student data from university databases for the registry as required by the ministry.
This document provides an overview of RDF Schema (RDFS), which extends RDF to enable the description of classes and properties. RDFS defines classes using rdfs:Class and subclasses using rdfs:subClassOf. Properties are instances of rdf:Property, and their domains and ranges can be specified using rdfs:domain and rdfs:range. RDFS provides a basic vocabulary for defining other application-specific vocabularies through classes, subclasses, and properties.
Pal gov.tutorial2.session12 2.architectural solutions for the integration issuesMustafa Jarrar
This document discusses two main architectural solutions for data integration issues: application-driven integration and data-driven integration. Application-driven integration uses middleware like web services or publish/subscribe architectures. Data-driven integration uses techniques like data consolidation, data warehousing, or virtual data integration to reconcile data schemas and queries. The document provides examples of architectures for each approach.
This document provides an introduction to the Resource Description Framework (RDF). RDF is a framework for describing resources on the web using Uniform Resource Identifiers (URIs) and properties. It represents data as a directed labeled graph consisting of triples with a subject, predicate, and object. Examples are provided to demonstrate how RDF can be used to describe resources and their properties. Key concepts explained include URIs, triples, and representing RDF data as a graph.
Pal gov.tutorial2.session1.xml basics and namespacesMustafa Jarrar
The document discusses XML (Extensible Markup Language) basics and namespaces. It provides an overview of XML, describing it as a protocol for containing and managing information by allowing users to create their own markup languages. The document also discusses the need for namespaces to avoid conflicts between element names and introduces the syntax for using namespaces, which involves associating namespaces with prefixes.
This document provides an outline for a tutorial on data integration and open information systems. The tutorial consists of 16 sessions over a total of approximately 40 hours. It will cover topics such as XML, RDF, OWL, data integration, linked data, and the semantic web. The intended learning outcomes include understanding data models, semantic web languages, integrating and querying heterogeneous data using techniques such as SPARQL and RDF. Students will gain practical skills in tools like Oracle Semantic Technology and Virtuoso for storing and querying RDF data. Attendance is mandatory for all sessions.
Pal gov.tutorial4.session1 1.needforsharedsemanticsMustafa Jarrar
This document provides an overview of the need for shared semantics and ontologies. It discusses how a lack of shared understanding limits communication between people and systems. XML provides syntax but not semantics, while standard vocabularies are often ambiguous. Ontologies aim to formally specify meaning and resolve discrepancies in semantics through shared conceptualizations. They allow computers and people to communicate meaningfully.
Producing, Publishing and Consuming Linked Data Three lessons from the Bio2RD...François Belleau
The document discusses three lessons learned from the Bio2RDF project about producing, publishing, and consuming linked data. Lesson 1 is that data transformation to RDF is an ETL task best done using frameworks like Talend. Lesson 2 is to publish semantic data using triplestores like Virtuoso and make SPARQL endpoints publicly available. Lesson 3 is that semantic data sources can be consumed in various ways, including SPARQL queries, HTTP requests, and SOAP services.
Pal gov.tutorial3.session2.xml ns and schemaMustafa Jarrar
Here are some examples of information item types in an XML document:
- The <book> element is an element information item that contains child elements and attributes.
- The "title" attribute of the <book> element is an attribute information item.
- The text "Learning XML" between the <title> tags is a character data information item.
- A comment like <!-- This is a book details --> is a comment information item.
- The XML declaration at the top <?xml version="1.0"?> provides version and encoding details and is also an information item.
- The document type declaration <!DOCTYPE book> is a document type information item.
- The entire XML document
Project Places, Periods and Properties for Linked Science discusses linking scientific data from the University of Münster (WWU) using linked open data principles. Currently, WWU's knowledge is siloed across organizations, terminology, formats and literature. The project aims to open these silos by allowing researchers to easily publish, find and reuse data and models. It will interconnect data about people, publications, projects, locations and time periods. This will make WWU's knowledge more transparent, accessible and foster transdisciplinary research. Several use cases are discussed, including archaeology, bioinformatics and social science interviews.
A Web 2.0 Personal Learning Environment for Classical Chinese PoetryRalf Klamma
A Web 2.0 Personal Learning Environmentfor Classical Chinese Poetry
Yiwei Cao 曹怡蔚, Ralf Klamma, Yan Gao 高岩, Rynson W.H. Lau 劉永雄, and Matthias Jarke
Informatik 5, RWTH Aachen University
Department of Computer Science, City University of Hong Kong
Aachen, Germany
ICWL 2009
20.08.2009
Learning Analytics for the Lifelong Long Tail LearnerRalf Klamma
Learning Analytics for the Lifelong Long Tail Learner
Ralf Klamma
RWTH Aachen University
Informatik 5 (DBIS)
CELSTEC, Heerlen, The Netherlands
February 24, 2011
Multimedia Processing on Multimedia Semantics and Multimedia ContextRalf Klamma
The 10thWorkshop on Multimedia Metadata (SeMuDaTe‘09)
Yiwei Cao, Ralf Klamma, and Dejan KovachevI
Informatik 5 (Information Systems), RWTH Aachen University
2.12.2009
Graz, Austria
Easysoft Limited do data access, we love data, nothing scares us.
Whether you have a data integration, data migration or data synchronisation project you need help with, get in touch with us sales@easysoft.com
The document discusses different approaches to integrating information from multiple systems, including:
1. Providing a uniform logical view of distributed data through approaches like mediated query systems, portals, federated database systems, and web services.
2. Realizing a common data storage through data warehouses and operational data stores that load and aggregate data from multiple sources.
3. Achieving integration through applications like workflow management systems that coordinate interactions between different systems and users.
Data integration ppt-bhawani nandan prasad - iim calcuttaBhawani N Prasad
The document provides an overview of an organization's data integration strategy. It discusses defining integration standards and architecture to enable data sharing across applications. The scope includes integrating various IT systems to support key business processes. A conceptual integration architecture is presented focusing on data integration, management and standardization. Business requirements for integration include consolidating data from various plant systems and applications into a common operating picture for operators. Integration technology options are also compared including enterprise information integration, extract transform and load, and enterprise application integration.
RDFa allows data embedded in web pages to be extracted by software agents by making the data embedding explicit through the use of attributes. The document provides an example of a simple web page marked up with RDFa to embed event data from a calendar ontology. An RDFa agent would be able to extract this data from the page.
This document discusses data schema integration, which involves identifying correspondences between elements in different data schemas that describe the same real-world concepts, and resolving conflicts between the schemas. The integration process includes schema transformation to homogenize the schemas, schema matching to discover correspondences, and schema integration to generate a unified schema and mapping rules between the integrated schema and source schemas. This resolves conflicts through classification, structural, descriptive and other transformations. Semi-automatic and manual methods can be used for the integration process.
Data integration involves providing unified access to data stored across multiple heterogeneous data sources. There are several data integration architectures including data warehouses, virtual mediators, and peer-to-peer integration. Key challenges in data integration include modeling the global schema, source schemas, and mappings between them, as well as reformulating queries over the global schema to retrieve answers from the source schemas. Languages for modeling schema mappings include GAV, LAV, and GLAV, with different advantages for query reformulation and modularity when new sources are added.
Lecture Notes by Mustafa Jarrar at Birzeit University, Palestine.
See the course webpage at: http://jarrar-courses.blogspot.com/2014/01/data-schema-integration.html and http://www.jarrar.info
you may also watch this lecture at: http://www.youtube.com/watch?v=VJtF_7ptln4
The lecture covers:
- Challenges of Data Schema Integration
- Framework for Schema Integration
- Schema Transformation
- Reverse Engineering
The document discusses database integration, which involves combining multiple existing databases with different schemas (called local conceptual schemas or LCSs) into a single integrated schema (called a global conceptual schema or GCS). It covers topics such as schema matching to find relationships between elements in different LCSs, schema mapping to translate between LCSs and the GCS, and methods for generating the GCS by combining parts of the LCSs. The goal is to enable queries and applications to interact with the distributed databases through a unified interface via the GCS.
Distributed database management systemsDhani Ahmad
This chapter discusses distributed database management systems (DDBMS). A DDBMS governs storage and processing of logically related data across interconnected computer systems. The chapter covers DDBMS components, levels of data and process distribution, transaction management, and design considerations like data fragmentation, replication, and allocation. Transparency and optimization techniques aim to make the distributed nature transparent to users.
Pal gov.tutorial4.session1 1.needforsharedsemanticsMustafa Jarrar
This tutorial discusses the need for shared semantics on the Internet and in open systems. As information systems become more interconnected, there needs to be agreement on the meaning and use of terms. XML provides syntax but not semantics, while standard vocabularies are often ambiguous. Ontologies provide precise definitions of terms and their relationships to allow computers and systems to truly understand each other. They formalize the concepts and semantic rules that humans use to distinguish similar things.
Pal gov.tutorial3.session3.xpath & xquery (lab1)Mustafa Jarrar
This tutorial document provides an overview of XPath and how it can be used to navigate an XML document by examining the different node types and axes in XPath that represent the hierarchical structure. It describes the XPath data model and how location path expressions use axes, node tests, and predicates to select nodes in an XML document based on their relationship to the context node.
This document describes a practical session on building a national student registry system for Palestine. It discusses setting up a central registry to manage student profiles and academic records across universities in Palestine. This will allow easy transfer of student records between universities and certification of diplomas by the Ministry of Higher Education. The registry can support various services, like generating transcripts for students (GTS). Students will individually use BPEL and Windows Workflow to invoke the GTS web service and print transcripts for given students based on their academic records in the central registry.
This document outlines a tutorial on process integration and service-oriented architectures. The 15-session tutorial covers topics including SOA, XML, REST, SOAP, WSDL, ESB, BPEL, and UDDI. It is part of a larger eGovernment project funded by the European Commission. The tutorial schedule spans 6 days and includes sessions led by different instructors, as well as hands-on labs. The goal is for students to gain knowledge and skills in designing, developing, deploying, and integrating applications and services using SOA principles and technologies.
This document provides information about a practical session on building a national student registry system in Palestine. The session aims to integrate student profile services from different universities. Students will select three existing services and draw integration plans to combine them into a single system. They will implement the integrated system using the provided ministry ontology. The final output is an integrated system that brings together student data from three universities. Each group will present their project for feedback.
SOAP is an XML-based protocol for exchanging structured information between applications, typically in web services. It relies on other protocols like HTTP for message negotiation and transmission. A SOAP message consists of an envelope containing an optional header and a mandatory body. The body contains application-specific content. SOAP defines a messaging framework and extensibility model to build web services. It also provides mechanisms for data representation, encoding, and error handling between services.
This document describes a practical session on implementing a National Student Registry using RESTful web services. The goal is to have universities provide student profile data to the Ministry of Education via services. Each student will individually implement and expose a RESTful service for the registry using skills from previous tutorials. They will reuse an ontology from earlier discussions and build a service to represent the registry, then consume their own service. The deliverables are the REST implementation of the National Student Registry and an application to consume that service.
Pal gov.tutorial4.session11.lab zinnarontologybasedwebservicesMustafa Jarrar
This document provides information about a practical session on building ontology-based web services. The session is part of a larger tutorial on ontology engineering and lexical semantics. Students will work in groups of three to model a governmental service and its processes. They will build web services for each sub-process and publish the WSDL of the services using terms that are defined in the Palestinian government ontology (Zinnar). Students can reuse services built by other groups. The goal is to implement an interoperable governmental service using standardized ontology terms.
Pal gov.tutorial4.session8 2.stepwisemethodologiesMustafa Jarrar
This document provides an overview of stepwise methodologies for ontology engineering. It discusses phases such as identifying the purpose and scope, building the ontology through capturing concepts and defining relationships, integrating existing ontologies, evaluating the ontology, and documenting it. The methodology proposes that building the ontology involves capturing concepts through brainstorming, organizing concepts, producing clear definitions, and defining taxonomies and properties. It emphasizes reaching consensus among those involved and reusing existing ontologies where possible. The goal is to develop ontologies that are clear, coherent, extensible, and reusable.
Pal gov.tutorial4.session3.lab bankcustomerontologyMustafa Jarrar
This document provides an overview of a tutorial on ontology engineering and lexical semantics. The tutorial aims to teach participants about developing ontologies, with a focus on building an ontology for bank customers. The tutorial will include sessions on ontology basics, tools, challenges, and multilingualism. It will guide students through practical exercises to build ontologies for populations, bank customers, legal persons, and using existing linguistic ontologies. The document outlines the learning objectives, session topics, and instructions for an exercise where each student will independently model the types of bank customers and their properties in an ontology.
Here are the key points about hyponymy relations in WordNet:
- A synset {x, x',...} is a hyponym of synset {y, y',...} if native English speakers accept sentences like "x is a (kind of) y". For example, "Table" is a hyponym of "Array" and "Array" is a hyponym of "Arrangement".
- Hyponymy generates a hierarchical structure where a hyponym inherits all the features of its superordinate concept and adds at least one distinguishing feature.
- Hyponymy is transitive, so if A is a hyponym of B and B is a hyponym of C
Pal gov.tutorial4.session5.lab ontologytoolsMustafa Jarrar
This document provides an overview of ontology tools that can be used to build ontologies. It discusses several popular and freely available tools, including Protégé, TopBraid, SWOOP, VisioModeler, NORMA, DogmaModeler, and how each one can be used to create and edit ontologies. It then focuses on using Protégé, providing a quick user guide on how to perform basic tasks like creating classes and object properties, adding domains and ranges, and viewing the ontology structure as a graph.
This document provides a tutorial on REST web services. It begins with an introduction to REST, covering key principles such as every element having an ID, things being linked, using standard HTTP methods, providing multiple representations, and communicating statelessly. The tutorial then discusses fundamental HTTP concepts, REST examples, design guidelines, documenting REST with WSDL and WADL, and a case study applying REST principles to a university student services API. It concludes with code samples in C# and Java for making HTTP requests, and an overview of other web service protocols like SOAP and OData.
This document provides an outline for a tutorial on ontology engineering and lexical semantics. The tutorial aims to teach participants how to build ontologies, tackle challenges in ontology engineering, and develop multilingual ontologies. It will include sessions on population ontologies, bank customer ontologies, legal person ontologies, ontology tools, and using existing linguistic ontologies like WordNets. Participants will learn about the Palestinian eGovernment interoperability framework called Zinnar and how to use ontologies in web services. The goal is to help participants gain knowledge and skills in ontology engineering, multilingual knowledge representation, and applying ontologies in eGovernment systems.
This document provides an overview of a practical session on implementing a Generate Transcript Service (GTS) for a National Student Registry system in Palestine. [The session instructs students to individually use MS Visual Studio to develop and invoke a GTS web service using WSDL that can generate transcripts for students based on their academic records stored in the centralized registry.] The registry integrates data from universities across Palestine to centrally manage student profiles, academic records, financial aid, and more for the Ministry of Higher Education. The practical aims to help students design services on top of the registry using SOA principles and web service standards.
Pal gov.tutorial4.session6 2.knowledge double-articulationMustafa Jarrar
The document discusses the Knowledge Double-Articulation theory for ontology engineering. It proposes that the meaning of a vocabulary should be doubly-articulated into a domain axiomatization and one or more application-kind axiomatizations. The domain axiomatization characterizes the intended meaning of terms at the community/domain level, while application axiomatizations focus on how the terms are useful for specific applications. This approach results in ontologies that are highly reusable, usable, easier to integrate, and simpler to maintain.
Similar to Pal gov.tutorial2.session15 1.linkeddata (16)
Clustering Arabic Tweets for Sentiment AnalysisMustafa Jarrar
Diab Abuaiadah, Dileep Rajendran, Mustafa Jarrar: Clustering Arabic Tweets for Sentiment Analysis. Proceedings of the 2017 IEEE/ACS 14th International Conference on Computer Systems and Applications. IEEE Computer Society. DOI 10.1109/AICCSA.2017.162
Classifying Processes and Basic Formal OntologyMustafa Jarrar
pdf http://www.jarrar.info/publications/JC17.pdf
Mustafa Jarrar and Werner Ceusters
ABSTRACT
Unlike what is the case for physical entities and other types of continuants, few process ontologies exist. This is not only because processes received less attention in the research community, but also because classifying them is challenging. Moreover, upper level categories or classification criteria to help in modelling and integrating lower level process ontologies have thus far not been developed or widely adopted. This paper proposes a basis for further classifying processes in the Basic Formal Ontology. The work is inspired by the aspectual characteristics of verbs such as homeomericity, cumulativity, telicity, atomicity, instantaneity and durativity. But whereas these characteristics have been proposed by linguists and philosophers of language from a linguistic perspective with a focus on how matters are described, our focus is on what is the case in reality thus providing an ontological perspective. This was achieved by first investigating the applicability of these characteristics to the top-level processes in the Gene Ontology, and then, where possible, deriving from the linguistic perspective relationships that are faithful to the ontological principles adhered to by the Basic Formal Ontology.
The goal of this course is to introduce students to ideas and techniques from discrete mathematics that are widely used in computer science. Ultimately, students are expected to understand and use (abstract) discrete structures that are the backbones of computer science. In particular, this class is meant to introduce logic, proofs, sets, functions, relations, counting, graphs and trees and with an emphasis on applications in computer science.
The document provides information about implementing and executing business processes using the Activiti framework. It discusses Activiti components and architecture, downloading and setting up the necessary software including Activiti, Java, Eclipse and Tomcat. It also demonstrates configuring a sample vacation request process in Activiti and exploring the process lifecycle. The document emphasizes hands-on practice for readers to understand business process automation using Activiti.
Business Process Design and Re-engineeringMustafa Jarrar
Lecture slides by Mustafa Jarrar at Birzeit University, Palestine.
Course Title: Data and Business Process Modeling
See the course webpage and video lectures at: http://jarrar-courses.blogspot.com/2015/01/data-and-business-process-modelling.html
and http://www.jarrar.info
The document provides instructions for two modeling projects using BPMN 2.0 - to model the processes of graduation clearance and faculty traveling permission at a university. It includes descriptions of the two business processes and tasks students to model each process in BPMN 2.0 in Signavio and submit the models by specific deadlines in April and May 2015.
The document provides an overview of descriptive constructs in BPMN 2.0 including activities, connecting objects, events, gateways, pools, lanes, artifacts, and data objects. It presents examples of processes for course enrollment and book borrowing. Key recommendations are that a process model should have a start and end event and all branches should be closed. The document is intended as lecture notes for a class on BPMN 2.0 descriptive constructs.
Introduction to Business Process ManagementMustafa Jarrar
The document provides an introduction to business process management concepts. It discusses what constitutes a process and gives examples. It also outlines the roles and challenges involved in process management. Finally, it introduces the business process management lifecycle, including modeling, improvement, automation, and monitoring of processes.
Lecture video by Mustafa Jarrar at Birzeit University, Palestine.
See the course webpage at: http://jarrar-courses.blogspot.com/2011/09/knowledgeengineering-fall2011.html
and http://www.jarrar.info
and on Youtube:
https://www.youtube.com/watch?v=GYmI37-0b5k&index=7&list=PLDEA50C29F3D28257
Lecture video by Mustafa Jarrar at Birzeit University, Palestine.
See the course webpage at: http://jarrar-courses.blogspot.com/2011/09/knowledgeengineering-fall2011.html
and http://www.jarrar.info
and on Youtube:
https://www.youtube.com/watch?v=GYmI37-0b5k&index=7&list=PLDEA50C29F3D28257
On Computer Science Trends and Priorities in PalestineMustafa Jarrar
On Computer Science Trends and Priorities in Palestine,
by Mustafa Jarrar
Computer Science
Birzeit University, Palestine
Personal Page: http://www.jarrar.info
At Workshop on ّIT Research Trends and Priorities
Islamic University of Gaza, Palestine
28 March, 2015
Lessons from Class Recording & Publishing of Eight Online CoursesMustafa Jarrar
Mustafa Jarrar presented lessons learned from recording and publishing eight of his online courses. He found that recording his lectures helped him improve his teaching materials and presentation. It also allowed students to watch lectures they missed or did not understand. Jarrar provided tips for effective recording, such as breaking lectures into short videos, adding titles and annotations, and working with students to help with equipment and uploading videos. Recording lectures benefited both professors and students by improving teaching quality and providing flexibility for students to learn.
Mustafa Jarrar, Nizar Habash, Diyam Akra, Nasser Zalmout: Building A Corpus For Palestinian Arabic: A Preliminary Study. In proceedings of the EMNLP 2014 Workshop on Arabic Natural Language Processing. Association for Computational Linguistics (ACL), Pages (18-27). October 25, 2014, Doha, Qatar. ISBN: 978-1-937284-96-1
Habash: Arabic Natural Language ProcessingMustafa Jarrar
This document provides an overview of Arabic natural language processing. It begins with an introduction to the Arabic script, including its alphabet, letter forms, diacritics, and encoding issues. It then discusses features of Modern Standard Arabic phonology and spelling, noting that Arabic spelling is mostly phonemic but can be ambiguous without diacritics. The document outlines challenges for processing Arabic text related to its orthography.
Adnan: Introduction to Natural Language Processing Mustafa Jarrar
This document provides an introduction to natural language processing (NLP). It discusses key topics in NLP including languages and intelligence, the goals of NLP, applications of NLP, and general themes in NLP like ambiguity in language and statistical vs rule-based methods. The document also previews specific NLP techniques that will be covered like part-of-speech tagging, parsing, grammar induction, and finite state analysis. Empirical approaches to NLP are discussed including analyzing word frequencies in corpora and addressing data sparseness issues.
Bouquet: SIERA Workshop on The Pillars of Horizon2020Mustafa Jarrar
The document summarizes key aspects of Horizon 2020, the European Union's research and innovation program from 2014 to 2020. It discusses the program's three main pillars of excellence in science, industrial leadership, and tackling societal challenges. It notes the increased focus on innovation and bringing ideas to market. It outlines the types of funding actions, eligibility requirements, evaluation criteria, and opportunities for participation by countries outside the EU like Palestine. The presentation aims to highlight opportunities for Birzeit University under Horizon 2020.
This document provides instructions for a project on using SPARQL and Oracle Semantic Technology to query RDF data. Students will convert marksheets into RDF tables, combine the tables, and load them into Oracle and a SPARQL endpoint. They will write queries to retrieve data from the graph, including simple queries and queries with paths of different lengths. Students will deliver a report including screenshots of the tables and queries with their results and descriptions.
This presentation includes basic of PCOS their pathology and treatment and also Ayurveda correlation of PCOS and Ayurvedic line of treatment mentioned in classics.
Executive Directors Chat Leveraging AI for Diversity, Equity, and InclusionTechSoup
Let’s explore the intersection of technology and equity in the final session of our DEI series. Discover how AI tools, like ChatGPT, can be used to support and enhance your nonprofit's DEI initiatives. Participants will gain insights into practical AI applications and get tips for leveraging technology to advance their DEI goals.
How to Setup Warehouse & Location in Odoo 17 InventoryCeline George
In this slide, we'll explore how to set up warehouses and locations in Odoo 17 Inventory. This will help us manage our stock effectively, track inventory levels, and streamline warehouse operations.
How to Build a Module in Odoo 17 Using the Scaffold MethodCeline George
Odoo provides an option for creating a module by using a single line command. By using this command the user can make a whole structure of a module. It is very easy for a beginner to make a module. There is no need to make each file manually. This slide will show how to create a module using the scaffold method.
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
ISO/IEC 27001, ISO/IEC 42001, and GDPR: Best Practices for Implementation and...PECB
Denis is a dynamic and results-driven Chief Information Officer (CIO) with a distinguished career spanning information systems analysis and technical project management. With a proven track record of spearheading the design and delivery of cutting-edge Information Management solutions, he has consistently elevated business operations, streamlined reporting functions, and maximized process efficiency.
Certified as an ISO/IEC 27001: Information Security Management Systems (ISMS) Lead Implementer, Data Protection Officer, and Cyber Risks Analyst, Denis brings a heightened focus on data security, privacy, and cyber resilience to every endeavor.
His expertise extends across a diverse spectrum of reporting, database, and web development applications, underpinned by an exceptional grasp of data storage and virtualization technologies. His proficiency in application testing, database administration, and data cleansing ensures seamless execution of complex projects.
What sets Denis apart is his comprehensive understanding of Business and Systems Analysis technologies, honed through involvement in all phases of the Software Development Lifecycle (SDLC). From meticulous requirements gathering to precise analysis, innovative design, rigorous development, thorough testing, and successful implementation, he has consistently delivered exceptional results.
Throughout his career, he has taken on multifaceted roles, from leading technical project management teams to owning solutions that drive operational excellence. His conscientious and proactive approach is unwavering, whether he is working independently or collaboratively within a team. His ability to connect with colleagues on a personal level underscores his commitment to fostering a harmonious and productive workplace environment.
Date: May 29, 2024
Tags: Information Security, ISO/IEC 27001, ISO/IEC 42001, Artificial Intelligence, GDPR
-------------------------------------------------------------------------------
Find out more about ISO training and certification services
Training: ISO/IEC 27001 Information Security Management System - EN | PECB
ISO/IEC 42001 Artificial Intelligence Management System - EN | PECB
General Data Protection Regulation (GDPR) - Training Courses - EN | PECB
Webinars: https://pecb.com/webinars
Article: https://pecb.com/article
-------------------------------------------------------------------------------
For more information about PECB:
Website: https://pecb.com/
LinkedIn: https://www.linkedin.com/company/pecb/
Facebook: https://www.facebook.com/PECBInternational/
Slideshare: http://www.slideshare.net/PECBCERTIFICATION
How to Add Chatter in the odoo 17 ERP ModuleCeline George
In Odoo, the chatter is like a chat tool that helps you work together on records. You can leave notes and track things, making it easier to talk with your team and partners. Inside chatter, all communication history, activity, and changes will be displayed.