1) The document proposes an approach to assist course creators in generating or restructuring courses by exploiting text mining techniques, semantic information from DBpedia, and linking educational resources.
2) The approach was implemented as a prototype that retrieves online courses, identifies key elements from text, formulates queries to other courses, and returns related courses to help creators generate mashups.
3) Preliminary tests on 265 computer science courses showed promising results, though future work is needed to improve similarity measures and generate concept maps between related courses.
MULTI-LEARNING SPECIAL SESSION / EDUCON 2018 / EMADRID TEAMeMadrid network
- The document presents the Computer Science Curricula Ontology (LOD-CS2013), which was created based on the Body of Knowledge of the IEEE and ACM Computer Science Curricula 2013.
- LOD-CS2013 aims to improve the usability and interoperability of computer science curricula. It defines the concepts, properties, and relationships between topics in the curricula.
- A general lifecycle for developing the ontology included specifying the concepts based on the curricula, formalizing the ontology, evaluating it, and allowing for continuous improvement.
Ontology-based Semantic Approach for Learning Object RecommendationIDES Editor
The main focus of this paper is to apply an ontologybased
approach for semantic learning object recommendation
towards personalized e-learning systems. Ontologies for
learner model, learning objects and semantic mapping rules
are proposed. The recommender can be able to provide
individually learning object by taking the learner preferences
and styles, which used to adjust or fine-tune in learning object
recommending process. In the proposed framework, we
demonstrated how the ontologies can be used to enable
machines to interpret and process learning resources in
recommendation system. The recommendation consists of four
steps: semantic mapping between learner and learning
objects, preference score calculation, learning object ranking
and recommending the learning object. As a result, a
personalized and most suitable learning object is
recommended to the learner.
The document proposes a layered model for authoring educational content with different levels of abstraction.
The lowest level is content, which can be learning objects. Above this is a structure level that provides a table of contents. Next is a task level that defines learning processes and activities. The highest level is conceptualization, which models the knowledge domain using ontologies and instructional templates.
This layered model separates concerns of content, structure, tasks and conceptualization/knowledge modeling. It aims to make authoring tools more flexible and the content more interoperable and adaptable to different contexts.
Constructing a Learner Centric Semantic Syllabus for Automatic Text Book Gen...Aliabbas Petiwala
The document discusses developing a semantic syllabus ontology to guide automatic textbook generation. It outlines key aspects of a learner-centric semantic syllabus such as collaborative active learning environments. The proposed ontology would represent syllabus topics and relationships to facilitate data integration and textbook customization for different learners. Future work is needed to specify content granularity and develop a book authoring tool integrated with an active learning community.
This document describes a proposed method for subontology-assisted web-based e-learning for resource management. Key points include:
1. Semantic mapping is used to integrate heterogeneous e-learning databases by mapping relational schemas to a global ontology.
2. Subontologies (SubOs) are context-specific portions of the full ontology that are evolved over time based on locality of resource reuse.
3. A SubO-based approach is used to achieve adaptive and efficient resource management and reuse by matching user requests to SubOs.
An effective method for semi automatic construction of domain module from ele...eSAT Journals
The document describes a method for semi-automatically generating a domain module from electronic textbooks. It uses techniques like natural language processing, ontologies, and heuristic reasoning. The domain module captures knowledge at two levels: a Learning Domain Ontology that represents topics and relationships between them, and a set of Learning Objects containing educational resources. The method involves preprocessing the textbook, analyzing its outline to generate an initial LDO, analyzing the full text to expand the LDO, and extracting Learning Objects. It was tested on an electronic textbook and the automatically generated knowledge was compared to a manually created domain module.
International Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
MULTI-LEARNING SPECIAL SESSION / EDUCON 2018 / EMADRID TEAMeMadrid network
- The document presents the Computer Science Curricula Ontology (LOD-CS2013), which was created based on the Body of Knowledge of the IEEE and ACM Computer Science Curricula 2013.
- LOD-CS2013 aims to improve the usability and interoperability of computer science curricula. It defines the concepts, properties, and relationships between topics in the curricula.
- A general lifecycle for developing the ontology included specifying the concepts based on the curricula, formalizing the ontology, evaluating it, and allowing for continuous improvement.
Ontology-based Semantic Approach for Learning Object RecommendationIDES Editor
The main focus of this paper is to apply an ontologybased
approach for semantic learning object recommendation
towards personalized e-learning systems. Ontologies for
learner model, learning objects and semantic mapping rules
are proposed. The recommender can be able to provide
individually learning object by taking the learner preferences
and styles, which used to adjust or fine-tune in learning object
recommending process. In the proposed framework, we
demonstrated how the ontologies can be used to enable
machines to interpret and process learning resources in
recommendation system. The recommendation consists of four
steps: semantic mapping between learner and learning
objects, preference score calculation, learning object ranking
and recommending the learning object. As a result, a
personalized and most suitable learning object is
recommended to the learner.
The document proposes a layered model for authoring educational content with different levels of abstraction.
The lowest level is content, which can be learning objects. Above this is a structure level that provides a table of contents. Next is a task level that defines learning processes and activities. The highest level is conceptualization, which models the knowledge domain using ontologies and instructional templates.
This layered model separates concerns of content, structure, tasks and conceptualization/knowledge modeling. It aims to make authoring tools more flexible and the content more interoperable and adaptable to different contexts.
Constructing a Learner Centric Semantic Syllabus for Automatic Text Book Gen...Aliabbas Petiwala
The document discusses developing a semantic syllabus ontology to guide automatic textbook generation. It outlines key aspects of a learner-centric semantic syllabus such as collaborative active learning environments. The proposed ontology would represent syllabus topics and relationships to facilitate data integration and textbook customization for different learners. Future work is needed to specify content granularity and develop a book authoring tool integrated with an active learning community.
This document describes a proposed method for subontology-assisted web-based e-learning for resource management. Key points include:
1. Semantic mapping is used to integrate heterogeneous e-learning databases by mapping relational schemas to a global ontology.
2. Subontologies (SubOs) are context-specific portions of the full ontology that are evolved over time based on locality of resource reuse.
3. A SubO-based approach is used to achieve adaptive and efficient resource management and reuse by matching user requests to SubOs.
An effective method for semi automatic construction of domain module from ele...eSAT Journals
The document describes a method for semi-automatically generating a domain module from electronic textbooks. It uses techniques like natural language processing, ontologies, and heuristic reasoning. The domain module captures knowledge at two levels: a Learning Domain Ontology that represents topics and relationships between them, and a set of Learning Objects containing educational resources. The method involves preprocessing the textbook, analyzing its outline to generate an initial LDO, analyzing the full text to expand the LDO, and extracting Learning Objects. It was tested on an electronic textbook and the automatically generated knowledge was compared to a manually created domain module.
International Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
This presentation discusses the current dilemma with respect to Open Educational Resources (OER) search. It introduces existing OER search methodologies and highlights their weaknesses. The Desirability framework for parametrically measuring the usefulness of an OER is also discussed. The desirability framework uses the D-index to measure the openness, accessibility and relevance of an OER. OERScout, a text mining based faceted search engine is introduced for improved OER search. It uses autonomously identified domain specific keywords, the D-index and faceted search to allow focused OER search.
This document discusses reusable learning objects (RLOs), which are small interactive e-learning modules designed to teach standalone learning objectives. RLOs can take various forms like animations, simulations, puzzles or quizzes. The document outlines factors to consider when designing RLOs like the subject matter, delivery method, and archiving approach. It also compares technologies for building RLOs such as Flash, JavaScript, Java applets, .NET, and third-party e-learning software. Java applets and .NET are highlighted as options that allow building complex interactive content that can be archived and delivered online or through other means.
A PROPOSED MULTI-DOMAIN APPROACH FOR AUTOMATIC CLASSIFICATION OF TEXT DOCUMENTSijsc
Classification is an important technique used in information retrieval. Supervised classification suffers
from certain limitations concerning the collection and labeling of the training dataset. When facing Multi-
Domain classification, multiple training datasets and classifiers are needed which is relatively difficult. In
this paper an unsupervised classification system is proposed that can manage the Multi-Domain
classification problem as well. It is a multi-domain system where each domain represented by an ontology.
A document is mapped on each ontology based on the weights of the mutual tokens between them with the
help of fuzzy sets, resulting in a mapping degree of the document with each domain. An experiment carried
out showing satisfying classification results with an improvement in the evaluation results of the proposed
system compared to Apache Lucene.
The document discusses research at UNED on enhancing authoring, modeling, and collaboration in e-learning environments. It outlines tools like ENLACE, PELICAN, and CARDS that allow for modeling and aggregation of educational content and learning objects from external tools. CARDS provides a metamodel for integrating outcomes from tools like concept mapping tools. PELICAN is an instructional design and collaboration platform that integrates external tools and supports adaptive contexts based on student performance. The research aims to automatically extract and classify learning objects from the web to enrich authored content and help students find relevant resources.
Using patterns to design technology enhanced learning scenarioseLearning Papers
This document discusses using design patterns to represent technology-enhanced learning scenarios. It reviews different mechanisms used to represent learning design issues, such as hypermedia models, ontologies, and educational modeling languages. The author proposes an approach using design patterns as they combine narrative representation with visualization and controlled vocabularies. Patterns are prepared by classifying them into categories like content managers, activity facilitators, and assessment producers. They are then applied to represent a specific learning scenario based on digital ink technologies.
Ontology learning techniques and applications computer science thesis writing...Tutors India
At Tutors India, we offer Computer science and Information Technology Research Guidance services – We deliver exceptional work where your dissertation will deserve publication without significant reworking or alternation.
For #Enquiry
https://www.tutorsindia.com
info@tutorsindia.com
(Whatsapp): +91-8754446690
(UK): +44-1143520021
Workshop on Learning Technology Standards for Agriculture and Rural Development (AgroLT 2008)
September 19, 2008, Athens, Greece
In conjunction with
4th International Conference on Information and Communication Technologies in Bio and Earth Sciences (HAICTA 2008)
A Survey on Text Mining-techniques and applicationRyota Eisaki
This document summarizes text mining techniques and applications. It discusses text mining processes like document gathering, pre-processing, transformation, feature selection, and pattern selection. It also describes text mining techniques including categorization, clustering, information extraction, information visualization, and natural language processing. Finally, it outlines applications of text mining in various domains such as business intelligence, bioinformatics, security, human resources, and web search enhancement.
ONTOLOGY VISUALIZATION PROTÉGÉ TOOLS – A REVIEW ijait
The document discusses ontology visualization tools in Protégé. It reviews four main visualization methods used in Protégé tools: indented list, node-link and tree, zoomable, and focus+context. It then examines specific Protégé tools that use each method, including their key features and limitations. The tools assessed are Protégé Class Browser (indented list), Protégé OntoViz and OntoSphere (node-link and tree), Jambalaya (zoomable), and Protégé TGVizTab (focus+context). The document concludes by summarizing and comparing the visualization characteristics of these Protégé tools.
Portable and Synchronized Distributed Learning Management System in Severe Ne...Fajar Purnama
Master thesis defense limited of Fajar Purnama, Graduate School of Science and Technology, Human Interface and Cyber Communication Laboratory.
Video: https://youtu.be/i8AERku88u8
Masters Thesis: https://www.publish0x.com/fajar-purnama-academics/portable-and-synchronized-distributed-learning-management-sy-xyvdwoz?a=4oeEw0Yb0B&tid=slideshare
A New Concept Extraction Method for Ontology Construction From Arabic TextCSCJournals
Ontology is one of the most popular representation model used for knowledge representation, sharing and reusing. The Arabic language has complex morphological, grammatical, and semantic aspects. Due to complexity of Arabic language, automatic Arabic terminology extraction is difficult. In addition, concept extraction from Arabic documents has been challenging research area, because, as opposed to term extraction, concept extraction are more domain related and more selective. In this paper, we present a new concept extraction method for Arabic ontology construction, which is the part of our ontology construction framework. A new method to extract domain relevant single and multi-word concepts in the domain has been proposed, implemented and evaluated. Our method combines linguistic, statistical information and domain knowledge. It first uses linguistic patterns based on POS tags to extract concept candidates, and then stop words filter is implemented to filter unwanted strings. To determine relevance of these candidates within the domain, different statistical measures and new domain relevance measure are implemented for first time for Arabic language. To enhance the performance of concept extraction, a domain knowledge will be integrated into the module. The concepts scores are calculated according to their statistical values and domain knowledge values. In order to evaluate the performance of the method, precision scores were calculated. The results show the high effectiveness of the proposed approach to extract concepts for Arabic ontology construction.
This document summarizes a workshop on data integration using ontologies. It discusses how data integration is challenging due to differences in schemas, semantics, measurements, units and labels across data sources. It proposes that ontologies can help with data integration by providing definitions for schemas and entities referred to in the data. Core challenges discussed include dealing with multiple synonyms for entities and relationships between biological entities that depend on context. The document advocates for shared community ontologies that can be extended and integrated to facilitate flexible and responsive data integration across multiple sources.
This summarizes an academic paper that proposes an automatic ontology creation method for classifying research papers. It uses text mining techniques like classification and clustering algorithms. It first builds a research ontology by extracting keywords and patterns from previous papers. It then uses a decision tree algorithm to classify new papers into disciplines defined in the ontology. The classified papers are then clustered based on similarities to group them. The method was tested on a dataset of 100 papers and achieved average precision of 85.7% for term-based and 89.3% for pattern-based keyword extraction.
Possibility of interdisciplinary research software engineering andnatural lan...Nakul Sharma
This document discusses the possibility of interdisciplinary research between software engineering and natural language processing. It provides a literature review of research papers from 2003 to 2014 related to applying tools and techniques from one field to the other. Some key areas discussed include generating UML diagrams from natural language text, developing ontologies to clarify meanings, and potential issues with joint research like determining complexity of sentences. The document proposes a flowchart for how artifacts could be analyzed using tasks from either field to enable interdisciplinary research.
Class Diagram Extraction from Textual Requirements Using NLP Techniquesiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Mining Users Rare Sequential Topic Patterns from Tweets based on Topic Extrac...IRJET Journal
This paper proposes a method to mine rare sequential topic patterns (URSTPs) from tweet data. It involves preprocessing tweets to extract topics, identifying user sessions, generating sequential topic pattern (STP) candidates, and selecting URSTPs based on rarity analysis. Experiments show the approach can identify special users and interpretable URSTPs, indicating users' characteristics. The paper aims to capture personalized and abnormal user behaviors through sequential relationships between extracted topics from successive tweets.
The increased potential of the ontologies to reduce the human interference has wide range of applications. This paper identifies requirements for an ontology development platform to innovate artificially intelligent web. To facilitate this process, RDF and OWL have been developed as standard formats for the sharing and integration of data and knowledge. The knowledge in the form of rich conceptual schemas called ontologies. Based on the framework, an architectural paradigm is put forward in view of ontology engineering and development of ontology applications and a development portal designed to support ontology engineering, content authoring and application development with a view to maximal scalability in size and complexity of semantic knowledge and flexible reuse of ontology models and ontology application processes in a distributed and collaborative engineering environment.
Ontology Based Approach for Semantic Information Retrieval SystemIJTET Journal
Abstract—The Information retrieval system is taking an important role in current search engine which performs searching operation based on keywords which results in an enormous amount of data available to the user, from which user cannot figure out the essential and most important information. This limitation may be overcome by a new web architecture known as the semantic web which overcome the limitation of the keyword based search technique called the conceptual or the semantic search technique. Natural language processing technique is mostly implemented in a QA system for asking user’s questions and several steps are also followed for conversion of questions to the query form for retrieving an exact answer. In conceptual search, search engine interprets the meaning of the user’s query and the relation among the concepts that document contains with respect to a particular domain that produces specific answers instead of showing lists of answers. In this paper, we proposed the ontology based semantic information retrieval system and the Jena semantic web framework in which, the user enters an input query which is parsed by Standford Parser then the triplet extraction algorithm is used. For all input queries, the SPARQL query is formed and further, it is fired on the knowledge base (Ontology) which finds appropriate RDF triples in knowledge base and retrieve the relevant information using the Jena framework.
In tech application-of_data_mining_technology_on_e_learning_material_recommen...Enhmandah Hemeelee
The document describes a recommendation system that applies data mining techniques to recommend e-learning materials. It proposes using LDAP for fast searching of materials across systems, JAXB for parsing content, and association rule mining and collaborative filtering for recommendations. A web spider collects content indexes from learning management systems and stores data in an LDAP directory. Users can search for related materials, and the system mines log data to associate frequently searched terms and recommend additional resources.
In tech application-of_data_mining_technology_on_e_learning_material_recommen...Enhmandah Hemeelee
The document describes a recommendation system that applies data mining techniques to recommend e-learning materials. It proposes using LDAP for fast searching of materials across systems, JAXB for parsing content, and association rule mining and collaborative filtering to generate recommendations. The system collects user activity data, analyzes it using Apriori algorithm to find related search terms and content, and stores results in an LDAP database to provide recommendations to users.
Comparative evaluation of four multi label classification algorithms in class...csandit
The classification of learning objects (LOs) enables users to search for, access, and reuse them
as needed. It makes e-learning as effective and efficient as possible. In this article the multilabel
learning approach is represented for classifying and ranking multi-labelled LOs, whereas
each LO might be associated with multiple labels as opposed to a single-label approach. A
comprehensive overview of the common fundamental multi-label classification algorithms and
metrics will be discussed. In this article, a new multi-labelled LOs dataset will be created and
extracted from ARIADNE Learning Object Repository. We experimentally train four effective
multi-label classifiers on the created LOs dataset and then, assess their performance based on
the results of 16 evaluation metrics. The result of this article will answer the question of: what is
the best multi-label classification algorithm for classifying multi-labelled LOs?
COMPARATIVE EVALUATION OF FOUR MULTI-LABEL CLASSIFICATION ALGORITHMS IN CLASS...cscpconf
The classification of learning objects (LOs) enables users to search for, access, and reuse them as needed. It makes e-learning as effective and efficient as possible. In this article the multilabel learning approach is represented for classifying and ranking multi-labelled LOs, whereas each LO might be associated with multiple labels as opposed to a single-label approach. A comprehensive overview of the common fundamental multi-label classification algorithms and metrics will be discussed. In this article, a new multi-labelled LOs dataset will be created and extracted from ARIADNE Learning Object Repository. We experimentally train four effective multi-label classifiers on the created LOs dataset and then, assess their performance based on the results of 16 evaluation metrics. The result of this article will answer the question of: what is the best multi-label classification algorithm for classifying multi-labelled LOs?
This presentation discusses the current dilemma with respect to Open Educational Resources (OER) search. It introduces existing OER search methodologies and highlights their weaknesses. The Desirability framework for parametrically measuring the usefulness of an OER is also discussed. The desirability framework uses the D-index to measure the openness, accessibility and relevance of an OER. OERScout, a text mining based faceted search engine is introduced for improved OER search. It uses autonomously identified domain specific keywords, the D-index and faceted search to allow focused OER search.
This document discusses reusable learning objects (RLOs), which are small interactive e-learning modules designed to teach standalone learning objectives. RLOs can take various forms like animations, simulations, puzzles or quizzes. The document outlines factors to consider when designing RLOs like the subject matter, delivery method, and archiving approach. It also compares technologies for building RLOs such as Flash, JavaScript, Java applets, .NET, and third-party e-learning software. Java applets and .NET are highlighted as options that allow building complex interactive content that can be archived and delivered online or through other means.
A PROPOSED MULTI-DOMAIN APPROACH FOR AUTOMATIC CLASSIFICATION OF TEXT DOCUMENTSijsc
Classification is an important technique used in information retrieval. Supervised classification suffers
from certain limitations concerning the collection and labeling of the training dataset. When facing Multi-
Domain classification, multiple training datasets and classifiers are needed which is relatively difficult. In
this paper an unsupervised classification system is proposed that can manage the Multi-Domain
classification problem as well. It is a multi-domain system where each domain represented by an ontology.
A document is mapped on each ontology based on the weights of the mutual tokens between them with the
help of fuzzy sets, resulting in a mapping degree of the document with each domain. An experiment carried
out showing satisfying classification results with an improvement in the evaluation results of the proposed
system compared to Apache Lucene.
The document discusses research at UNED on enhancing authoring, modeling, and collaboration in e-learning environments. It outlines tools like ENLACE, PELICAN, and CARDS that allow for modeling and aggregation of educational content and learning objects from external tools. CARDS provides a metamodel for integrating outcomes from tools like concept mapping tools. PELICAN is an instructional design and collaboration platform that integrates external tools and supports adaptive contexts based on student performance. The research aims to automatically extract and classify learning objects from the web to enrich authored content and help students find relevant resources.
Using patterns to design technology enhanced learning scenarioseLearning Papers
This document discusses using design patterns to represent technology-enhanced learning scenarios. It reviews different mechanisms used to represent learning design issues, such as hypermedia models, ontologies, and educational modeling languages. The author proposes an approach using design patterns as they combine narrative representation with visualization and controlled vocabularies. Patterns are prepared by classifying them into categories like content managers, activity facilitators, and assessment producers. They are then applied to represent a specific learning scenario based on digital ink technologies.
Ontology learning techniques and applications computer science thesis writing...Tutors India
At Tutors India, we offer Computer science and Information Technology Research Guidance services – We deliver exceptional work where your dissertation will deserve publication without significant reworking or alternation.
For #Enquiry
https://www.tutorsindia.com
info@tutorsindia.com
(Whatsapp): +91-8754446690
(UK): +44-1143520021
Workshop on Learning Technology Standards for Agriculture and Rural Development (AgroLT 2008)
September 19, 2008, Athens, Greece
In conjunction with
4th International Conference on Information and Communication Technologies in Bio and Earth Sciences (HAICTA 2008)
A Survey on Text Mining-techniques and applicationRyota Eisaki
This document summarizes text mining techniques and applications. It discusses text mining processes like document gathering, pre-processing, transformation, feature selection, and pattern selection. It also describes text mining techniques including categorization, clustering, information extraction, information visualization, and natural language processing. Finally, it outlines applications of text mining in various domains such as business intelligence, bioinformatics, security, human resources, and web search enhancement.
ONTOLOGY VISUALIZATION PROTÉGÉ TOOLS – A REVIEW ijait
The document discusses ontology visualization tools in Protégé. It reviews four main visualization methods used in Protégé tools: indented list, node-link and tree, zoomable, and focus+context. It then examines specific Protégé tools that use each method, including their key features and limitations. The tools assessed are Protégé Class Browser (indented list), Protégé OntoViz and OntoSphere (node-link and tree), Jambalaya (zoomable), and Protégé TGVizTab (focus+context). The document concludes by summarizing and comparing the visualization characteristics of these Protégé tools.
Portable and Synchronized Distributed Learning Management System in Severe Ne...Fajar Purnama
Master thesis defense limited of Fajar Purnama, Graduate School of Science and Technology, Human Interface and Cyber Communication Laboratory.
Video: https://youtu.be/i8AERku88u8
Masters Thesis: https://www.publish0x.com/fajar-purnama-academics/portable-and-synchronized-distributed-learning-management-sy-xyvdwoz?a=4oeEw0Yb0B&tid=slideshare
A New Concept Extraction Method for Ontology Construction From Arabic TextCSCJournals
Ontology is one of the most popular representation model used for knowledge representation, sharing and reusing. The Arabic language has complex morphological, grammatical, and semantic aspects. Due to complexity of Arabic language, automatic Arabic terminology extraction is difficult. In addition, concept extraction from Arabic documents has been challenging research area, because, as opposed to term extraction, concept extraction are more domain related and more selective. In this paper, we present a new concept extraction method for Arabic ontology construction, which is the part of our ontology construction framework. A new method to extract domain relevant single and multi-word concepts in the domain has been proposed, implemented and evaluated. Our method combines linguistic, statistical information and domain knowledge. It first uses linguistic patterns based on POS tags to extract concept candidates, and then stop words filter is implemented to filter unwanted strings. To determine relevance of these candidates within the domain, different statistical measures and new domain relevance measure are implemented for first time for Arabic language. To enhance the performance of concept extraction, a domain knowledge will be integrated into the module. The concepts scores are calculated according to their statistical values and domain knowledge values. In order to evaluate the performance of the method, precision scores were calculated. The results show the high effectiveness of the proposed approach to extract concepts for Arabic ontology construction.
This document summarizes a workshop on data integration using ontologies. It discusses how data integration is challenging due to differences in schemas, semantics, measurements, units and labels across data sources. It proposes that ontologies can help with data integration by providing definitions for schemas and entities referred to in the data. Core challenges discussed include dealing with multiple synonyms for entities and relationships between biological entities that depend on context. The document advocates for shared community ontologies that can be extended and integrated to facilitate flexible and responsive data integration across multiple sources.
This summarizes an academic paper that proposes an automatic ontology creation method for classifying research papers. It uses text mining techniques like classification and clustering algorithms. It first builds a research ontology by extracting keywords and patterns from previous papers. It then uses a decision tree algorithm to classify new papers into disciplines defined in the ontology. The classified papers are then clustered based on similarities to group them. The method was tested on a dataset of 100 papers and achieved average precision of 85.7% for term-based and 89.3% for pattern-based keyword extraction.
Possibility of interdisciplinary research software engineering andnatural lan...Nakul Sharma
This document discusses the possibility of interdisciplinary research between software engineering and natural language processing. It provides a literature review of research papers from 2003 to 2014 related to applying tools and techniques from one field to the other. Some key areas discussed include generating UML diagrams from natural language text, developing ontologies to clarify meanings, and potential issues with joint research like determining complexity of sentences. The document proposes a flowchart for how artifacts could be analyzed using tasks from either field to enable interdisciplinary research.
Class Diagram Extraction from Textual Requirements Using NLP Techniquesiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Mining Users Rare Sequential Topic Patterns from Tweets based on Topic Extrac...IRJET Journal
This paper proposes a method to mine rare sequential topic patterns (URSTPs) from tweet data. It involves preprocessing tweets to extract topics, identifying user sessions, generating sequential topic pattern (STP) candidates, and selecting URSTPs based on rarity analysis. Experiments show the approach can identify special users and interpretable URSTPs, indicating users' characteristics. The paper aims to capture personalized and abnormal user behaviors through sequential relationships between extracted topics from successive tweets.
The increased potential of the ontologies to reduce the human interference has wide range of applications. This paper identifies requirements for an ontology development platform to innovate artificially intelligent web. To facilitate this process, RDF and OWL have been developed as standard formats for the sharing and integration of data and knowledge. The knowledge in the form of rich conceptual schemas called ontologies. Based on the framework, an architectural paradigm is put forward in view of ontology engineering and development of ontology applications and a development portal designed to support ontology engineering, content authoring and application development with a view to maximal scalability in size and complexity of semantic knowledge and flexible reuse of ontology models and ontology application processes in a distributed and collaborative engineering environment.
Ontology Based Approach for Semantic Information Retrieval SystemIJTET Journal
Abstract—The Information retrieval system is taking an important role in current search engine which performs searching operation based on keywords which results in an enormous amount of data available to the user, from which user cannot figure out the essential and most important information. This limitation may be overcome by a new web architecture known as the semantic web which overcome the limitation of the keyword based search technique called the conceptual or the semantic search technique. Natural language processing technique is mostly implemented in a QA system for asking user’s questions and several steps are also followed for conversion of questions to the query form for retrieving an exact answer. In conceptual search, search engine interprets the meaning of the user’s query and the relation among the concepts that document contains with respect to a particular domain that produces specific answers instead of showing lists of answers. In this paper, we proposed the ontology based semantic information retrieval system and the Jena semantic web framework in which, the user enters an input query which is parsed by Standford Parser then the triplet extraction algorithm is used. For all input queries, the SPARQL query is formed and further, it is fired on the knowledge base (Ontology) which finds appropriate RDF triples in knowledge base and retrieve the relevant information using the Jena framework.
In tech application-of_data_mining_technology_on_e_learning_material_recommen...Enhmandah Hemeelee
The document describes a recommendation system that applies data mining techniques to recommend e-learning materials. It proposes using LDAP for fast searching of materials across systems, JAXB for parsing content, and association rule mining and collaborative filtering for recommendations. A web spider collects content indexes from learning management systems and stores data in an LDAP directory. Users can search for related materials, and the system mines log data to associate frequently searched terms and recommend additional resources.
In tech application-of_data_mining_technology_on_e_learning_material_recommen...Enhmandah Hemeelee
The document describes a recommendation system that applies data mining techniques to recommend e-learning materials. It proposes using LDAP for fast searching of materials across systems, JAXB for parsing content, and association rule mining and collaborative filtering to generate recommendations. The system collects user activity data, analyzes it using Apriori algorithm to find related search terms and content, and stores results in an LDAP database to provide recommendations to users.
Comparative evaluation of four multi label classification algorithms in class...csandit
The classification of learning objects (LOs) enables users to search for, access, and reuse them
as needed. It makes e-learning as effective and efficient as possible. In this article the multilabel
learning approach is represented for classifying and ranking multi-labelled LOs, whereas
each LO might be associated with multiple labels as opposed to a single-label approach. A
comprehensive overview of the common fundamental multi-label classification algorithms and
metrics will be discussed. In this article, a new multi-labelled LOs dataset will be created and
extracted from ARIADNE Learning Object Repository. We experimentally train four effective
multi-label classifiers on the created LOs dataset and then, assess their performance based on
the results of 16 evaluation metrics. The result of this article will answer the question of: what is
the best multi-label classification algorithm for classifying multi-labelled LOs?
COMPARATIVE EVALUATION OF FOUR MULTI-LABEL CLASSIFICATION ALGORITHMS IN CLASS...cscpconf
The classification of learning objects (LOs) enables users to search for, access, and reuse them as needed. It makes e-learning as effective and efficient as possible. In this article the multilabel learning approach is represented for classifying and ranking multi-labelled LOs, whereas each LO might be associated with multiple labels as opposed to a single-label approach. A comprehensive overview of the common fundamental multi-label classification algorithms and metrics will be discussed. In this article, a new multi-labelled LOs dataset will be created and extracted from ARIADNE Learning Object Repository. We experimentally train four effective multi-label classifiers on the created LOs dataset and then, assess their performance based on the results of 16 evaluation metrics. The result of this article will answer the question of: what is the best multi-label classification algorithm for classifying multi-labelled LOs?
Although of the semantic web technologies utilization in the learning development field is a new research area, some authors have already proposed their idea of how an effective that operate. Specifically, from analysis of the literature in the field, we have identified three different types of existing applications that actually employ these technologies to support learning. These applications aim at: Enhancing the learning objects reusability by linking them to an ontological description of the domain, or, more generally, describe relevant dimension of the learning process in an ontology, then; providing a comprehensive authoring system to retrieve and organize web material into a learning course, and constructing advanced strategies to present annotated resources to the user, in the form of browsing facilities, narrative generation and final rendering of a course. On difference with the approaches cited above, here we propose an approach that is modeled on narrative studies and on their transposition in the digital world. In the rest of the paper, we present the theoretical basis that inspires this approach, and show some examples that are guiding our implementation and testing of these ideas within e-learning. By emerging the idea of the ontologies are recognized as the most important component in achieving semantic interoperability of e-learning resources. The benefits of their use have already been recognized in the learning technology community. In order to better define different aspects of ontology applications in e-learning, researchers have given several classifications of ontologies. We refer to a general one given in that differentiates between three dimensions ontologies can describe: content, context, and structure. Most of the present research has been dedicated to the first group of ontologies. A well-known example of such an ontology is based on the ACM Computer Classification System (ACM CCS) and defined by Resource Description Framework Schema (RDFS). It’s used in the MOODLE to classify learning objects with a goal to improve searching. The chapter will cover the terms of the semantic web and e-learning systems design and management in e-learning (MOODLE) and some of studies depend on e-learning and semantic web, thus the tools will be used in this paper, and lastly we shall discuss the expected contribution. The special attention will be putted on the above topics.
This document discusses applying semantic web technologies to enhance the services of e-learning systems. It proposes developing a semantic learning management system (S-LMS) based on technologies like XML, RDF, OWL and SPARQL to automate and accurately search for information on e-learning systems like Moodle. The S-LMS would add semantic capabilities to allow students to search for learning resources based on semantics and provide personalized, customized content tailored to individual needs. It presents applying ontologies and metadata to Moodle in order to define domains and describe learning content in a way that improves search, interoperability and reusability of educational resources.
This work presents a data architecture based on semantic web technologies that support to the inclusion of open materials in massive online courses. The framework provides transparent access to RDF data sources for Open Educational Resources stored in OpenCourseWare repositories.
Speaker(s): Nelson Piedra and Edmundo Tovar
Profile-based Dataset Recommendation for RDF Data Linking Mohamed BEN ELLEFI
This document summarizes Mohamed Ben Ellefi's PhD thesis defense on profile-based dataset recommendation for RDF data linking. The thesis proposes two approaches: a topic profile-based approach and an intensional profile-based approach. The topic profile-based approach models datasets as topics and recommends target datasets based on similarity between source and target topic profiles, achieving an average recall of 81% and reducing the search space by 86%. The approach shows better performance than baselines but needs improvement on precision.
A SEMANTIC BASED APPROACH FOR KNOWLEDGE DISCOVERY AND ACQUISITION FROM MULTIP...IJwest
This document describes a semantic-based approach for knowledge discovery and information extraction from multiple web pages using ontologies. It presents a model for storing web content in an organized, structured RDF format. Information extraction techniques and developed ontologies can then discover new knowledge with minimal time compared to manual efforts. The paper details two experiments applying this approach. Experiment 1 extracts staff profiles from web pages into RDF, discovering related research colleagues. Experiment 2 extracts student data from HTML tables into XML/RDF, enabling faster querying and analysis versus manual parsing. The approach effectively organizes unstructured web data for knowledge inference and acquisition.
CNI fall 2009 enhanced publications john_doove-SURFfoundationJohn Doove
- SURF is an organization in the Netherlands that works to improve ICT infrastructure for higher education and research.
- SURF is working on projects to develop "enhanced publications" which combine traditional publications like text with additional materials like data, maps, images and annotations.
- Several projects have been funded to create enhanced publications in fields like archaeology and psychology. Challenges include presentation, identification, long-term preservation and developing tools and infrastructure to support enhanced publications.
- Moving forward, SURF will work on developing repository infrastructure to store and share enhanced publications, creating guidelines and incentivizing their creation through things like legal reports and reward systems.
This document presents a new method for extracting class diagrams from textual requirements using natural language processing (NLP) techniques. It proposes the Requirements Analysis and Class diagram Extraction (RACE) system, which uses tools like the OpenNLP parser, a stemming algorithm, and WordNet to extract concepts and identify classes, attributes and relationships. The RACE system applies heuristic rules and a domain ontology to the output of the NLP tools to refine and finalize the extracted class diagram. The paper concludes that the RACE system demonstrates the effective use of NLP techniques to automate the extraction of class diagrams from informal natural language requirements specifications.
This presentation provides some thoughts on OER search. It includes a brief background on the metadata recommendation for OCW Consortium members from 2006. It also provides a bit of background on educational metadata and resource repositories. This presentation was prepared for a OER Search meeting hosted by Google on December 1, 2010. (This presentation replaces the draft presentation that had received 120 views.)
This document proposes using Word2Vec and decision trees to extract keywords from textual documents and classify the documents. It reviews related work on keyword extraction and text classification techniques. The proposed approach involves preprocessing text, representing words as vectors with Word2Vec, calculating frequently occurring keywords for each category, and using decision trees to classify documents based on keyword similarity. Experiments using different preprocessing and Word2Vec settings achieved an F-score of up to 82% for document classification.
Novel Database-Centric Framework for Incremental Information Extractionijsrd.com
Information extraction (IE) has been an active research area that seeks techniques to uncover information from a large collection of text. IE is the task of automatically extracting structured information from unstructured and/or semi structured machine-readable documents. In most of the cases this activity concerns processing human language texts by means of natural language processing (NLP). Recent activities in document processing like automatic annotation and content extraction could be seen as information extraction. Many applications call for methods to enable automatic extraction of structured information from unstructured natural language text. Due to the inherent challenges of natural language processing, most of the existing methods for information extraction from text tend to be domain specific. In this project a new paradigm for information extraction. In this extraction framework, intermediate output of each text processing component is stored so that only the improved component has to be deployed to the entire corpus. Extraction is then performed on both the previously processed data from the unchanged components as well as the updated data generated by the improved component. Performing such kind of incremental extraction can result in a tremendous reduction of processing time and there is a mechanism to generate extraction queries from both labeled and unlabeled data. Query generation is critical so that casual users can specify their information needs without learning the query language.
Architecture of an ontology based domain-specific natural language question a...IJwest
The document summarizes the architecture of an ontology-based domain-specific natural language question answering system. The proposed architecture defines four main modules: 1) question processing which analyzes and classifies questions and reformulates queries, 2) document retrieval which retrieves relevant documents, 3) document processing which processes retrieved documents, and 4) answer extraction which extracts and generates responses. Natural language processing techniques and ontologies are used to analyze questions and documents and extract relationships and answers. The system aims to generate concise, specific answers to natural language questions in a given domain and achieved 94% accuracy in testing.
A Reuse-based Lightweight Method for Developing Linked Data Ontologies and Vo...María Poveda Villalón
The document proposes a lightweight methodology called LOT (Linked Open Terms) for developing Linked Data ontologies and vocabularies in a reusable way. The methodology is data-driven and focuses on ontology search, selection, integration, completion and evaluation activities. It provides guidelines for reusing existing terms and linking them according to Linked Data principles while keeping the processes lightweight. The methodology is intended to help domain experts create ontologies and vocabularies for publishing data on the semantic web in an interoperable way without requiring extensive knowledge engineering expertise. Future work involves providing more detailed guidelines, examples, and connecting existing tools to support each step of the methodology.
French machine reading for question answeringAli Kabbadj
This paper proposes to unlock the main barrier to machine reading and comprehension French natural language texts. This open the way to machine to find to a question a precise answer buried in the mass of unstructured French texts. Or to create a universal French chatbot. Deep learning has produced extremely promising results for various tasks in natural language understanding particularly topic classification, sentiment analysis, question answering, and language translation. But to be effective Deep Learning methods need very large training da-tasets. Until now these technics cannot be actually used for French texts Question Answering (Q&A) applications since there was not a large Q&A training dataset. We produced a large (100 000+) French training Dataset for Q&A by translating and adapting the English SQuAD v1.1 Dataset, a GloVe French word and character embed-ding vectors from Wikipedia French Dump. We trained and evaluated of three different Q&A neural network ar-chitectures in French and carried out a French Q&A models with F1 score around 70%.
Similar to MULTI-LEARNING SPECIAL SESSION / EDUCON 2018 / EMADRID TEAM (20)
Recognizing Lifelong Learning Competences: A Report of Two Cases - Edmundo TovareMadrid network
The document discusses two European projects - STEMSOFT and TEASPILS - that aim to develop micro-credentials for lifelong learning. STEMSOFT focuses on developing soft skills for STEM professionals through open online courses, while TEASPILS uses IoT planters to teach environmental awareness. Both projects map learning outcomes to competencies frameworks and plan to pilot short courses to certify skills acquisition through micro-credentials. The document also outlines the European policy context around micro-credentials and lifelong learning, and how the projects aim to address skills gaps through flexible, targeted training opportunities.
Recognition of learning: Status, experiences and challenges - Carlos Delgado ...eMadrid network
1. The document discusses recognition of learning, experiences, and challenges. It describes an eMadrid Network special session on this topic with presentations from various universities.
2. The UC3M presentation focuses on recognizing the value of recognition in education. It discusses formats for recognition like badges and credentials and uses cases at UC3M involving competitions, gamification, and digital credentials.
3. Recognizing learning is important for motivation and signaling achievement. Recognition elements should be integrated into instructional design similar to activities and assessments.
Bootstrapping serious games to assess learning through analytics - Baltasar F...eMadrid network
This document summarizes research on using serious games to assess learning through analytics. It discusses how games can be validated using pre-post tests to ensure effectiveness and provide training data for machine learning models. Interaction data from validated games can then train models to predict learning from gameplay without exams. The researchers developed tools like a game analytics tracker, validation tool, and analysis tool to facilitate collecting interaction data, validating games, and analyzing results. Their authoring tool integrates these analytics capabilities. Future work will integrate machine learning models into the validation tool to directly provide assessment scores based on interaction data. The goal is to close the assessment loop for serious games.
Meta-review of recognition of learning in LMS and MOOCs - Ruth CoboseMadrid network
The meta-review examines 10 studies that provide overviews of recognition of learning techniques in learning management systems (LMSs) and massive open online courses (MOOCs). The studies were published between 2017-2021 and included reviews, experiences, and challenges. Most focused on MOOCs and used badges for recognition. Results showed techniques like gamification and badges positively impact motivation and engagement. Limitations included short study periods and small samples. Future work could study applications over longer periods, combine data types, and consider diverse stakeholders and environments.
The document announces that Abdallah Yusuf Al-Zoubi, Manuel Castro, Fadi Shahroury and Elio Sancristobal received the Best Paper Award in the category of Innovation Engineering Education from the IEEE Global Engineering Education Conference held from May 11-14, 2023 at the American University of Kuwait in Salmiya, Kuwait for their paper titled "Impact of Remate Labs in Preparing Students for Work 4.0: The Story at Princess Sumaya University for Technology."
Seminario eMadrid_Curso MOOC_Antonio de Nebrija_Apología del saber.pptx.pdfeMadrid network
Este documento resume un curso MOOC sobre Antonio de Nebrija diseñado por la Universidad Nebrija. El curso conmemora el 500 aniversario de Nebrija y tiene como objetivo divulgar su figura y legado. Consiste en 6 módulos y 30 horas de contenido sobre la vida y obra de Nebrija, así como sobre temas como la pedagogía, la gramática del español y la literatura. El curso tuvo una gran participación con 1.540 inscritos de 62 nacionalidades y recibió buenas valoraciones.
This document discusses digital education initiatives at Politehnica University of Timisoara. It describes creating open educational resources (OERs) through collaboration between students and faculty. Students research topics and use multimedia tools to create OERs that are peer-reviewed and published with Creative Commons licenses for reuse. The document also outlines virtual mobility programs that improve students' digital skills through international collaboration projects using virtual reality tools and blogging.
The document discusses challenges in establishing digital credentials for learning achievements that were investigated by the DiBiHo research project. It identifies key challenges such as technical interoperability, credential revocation, and privacy-enhancing cryptography. A proof of concept was created to test proposed solutions for these challenges. The presentation will discuss the identified challenges, proposed approaches, and remaining open questions regarding digital credentials.
The document discusses the evolution of MOOC certification and credentialing from stage 1 of course certificates to stage 4 of accredited learning pathways. It outlines Federica's experience with early partnerships providing certification via Coursera and edX courses. Federica has since developed an in-house system awarding certificates and badges for its own courses. The document also covers recent European trends in microcredentials and Federica's key partnerships in Italy providing certification for public sector training.
The document discusses European Digital Credentials for Learning, which aims to empower citizens to own credentials that can be easily shared across Europe. The initiative seeks to reduce market fragmentation, create an EU skills data space, and remove barriers to credential recognition. The infrastructure will include standards, services, and software to allow credentials to be issued, stored, verified, and shared digitally. This framework aims to capture all types of learning and be interoperable, multilingual, and applicable across one's career. It is a central part of the EU's agenda to support lifelong learning and labor mobility.
2022_12_16 «“La informática en la educación escolar en Europa”, informe Euryd...eMadrid network
El informe resume la enseñanza de la informática en los sistemas educativos europeos, incluyendo cuándo se introduce la asignatura, cómo se distribuye a lo largo de las etapas educativas, los contenidos abordados y la preparación del profesorado. Algunos de los hallazgos clave son que la mayoría de países comienzan la informática en primaria o secundaria, cubren sobre todo algoritmos y programación, y existe escasez de profesores especializados en la asignatura.
2022_12_16 «Informatics – A Fundamental Discipline for the 21st Century»eMadrid network
The document provides an overview of efforts to establish informatics as a fundamental discipline in school education across Europe. It discusses the Informatics for All coalition which developed an Informatics Reference Framework for School to advocate for including informatics in curriculums. The framework defines 11 core topics and was informed by broad consultation. The status of informatics in schools across Europe is then analyzed according to this framework, finding most systems integrate it into other subjects rather than as a standalone discipline. Informatics is positioned as a new fundamental competence and language for all students akin to mathematics and languages.
2022_11_11 «AI and ML methods for Multimodal Learning Analytics»eMadrid network
This document discusses using multimodal data and machine learning methods for analyzing learning across multiple contexts. It describes several studies that collected eye tracking, physiological, video, and other data from participants in contexts like playing Pacman, self-assessment tests, debugging programs, educational games, and collaborative concept mapping. Machine learning models were developed to predict outcomes like test scores, effort, and performance using features from the multimodal data. The document discusses the value of collecting multimodal data, developing explainable AI pipelines, and generalizing models across different learning contexts and tasks. It concludes by considering opportunities for using online learning system logs and designing more similar learning contexts.
2022_11_11 «The promise and challenges of Multimodal Learning Analytics»eMadrid network
1. The document discusses three conceptualizations of multimodal learning analytics (MMLA): MMLA to automate human tasks, augment teaching and learning practices, and as a research methodology.
2. It examines what modalities of data are used in MMLA, including video/audio data, eye tracking data, physiological sensors, and location sensing. Machine learning has been applied to MMLA tasks like classifying collaboration.
3. Challenges of MMLA include connecting findings to learning theory, addressing ethics concerns like privacy and surveillance, and determining what behaviors are considered good or bad in education. Students have mixed reactions to being analyzed by MMLA.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
MULTI-LEARNING SPECIAL SESSION / EDUCON 2018 / EMADRID TEAM
1. Using Text Mining and Linked Open Data to
assist the Mashup of Educational Resources
Santa Vallejo-Figueroa
Miguel Rodriguez-Artacho
Manuel Castor-Gil
Elio San Cristobal
IEEE Global Engineering Education Conference
Abril 2018, Santa Cruz de Tenerife
Universidad Nacional de Educación a Distancia
(UNED)
2. 2
Context (1)
Creation: hard and time-consuming
task
There not exist standards
Necessity of OERs is increasing
Several institutions promote the
generation and use of OERs
The integration of learning contents
is a niche for reusing of such
contents
We refer to an online course as an
OER
4. 4
Method
Premises for the method:
1) Exists an OER repository
containing online courses
2) Courses have a basic structure, as
minimum a textual description
3) Instructional aspects of contents
are not taken into account
4) The human creator will get
suggested resources to be related
and integrated according to final
decisions of the creator.
Aim 1: to exploit
existent LOD
information for
integrating contents
for OERs
Aim 2: to assist in the
creation of an OER
(course)
7. 7
Method - Text processing (1)
Named Entities (NE) and relevant
words are detected for retrieved
courses
A Named Entity is a universal-known
word with a unique meaning, such
as persons, locations, organizations,
etc.
Relevant words are keywords or
concepts to “describe” a text
A concept is a semantic class (group)
of terms sharing a similar idea.
A keyword is a term with a number
of occurrences in sentences.
A concept can group words and
keywords
8. 8
Method - Text processing (2)
Semantic information from the
DBpedia knowledge base is used
An OER is represented by:
a syntactic layer (textual
description), and
a semantic layer (sets of NEs
and relevant words)
The semantic layer points to
resources from DBpedia
Each course is indexed by using a
text search engine (Lucene)
8
9. 9
Method - Query Generation
A course is retrieved by using
a text fragment or set of
words
NEs and/or relevant words
are identified from input text
These are used to formulate
a simple text query over the
text search engine
No SPARQL queries are
required
9
10. 10
Method - Query Processing (1)
The most relevant courses are
retrieved from the semantic
index
For syntactic search, NEs,
concepts, and keywords from
query are used for searching on
fields ([NEs], [concepts],
[keywords])
Only NEs, concepts, and
keywords are used from the
query
A list of retrieved courses is
ranked according to its similarity
respect to the query
10
11. 11
Method - Query Processing (2)
For semantic search, a
matching of relationships
between NEs/concepts of the
query against NEs/concepts of
each retrieved course is made
The relations (graph) that each
NE/concept has in the
DBpedia are explored
The set of graphs of the query
and the set of graphs of each
retrieved course are obtained
Those courses with greater
matching to the query are
ranked as a result
11
12. 12
Method - Results Processing
The relationships between
the NEs/concepts of retrieved
courses are given to the
creator
Per each retrieved course its
NEs/concepts are connected
by means of a concept map
The main idea is to represent
how NEs/concepts from
retrieved courses are related.
The creator can generate a big
picture about the mashup of
OERs and DBpedia resources
12
13. 13
Preliminary results (1)
An implementation was developed on an 8GB RAM
Linux machine by using Java (web application), DBpedia
SpotLight, KeyGraph, Lucene, and MySQL
A total number of 265 courses were retrieved from the
UK Open University
The application was tested by queries about Parallel
Computing, Database, Software, Computer Aided
Software Engineering, Data Structures, and Operating
Systems
16. 16
Preliminary results (4)
At this moment only the
syntactic and semantic
search has been
implemented
We are working on a
similarity measure for
distinguising OERs with
similar names
The concept maps are not
generated yet
The implementation shows
the feasibility of this
approach
17. 17
Conclusions (1)
This work proposes an approach to assist to the human
creator in the generation or re-structuring of courses
A course is an educational resource
The approach exploits:
Text mining techniques to identity key elements
from text
Semantic linked information from the DBpedia
knowledge base
Stages of the method
Getting OERs
Text processing
Query generation
Query processing
Results processing
18. 18
Conclusions (2)
The approach was implemented as a prototype
showing promising results
Real experimentation
265 online courses related to Computer Science were
retrieved from the UK Open University
Future work
Enhance the semantic similarity measure
The generation of concept maps
19. Using Text Mining and Linked Open Data to
assist the Mashup of Educational Resources
Santa Vallejo-Figueroa
Miguel Rodriguez-Artacho
Manuel Castor-Gil
Elio San Cristobal
IEEE Global Engineering Education Conference
Abril 2018, Santa Cruz de Tenerife
Universidad Nacional de Educación a Distancia
(UNED)
Thanks!
Editor's Notes
This is a work developed in a colaboration between Santa Vallejo-Figueroa, Miguel Rodriguez-Artacho, Manuel Castor-Gil, and I
It is about the intersection of Educational Resources, Text Mining, and the Semantic Web (Linked Open Data)
- As we know, Open Educational Resources are very useful means for facilitating teaching and learning tasks
- But its creation poses very challenges not only from Instructional point of view, the integration of tools and standards for its creation is a very hard and time-consuming task.
- Although, by definition, Open Educational Resources must be open, accessible and reusable means, there not exist standard technologies for this purpose.
However, according to its philosophy, more and more OERs are required in many areas for teaching and learning.
Many researchers and institutions are interested on the generation, distribution and use of OERs. International initiatives are evidence of this interest: Open Universitues, Coursera, Udacity, Stanford University, MIT, etc.
Regardless the Instructional requirements, the integrattion of learning contents is a tendency and challenge to promote the reuse of existent "base" learning materials.
These "base" materials can come from diverse sources of information: another kind of repositories, knowledge bases, dictionaries, etc.
In this work, we refer to an online course as an OER, which is created by a human creator
One of the most applied approaches for publishing, reusing information in several domains is the Semantic Web through the Linked Open Data initiative.
The advantage of this approach is that information is well-structured and well-defined in that way it can be consumed by human and computer applications. It represents the information based on semantics.
However, apply Linked Open Data on OERs domains is not easy because must exist a correspondence between information to be represented and the manner how is organized (taking into account the semantics)
The general premise is that well represented and organized information of OERs facilitate its search and consequently its re-use
The proposed work exploits the organization and contents of information of a Linked Open Data knowledge base (DBpedia) to suggest core-components (from LOD) to human creators of OERs
Our intention is extract and integrate information from LOD to assit human creators of OERs
For the proposed approach, the following premises are considered
- It exists a repository of courses (not necessarly OERs). The elements of this repository are used to populate the initial knowledge base. Information from courses will feed the search of LOD resources.
- The courses in the repository have as minimum a textual description, which is used to extract information from courses
- In the integration of information no Instructional elements are taken into account. The human creator is responsible of these Instructional tasks
- The result of the approach is a set of LOD resources which the human creator can integrate in a new one course.
This is the general architecture of our method
Each component of this architecture is next described
- In the first stage of the approach, and only once, online courses (OERs) are retrieved from a repository. This can be one online repository, like UK Open University, MIT OpenCourseWare.
- The courses can be retrieved by using SPARL queries if its a LOD repository, SQL queries if is a relational-based repository, or text queries if is a web-based repository
- We use SPARQL queries for retrieving courses from the UK Open University
- The key elements in the core of this approach are Named Entities and relevant words.
- Named Entities and relevant words are extracted from the textual description of each course. For this the text is processed to detect them.
- A Named Entity is a universally-well-know word which meaning is unique: persons, locations, acronyms, etc.
- Relevant words can be concepts (general words -semantic classes- relating specific words) or keywords (repeated words)
- A concept may include a group of keywords
- For the processing of text we use the DBpedia knowledge base, exploiting its semantic information.
- DBpedia is the largest knowledge base from the Linked Open Data initiative, it contains well-structured information from Wikipedia. Such information is annoted semantically.
- An OER (course) is represented by means of two layers: syntactic (textual description) and semantic (sets of NEs and relevant words)
- For the semantic layer, each OER contains references (URLs) to resources from DBpedia. In that way each resource is accesible from the course.
- Based on both layers, each course is indexed by using a text search engine (in our case Lucene)
- Note that RDF is not employed in the representation, only syntactic and semantic contents
- After each course is indexed, the system is ready for querying it
- The idea in this component is the human creator can search, based on a input text, related courses of its interests
- The creator can introduced a fragment of text (article, news, webpage), or one or more sentence
- From the input text NEs and relevant words are identified by the same component of Text Processing
- Based on the NEs or relevant words, a simpre text is executed over the search engine
- Note that here no SPARQL queries are required
- The most relevant courses are retrieved from the index by using NEs and relevant words
- From these, two type of searches are executed: syntactic and semantic
- For syntactic search each NE, concept and keyword identified in the input text is searched on its corresponding field in the index.
- As a result, a list of retrieved courses is ranked according to its similarity to the query
- By the moment, the similarity takes into account the matching in the following order: more importance to NEs, then to concepts, and finally to keywords
- For the semantic search the approach makes a matching between the NEs and concepts of the query and the NEs and concepts of each retrieved course.
- Note that in this search keywords are not used because only NEs and concepts have semantic meaning in DBpedia
- The approach takes advantage that DBpedia is organized by means of a graph of NEs and concepts
- Thus, a NE or concept has connections in form of subgraph in DBpedia, where NEs and concepts are nodes,
and an edge is a relation to other NE/concept. These relations take the form of a RDF triple (node, relation, node), that is Subject, Predicate, Object
- Two sets of subgraphs are retrieved from DBpedia, one for the query of creator, and the other for each retrieved course
- The subgraph of the query is compared with the subgraph of each retrieved course. Those courses with greater matching are ranked as a result
- Once the courses are retrieved and ranked, the results must be presented to the creator for using them
- The idea of this module is exploits the advantage of associations of DBpedia resources (NEs and concepts) to provide a better understanding of results to the creator.
- For this, for each retrieved course its corresponding graph is represented by means of a concept map
- A concept map is an ideal means for transmitting the semantic meaning of a course (NEs and concepts)
- Thus, from the constructivist learning approach, the creator can generate or re-structure a course following the key elements from existing courses.
- This will represent a big picture about the mashup of OERs and DBpedia resources.
- The approach has been implemented by means of a Java application integrating DBpedia (knowledge base), SpotLight (tool for identtify NEs and concepts), KeyGraph (tool for identtify keywords), Lucene (text search engine) and MySQL (database engine for storing original information of online courses)
- 265 courses related to Computer Science were retrieved from the repository of the UK Open University
- Several queries were executed on the implemented application: Parallel Computing, Database, Software, Computer Aided Software Engineering, Data Structures, and Operating Systems
Here are presented some intermediate results within each stage
First, SPARQL queries are performed for getting OERs from an existent repository
Then, the textual description of the course is processed to identify Named Entities and concepts (in bold)
In the Query Genertation the iinput text is processeed to identity Named Entities and concepts (in bold)
Relevant words (Named Entities (NE) and concepts (CO) ) are searched (appearance) in Syntactic search
In Semantic search, related resources from DBpedia are identified for Named Entities and concepts
This is the result for the query “Computer Aided Software Engineering”
As we can see, the implementation must be enhanced for a better ranking of resources. Here appear two different courses with the same name “Software Engineering”. We are working on this.
- The entire approach has not been implemented yet
- Concept maps are not yet generated
- At this moment we are working for improving the similarity measure for a better ranking of results
- By the moment, the obtained results are promising and demonstrate the feasibility of the proposed approach
- As conclusions we can summarize the following
- The main idea of the proposed approach is to assist to the human creator in the generation of courses from scratch or re-structure an existing
- The approach takes advantage of:
a) Text mining techniques for identitying key elements from text (NEs and relevants -concepts and keywords-)
b) Semantic linked information from DBpedia knowledge base, which is the largest knowledge base from the Linked Open Data initiative, it contains well-structured semantic information which is used for linking information from courses
For technical purposes, we denote an educational resource as a course.
We not take into account educational aspects from such courses.
We use the textual description of the course, considering as valid and correct such description
The stages of the method are: Getting OERs, Text processing, Query generation, Query processing, and Results processing
The approach was implemented as a Java web application by using open source libraries
Although the entire approach has not been implemented, the obtained results are promising
The prototype was tested on a real scenario working on 265 online courses related to Computer Science from the UK Open University
At this moment the prototype does not implement the entire method
We are working on the last two stages: Query processing, and Results processing
On Query processing we want enhance the ranking of resulting courses by adapting the semantic similarity measure
The stage Results processing is not yet implemented for the generation of concept maps