This document proposes a framework for preparing open-corpus content to support adaptive hypermedia systems. The framework includes three components: 1) structural analysis to segment web pages and remove redundant information, 2) statistical analysis to extract concepts using techniques like hidden Markov models, and 3) intelligent slicing to fulfill specific information requests from adaptive systems by retrieving and tailoring open-corpus content. The goal is to leverage existing open web content for adaptive systems by automatically preparing and enriching content with metadata in a format agnostic to any specific system.
A Grid-Enabled Infrastructure for Resource Sharing, E-Learning, Searching and...ijgca
In the recent years, service-based approaches for sharing of data among repositories and online learning
are rising to prominence because of their potential to meet the requirements in the area of high
performance computing. Developing education based grid services and assuring high availability,
reliability and scalability are demanding in web service architectures. On the other hand, grid computing
provides flexibility towards aggregating distributed CPU, memory, storage, data and supports large
number of distributed resource sharing to provide the full potential for education like applications to share
the knowledge that can be attainable on any single system. However, the literature shows that the potential
of grid resources for educational purposes is not being utilized yet. In this paper, an education based grid
framework architecture that provides promising platform to support sharing of geographically dispersed
learning content among universities is developed. It allows students, faculty and researchers to share and
gain knowledge in their area of interest by using e-learning, searching and distributed repository services
among universities from anywhere, anytime. Globus toolkit 5.2.5 (GTK) software is used as grid
middleware that provides resource access, discovery and management, data movement, security, and so
forth. Furthermore, this work uses the OGSA-DAI that provides database access and operations. The
resulting infrastructure enables users to discover education services and interact with them using the grid
portal.
The document proposes creating a digital library at Anonymous University using the Dublin Core metadata standard and Greenstone digital library software. It recommends training library staff on Dublin Core, the controlled vocabularies LCNAF and DCT, and assigning roles for the project such as project manager, digital manager, curator, and digitization staff. It also outlines plans for metadata elements, training procedures, collection assessment, and ensuring quality control of the digital library materials and records.
A novel method for generating an elearning ontologyIJDKP
The Semantic Web provides a common framework that allows data to be shared and reused across
applications, enterprises, and community boundaries. The existing web applications need to express
semantics that can be extracted from users' navigation and content, in order to fulfill users' needs. Elearning
has specific requirements that can be satisfied through the extraction of semantics from learning
management systems (LMS) that use relational databases (RDB) as backend. In this paper, we propose
transformation rules for building owl ontology from the RDB of the open source LMS Moodle. It allows
transforming all possible cases in RDBs into ontological constructs. The proposed rules are enriched by
analyzing stored data to detect disjointness and totalness constraints in hierarchies, and calculating the
participation level of tables in n-ary relations. In addition, our technique is generic; hence it can be applied
to any RDB.
When we look at the rapid growth of scientific databases on the Internet in the past decade, we tend to take the accessibility and provenance of the data for granted. As we see a future of increased database integration, the licensing of the data may be a hurdle that hampers progress and usability. We have formulated four rules for licensing data for open drug discovery, which we propose as a starting point for consideration by databases and for their ultimate adoption. This work could also be extended to the computational models derived from such data. We suggest that scientists in the future will need to consider data licensing before they embark upon re-using such content in databases they construct themselves.
This document is the proposal for the TRACKOER project that is supported by the JISC Open Educational Resources Rapid Innovation programme (OERRI). In TRACKOER we are developing potential solutions to how to keep track of content in the open. We are looking both at ways to follow content as it moves from one server to another and then gets reused, and at how to capture other changes that people may make with cut and paste editing. The rationale for the project is to understand whether content gets reused but it also offers a model that could help track other activity around shared content. More about the project progress is available via http://track.olnet.org/ and the project blog at http://cloudworks.ac.uk/tag/view/TrackOER .
VIZ-VIVO: Towards Visualizations-driven Linked Data NavigationMuhammad Javed
Paper published in ISWC correlated workshop VOILA 2016.
Abstract: Scholars@Cornell is a new project of Cornell University Library (CUL) that provides linked data and novel visualizations of the scholarly record. Our goal is to enable easy discovery of explicit and latent patterns that can reveal high-impact research areas, the dynamics of scholarly collaboration, and expertise of faculty and researchers. We describe VIZ-VIVO, an extension for the VIVO framework that enables end-user exploration of a scholarly knowledge-base through a configurable set of data-driven visualizations. Unlike systems that provide web pages of researcher profiles using lists and directory-style metaphors, our work explores the power of visual metaphors for navigating a rich semantic network of scholarly data modeled with the VIVO-ISF ontology. We produce dynamic web pages using D3 visualizations and bridge the user experience layer with the underlying semantic triple-store layer. Our selection of visual metaphors enables end users to start with the big picture of scholarship and navigate to individuals faculty and researchers within a macro visual context. The D3-enabled interactive environment can guide the user through a sea of scholarly data depending on the questions the user wishes to answer. In this paper, we discuss our process for selection, design, and development of an initial set of visualizations as well as our approach to the underlying technical architecture. By engaging an initial set of pilot partners we are evaluating the use of these data-driven visualizations by multiple stakeholders, including faculty, students, librarians, administrators, and the public.
This study analyzes distributed database solutions to address scalability, availability and performance issues facing the Brazilian government's Portal do Aluno project. The project aims to create an educational portal for 5 million students but faces challenges from relational databases becoming single points of failure. To solve this, the study explores NoSQL distributed databases offering horizontal scalability while relaxing consistency requirements. It finds that a key-value or document-oriented database best meets the project's needs to scale out reads and writes across multiple nodes with high availability.
The document discusses the PRELIDA project which aims to identify differences between linked data and digital preservation communities and analyze gaps between the two. The objectives are to collect use cases of long-term preservation of linked data and identify challenges of applying existing preservation approaches to linked data. Issues discussed include differences in preservation requirements for linked data versus other data types and whether linked data preservation can be viewed as a special case of web archiving.
A Grid-Enabled Infrastructure for Resource Sharing, E-Learning, Searching and...ijgca
In the recent years, service-based approaches for sharing of data among repositories and online learning
are rising to prominence because of their potential to meet the requirements in the area of high
performance computing. Developing education based grid services and assuring high availability,
reliability and scalability are demanding in web service architectures. On the other hand, grid computing
provides flexibility towards aggregating distributed CPU, memory, storage, data and supports large
number of distributed resource sharing to provide the full potential for education like applications to share
the knowledge that can be attainable on any single system. However, the literature shows that the potential
of grid resources for educational purposes is not being utilized yet. In this paper, an education based grid
framework architecture that provides promising platform to support sharing of geographically dispersed
learning content among universities is developed. It allows students, faculty and researchers to share and
gain knowledge in their area of interest by using e-learning, searching and distributed repository services
among universities from anywhere, anytime. Globus toolkit 5.2.5 (GTK) software is used as grid
middleware that provides resource access, discovery and management, data movement, security, and so
forth. Furthermore, this work uses the OGSA-DAI that provides database access and operations. The
resulting infrastructure enables users to discover education services and interact with them using the grid
portal.
The document proposes creating a digital library at Anonymous University using the Dublin Core metadata standard and Greenstone digital library software. It recommends training library staff on Dublin Core, the controlled vocabularies LCNAF and DCT, and assigning roles for the project such as project manager, digital manager, curator, and digitization staff. It also outlines plans for metadata elements, training procedures, collection assessment, and ensuring quality control of the digital library materials and records.
A novel method for generating an elearning ontologyIJDKP
The Semantic Web provides a common framework that allows data to be shared and reused across
applications, enterprises, and community boundaries. The existing web applications need to express
semantics that can be extracted from users' navigation and content, in order to fulfill users' needs. Elearning
has specific requirements that can be satisfied through the extraction of semantics from learning
management systems (LMS) that use relational databases (RDB) as backend. In this paper, we propose
transformation rules for building owl ontology from the RDB of the open source LMS Moodle. It allows
transforming all possible cases in RDBs into ontological constructs. The proposed rules are enriched by
analyzing stored data to detect disjointness and totalness constraints in hierarchies, and calculating the
participation level of tables in n-ary relations. In addition, our technique is generic; hence it can be applied
to any RDB.
When we look at the rapid growth of scientific databases on the Internet in the past decade, we tend to take the accessibility and provenance of the data for granted. As we see a future of increased database integration, the licensing of the data may be a hurdle that hampers progress and usability. We have formulated four rules for licensing data for open drug discovery, which we propose as a starting point for consideration by databases and for their ultimate adoption. This work could also be extended to the computational models derived from such data. We suggest that scientists in the future will need to consider data licensing before they embark upon re-using such content in databases they construct themselves.
This document is the proposal for the TRACKOER project that is supported by the JISC Open Educational Resources Rapid Innovation programme (OERRI). In TRACKOER we are developing potential solutions to how to keep track of content in the open. We are looking both at ways to follow content as it moves from one server to another and then gets reused, and at how to capture other changes that people may make with cut and paste editing. The rationale for the project is to understand whether content gets reused but it also offers a model that could help track other activity around shared content. More about the project progress is available via http://track.olnet.org/ and the project blog at http://cloudworks.ac.uk/tag/view/TrackOER .
VIZ-VIVO: Towards Visualizations-driven Linked Data NavigationMuhammad Javed
Paper published in ISWC correlated workshop VOILA 2016.
Abstract: Scholars@Cornell is a new project of Cornell University Library (CUL) that provides linked data and novel visualizations of the scholarly record. Our goal is to enable easy discovery of explicit and latent patterns that can reveal high-impact research areas, the dynamics of scholarly collaboration, and expertise of faculty and researchers. We describe VIZ-VIVO, an extension for the VIVO framework that enables end-user exploration of a scholarly knowledge-base through a configurable set of data-driven visualizations. Unlike systems that provide web pages of researcher profiles using lists and directory-style metaphors, our work explores the power of visual metaphors for navigating a rich semantic network of scholarly data modeled with the VIVO-ISF ontology. We produce dynamic web pages using D3 visualizations and bridge the user experience layer with the underlying semantic triple-store layer. Our selection of visual metaphors enables end users to start with the big picture of scholarship and navigate to individuals faculty and researchers within a macro visual context. The D3-enabled interactive environment can guide the user through a sea of scholarly data depending on the questions the user wishes to answer. In this paper, we discuss our process for selection, design, and development of an initial set of visualizations as well as our approach to the underlying technical architecture. By engaging an initial set of pilot partners we are evaluating the use of these data-driven visualizations by multiple stakeholders, including faculty, students, librarians, administrators, and the public.
This study analyzes distributed database solutions to address scalability, availability and performance issues facing the Brazilian government's Portal do Aluno project. The project aims to create an educational portal for 5 million students but faces challenges from relational databases becoming single points of failure. To solve this, the study explores NoSQL distributed databases offering horizontal scalability while relaxing consistency requirements. It finds that a key-value or document-oriented database best meets the project's needs to scale out reads and writes across multiple nodes with high availability.
The document discusses the PRELIDA project which aims to identify differences between linked data and digital preservation communities and analyze gaps between the two. The objectives are to collect use cases of long-term preservation of linked data and identify challenges of applying existing preservation approaches to linked data. Issues discussed include differences in preservation requirements for linked data versus other data types and whether linked data preservation can be viewed as a special case of web archiving.
Applications of xml, semantic web or linked data in Library/Information Servi...Nurhazman Abdul Aziz
Applications of XML, Semantic Web & Linked Data in Library/Information Services & Skills needed by System Librarians.
H6716 (Internet & Web Technologies) & K6224 (Internet Technologies & Applications)
Semester 2 – 2011/2012
Hazman Aziz, Librarian (Library Technology & Systems)
Amirrudin Dahlan, Senior IT Specialist (Center for IT & Services)
Nanyang Technological University
Enabling Accessible Resource Access via Service ProvidersAlexander Haffner
Libraries have become digitized and are using information technology for storing and managing their
resource inventory. Additional metadata are used for describing properties of non-digital assets as well
as of purely digital resources. In particular, as the amount of digital resources increases, there is
demand for centralized services for searching and distribution of content.
Stakeholders of libraries and publishing industry already have made progress in areas of archival
strategies and standardization of preservation strategies. A variety of metadata standards and
exchange protocols enable service providers to offer a single access point to resources. However,
there is still an increased demand for improvement of resource organization and enhanced quality
particularly in terms of accessibility. This paper presents strategies to increase accessibility of
resources as a valuable step towards access-for-all. For consideration of accessibility within the
publishing chain we analyse the whole processing chain and identify stakeholders such as national
libraries and their corresponding responsibilities for ingest, archival storage and dissemination of
digital resources.
One Standard to rule them all?: Descriptive Choices for Open EducationR. John Robertson
R. John Robertson1, Lorna Campbell1, Phil Barker2, Li Yuan3, and Sheila MacNeill1
1Centre for Academic Practice and Learning Enhancement, University of Strathclyde, 2Institute for Computer Based Learning, Heriot-Watt University 3Institute for Cybernetic Education, University of Bolton
Drawing on our experience of supporting a nationwide Open Educational Resources programme (the UKOER programme), this presentation will consider the diverse range of approaches to describing OERs that have emerged across the programme and their impact on resource sharing, workflows, and an aggregate view of the resources.
Due to the diverse nature of the projects in the programme, ranging from individual educators to discipline-based consortia and institutions, it was apparent that no one technical or descriptive solution would fit all. Consequently projects were mandated to supply only a limited amount of descriptive information (programme tag, author, title, date, url, file format, file size, rights) with some additional information suggested (language, subject classifications, keywords, tags, comments, description). Projects were free to choose how this information should be encoded (if at all), stored, and shared.
In response, the projects have taken many different approaches to the description and management of resources. These range from using traditional highly structured and detailed metadata standards to approaches using whatever descriptions are supported by particular web2.0 applications. This experimental approach to resource description offers the wider OER community an opportunity to examine and assess the implications of different strategies for resource description and management
This paper illustrates a number of examples of projects’ approaches to description, noting the workflows and effort involved. We will consider the relationship of the choice of tool (repository, web2.0 application, VLE) to the choice of standards; and the relationship between local requirements and those of the wider community.
We will consider the impact of those choices on the dissemination and discoverability of resources. For example, the implications of resource description choices for discovery services which draw on multiple sources of OERs.
Flexibility in Metadata Schemes and Standardisation: the Case of CMDI and DAN...vty
Presentation at ISKO Knowledge Organisation Research Observatory. RESEARCH REPOSITORIES AND DATAVERSE: NEGOTIATING METADATA, VOCABULARIES AND DOMAIN NEEDS
There are three related projects aimed at standardizing open access metadata: 1) The Open Access Spectrum guide published by PLOS, SPARC, and OASPA which identifies the components of open access on a spectrum from fully open to closed. 2) The Vocabularies for Open Access project by Jisc which aims to develop consensus vocabularies for key open access terms. 3) A NISO initiative to develop standardized metadata and indicators for open access status at the article level to provide information on access and reuse rights. As open access becomes more ubiquitous, standardizing this metadata is important for various stakeholders like funders, publishers, authors and readers.
Dynamic Metadata Management in Semantic File SystemsIJERA Editor
The progression in data capacity and difficulty inflicts great challenges for file systems. To address these contests, an inventive namespace management scheme is in distracted need to deliver both the ease and competence of data access. For scalability, each server makes only local, autonomous decisions about relocation for load balancing. Associative access is provided by a traditional extension to present tree-structured file system conventions, and by protocols that are intended specifically for content based access.Rapid attribute-based accesstofile system contents is fulfilled by instinctive extraction and indexing of key properties of file system objects. The programmed indexing of files and calendars is called “semantic” because user programmable transducers use data about the semantics of efficient file system objects to extract the properties for indexing. Tentative results from a semantic file system execution support the thesis that semantic file systems present a more active storage abstraction than do traditional tree planned file systems for data sharing and command level programming. Semantic file system is executed as a middleware in predictable file systems and works orthogonally with categorized directory trees. The semantic relationships and file groups recognized in file systems can also be used to facilitate file prefetching among other system-level optimizations. All-encompassing trace-driven experiments on our sample implementation validate the efficiency and competence.
Manuel Noya talks about the science-industry relationship driven by competitive intelligence and how to surf emerging technologies
Workshop title:TDM unlocking a goldmine of information
Training overview:
Text and Data Mining (TDM) is a natural ‘next step’ in open science. It can lead to new and unexpected discoveries and increase the impact of publications and repositories. This workshop showcases examples of successful TDM and infrastructural solutions for researchers. We will also discuss what is needed to make most of infrastructures and how publishers and repositories can open up their content.
DAY 2 - PARALLEL SESSION 4 & 5
Key developments in electronic delivery in LIS 2005-2008Catherine Ebenezer
This document summarizes key developments in electronic delivery in libraries from 2005-2008. It discusses trends toward user preferences for more seamless, convenient access to full text content. Technologies like federated search, portals, web services, and RSS were innovating how libraries delivered information to users. The National Library for Health in the UK was working to develop new services using these technologies and integrate library resources into clinical systems. Remote library services were also expanding to provide more online reference, document delivery, and loan management.
Simple Knowledge Organization System (SKOS) in the Context of Semantic Web De...gardensofmeaning
The document discusses the development and use of SKOS (Simple Knowledge Organization System) for representing knowledge organization systems like thesauri and classification schemes as structured data on the semantic web. It describes how LCSH (Library of Congress Subject Headings) has been converted to RDF using SKOS and published as linked open data. It suggests further steps like linking LCSH to other metadata and developing RDF representations of additional bibliographic schemas.
Este documento introduce la teoría de los estilos de aprendizaje y la enseñanza basada en estilos de aprendizaje. Explica los cuatro estilos de aprendizaje principales - activo, reflexivo, teórico y pragmático - y cómo cada uno aprende mejor. También describe posibles dificultades de aprendizaje para estudiantes con una preferencia alta en cada estilo y aspectos de la enseñanza que pueden interferir. El objetivo es ayudar a los maestros a comprender mejor cómo aprenden sus estudiantes y adapt
Este documento resume la historia y el desarrollo de la medicina a través de los tiempos. Comenzó en la prehistoria con el uso de plantas y minerales por chamanes y sacerdotes. Hipócrates en la antigua Grecia y médicos musulmanes como Avicena contribuyeron al desarrollo temprano de la medicina. En los siglos XIX y XX hubo grandes avances, incluyendo el descubrimiento de los antibióticos, vacunas y el mapeo del genoma humano. Hoy en día, la medicina se
La netiqueta son reglas de cortesía para comunicarse en Internet. Estas reglas promueven ser amable, despertar el interés del público dirigiéndose al grupo, usar detalles específicos, parafrasiar y analizar ideas de otros, analizar sus propias ideas, y terminar el mensaje de forma que promueva la discusión. Siguiendo estas normas se fomenta una comunicación efectiva y respetuosa en línea.
The document summarizes observations of an English class for high school students in Mexico. It describes 3 main activities: 1) Checking homework through repetition. 2) Group work creating dialogs using past perfect/continuous tenses. 3) Book exercises with repetition of words. It analyzes the class through theories of language acquisition, noting the teacher used behaviorism and total physical response. It suggests adding a game to motivate distracted students and help shy students participate more.
El documento discute las cuatro posibilidades de comunicación en el entorno educativo incorporando las nuevas tecnologías de la información y la comunicación (NNTT). Estas cuatro posibilidades son: 1) mismo tiempo y mismo lugar, como la enseñanza en el aula; 2) mismo tiempo y distinto lugar, como la enseñanza fuera del aula; 3) distinto tiempo y mismo lugar, como la enseñanza en el aula sin seguir horarios; y 4) distinto tiempo y distinto lugar, como la enseñanza aprovechando las mejores oportunidades sin
Este documento presenta una introducción a Excel 2007. Explica los capítulos sobre conceptos básicos e introducción, conceptos básicos como elementos de Excel, y formatos de celdas, alineación y bordes. También muestra los pasos para crear, guardar y abrir libros de Excel, e identifica las herramientas principales como la barra de título, barra de fórmulas, y celdas, filas y columnas.
El documento presenta información sobre dos estudiantes de comunicación social y periodismo de la Universidad Tolima en Colombia: Harold Cáceres y John Ramirez. Ambos estudiantes cursan el semestre A de 2010 y sus áreas de interés incluyen el fotoperiodismo. John considera que el periodismo es importante para mantenerse actualizado sobre los eventos del entorno, mientras que a Harold le interesan los medios audiovisuales.
Ontology-Oriented Inference-Based Learning Content Management System dannyijwest
The world is witnessing the electronic revolution in many fields of life such as health, education,
government and commerce. E-learning is considered one of the hot topics in the e-revolution as it brings
with it rapid change and greater opportunities to increase learning ability in colleges and schools. The
fields of Learning Management Systems (LMS) and Learning Content Management Systems (LCMS) are
full of open source and commercial products, however LCMS systems in general inherit the drawbacks of
information system such as weakness in user expected information retrieval and semantic modelling and
searching of contents & courses. In this paper, we propose a new prototype of LCMS that uses the
Semantic Web technologies and Ontology Reasoner with logical rules, as an inference engine to satisfy the
constraints and criteria specified by a user, and retrieves relevant content from the domain ontology in an
organized fashion. This enables construction of a user-specific course, by semantic querying for topics of
interest. We present the development of an Ontology-oriented Inference-based Learning Content
Management System OILCMS, its architecture, conception and strengths.
Ontology-Oriented Inference-Based Learning Content Management System dannyijwest
The world is witnessing the electronic revolution in many fields of life such as health, education,
government and commerce. E-learning is considered one of the hot topics in the e-revolution as it brings
with it rapid change and greater opportunities to increase learning ability in colleges and schools. The
fields of Learning Management Systems (LMS) and Learning Content Management Systems (LCMS) are
full of open source and commercial products, however LCMS systems in general inherit the drawbacks of
information system such as weakness in user expected information retrieval and semantic modelling and
searching of contents & courses. In this paper, we propose a new prototype of LCMS that uses the
Semantic Web technologies and Ontology Reasoner with logical rules, as an inference engine to satisfy the
constraints and criteria specified by a user, and retrieves relevant content from the domain ontology in an
organized fashion. This enables construction of a user-specific course, by semantic querying for topics of
interest. We present the development of an Ontology-oriented Inference-based Learning Content
Management System OILCMS, its architecture, conception and strengths.
The world is witnessing the electronic revolution in many fields of life such as health, education, government and commerce. E-learning is considered one of the hot topics in the e-revolution as it brings with it rapid change and greater opportunities to increase learning ability in colleges and schools. The fields of Learning Management Systems (LMS) and Learning Content Management Systems (LCMS) are full of open source and commercial products, however LCMS systems in general inherit the drawbacks of information system such as weakness in user expected information retrieval and semantic modelling and searching of contents & courses. In this paper, we propose a new prototype of LCMS that uses the Semantic Web technologies and Ontology Reasoner with logical rules, as an inference engine to satisfy the constraints and criteria specified by a user, and retrieves relevant content from the domain ontology in an organized fashion. This enables construction of a user-specific course, by semantic querying for topics of interest. We present the development of an Ontology-oriented Inference-based Learning Content Management System OILCMS, its architecture, conception and strengths.
1. The document discusses a model for using open source software to integrate institutional resources like publications and research outputs into a unified system accessible through a library. It presents opportunities for libraries to better manage digital assets and knowledge over the long term.
2. Key aspects of the model include using discovery tools like VuFind to search across resources from the library catalog and institutional repository. Systems like Koha and DSpace are proposed to manage the catalog and repository respectively, with their integration providing a sustainable solution.
3. The open source approach allows institutions to make better use of their research and ensures long-term access to content through the community support of these systems compared to proprietary alternatives.
This document discusses applying semantic web technologies to enhance the services of e-learning systems. It proposes developing a semantic learning management system (S-LMS) based on technologies like XML, RDF, OWL and SPARQL to automate and accurately search for information on e-learning systems like Moodle. The S-LMS would add semantic capabilities to allow students to search for learning resources based on semantics and provide personalized, customized content tailored to individual needs. It presents applying ontologies and metadata to Moodle in order to define domains and describe learning content in a way that improves search, interoperability and reusability of educational resources.
Applications of xml, semantic web or linked data in Library/Information Servi...Nurhazman Abdul Aziz
Applications of XML, Semantic Web & Linked Data in Library/Information Services & Skills needed by System Librarians.
H6716 (Internet & Web Technologies) & K6224 (Internet Technologies & Applications)
Semester 2 – 2011/2012
Hazman Aziz, Librarian (Library Technology & Systems)
Amirrudin Dahlan, Senior IT Specialist (Center for IT & Services)
Nanyang Technological University
Enabling Accessible Resource Access via Service ProvidersAlexander Haffner
Libraries have become digitized and are using information technology for storing and managing their
resource inventory. Additional metadata are used for describing properties of non-digital assets as well
as of purely digital resources. In particular, as the amount of digital resources increases, there is
demand for centralized services for searching and distribution of content.
Stakeholders of libraries and publishing industry already have made progress in areas of archival
strategies and standardization of preservation strategies. A variety of metadata standards and
exchange protocols enable service providers to offer a single access point to resources. However,
there is still an increased demand for improvement of resource organization and enhanced quality
particularly in terms of accessibility. This paper presents strategies to increase accessibility of
resources as a valuable step towards access-for-all. For consideration of accessibility within the
publishing chain we analyse the whole processing chain and identify stakeholders such as national
libraries and their corresponding responsibilities for ingest, archival storage and dissemination of
digital resources.
One Standard to rule them all?: Descriptive Choices for Open EducationR. John Robertson
R. John Robertson1, Lorna Campbell1, Phil Barker2, Li Yuan3, and Sheila MacNeill1
1Centre for Academic Practice and Learning Enhancement, University of Strathclyde, 2Institute for Computer Based Learning, Heriot-Watt University 3Institute for Cybernetic Education, University of Bolton
Drawing on our experience of supporting a nationwide Open Educational Resources programme (the UKOER programme), this presentation will consider the diverse range of approaches to describing OERs that have emerged across the programme and their impact on resource sharing, workflows, and an aggregate view of the resources.
Due to the diverse nature of the projects in the programme, ranging from individual educators to discipline-based consortia and institutions, it was apparent that no one technical or descriptive solution would fit all. Consequently projects were mandated to supply only a limited amount of descriptive information (programme tag, author, title, date, url, file format, file size, rights) with some additional information suggested (language, subject classifications, keywords, tags, comments, description). Projects were free to choose how this information should be encoded (if at all), stored, and shared.
In response, the projects have taken many different approaches to the description and management of resources. These range from using traditional highly structured and detailed metadata standards to approaches using whatever descriptions are supported by particular web2.0 applications. This experimental approach to resource description offers the wider OER community an opportunity to examine and assess the implications of different strategies for resource description and management
This paper illustrates a number of examples of projects’ approaches to description, noting the workflows and effort involved. We will consider the relationship of the choice of tool (repository, web2.0 application, VLE) to the choice of standards; and the relationship between local requirements and those of the wider community.
We will consider the impact of those choices on the dissemination and discoverability of resources. For example, the implications of resource description choices for discovery services which draw on multiple sources of OERs.
Flexibility in Metadata Schemes and Standardisation: the Case of CMDI and DAN...vty
Presentation at ISKO Knowledge Organisation Research Observatory. RESEARCH REPOSITORIES AND DATAVERSE: NEGOTIATING METADATA, VOCABULARIES AND DOMAIN NEEDS
There are three related projects aimed at standardizing open access metadata: 1) The Open Access Spectrum guide published by PLOS, SPARC, and OASPA which identifies the components of open access on a spectrum from fully open to closed. 2) The Vocabularies for Open Access project by Jisc which aims to develop consensus vocabularies for key open access terms. 3) A NISO initiative to develop standardized metadata and indicators for open access status at the article level to provide information on access and reuse rights. As open access becomes more ubiquitous, standardizing this metadata is important for various stakeholders like funders, publishers, authors and readers.
Dynamic Metadata Management in Semantic File SystemsIJERA Editor
The progression in data capacity and difficulty inflicts great challenges for file systems. To address these contests, an inventive namespace management scheme is in distracted need to deliver both the ease and competence of data access. For scalability, each server makes only local, autonomous decisions about relocation for load balancing. Associative access is provided by a traditional extension to present tree-structured file system conventions, and by protocols that are intended specifically for content based access.Rapid attribute-based accesstofile system contents is fulfilled by instinctive extraction and indexing of key properties of file system objects. The programmed indexing of files and calendars is called “semantic” because user programmable transducers use data about the semantics of efficient file system objects to extract the properties for indexing. Tentative results from a semantic file system execution support the thesis that semantic file systems present a more active storage abstraction than do traditional tree planned file systems for data sharing and command level programming. Semantic file system is executed as a middleware in predictable file systems and works orthogonally with categorized directory trees. The semantic relationships and file groups recognized in file systems can also be used to facilitate file prefetching among other system-level optimizations. All-encompassing trace-driven experiments on our sample implementation validate the efficiency and competence.
Manuel Noya talks about the science-industry relationship driven by competitive intelligence and how to surf emerging technologies
Workshop title:TDM unlocking a goldmine of information
Training overview:
Text and Data Mining (TDM) is a natural ‘next step’ in open science. It can lead to new and unexpected discoveries and increase the impact of publications and repositories. This workshop showcases examples of successful TDM and infrastructural solutions for researchers. We will also discuss what is needed to make most of infrastructures and how publishers and repositories can open up their content.
DAY 2 - PARALLEL SESSION 4 & 5
Key developments in electronic delivery in LIS 2005-2008Catherine Ebenezer
This document summarizes key developments in electronic delivery in libraries from 2005-2008. It discusses trends toward user preferences for more seamless, convenient access to full text content. Technologies like federated search, portals, web services, and RSS were innovating how libraries delivered information to users. The National Library for Health in the UK was working to develop new services using these technologies and integrate library resources into clinical systems. Remote library services were also expanding to provide more online reference, document delivery, and loan management.
Simple Knowledge Organization System (SKOS) in the Context of Semantic Web De...gardensofmeaning
The document discusses the development and use of SKOS (Simple Knowledge Organization System) for representing knowledge organization systems like thesauri and classification schemes as structured data on the semantic web. It describes how LCSH (Library of Congress Subject Headings) has been converted to RDF using SKOS and published as linked open data. It suggests further steps like linking LCSH to other metadata and developing RDF representations of additional bibliographic schemas.
Este documento introduce la teoría de los estilos de aprendizaje y la enseñanza basada en estilos de aprendizaje. Explica los cuatro estilos de aprendizaje principales - activo, reflexivo, teórico y pragmático - y cómo cada uno aprende mejor. También describe posibles dificultades de aprendizaje para estudiantes con una preferencia alta en cada estilo y aspectos de la enseñanza que pueden interferir. El objetivo es ayudar a los maestros a comprender mejor cómo aprenden sus estudiantes y adapt
Este documento resume la historia y el desarrollo de la medicina a través de los tiempos. Comenzó en la prehistoria con el uso de plantas y minerales por chamanes y sacerdotes. Hipócrates en la antigua Grecia y médicos musulmanes como Avicena contribuyeron al desarrollo temprano de la medicina. En los siglos XIX y XX hubo grandes avances, incluyendo el descubrimiento de los antibióticos, vacunas y el mapeo del genoma humano. Hoy en día, la medicina se
La netiqueta son reglas de cortesía para comunicarse en Internet. Estas reglas promueven ser amable, despertar el interés del público dirigiéndose al grupo, usar detalles específicos, parafrasiar y analizar ideas de otros, analizar sus propias ideas, y terminar el mensaje de forma que promueva la discusión. Siguiendo estas normas se fomenta una comunicación efectiva y respetuosa en línea.
The document summarizes observations of an English class for high school students in Mexico. It describes 3 main activities: 1) Checking homework through repetition. 2) Group work creating dialogs using past perfect/continuous tenses. 3) Book exercises with repetition of words. It analyzes the class through theories of language acquisition, noting the teacher used behaviorism and total physical response. It suggests adding a game to motivate distracted students and help shy students participate more.
El documento discute las cuatro posibilidades de comunicación en el entorno educativo incorporando las nuevas tecnologías de la información y la comunicación (NNTT). Estas cuatro posibilidades son: 1) mismo tiempo y mismo lugar, como la enseñanza en el aula; 2) mismo tiempo y distinto lugar, como la enseñanza fuera del aula; 3) distinto tiempo y mismo lugar, como la enseñanza en el aula sin seguir horarios; y 4) distinto tiempo y distinto lugar, como la enseñanza aprovechando las mejores oportunidades sin
Este documento presenta una introducción a Excel 2007. Explica los capítulos sobre conceptos básicos e introducción, conceptos básicos como elementos de Excel, y formatos de celdas, alineación y bordes. También muestra los pasos para crear, guardar y abrir libros de Excel, e identifica las herramientas principales como la barra de título, barra de fórmulas, y celdas, filas y columnas.
El documento presenta información sobre dos estudiantes de comunicación social y periodismo de la Universidad Tolima en Colombia: Harold Cáceres y John Ramirez. Ambos estudiantes cursan el semestre A de 2010 y sus áreas de interés incluyen el fotoperiodismo. John considera que el periodismo es importante para mantenerse actualizado sobre los eventos del entorno, mientras que a Harold le interesan los medios audiovisuales.
Ontology-Oriented Inference-Based Learning Content Management System dannyijwest
The world is witnessing the electronic revolution in many fields of life such as health, education,
government and commerce. E-learning is considered one of the hot topics in the e-revolution as it brings
with it rapid change and greater opportunities to increase learning ability in colleges and schools. The
fields of Learning Management Systems (LMS) and Learning Content Management Systems (LCMS) are
full of open source and commercial products, however LCMS systems in general inherit the drawbacks of
information system such as weakness in user expected information retrieval and semantic modelling and
searching of contents & courses. In this paper, we propose a new prototype of LCMS that uses the
Semantic Web technologies and Ontology Reasoner with logical rules, as an inference engine to satisfy the
constraints and criteria specified by a user, and retrieves relevant content from the domain ontology in an
organized fashion. This enables construction of a user-specific course, by semantic querying for topics of
interest. We present the development of an Ontology-oriented Inference-based Learning Content
Management System OILCMS, its architecture, conception and strengths.
Ontology-Oriented Inference-Based Learning Content Management System dannyijwest
The world is witnessing the electronic revolution in many fields of life such as health, education,
government and commerce. E-learning is considered one of the hot topics in the e-revolution as it brings
with it rapid change and greater opportunities to increase learning ability in colleges and schools. The
fields of Learning Management Systems (LMS) and Learning Content Management Systems (LCMS) are
full of open source and commercial products, however LCMS systems in general inherit the drawbacks of
information system such as weakness in user expected information retrieval and semantic modelling and
searching of contents & courses. In this paper, we propose a new prototype of LCMS that uses the
Semantic Web technologies and Ontology Reasoner with logical rules, as an inference engine to satisfy the
constraints and criteria specified by a user, and retrieves relevant content from the domain ontology in an
organized fashion. This enables construction of a user-specific course, by semantic querying for topics of
interest. We present the development of an Ontology-oriented Inference-based Learning Content
Management System OILCMS, its architecture, conception and strengths.
The world is witnessing the electronic revolution in many fields of life such as health, education, government and commerce. E-learning is considered one of the hot topics in the e-revolution as it brings with it rapid change and greater opportunities to increase learning ability in colleges and schools. The fields of Learning Management Systems (LMS) and Learning Content Management Systems (LCMS) are full of open source and commercial products, however LCMS systems in general inherit the drawbacks of information system such as weakness in user expected information retrieval and semantic modelling and searching of contents & courses. In this paper, we propose a new prototype of LCMS that uses the Semantic Web technologies and Ontology Reasoner with logical rules, as an inference engine to satisfy the constraints and criteria specified by a user, and retrieves relevant content from the domain ontology in an organized fashion. This enables construction of a user-specific course, by semantic querying for topics of interest. We present the development of an Ontology-oriented Inference-based Learning Content Management System OILCMS, its architecture, conception and strengths.
1. The document discusses a model for using open source software to integrate institutional resources like publications and research outputs into a unified system accessible through a library. It presents opportunities for libraries to better manage digital assets and knowledge over the long term.
2. Key aspects of the model include using discovery tools like VuFind to search across resources from the library catalog and institutional repository. Systems like Koha and DSpace are proposed to manage the catalog and repository respectively, with their integration providing a sustainable solution.
3. The open source approach allows institutions to make better use of their research and ensures long-term access to content through the community support of these systems compared to proprietary alternatives.
This document discusses applying semantic web technologies to enhance the services of e-learning systems. It proposes developing a semantic learning management system (S-LMS) based on technologies like XML, RDF, OWL and SPARQL to automate and accurately search for information on e-learning systems like Moodle. The S-LMS would add semantic capabilities to allow students to search for learning resources based on semantics and provide personalized, customized content tailored to individual needs. It presents applying ontologies and metadata to Moodle in order to define domains and describe learning content in a way that improves search, interoperability and reusability of educational resources.
Although of the semantic web technologies utilization in the learning development field is a new research area, some authors have already proposed their idea of how an effective that operate. Specifically, from analysis of the literature in the field, we have identified three different types of existing applications that actually employ these technologies to support learning. These applications aim at: Enhancing the learning objects reusability by linking them to an ontological description of the domain, or, more generally, describe relevant dimension of the learning process in an ontology, then; providing a comprehensive authoring system to retrieve and organize web material into a learning course, and constructing advanced strategies to present annotated resources to the user, in the form of browsing facilities, narrative generation and final rendering of a course. On difference with the approaches cited above, here we propose an approach that is modeled on narrative studies and on their transposition in the digital world. In the rest of the paper, we present the theoretical basis that inspires this approach, and show some examples that are guiding our implementation and testing of these ideas within e-learning. By emerging the idea of the ontologies are recognized as the most important component in achieving semantic interoperability of e-learning resources. The benefits of their use have already been recognized in the learning technology community. In order to better define different aspects of ontology applications in e-learning, researchers have given several classifications of ontologies. We refer to a general one given in that differentiates between three dimensions ontologies can describe: content, context, and structure. Most of the present research has been dedicated to the first group of ontologies. A well-known example of such an ontology is based on the ACM Computer Classification System (ACM CCS) and defined by Resource Description Framework Schema (RDFS). It’s used in the MOODLE to classify learning objects with a goal to improve searching. The chapter will cover the terms of the semantic web and e-learning systems design and management in e-learning (MOODLE) and some of studies depend on e-learning and semantic web, thus the tools will be used in this paper, and lastly we shall discuss the expected contribution. The special attention will be putted on the above topics.
An Incremental Method For Meaning Elicitation Of A Domain OntologyAudrey Britton
This document describes MELIS (Meaning Elicitation and Lexical Integration System), a tool developed to support the annotation of data sources with conceptual and lexical information during the ontology generation process in MOMIS (Mediator envirOnment for Multiple Information Systems). MELIS improves upon MOMIS by enabling more automated annotation of data sources and by extracting relationships between lexical elements using lexical and domain knowledge to provide a richer semantics. The document outlines the MOMIS architecture integrated with MELIS and explains how MELIS supports key steps in the ontology creation process, including source annotation, common thesaurus generation, and relationship extraction.
A Domain Based Approach to Information Retrieval in Digital Libraries - Rotel...University of Bari (Italy)
The current abundance of electronic documents requires automatic techniques that support the users in understanding their content and extracting useful information. To this aim, improving the retrieval performance must necessarily go beyond simple lexical interpretation of the user queries, and pass through an understanding of their semantic content and aims. It goes without saying that any digital library would take enormous advantage from the availability of effective Information Retrieval techniques to provide to their users. This paper proposes an approach to Information Retrieval based on a correspondence of the domain of discourse between the query and the documents in the repository. Such an association is based on standard general-purpose linguistic resources (WordNet and WordNet Domains) and on a novel similarity assessment technique. Although the work is at a preliminary stage, interesting initial results suggest to go on extending and improving the approach.
This document discusses enabling technologies for interoperability between geographic information systems (GIS). It addresses problems at the syntactic, structural, and semantic levels of integration that must be solved to achieve fully interoperable GIS. At the syntactic level, standards like XML are used to integrate different data types. At the structural level, mediator systems use mapping rules to integrate heterogeneous data structures. The most difficult problem is semantic integration, where the meanings and contexts of concepts must be resolved. Ontologies and semantic modeling with XML and RDF can help describe information semantically and perform semantic translation between contexts to enable intelligent information integration.
In tech application-of_data_mining_technology_on_e_learning_material_recommen...Enhmandah Hemeelee
The document describes a recommendation system that applies data mining techniques to recommend e-learning materials. It proposes using LDAP for fast searching of materials across systems, JAXB for parsing content, and association rule mining and collaborative filtering to generate recommendations. The system collects user activity data, analyzes it using Apriori algorithm to find related search terms and content, and stores results in an LDAP database to provide recommendations to users.
In tech application-of_data_mining_technology_on_e_learning_material_recommen...Enhmandah Hemeelee
The document describes a recommendation system that applies data mining techniques to recommend e-learning materials. It proposes using LDAP for fast searching of materials across systems, JAXB for parsing content, and association rule mining and collaborative filtering for recommendations. A web spider collects content indexes from learning management systems and stores data in an LDAP directory. Users can search for related materials, and the system mines log data to associate frequently searched terms and recommend additional resources.
Data Integration in Multi-sources Information Systemsijceronline
International Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
Here are the key points about using content-based filtering techniques:
- Content-based filtering relies on analyzing the content or description of items to recommend items similar to what the user has liked in the past. It looks for patterns and regularities in item attributes/descriptions to distinguish highly rated items.
- The item content/descriptions are analyzed automatically by extracting information from sources like web pages, or entered manually from product databases.
- It focuses on objective attributes about items that can be extracted algorithmically, like text analysis of documents.
- However, personal preferences and what makes an item appealing are often subjective qualities not easily extracted algorithmically, like writing style or taste.
- So while content-based filtering can
Generic Algorithm based Data Retrieval Technique in Data MiningAM Publications,India
This system Hybrid extraction of robust model (GA), a dynamic XAML based mechanism for the adaptive management and reuse of e-learning resources in a distributed environment like the Web. This proposed system argues that to achieve the on-demand semantic-based resource management for Web-based e-learning, one should go beyond using domain ontology’s statically. So the propose XAML based matching process involves semantic mapping has done on both the open dataset and closed dataset mechanism to integrate e-learning databases by using ontology semantics. It defines context-specific portions from the whole ontology as optimized data and proposes an XAML based resource reuse approach by using an evolution algorithm. It explains the context aware based evolution algorithm for dynamic e-learning resource reuse in detail. This system is going to conduct a simulation experiment and evaluate the proposed approach with a xaml based e-learning scenario. The proposed approach for matching process in web cluster databases from different database servers can be easily integrated and deliver highly dimensional e-learning resource management and reuse is far from being mature. However, e-learning is also a widely open research area, and there is still much room for improvement on the method. This research mechanism includes 1) improving the proposed evolution approach by making use of and comparing different evolutionary algorithms, 2) applying the proposed approach to support more applications, and 3) extending to the situation with multiple e-learning systems or services.
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
An approach for transforming of relational databases to owl ontologyIJwest
Rapid growth of documents, web pages, and other types of text content is a huge challenge for the modern content management systems. One of the problems in the areas of information storage and retrieval is the lacking of semantic data. Ontologies can present knowledge in sharable and repeatedly usable manner and provide an effective way to reduce the data volume overhead by encoding the structure of a particular domain. Metadata in relational databases can be used to extract ontology from database in a special domain. According to solve the problem of sharing and reusing of data, approaches based on transforming relational database to ontology are proposed. In this paper we propose a method for automatic ontology construction based on relational database. Mining and obtaining further components from relational database leads to obtain knowledge with high semantic power and more expressiveness. Triggers are one of the database components which could be transformed to the ontology model and increase the amount of power and expressiveness of knowledge by presenting part of the knowledge dynamically.
Iaetsd hierarchical fuzzy rule based classificationIaetsd Iaetsd
This document discusses a hierarchical fuzzy rule-based classification system using genetic rule selection to filter unwanted messages from online social networks. It aims to improve performance on imbalanced data sets by increasing granularity of fuzzy partitions at class boundaries. The system uses a neural network learning model and genetic algorithm for rule selection to build an accurate and compact fuzzy rule-based model. It analyzes challenges in classifying short texts from social media posts and reviews related work on content-based filtering and policy-based personalization for social networks. The document also discusses issues with imbalanced data sets and proposes oversampling the minority class using SMOTE (Synthetic Minority Over-sampling Technique) as a preprocessing step to address class imbalance problems.
INTELLIGENT INFORMATION RETRIEVAL WITHIN DIGITAL LIBRARY USING DOMAIN ONTOLOGYcscpconf
A digital library is a type of information retrieval (IR) system. The existing information retrieval
methodologies generally have problems on keyword-searching. We proposed a model to solve
the problem by using concept-based approach (ontology) and metadata case base. This model
consists of identifying domain concepts in user’s query and applying expansion to them. The
system aims at contributing to an improved relevance of results retrieved from digital libraries
by proposing a conceptual query expansion for intelligent concept-based retrieval. We need to
import the concept of ontology, making use of its advantage of abundant semantics and
standard concept. Domain specific ontology can be used to improve information retrieval from
traditional level based on keyword to the lay based on knowledge (or concept) and change the
process of retrieval from traditional keyword matching to semantics matching. One approach is
query expansion techniques using domain ontology and the other would be introducing a case
based similarity measure for metadata information retrieval using Case Based Reasoning
(CBR) approach. Results show improvements over classic method, query expansion using
general purpose ontology and a number of other approaches.
The document discusses two approaches to building information systems: the single-level methodology and the two-level model methodology. The two-level model separates domain knowledge from metadata about the structural representation of that knowledge. This allows knowledge to change without requiring software redesign. The openEHR two-level architecture separates information semantics from knowledge concepts. It follows the RM/ODP reference model approach and allows knowledge to be instantiated at runtime rather than hardcoded. The document argues this makes systems more adaptable, interoperable, and able to keep up with changing healthcare knowledge.
Bridging the gap between the semantic web and big data: answering SPARQL que...IJECEIAES
Nowadays, the database field has gotten much more diverse, and as a result, a variety of non-relational (NoSQL) databases have been created, including JSON-document databases and key-value stores, as well as extensible markup language (XML) and graph databases. Due to the emergence of a new generation of data services, some of the problems associated with big data have been resolved. In addition, in the haste to address the challenges of big data, NoSQL abandoned several core databases features that make them extremely efficient and functional, for instance the global view, which enables users to access data regardless of how it is logically structured or physically stored in its sources. In this article, we propose a method that allows us to query non-relational databases based on the ontology-based access data (OBDA) framework by delegating SPARQL protocol and resource description framework (RDF) query language (SPARQL) queries from ontology to the NoSQL database. We applied the method on a popular database called Couchbase and we discussed the result obtained.
Similar to A Framework for Content Preparation to Support Open-Corpus Adaptive Hypermedia (20)
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsDianaGray10
Join us to learn how UiPath Apps can directly and easily interact with prebuilt connectors via Integration Service--including Salesforce, ServiceNow, Open GenAI, and more.
The best part is you can achieve this without building a custom workflow! Say goodbye to the hassle of using separate automations to call APIs. By seamlessly integrating within App Studio, you can now easily streamline your workflow, while gaining direct access to our Connector Catalog of popular applications.
We’ll discuss and demo the benefits of UiPath Apps and connectors including:
Creating a compelling user experience for any software, without the limitations of APIs.
Accelerating the app creation process, saving time and effort
Enjoying high-performance CRUD (create, read, update, delete) operations, for
seamless data management.
Speakers:
Russell Alfeche, Technology Leader, RPA at qBotic and UiPath MVP
Charlie Greenberg, host
ScyllaDB is making a major architecture shift. We’re moving from vNode replication to tablets – fragments of tables that are distributed independently, enabling dynamic data distribution and extreme elasticity. In this keynote, ScyllaDB co-founder and CTO Avi Kivity explains the reason for this shift, provides a look at the implementation and roadmap, and shares how this shift benefits ScyllaDB users.
In the realm of cybersecurity, offensive security practices act as a critical shield. By simulating real-world attacks in a controlled environment, these techniques expose vulnerabilities before malicious actors can exploit them. This proactive approach allows manufacturers to identify and fix weaknesses, significantly enhancing system security.
This presentation delves into the development of a system designed to mimic Galileo's Open Service signal using software-defined radio (SDR) technology. We'll begin with a foundational overview of both Global Navigation Satellite Systems (GNSS) and the intricacies of digital signal processing.
The presentation culminates in a live demonstration. We'll showcase the manipulation of Galileo's Open Service pilot signal, simulating an attack on various software and hardware systems. This practical demonstration serves to highlight the potential consequences of unaddressed vulnerabilities, emphasizing the importance of offensive security practices in safeguarding critical infrastructure.
GlobalLogic Java Community Webinar #18 “How to Improve Web Application Perfor...GlobalLogic Ukraine
Під час доповіді відповімо на питання, навіщо потрібно підвищувати продуктивність аплікації і які є найефективніші способи для цього. А також поговоримо про те, що таке кеш, які його види бувають та, основне — як знайти performance bottleneck?
Відео та деталі заходу: https://bit.ly/45tILxj
Dandelion Hashtable: beyond billion requests per second on a commodity serverAntonios Katsarakis
This slide deck presents DLHT, a concurrent in-memory hashtable. Despite efforts to optimize hashtables, that go as far as sacrificing core functionality, state-of-the-art designs still incur multiple memory accesses per request and block request processing in three cases. First, most hashtables block while waiting for data to be retrieved from memory. Second, open-addressing designs, which represent the current state-of-the-art, either cannot free index slots on deletes or must block all requests to do so. Third, index resizes block every request until all objects are copied to the new index. Defying folklore wisdom, DLHT forgoes open-addressing and adopts a fully-featured and memory-aware closed-addressing design based on bounded cache-line-chaining. This design offers lock-free index operations and deletes that free slots instantly, (2) completes most requests with a single memory access, (3) utilizes software prefetching to hide memory latencies, and (4) employs a novel non-blocking and parallel resizing. In a commodity server and a memory-resident workload, DLHT surpasses 1.6B requests per second and provides 3.5x (12x) the throughput of the state-of-the-art closed-addressing (open-addressing) resizable hashtable on Gets (Deletes).
In our second session, we shall learn all about the main features and fundamentals of UiPath Studio that enable us to use the building blocks for any automation project.
📕 Detailed agenda:
Variables and Datatypes
Workflow Layouts
Arguments
Control Flows and Loops
Conditional Statements
💻 Extra training through UiPath Academy:
Variables, Constants, and Arguments in Studio
Control Flow in Studio
Northern Engraving | Modern Metal Trim, Nameplates and Appliance PanelsNorthern Engraving
What began over 115 years ago as a supplier of precision gauges to the automotive industry has evolved into being an industry leader in the manufacture of product branding, automotive cockpit trim and decorative appliance trim. Value-added services include in-house Design, Engineering, Program Management, Test Lab and Tool Shops.
QR Secure: A Hybrid Approach Using Machine Learning and Security Validation F...AlexanderRichford
QR Secure: A Hybrid Approach Using Machine Learning and Security Validation Functions to Prevent Interaction with Malicious QR Codes.
Aim of the Study: The goal of this research was to develop a robust hybrid approach for identifying malicious and insecure URLs derived from QR codes, ensuring safe interactions.
This is achieved through:
Machine Learning Model: Predicts the likelihood of a URL being malicious.
Security Validation Functions: Ensures the derived URL has a valid certificate and proper URL format.
This innovative blend of technology aims to enhance cybersecurity measures and protect users from potential threats hidden within QR codes 🖥 🔒
This study was my first introduction to using ML which has shown me the immense potential of ML in creating more secure digital environments!
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
Lee Barnes - Path to Becoming an Effective Test Automation Engineer.pdfleebarnesutopia
So… you want to become a Test Automation Engineer (or hire and develop one)? While there’s quite a bit of information available about important technical and tool skills to master, there’s not enough discussion around the path to becoming an effective Test Automation Engineer that knows how to add VALUE. In my experience this had led to a proliferation of engineers who are proficient with tools and building frameworks but have skill and knowledge gaps, especially in software testing, that reduce the value they deliver with test automation.
In this talk, Lee will share his lessons learned from over 30 years of working with, and mentoring, hundreds of Test Automation Engineers. Whether you’re looking to get started in test automation or just want to improve your trade, this talk will give you a solid foundation and roadmap for ensuring your test automation efforts continuously add value. This talk is equally valuable for both aspiring Test Automation Engineers and those managing them! All attendees will take away a set of key foundational knowledge and a high-level learning path for leveling up test automation skills and ensuring they add value to their organizations.
This talk will cover ScyllaDB Architecture from the cluster-level view and zoom in on data distribution and internal node architecture. In the process, we will learn the secret sauce used to get ScyllaDB's high availability and superior performance. We will also touch on the upcoming changes to ScyllaDB architecture, moving to strongly consistent metadata and tablets.
"$10 thousand per minute of downtime: architecture, queues, streaming and fin...Fwdays
Direct losses from downtime in 1 minute = $5-$10 thousand dollars. Reputation is priceless.
As part of the talk, we will consider the architectural strategies necessary for the development of highly loaded fintech solutions. We will focus on using queues and streaming to efficiently work and manage large amounts of data in real-time and to minimize latency.
We will focus special attention on the architectural patterns used in the design of the fintech system, microservices and event-driven architecture, which ensure scalability, fault tolerance, and consistency of the entire system.
AppSec PNW: Android and iOS Application Security with MobSFAjin Abraham
Mobile Security Framework - MobSF is a free and open source automated mobile application security testing environment designed to help security engineers, researchers, developers, and penetration testers to identify security vulnerabilities, malicious behaviours and privacy concerns in mobile applications using static and dynamic analysis. It supports all the popular mobile application binaries and source code formats built for Android and iOS devices. In addition to automated security assessment, it also offers an interactive testing environment to build and execute scenario based test/fuzz cases against the application.
This talk covers:
Using MobSF for static analysis of mobile applications.
Interactive dynamic security assessment of Android and iOS applications.
Solving Mobile app CTF challenges.
Reverse engineering and runtime analysis of Mobile malware.
How to shift left and integrate MobSF/mobsfscan SAST and DAST in your build pipeline.
"What does it really mean for your system to be available, or how to define w...Fwdays
We will talk about system monitoring from a few different angles. We will start by covering the basics, then discuss SLOs, how to define them, and why understanding the business well is crucial for success in this exercise.
"What does it really mean for your system to be available, or how to define w...
A Framework for Content Preparation to Support Open-Corpus Adaptive Hypermedia
1. A Framework for Content Preparation to Support
Open-Corpus Adaptive Hypermedia
Killian Levacher, Éamonn Hynes, Séamus Lawless, Alexander O'Connor,
Vincent Wade
Centre for Next Generation Localisation, Knowledge & Data Engineering Group,
School of Computer Science and Statistics, Trinity College, Dublin, Ireland
{Killian.Levacher, hynese, Seamus.Lawless, Alex.OConnor,
Vincent.Wade}@cs.tcd.ie
Abstract. A key impediment for enabling the mainstream adoption of
Adaptive Hypermedia for web applications and corporate websites is the
diculty in repurposing existing content for such delivery systems. This
paper proposes a novel framework for open-corpus content preparation,
making it usable for adaptive hypermedia systems. The proposed frame-
work processes documents drawn from both open (i.e. web) and closed
corpora, producing coherent conceptual sections of text with associated
descriptive metadata. The solution bridges the gap between information
resources and information requirements of adaptive systems by adopting
state-of-the-art information extraction and structural content analysis
techniques. The result is an on-demand provision of tailored, atomic in-
formation objects called slices. The challenges associated with open
corpus content reusability are addressed with the aim of improving the
scalability and interoperability of adaptive systems. This paper proposes
an initial architecture for such a framework in addition to reviews of
associated technologies.
Key words: Open Corpus Content, Adaptive Hypermedia, Statistical
Content Analysis, Structural Content Analysis
1 Introduction
The increasing pervasiveness of the internet is fundamentally changing how peo-
ple author, interact with and consume content. There are early signs of a shift
in the way digital content is created, from the linear authoring and publication
of material towards the aggregation and re-use of existing content from various
disparate sources.
Adaptive Hypermedia Systems (AHS) have traditionally attempted to de-
liver dynamically adapted and personalised presentations to users through the
sequencing of recongurable pieces of information.
While the eectiveness and benets of such systems have been proven in
numerous studies [1][2], a major obstacle to their widespread adoption relates to
2. the restriction of suitable content available to provide such adaptivity in terms
of volume, granularity, style and meta-data.
One of the key impediments in oering a wide range of information objects is,
for example, the considerable amount of manual eort involved in creating these
resources. Information object content consists of either pre-existing documents
[3], or of content created by small groups of individuals [4], intentionally authored
for a particular system [5]. This generates both scalability and interoperability
issues, which prevents the mainstream adoption of adaptive systems.
Metadata standards such as LOM [6] attempt to provide generic content
packaging methods with the aim of enabling the interoperability of information
objects within AHSs. Although they implement usage-agnostic principles, their
development is very time consuming due to the complexity involved [7]. Fur-
thermore, they also require AHSs to comply with a specic content structure in
order to make full usage of these resources.
In parallel with these developments, a wealth of information is now accessible
on the web, in digital repositories and as part of library initiatives. However, due
to its heterogeneity, this information is not suited to direct usage by AHSs in its
present form. It is available in several languages and within large documents with
very limited meta-data. These objects are also generally very coarse-grained and
correlated with unnecessary noise such as navigation bars, advertisements etc.
Automated content preparation techniques are therefore necessary to provide
scalable solutions to AHS specic content requirements.
Our goal is to provide AHSs with the ability to benet from the wealth
of knowledge already accessible on the web and in digital libraries by bridging
the gap between these information resources and the information requirements
of adaptive systems. We propose the creation of a service that can tailor open-
corpus content for use in AHSs without requiring such systems to adopt a generic
content structure. This system would hence resolve scalability issues in content
authoring by providing AHSs with a large array of information objects. More-
over, the inclusion of open corpus information would provide up-to-date content
in a wide variety of styles and structures.
We aim to combine several disciplines such as Information Retrieval, Adap-
tive Web and Information Extraction in order to provide an on-demand informa-
tion object service. This service will harvest open corpus information, segment
it structurally and automatically augment it with successive layers of internal
metadata. Any AHS could then provide this service with a request that includes
requirements for an information object which would in turn fetch and tailor this
content to its specication.
This paper is structured as follows: Section 2 will present the key challenges
addressed within this research in relation to AHS information requirements as
well as structural and statistical text analysis. The framework we envisage de-
veloping is presented in detail in section 3, along with the overall workow pro-
duced. Section 4 will present the application area of such a technology and how
we intend to evaluate it. Finally, section 5 will conclude and present the road
3. map ahead. This work is applicable to any type of content, however some of the
examples presented in this paper will be based on adaptive e-learning systems.
2 Key Challenges and Related Work
Throughout the evolution of adaptive systems, technologies introduced have
moved from original all in one solutions, towards increasing modularity whereby
AHSs become consumers of domain, pedagogical and user models [8]. This has
the eect of leaving such systems with a core adaptive engine capable of deal-
ing with a wider range of loosely coupled models that are integrated as desired.
Content, however, is still very tightly coupled to these engines and as a result
strongly impedes the general adoption of these systems.
Most AH systems, up until recently have operated in closed document spaces
with content specically authored for their usage [4], hence obstructing interop-
erability, both by accepting only a narrow eld of content, and by motivating the
generation of content in a highly-specic format. As a result adaptive systems
also encounter scalability issues due to the limited amount of content available
to adapt on, arising from the small amount of manual contributions available.
Open corpus content is increasingly seen as providing a solution to these is-
sues [9]. However, most systems incorporating this type of content have, for the
moment, mainly focused on linking it with internal content as alternative ex-
ploration paths. Those incorporating it fully into their system [10][11] require
manual mark-up of such content with specic meta-data. Schemas such as LOM
(in the area of e-learning) and IMS packagings, attempt to provide usage-agnostic
solutions, however they require a lot of development eort thus prohibiting scal-
ability [7]. Moreover, AHSs must comply with specic content structures so as to
avail of these resources. In order to leverage the full potential of open corpus re-
sources, adaptive systems need to incorporate and correlate this type of content
fully into existing systems using structure-agnostic and automated approaches.
Our approach, on the other hand, moves away from the packaging model
altogether by considering content as a pluggable feature of adaptive systems to
the same extent as user or domain models. A service providing open corpus con-
tent automatically in a form and format customised for each requesting adaptive
system decouple this feature from current systems, disentangling it from content
provision and hence providing a solution to both scalability and interoperability
issues.
Fully integrating information harvested over the web within adaptive presen-
tations, on the other hand, generates new issues. Web pages are usually written
as stand-alone content without any re-composition purposes. In addition to the
main content, they usually present navigation bars, advertisement banners and
irrelevant features from various sources that must be removed prior to proper
re-use.
Structural segmentation will therefore be an initial fundamental requirement
in tailoring open corpus content for use within adaptive systems. A large propor-
4. tion of pages harvested will need to be stripped of any redundant information
that might prevent its adequate re-use in adaptive systems.
The problem of structural segmentation has already been addressed from
dierent perspectives mainly with the purpose of tailoring existing content for
dierent displays [12]. Most of the techniques up until recently have focused
on analyzing the DOM structure of a HTML page as a basis for segmentation.
Several attempts, for instance, focus on removing templates from pages by iden-
tifying common DOM sub-trees [13] or using isotonic regression [14]. Machine
Learning algorithms are also used to decide which pair of tags should coincide
with suitable segmentation points. However, it has become increasingly popu-
lar to separate any formatting style from within HTML tags using Cascading
Style Sheets (CSS) and Javascript, thus increasing the heterogeneity of rendered
content from similar HTML trees [15]. For this reason, such techniques do not
appear to be an adequate solution when dealing with a large set of documents
as diverse as the World Wide Web.
Vision-based techniques, using entropy reduction [12], or techniques such as
VIP algorithms [16] on the other hand, partition a page based on its rendering.
These have the benet of covering a wider range of pages regardless of their
HTML trees. But their usage within systems dealing with large volume of data is
questionable since rendering must be performed prior to any analysis, inevitably
producing delays. Moreover, they completely ignore DOM features altogether
which do provide structural clues where rendered surfaces appear similar [12].
This part of our research will therefore seek to provide a solution to this
paradigm that scales both in terms of processing speed as well as structural
segmentation precision within a wide heterogeneous range of pages. The proce-
dure will additionally need to incorporate the notion of granularity in order to
keep cohesive chunks of text together to provide meaningful input to subsequent
semantic analysis.
Within the statistical analysis component of the framework, it is important
to extract concepts with high precision/recall and eciently represent these con-
cepts in such a way as to make them as interoperable and re-usable as possible.
Brusilovosky et al. [9] discuss the lack of re-usability and interoperability between
adaptive systems and how current adaptive hypermedia systems are restricted
to closed corpora by the nature of their design. This framework architecture
is based on the idea that open corpus adaptive hypermedia is feasible: the ef-
fort will be in moving away from tables of document-term frequencies towards
a richer semantic representation of data that is uncomplicated, easy-to-use and
highly interoperable.
3 Framework for Content Preparation
The framework proposed in this research is divided into three separate com-
ponents as shown in gure 1. Each part of the framework executes a specic
task on the open corpus content. A specic layer of meta-data is subsequently
appended, enriching it with structural and semantic clues for further analysis at
5. each step. The rst two stages of the pipeline are executed pre-emptively prior
to any client request while the intelligent slicing task is executed at run time.
The concept of a slice is an abstract notion representing a stand-alone piece
of information, originally part of an existing document, extracted and segmented
to full a specic information request. A slice can be atomic or composed of
other slices. It possesses its own set of properties (or meta-data) and can inherit
those of others. A slice is virtual in the sense that it only exists temporarily to
full a client request. It is specic to each information request and represents a
subjective perspective on a particular piece of a document and its description.
The degree of complexity of a slice will match the requirements of the system
which requests it.
The rest of this section will present in detail each stage of the framework.
The structural analysis of open corpus content will initially be described followed
by the statistical analysis of these resources. Finally, the overall work-ow be-
tween this framework and AHSs clients will be outlined using intelligent slicing
technology.
Fig. 1. Content analysis methodology
3.1 Structural Analysis
In order to provide the initial structural segmentation needed to create slices,
this research will attempt to combine benets of both DOM- and vision-based
segmentation techniques. The process will initially divide pages in atomic blocks
using relevant HTML tags similar to [15]. Following this extensive break down,
the page will be reconstructed, aggregating relevant blocks together with similar
inherent textual properties. Changes in text ow (short phrases in navigation
bars vs. long sentences in text), for instance, will be considered as segment
delineations.
6. The notion of text density, originating from the eld of computer vision will
also be applied to these pages using the ratio between the number of words and
the space occupied on the page [15] as an additional hint for section divisions.
Furthermore, the use of linguistic analysis techniques such as style boundary
detection will be considered as additional boundary indication.
3.2 Statistical Analysis
In this section we discuss methods for statistically analysing text and extracting
information such that it can be used as part of the wider framework; it serv-
ing as a substrate for the incorporation of several dierent statistical analysis
techniques. In order for the proposed framework to be able to interact naturally
with information seekers, it must cope with inherent ambivalence, impreciseness
or even vagueness of requests and deal with them in a structured and systematic
fashion. For these reasons, statistical methods oer a wealth of neat techniques
for dealing with the underlying complex characteristics of language.
Work currently underway includes looking at methods for the automatic
extraction of concepts from text. The vision is to extract and create a conceptual
layer over the text which is amenable to ecient searching and logical reasoning.
We move away from working only on closed-corpus texts and go some way to
addressing the huge challenge that is the open domain problem [9] by making
our algorithms as generalised as possible and not restricted to any particular
domain. One method for identifying concepts from text is to use supervised
learning methods. We use HMMs (Hidden Markov Models) to extract coherent,
relevant passages from text [17]. This method is particularly suited to extraction
from expositional texts as opposed to highly expressive literary texts (i.e. poetry,
literary prose, etc.) which have high entropy vocabularies and diverse language
use. Jiang et al. report very high precision/recall compared to other methods and
their work ts in nicely with the idea of extracting concepts/coherent relevant
passages of text.
The basic idea is as follows: given some a priori concept that is to be extracted
from a text, several passages identied as being associated with that particular
concept are used as training data for the HMM classier. Once a reasonable
amount of training data has been collected (enough data such that the classier's
train and test error converges towards the desired performance), the HMM can
then make soft decisions on a text stream and identify coherent relevant passages
of varying length with very high accuracy.
The topology of the HMM is shown in gure 2 below: the states B1, B2 and
B3 are the background states, state E is an end state and state R is a relevant
state. The 5 state HMM structure allows the system to emit a relevant symbol,
move to a background state and then re-emit a relevant symbol at some later
point before reaching the E state. The output observation probabilities can be
extracted directly from the training corpora, and the objective is to learn the
transition probabilities so as to maximise the likelihood of observing a relevant
passage given a concept.
7. Fig. 2. HMM topology
On a synthesised corpus consisting of a number of article abstracts (coherent
relevant passages) stitched together, with each abstract having an associated
label, very high precision and recall values were reported [17]. It remains to be
seen how well the HMM method performs on non-synthetic corpora and work is
currently underway to explore this.
Several extensions to the work of Jiang et al. have also been explored, includ-
ing the incorporation of anaphora to resolve pronouns and augment the presence
of important entities such as people and places. In addition to the uni-gram lexi-
cal features, more elaborate query expansion methods and semantic analysis has
been looked at to extract more elaborate features for the classier to work from.
The incorporation of this additional functionality is expected to yield higher
performance in terms of precision and recall on the synthetic document set and
more reliable performance on real corpora. Instead of just a single-dimension
vector with just one feature (i.e. a stemmed word), the word is represented
as a slightly higher dimensional vector (i.e. the stemmed word with additional
features) that helps improve the overall accuracy of the classier.
Word Sense Clustering Another method used to extract conceptual enti-
ties from text is word sense clustering. Lexical resources such as WordNet [18]
provide ne-grained sense descriptions for a huge number of words in the En-
glish language. However, such ne-grained detail is not always required: it is
possible to cluster together word senses and come up with a looser word sense
that abstracts away a lot of the detail and is more compatible with the frame-
work as described in this paper. Snow et al. [19] describe a supervised learning
method for merging word senses using a SVM-based (Support Vector Machine)
clustering algorithm. This algorithm yields senses at a slightly higher level of
granularity and is more compatible with corporation into a conceptual layer
representation. Use of latent semantic indexing techniques is another powerful
and elegant method for dealing with synonymy and polysemy whilst at the same
time, automatically generating concepts, or clusters of word senses [20]. In la-
tent semantic indexing, each document is represented as a vector of terms, with
8. each element of the vector having an associated weight; some function of the
term frequency. Dimensionality reduction techniques are then applied to each of
these document vectors, resulting in a set of vectors that point in the direction
of semantically similar words. Each term in a document can then be projected
onto the reduced-dimensionality semantic space, where the dominant vector(s)
correspond to the concept that the term is most associated with. Probabilistic
latent semantic indexing techniques [21] provide an even more accurate means
of mapping from terms to concepts and such techniques will be used extensively
for the statistical analysis component of the framework.
Representing content in the form of a semantic map is thought to mimic the
way the human mind classies linguistic categories. While there is no direct link
between semantic maps and the representation of concepts in the human mind,
it is an objective starting point for investigation of the eciency of the semantic
map representation on human cognition.
3.3 Content Preparation Work-Flow
The framework architecture presented in gure 3 belongs to the core functional
unit of a web service that delivers on-demand slices to adaptive systems.
As a content provider, the framework will rst of all require domain specic
data gathered previously by pre-emptive crawls in areas of interest to adaptive
engine consumers. We envisage as an initial step to use the OCCS harvesting
system [22], developed previously, as a content gathering tool in order to boot-
strap this framework with a renewable wide range of content. The system will
be content harvester agnostic, hence any other content gathering system could
be plugged in if required, such as private repositories for instance.
Once suitable content is available, the framework described previously will
rstly remove any redundant content, then hand over only relevant information
to our semantic analysis unit, and then create a conceptual map of the data.
These two steps will produce a large amount of internal meta-data, both from a
semantic and structural point of view.
At this stage, the service will be ready to receive content requests from a
diverse range of consumers through conceptual queries which the service will at-
tempt to map as closely as possible to the internal meta-data gathered previously
on its content. The intelligent slicing unit will be responsible for matching re-
quests to available data within the system and slicing information objects to the
nest semantic grain possible matching the query. Semantic maps constructed a
priori will be used to produce a personalised slice for the adaptive systems. Se-
mantic technologies (OWL and SPARQL) will form the underlying mechanism
navigating through pre-emptive semantic mappings.
4 Evaluation and Roadmap
This section provides a brief summary of what are the main benets of this
framework in comparison to traditional systems. The limitations of this archi-
tecture are also discussed along with possible solutions. This will result in a
9. Fig. 3. Content analysis architecture
roadmap which focuses on an initial empirical evaluation of critical components,
its implementation and the overall evaluation of the framework.
The content preparation framework discussed in this paper aims at bridging
the gap between an increasingly large amount of open-corpus resources available
on the web and AHS information requirements. The automated content slicing
approach proposed oers a solution to the authorship scalability issues currently
impeding the mainstream adoption of adaptive web technologies.
This method removes any content packaging interoperability issues by local-
ising slices to the requirements of the AHS, instead of the AHS complying to a
specic format. Moreover, the pipelined architecture of this framework enables
the incorporation of additional plug-able meta-data generation services to the
system, thus removing the need for unscalable hand-made meta-data.
However, this novel content preparation solution raises new challenges. At
a component level, several technological challenges exist. To the best of our
knowledge, no existing structural segmentation approach currently fullls all of
our requirements. For this reason, an initial evaluation of existing algorithms will
be conducted, resulting in the combination of several techniques. Additionally,
the semantic analysing step of this framework seen as the critical stage of this
pipeline. An empirical evaluation of this step will therefore be needed so as to
measure the quality and size of the meta-data created by this component. This
process is critical since it will subsequently determine methods available to match
slices with content requests. As a result, the content precision of slices generated
will depend on the successive layers of the automated meta-data available. It is
possible, in addition, that this precision could be improved by the use of an a
posteriori manual crowd sourcing renement of the slices.
As part of the Centre for Next Generation Localisation (CNGL), an initial
implementation of this framework within a personalised multi-lingual customer
care (PMCC) system is planned. A number of framework level evaluations will be
undertaken in partnership with our industrial partners. An empirical evaluation
of automatic meta-data generation performance will be conducted in order to
estimate the degree of reliance towards automatic slice generation. These initial
10. steps will be necessary to ultimately provide a reliable foundation for the PMCC
application envisaged.
5 Discussion
This paper described how restrictions in the provision of suitable content for
Adaptive Hypermedia Systems currently impedes their full mainstream adoption
due to interoperability and scalability authoring issues. The decoupling of con-
tent infrastructures from Adaptive Systems along with the use of automatically
tailored open corpus resources was presented as a solution to this predicament.
Consequently, the architecture for a content provision service that lls the
gap between open corpus information resources and Adaptive System informa-
tion requirements, was outlined. To the best of our knowledge, this framework
represents the rst attempt to produce a pipelined, client agnostic, information
preparation service that produces virtual content slices for adaptive systems.
Other innovations in the framework architecture include the modular and plug-
gable open corpus analysers, which produce successive layers of internal meta-
data information. The combination of these components and how they integrate
coherently to produce the framework is also an aspect under investigation.
An initial prototype of this framework is currently under development and
aims towards integrating fully with a customer care information system that is
currently being deployed as part of the Centre for Next Generation Localisation
project.
Acknowledgements This research is supported by the Science Foundation
Ireland (grant 07/CE/I1142) as part of the Centre for Next Generation Locali-
sation (www.cngl.ie) at Trinity College, Dublin. The authors wish to thank the
anonymous reviewers for their comments.
References
1. Brusilovsky, P.: Adaptive Navigation Support in Educational Hypermedia: An
Evaluation of the ISIS-Tutor. Journal of Computing and Information Technology
(1998)
2. Hook, K.: Evaluating the utility and usability of an adaptive hypermedia system.
International Conference on Intelligent User Interfaces (1997)
3. Henze, N., Nejdl, W.: Adaptivity in the KBS Hyperbook System. 2nd Workshop
on Adaptive Systems and User Modeling on the WWW (1999)
4. Dieberger, A., Guzdial, M.: Coweb - Experiences with Collaborative Web Spaces.
Interacting with Social Information Spaces (2002)
5. De Bra, P.: Teaching Hypertext and Hypermedia through the Web. Journal of
Universal Computer Science 2 (1996)
6. IMS Global Learning Consortium: Learning Object Metadata (LOM) Information
Model 1.2.2 (2009)
11. 7. Farrell, R.G., Liburd, S.D., Thomas, J.C.: Dynamic Assembly of Learning Objects.
In: Proceedings of the 13th International World Wide Web Conference on Alternate
Track Papers Posters. (2004) 162169
8. Conlan, O., Wade, V., Bruen, C., Gargan, M.: Multi-model, Metadata Driven Ap-
proach to Adaptive Hypermedia Services for Personalized eLearning. In: Adaptive
Hypermedia and Adaptive Web-Based Systems. Springer-Verlag (2002) 100111
9. Brusilovsky, P., Henze, N.: The Adaptive Web. In: Open Corpus Adaptive Edu-
cational Hypermedia. Springer (2007) 671696
10. Henze, N., Lenski, W., Wette-Roch, E.: Adaptation in Open Corpus Hypermedia.
Journal of Articial Intelligence in Education 12 (2001) 325350
11. Carmona, C., Bueno, D., Guzman, E., Conejo, R.: SIGUE: Making Web Courses
Adaptive. In: Adaptive Hypermedia and Adaptive Web-Based Systems. (2002)
12. Baluja, S.: Browsing on Small Screens: Recasting Web-Page Segmentation into
an Ecient Machine Learning Framework. In: Proceedings of WWW-2006, ACM
(2006) 3342
13. Vieira, K., da Silva, A.S., Pinto, N., de Moura, E.S., Cavalcanti, J.M., Freire,
J.: A Fast and Robust Method for Web Page Template Detection and Removal.
In: Proceedings of the 15th ACM International Conference on Information and
Knowledge Management, ACM (2006) 258267
14. Chakrabarti, D., Kumar, R., Punera, K.: Page-Level Template Detection via Iso-
tonic Smoothing. In: Proceedings of the 16th International Conference on World
Wide Web, ACM (2007) 6170
15. Kohlsch, C., Nejdl, W.: A Densitometric Approach to Web Page Segmentation.
In: Proceedings of the 17th ACM Conference on Information and Knowledge Man-
agement, ACM (2008) 11731182
16. Cai, D., Yu, S., Wen, J.R., Ma, W.Y.: Extracting Content Structure for Web Pages
Based on Visual Representation. Springer (2003)
17. Jiang, J., Zhai, C.: Extraction of Coherent Relevant Passages using Hidden Markov
Models. ACM Transactions on Information Systems (TOIS) 24 (2006) 295319
18. Fellbaum, C.: WordNet: An Electronic Lexical Database. MIT Press (1998)
19. Snow, R., Prakash, S., Jurafsky, D., Ng, A.Y.: Learning to Merge Word Senses.
In: EMNLP. (2007)
20. Deerwester, S., Dumais, S., Furnas, G., Landauer, T., Harshaman, R.: Indexing
by Latent Semantic Analysis. Journal of the American Society for Information
Science 41 (1990) 391407
21. Hofmann, T.: Probabilistic Latent Semantic Indexing. In: 22nd Annual Inter-
national ACM SIGIR Conference on Research and Development in Information
Retrieval. (1999)
22. Lawless, S., Hederman, L., Wade, V.: Occs: Enabling the Dynamic Discovery,
Harvesting and Delivery of Educational Content from Open Corpus Sources. In:
Eighth IEEE International Conference on Advanced Learning Technologies, IEEE
Computer Society (2008) 676678