This document provides an overview of Medici 2, a scalable content management system for cultural heritage datasets. Medici 2 allows for the digitization, storage, management and analysis of large cultural heritage collections. It provides tools for dataset preprocessing, automatic metadata extraction, user metadata input, and visualization of data formats like images, 3D models, videos and Reflectance Transformation Imaging (RTI). The document describes Medici 2's architecture, which includes a web server, preprocessors/extractors, previewers, and database. It also covers Medici 2's metadata handling and provides an example of its dataset upload and preprocessing workflow.
The huge volume of text documents available on the internet has made it difficult to find valuable
information for specific users. In fact, the need for efficient applications to extract interested knowledge
from textual documents is vitally important. This paper addresses the problem of responding to user
queries by fetching the most relevant documents from a clustered set of documents. For this purpose, a
cluster-based information retrieval framework was proposed in this paper, in order to design and develop
a system for analysing and extracting useful patterns from text documents. In this approach, a pre-
processing step is first performed to find frequent and high-utility patterns in the data set. Then a Vector
Space Model (VSM) is performed to represent the dataset. The system was implemented through two main
phases. In phase 1, the clustering analysis process is designed and implemented to group documents into
several clusters, while in phase 2, an information retrieval process was implemented to rank clusters
according to the user queries in order to retrieve the relevant documents from specific clusters deemed
relevant to the query. Then the results are evaluated according to evaluation criteria. Recall and Precision
(P@5, P@10) of the retrieved results. P@5 was 0.660 and P@10 was 0.655.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
The huge volume of text documents available on the internet has made it difficult to find valuable
information for specific users. In fact, the need for efficient applications to extract interested knowledge
from textual documents is vitally important. This paper addresses the problem of responding to user
queries by fetching the most relevant documents from a clustered set of documents. For this purpose, a
cluster-based information retrieval framework was proposed in this paper, in order to design and develop
a system for analysing and extracting useful patterns from text documents. In this approach, a pre-
processing step is first performed to find frequent and high-utility patterns in the data set. Then a Vector
Space Model (VSM) is performed to represent the dataset. The system was implemented through two main
phases. In phase 1, the clustering analysis process is designed and implemented to group documents into
several clusters, while in phase 2, an information retrieval process was implemented to rank clusters
according to the user queries in order to retrieve the relevant documents from specific clusters deemed
relevant to the query. Then the results are evaluated according to evaluation criteria. Recall and Precision
(P@5, P@10) of the retrieved results. P@5 was 0.660 and P@10 was 0.655.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
Study on potential capabilities of a nodb systemijitjournal
There is a need of optimal data to query processing technique to handle the increasing database size,
complexity, diversity of use. With the introduction of commercial website, social network, expectations are
that the high scalability, more flexible database will replace the RDBMS. Complex application and Big
Table require highly optimized queries. Users are facing the increasing bottlenecks in their data analysis. A
growing part of the database community recognizes the need for significant and fundamental changes to
database design. A new philosophy for creating database systems called noDB aims at minimizing the datato-
query time, most prominently by removing the need to load data before launching queries. That will
process queries without any data preparation or loading steps. There may not need to store data. User can
pipe raw data from websites, DBs, excel sheets into two promise sample inputs without storing anything.
This study is based on PostgreSQL systems. A series of the baseline experiment are executed to evaluate the
Performance of this system as per -a. Data loading cost, b-Query processing timing, c-Avoidance of
Collision and Deadlock, d-Enabling the Big data storage and e-Optimize query processing etc. The study
found significant possible capabilities of noDB system over the traditional database management system.
Big Data analytics is common in many business domains, for example in financial sector for savings portfolio analysis, in government agencies, scientific research and insurance providers, to name a few. The uses of Big Data range from generating simple reports to executing complex analytical workloads. The increases in the amount of data being stored and processed in these domains expose many challenges with respect to scalable processing of analytical queries. Massively Parallel Processing (MPP) databases address these challenges by distributing storage and query processing across multiple compute nodes and distributed processes in parallel, usually in a shared-nothing architecture. Today, new technologies are shaping up the way platforms for the Internet of services are designed and managed. This technology is the containers, like Docker[1,2] and LXC[3,4,5]. The use of container as base technology for large-scale distributed systems opens many challenges in the area of resource management at run-time, for example: auto-scaling, optimal deployment and monitoring.
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
Flexibility in Metadata Schemes and Standardisation: the Case of CMDI and DAN...vty
Presentation at ISKO Knowledge Organisation Research Observatory. RESEARCH REPOSITORIES AND DATAVERSE: NEGOTIATING METADATA, VOCABULARIES AND DOMAIN NEEDS
More number of users requires cloud services to transfer private data like electronic health records and financial transaction records. A cloud computing services offers several flavors of virtual machines to handle large scale datasets. But centralized approaches are difficult in handling of large datasets. Data anonymization is used for privacy preservation techniques. It is challenged to manage and process such large-scale data within a cloud application. A scalable two-phase top-down specialization (TDS) approach to anonymize large-scale data sets using the Map Reduce framework on cloud. It is used to investigate the scalability problem of large-scale data anonymization techniques. These approaches deliberately design a group of innovative Map Reduce jobs to concretely accomplish the specialization computation in a highly scalable way. The Top-Down Specialization process speeds up the specialization process because indexing structure avoids frequently scanning entire data sets and storing statistical results.
Personalized Multimedia Web Services in Peer to Peer Networks Using MPEG-7 an...University of Piraeus
Multimedia information has been increased in the recent years while new content delivery services enhanced with personalization functionalities are provided to users. Several standards are proposed for the representation and retrieval of multimedia content. This paper makes an overview of the available standards and technologies. Furthermore a prototype semantic P2P architecture is presented which delivers personalized audio information. The metadata which support personalization are separated in two categories: the metadata describing user preferences stored at each user and the resource adaptation metadata stored at the P2P network’s web services. The multimedia models MPEG-21 and MPEG-7 are used to describe metadata information and the Web Ontology Language (OWL) to produce and manipulate ontological descriptions. SPARQL is used for querying the OWL ontologies. The MPEG Query Format (MPQF) is also used, providing a well-known framework for applying queries to the metadata and to the ontologies.
DATABASE SYSTEMS PERFORMANCE EVALUATION FOR IOT APPLICATIONSijdms
ABSTRACT
The amount of data stored in IoT databases increases as the IoT applications extend throughout smart city appliances, industry and agriculture. Contemporary database systems must process huge amounts of sensory and actuator data in real-time or interactively. Facing this first wave of IoT revolution, database vendors struggle day-by-day in order to gain more market share, develop new capabilities and attempt to overcome the disadvantages of previous releases, while providing features for the IoT.
There are two popular database types: The Relational Database Management Systems and NoSQL databases, with NoSQL gaining ground on IoT data storage. In the context of this paper these two types are examined. Focusing on open source databases, the authors experiment on IoT data sets and pose an answer to the question which one performs better than the other. It is a comparative study on the performance of the commonly market used open source databases, presenting results for the NoSQL MongoDB database and SQL databases of MySQL and PostgreSQL
COMPLETE END-TO-END LOW COST SOLUTION TO A 3D SCANNING SYSTEM WITH INTEGRATED...ijcsit
3D reconstruction is a technique used in computer vision which has a wide range of applications in
areas like object recognition, city modelling, virtual reality, physical simulations, video games and
special effects. Previously, to perform a 3D reconstruction, specialized hardwares were required.
Such systems were often very expensive and was only available for industrial or research purpose.
With the rise of the availability of high-quality low cost 3D sensors, it is now possible to design
inexpensive complete 3D scanning systems. The objective of this work was to design an acquisition and
processing system that can perform 3D scanning and reconstruction of objects seamlessly. In addition,
the goal of this work also included making the 3D scanning process fully automated by building and
integrating a turntable alongside the software. This means the user can perform a full 3D scan only by
a press of a few buttons from our dedicated graphical user interface. Three main steps were followed
to go from acquisition of point clouds to the finished reconstructed 3D model. First, our system
acquires point cloud data of a person/object using inexpensive camera sensor. Second, align and
convert the acquired point cloud data into a watertight mesh of good quality. Third, export the
reconstructed model to a 3D printer to obtain a proper 3D print of the model.
CYBER INFRASTRUCTURE AS A SERVICE TO EMPOWER MULTIDISCIPLINARY, DATA-DRIVEN S...ijcsit
In supporting its large scale, multidisciplinary scientific research efforts across all the university campuses and by the research personnel spread over literally every corner of the state, the state of Nevada needs to build and leverage its own Cyber infrastructure. Following the well-established as-a-service model, this state-wide Cyber infrastructure that consists of data acquisition, data storage, advanced instruments, visualization, computing and information processing systems, and people, all seamlessly linked together through a high-speed network, is designed and operated to deliver the benefits of Cyber infrastructure-as-aService (CaaS).There are three major service groups in this CaaS, namely (i) supporting infrastructural
services that comprise sensors, computing/storage/networking hardware, operating system, management tools, virtualization and message passing interface (MPI); (ii) data transmission and storage services that provide connectivity to various big data sources, as well as cached and stored datasets in a distributed
storage backend; and (iii) processing and visualization services that provide user access to rich processing and visualization tools and packages essential to various scientific research workflows. Built on commodity hardware and open source software packages, the Southern Nevada Research Cloud(SNRC)and a data repository in a separate location constitute a low cost solution to deliver all these services around CaaS. The service-oriented architecture and implementation of the SNRC are geared to encapsulate as much detail of big data processing and cloud computing as possible away from end users; rather scientists only need to learn and access an interactive web-based interface to conduct their collaborative, multidisciplinary, dataintensive research. The capability and easy-to-use features of the SNRC are demonstrated through a use case that attempts to derive a solar radiation model from a large data set by regression analysis.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Grid scheduling is a process of mapping Grid jobs to resources over multiple administrative domains.
A Grid job can be split into many small tasks.
The scheduler has the responsibility of selecting resources and scheduling jobs in such a way that the user and application requirements are met,in terms of overall execution time (throughput) and cost of the resources utilized.
Semantic Interoperability in Infocosm: Beyond Infrastructural and Data Intero...Amit Sheth
Amit Sheth, Keynote: International Conference on Interoperating Geographic Systems (Interop’97), Santa Barbara, December 3-4 1997.
Related technical paper: http://knoesis.org/library/resource.php?id=00230
Dynamic Resource Provisioning with Authentication in Distributed DatabaseEditor IJCATR
Data center have the largest consumption amounts of energy in sharing the power. The public cloud workloads of different
priorities and performance requirements of various applications [4]. Cloud data center have capable of sensing an opportunity to present
different programs. In my proposed construction and the name of the security level of imperturbable privacy leakage rarely distributed
cloud system to deal with the persistent characteristics there is a substantial increases and information that can be used to augment the
profit, retrenchment overhead or both. Data Mining Analysis of data from different perspectives and summarizing it into useful
information is a process. Three empirical algorithms have been proposed assignments estimate the ratios are dissected theoretically and
compared using real Internet latency data recital of testing methods
Big data is a term which refers to those data sets or combinations of data sets whose volume,
complexity, and rate of growth make them difficult to be captured, managed, processed or analysed by
traditional tools and technologies. Big data is relatively a new concept that includes huge quantities of data,
social media analytics and real time data. Over the past era, there have been a lot of efforts and studies are
carried out in growing proficient tools for performing various tasks in big data. Due to the large and
complex collection of datasets it is difficult to process on traditional data processing applications. This
concern turns to be further mandatory for producing various tools in big data. In this survey, a various
collection of big data tools are illustrated and also analysed with the salient features.
SYSTEMATIC LITERATURE REVIEW ON RESOURCE ALLOCATION AND RESOURCE SCHEDULING I...ijait
he objective the work is intend to highlight the key features and afford finest future directions in the
research community of Resource Allocation, Resource Scheduling and Resource management from 2009 to
2016. Exemplifying how research on Resource Allocation, Resource Scheduling and Resource management
has progressively increased in the past decade by inspecting articles, papers from scientific and standard
publications. Survey materialized in three fold process. Firstly, investigate on the amalgamation of
Resource Allocation, Resource Scheduling and then proceeded with Resource management. Secondly, we
performed a structural analysis on different author’s prominent contributions in the form of tabulation by
categories and graphical representation. Thirdly, huddle with conceptual similarity in the field and also
impart a summary on all resource allocations. In cloud computing environments, there are two players:
cloud providers and cloud users. On one hand, providers hold massive computing resources in their large
datacenters and rent resources out to users on a per-usage basis. On the other hand, there are users who
Leveraging Open Source Technologies to Enable Scientific Archiving and Discovery; Steve Hughes, NASA; Data Publication Repositories
The 2nd Research Data Access and Preservation (RDAP) Summit
An ASIS&T Summit
March 31-April 1, 2011 Denver, CO
In cooperation with the Coalition for Networked Information
http://asist.org/Conferences/RDAP11/index.html
Study on potential capabilities of a nodb systemijitjournal
There is a need of optimal data to query processing technique to handle the increasing database size,
complexity, diversity of use. With the introduction of commercial website, social network, expectations are
that the high scalability, more flexible database will replace the RDBMS. Complex application and Big
Table require highly optimized queries. Users are facing the increasing bottlenecks in their data analysis. A
growing part of the database community recognizes the need for significant and fundamental changes to
database design. A new philosophy for creating database systems called noDB aims at minimizing the datato-
query time, most prominently by removing the need to load data before launching queries. That will
process queries without any data preparation or loading steps. There may not need to store data. User can
pipe raw data from websites, DBs, excel sheets into two promise sample inputs without storing anything.
This study is based on PostgreSQL systems. A series of the baseline experiment are executed to evaluate the
Performance of this system as per -a. Data loading cost, b-Query processing timing, c-Avoidance of
Collision and Deadlock, d-Enabling the Big data storage and e-Optimize query processing etc. The study
found significant possible capabilities of noDB system over the traditional database management system.
Big Data analytics is common in many business domains, for example in financial sector for savings portfolio analysis, in government agencies, scientific research and insurance providers, to name a few. The uses of Big Data range from generating simple reports to executing complex analytical workloads. The increases in the amount of data being stored and processed in these domains expose many challenges with respect to scalable processing of analytical queries. Massively Parallel Processing (MPP) databases address these challenges by distributing storage and query processing across multiple compute nodes and distributed processes in parallel, usually in a shared-nothing architecture. Today, new technologies are shaping up the way platforms for the Internet of services are designed and managed. This technology is the containers, like Docker[1,2] and LXC[3,4,5]. The use of container as base technology for large-scale distributed systems opens many challenges in the area of resource management at run-time, for example: auto-scaling, optimal deployment and monitoring.
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
Flexibility in Metadata Schemes and Standardisation: the Case of CMDI and DAN...vty
Presentation at ISKO Knowledge Organisation Research Observatory. RESEARCH REPOSITORIES AND DATAVERSE: NEGOTIATING METADATA, VOCABULARIES AND DOMAIN NEEDS
More number of users requires cloud services to transfer private data like electronic health records and financial transaction records. A cloud computing services offers several flavors of virtual machines to handle large scale datasets. But centralized approaches are difficult in handling of large datasets. Data anonymization is used for privacy preservation techniques. It is challenged to manage and process such large-scale data within a cloud application. A scalable two-phase top-down specialization (TDS) approach to anonymize large-scale data sets using the Map Reduce framework on cloud. It is used to investigate the scalability problem of large-scale data anonymization techniques. These approaches deliberately design a group of innovative Map Reduce jobs to concretely accomplish the specialization computation in a highly scalable way. The Top-Down Specialization process speeds up the specialization process because indexing structure avoids frequently scanning entire data sets and storing statistical results.
Personalized Multimedia Web Services in Peer to Peer Networks Using MPEG-7 an...University of Piraeus
Multimedia information has been increased in the recent years while new content delivery services enhanced with personalization functionalities are provided to users. Several standards are proposed for the representation and retrieval of multimedia content. This paper makes an overview of the available standards and technologies. Furthermore a prototype semantic P2P architecture is presented which delivers personalized audio information. The metadata which support personalization are separated in two categories: the metadata describing user preferences stored at each user and the resource adaptation metadata stored at the P2P network’s web services. The multimedia models MPEG-21 and MPEG-7 are used to describe metadata information and the Web Ontology Language (OWL) to produce and manipulate ontological descriptions. SPARQL is used for querying the OWL ontologies. The MPEG Query Format (MPQF) is also used, providing a well-known framework for applying queries to the metadata and to the ontologies.
DATABASE SYSTEMS PERFORMANCE EVALUATION FOR IOT APPLICATIONSijdms
ABSTRACT
The amount of data stored in IoT databases increases as the IoT applications extend throughout smart city appliances, industry and agriculture. Contemporary database systems must process huge amounts of sensory and actuator data in real-time or interactively. Facing this first wave of IoT revolution, database vendors struggle day-by-day in order to gain more market share, develop new capabilities and attempt to overcome the disadvantages of previous releases, while providing features for the IoT.
There are two popular database types: The Relational Database Management Systems and NoSQL databases, with NoSQL gaining ground on IoT data storage. In the context of this paper these two types are examined. Focusing on open source databases, the authors experiment on IoT data sets and pose an answer to the question which one performs better than the other. It is a comparative study on the performance of the commonly market used open source databases, presenting results for the NoSQL MongoDB database and SQL databases of MySQL and PostgreSQL
COMPLETE END-TO-END LOW COST SOLUTION TO A 3D SCANNING SYSTEM WITH INTEGRATED...ijcsit
3D reconstruction is a technique used in computer vision which has a wide range of applications in
areas like object recognition, city modelling, virtual reality, physical simulations, video games and
special effects. Previously, to perform a 3D reconstruction, specialized hardwares were required.
Such systems were often very expensive and was only available for industrial or research purpose.
With the rise of the availability of high-quality low cost 3D sensors, it is now possible to design
inexpensive complete 3D scanning systems. The objective of this work was to design an acquisition and
processing system that can perform 3D scanning and reconstruction of objects seamlessly. In addition,
the goal of this work also included making the 3D scanning process fully automated by building and
integrating a turntable alongside the software. This means the user can perform a full 3D scan only by
a press of a few buttons from our dedicated graphical user interface. Three main steps were followed
to go from acquisition of point clouds to the finished reconstructed 3D model. First, our system
acquires point cloud data of a person/object using inexpensive camera sensor. Second, align and
convert the acquired point cloud data into a watertight mesh of good quality. Third, export the
reconstructed model to a 3D printer to obtain a proper 3D print of the model.
CYBER INFRASTRUCTURE AS A SERVICE TO EMPOWER MULTIDISCIPLINARY, DATA-DRIVEN S...ijcsit
In supporting its large scale, multidisciplinary scientific research efforts across all the university campuses and by the research personnel spread over literally every corner of the state, the state of Nevada needs to build and leverage its own Cyber infrastructure. Following the well-established as-a-service model, this state-wide Cyber infrastructure that consists of data acquisition, data storage, advanced instruments, visualization, computing and information processing systems, and people, all seamlessly linked together through a high-speed network, is designed and operated to deliver the benefits of Cyber infrastructure-as-aService (CaaS).There are three major service groups in this CaaS, namely (i) supporting infrastructural
services that comprise sensors, computing/storage/networking hardware, operating system, management tools, virtualization and message passing interface (MPI); (ii) data transmission and storage services that provide connectivity to various big data sources, as well as cached and stored datasets in a distributed
storage backend; and (iii) processing and visualization services that provide user access to rich processing and visualization tools and packages essential to various scientific research workflows. Built on commodity hardware and open source software packages, the Southern Nevada Research Cloud(SNRC)and a data repository in a separate location constitute a low cost solution to deliver all these services around CaaS. The service-oriented architecture and implementation of the SNRC are geared to encapsulate as much detail of big data processing and cloud computing as possible away from end users; rather scientists only need to learn and access an interactive web-based interface to conduct their collaborative, multidisciplinary, dataintensive research. The capability and easy-to-use features of the SNRC are demonstrated through a use case that attempts to derive a solar radiation model from a large data set by regression analysis.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Grid scheduling is a process of mapping Grid jobs to resources over multiple administrative domains.
A Grid job can be split into many small tasks.
The scheduler has the responsibility of selecting resources and scheduling jobs in such a way that the user and application requirements are met,in terms of overall execution time (throughput) and cost of the resources utilized.
Semantic Interoperability in Infocosm: Beyond Infrastructural and Data Intero...Amit Sheth
Amit Sheth, Keynote: International Conference on Interoperating Geographic Systems (Interop’97), Santa Barbara, December 3-4 1997.
Related technical paper: http://knoesis.org/library/resource.php?id=00230
Dynamic Resource Provisioning with Authentication in Distributed DatabaseEditor IJCATR
Data center have the largest consumption amounts of energy in sharing the power. The public cloud workloads of different
priorities and performance requirements of various applications [4]. Cloud data center have capable of sensing an opportunity to present
different programs. In my proposed construction and the name of the security level of imperturbable privacy leakage rarely distributed
cloud system to deal with the persistent characteristics there is a substantial increases and information that can be used to augment the
profit, retrenchment overhead or both. Data Mining Analysis of data from different perspectives and summarizing it into useful
information is a process. Three empirical algorithms have been proposed assignments estimate the ratios are dissected theoretically and
compared using real Internet latency data recital of testing methods
Big data is a term which refers to those data sets or combinations of data sets whose volume,
complexity, and rate of growth make them difficult to be captured, managed, processed or analysed by
traditional tools and technologies. Big data is relatively a new concept that includes huge quantities of data,
social media analytics and real time data. Over the past era, there have been a lot of efforts and studies are
carried out in growing proficient tools for performing various tasks in big data. Due to the large and
complex collection of datasets it is difficult to process on traditional data processing applications. This
concern turns to be further mandatory for producing various tools in big data. In this survey, a various
collection of big data tools are illustrated and also analysed with the salient features.
SYSTEMATIC LITERATURE REVIEW ON RESOURCE ALLOCATION AND RESOURCE SCHEDULING I...ijait
he objective the work is intend to highlight the key features and afford finest future directions in the
research community of Resource Allocation, Resource Scheduling and Resource management from 2009 to
2016. Exemplifying how research on Resource Allocation, Resource Scheduling and Resource management
has progressively increased in the past decade by inspecting articles, papers from scientific and standard
publications. Survey materialized in three fold process. Firstly, investigate on the amalgamation of
Resource Allocation, Resource Scheduling and then proceeded with Resource management. Secondly, we
performed a structural analysis on different author’s prominent contributions in the form of tabulation by
categories and graphical representation. Thirdly, huddle with conceptual similarity in the field and also
impart a summary on all resource allocations. In cloud computing environments, there are two players:
cloud providers and cloud users. On one hand, providers hold massive computing resources in their large
datacenters and rent resources out to users on a per-usage basis. On the other hand, there are users who
Leveraging Open Source Technologies to Enable Scientific Archiving and Discovery; Steve Hughes, NASA; Data Publication Repositories
The 2nd Research Data Access and Preservation (RDAP) Summit
An ASIS&T Summit
March 31-April 1, 2011 Denver, CO
In cooperation with the Coalition for Networked Information
http://asist.org/Conferences/RDAP11/index.html
Designing and configuring context-aware semantic web applicationsTELKOMNIKA JOURNAL
Context-aware services are attracting attention of world as the use of web services are rapidly growing. We designed an architecture of context-aware semantic web which provides on demand flexibility and scalability in extracting and mining the research papers from well-known digital libraries i.e. ACM, IEEE and SpringerLink. This paper proposes a context-aware administrations system, which supports programmed revelation and incorporation of setting dependent on Semantic Web administrations. This work has been done using the python programming language with a dedicated library for the semantic web analysis named as “Cubic-Web” on any defined dataset, in our case as we have used a dataset for extracting and studying several publications to measure the impact of context aware semantic web application on the results. We have found the average recall and averge accuracy for all the context aware research journals in our research work. Moreover, as this study is limited journal documents, other future studies can be approached by examining different types of publications using this advance research. An efficient system has been designed considering the parameters of research article meta-data to find out the papers from the web using semantic web technology. Parameters like year of publication, type of publication, number of contributors, evaluation methods and analysis method used in publication. All this data has been extracted using the designed context-aware semantic web technology.
Open Source Platforms Integration for the Development of an Architecture of C...Eswar Publications
The goal of the Internet of Things (IoT) is to achieve the interconnection and interaction of all kind of everyday
objects. IoT architecture can be implemented in various ways. This paper presents a way to mount an IoT architecture using open source hardware and software platforms and shows that this is a viable option to collect information through various sensors and present it through a web page.
As the volume and complexity of data from myriad Earth Observing platforms, both remote sensing and in-situ increases so does the demand for access to both data and information products from these data. The audience no longer is restricted to an investigator team with specialist science credentials. Non-specialist users from scientists from other disciplines, science-literate public, to teachers, to the general public and decision makers want access. What prevents them from this access to resources? It is the very complexity and specialist developed data formats, data set organizations and specialist terminology. What can be done in response? We must shift the burden from the user to the data provider. To achieve this our developed data infrastructures are likely to need greater degrees of internal code and data structure complexity to achieve (relatively) simpler end-user complexity. Evidence from numerous technical and consumer markets supports this scenario. We will cover the elements of modern data environments, what the new use cases are and how we can respond to them.
Message Oriented Middleware for Library’s Metadata ExchangeTELKOMNIKA JOURNAL
Library is one of the important tools in the development of science to store various intellectual properties. Currently most libraries are managed by standalone systems and are not equipped with data exchange facilities with other libraries for sharing information. Sharing of information between libraries can be done with integration metadata owned library. In this research, the integration architecture of metadata exchange is done with Message Oriented Middleware (MOM) technology. This MOM redeems the collection metadata that matches the standard Dublin Core format. In this research, database structure, MOM structure and set of rules to perform data sharing process. With the proposed MOM architectural design is expected to search process information between libraries will become easier and cheaper.
The best way to publish and share research data is with a research data repository. A repository is an online database that allows research data to be preserved across time and helps others find it.
This presentation gives details on technologies and approaches towards exploiting Linked Data by building LD applications. In particular, it gives an overview of popular existing applications and introduces the main technologies that support implementation and development. Furthermore, it illustrates how data exposed through common Web APIs can be integrated with Linked Data in order to create mashups.
This presentation will provide an overview of issues in digital preservation. Presentation was delivered during the joint DPE/Planets/CAPAR/nestor training event, ‘The Preservation challenge: basic concepts and practical applications’ (Barcelona, March 2009)
1. Medici 2: A Scalable CMS for Cultural Heritage
Datasets
Motivation, Capabilities, Future Directions
Constantinos Sophocleous
Computation-based Science and
Technology Research Center
The Cyprus Institute
Nicosia, Cyprus
Luigi Marini
National Center for Supercomputing
Applications
University of Illinois at Urbana-
Champaign
Urbana, Illinois, USA
Ropertos Georgiou
Science and Technology in
Archaeology Research Center
The Cyprus Institute
Nicosia, Cyprus
Mohammed Elfarargy
International School of Information Science
Bibliotheca Alexandrina
Alexandria, Egypt
Abstract— Digitizing large collections of Cultural Heritage
(CH) resources and providing tools for their management,
analysis and visualization is critical to CH research. A key
element in achieving the above goal is to provide user-friendly
software offering an abstract interface for interaction with a
variety of digital content types. To address these needs, the
Medici content management system is being developed in a
collaborative effort between the National Center for
Supercomputing Applications (NCSA) at the University of
Illinois at Urbana-Champaign, Bibliotheca Alexandrina (BA) in
Egypt, and the Cyprus Institute (CyI). The project is pursued in
the framework of European Project LinkSCEEM2 and
supported by work funded through the U.S. National Science
Foundation (NSF), the U.S. National Archives and Records
Administration (NARA), the U.S. National Institutes of Health
(NIH), the U.S. National Endowment for the Humanities (NEH),
the U.S. Office of Naval Research (ONR), the U.S. Environmental
Protection Agency (EPA) as well as other private sector efforts.
Medici is a Web 2.0 environment integrating analysis tools for
the auto-curation of un-curated digital data, allowing automatic
processing of input (CH) datasets, and visualization of both data
and collections. It offers a simple user interface for dataset
preprocessing, previewing, automatic metadata extraction, user
input of metadata and provenance support, storage, archiving
and management, representation and reproduction.
Building on previous experience (Medici 1), NCSA, and CyI
are working towards the improvement of the technical,
performance and functionality aspects of the system. The current
version of Medici (Medici 2) is the result of these efforts. It is a
scalable, flexible, robust distributed framework with wide data
format support (including 3D models and Reflectance
Transformation Imaging-RTI) and metadata functionality.
We provide an overview of Medici 2’s current features
supported by representative use cases as well as a discussion of
future development directions.
Kenton McHenry
National Center for Supercomputing Applications
University of Illinois at Urbana-Champaign
Urbana, Illinois, USA
Index Terms—Medici, Archaeology, Cultural Heritage,
Content Based Analysis, Un-curated Data, Auto-curation, Web
2D, Web 3D, Reflectance Transformation Imaging (RTI), RTI-
to-3D.
I. MOTIVATION – INTRODUCTION
CH research is producing huge amounts of data at an ever-
increasing rate and size due to the continuous development of
more advanced and detailed data extraction technologies and
methods. Such research efforts include the pilot projects
undertaken by the Cyprus Institute during the establishment of
an advanced imaging center for CH and archaeology as part of
the LinkSCEEM-2 Cultural Heritage Work Package. The three
pilot projects used cutting-edge visualization technologies such
as RTI. They generated scientific digital data of archaeological
objects and artifacts, such as cylinder seals, ceramics, jewelry
and other museum collections, as well as wall paintings and
icons from the world known medieval churches of Cyprus.
In these contexts, the digital documentation, storage and
management of the aforementioned data as well as providing
easy access to researchers is becoming ever-more critical.
Flexible database storage, processing, metadata extraction and
presentation frameworks are needed in order to satisfy these
requirements. The development of such database/data
repository systems establishes a long-term solution for
accessing and storing various kinds of CH data and metadata.
These systems must be flexible in terms of being able to have
additional data formats support incorporated over time, thus
ensuring their continuous update with latest technological
developments; thus maintaining the CH scientific
communities’ interest.
The developed framework will be complementary to
repositories such as the Inscriptifact database and online library
[1]. Having a specialized resource repository for the needs of
2. specific datasets (such as Inscriptifact) must be coupled with a
general digital storage and management framework also
capable of handling types of data that are not the focus of
Inscriptifact. This collaboration will immediately allow the
scientific exploration of newly generated CH datasets by an
already established user community.
Medici has and is being developed to address the above. It
is a Web 2.0-based general multimedia content management
system capable of semantic content management and
service/cloud-based workflows. It supports a broad range of
throughput-intensive research techniques and community data
management. Medici provides scalable storage and media
processing, straightforward user interfaces, social annotation
capabilities, preprocessing and previewing, metadata extension
and manipulation, provenance support and citable persistent
data references [2].
Through Medici, users can upload datasets in a variety of
formats, including 3D and RTI formats widely used in cultural
heritage research and collections [3][4]. Dataset metadata is
obtained by extraction services run as data preprocessors and
presented to the user along with metadata input by the
community for each dataset. Utilizing these extractors Medici
is capable of deriving semantically meaningful features for
content-based comparison between datasets (e.g. textual data
from handwritten text). Previews of large datasets in a variety
of formats are also extracted and viewed to avoid the need of
(down) loading the whole content on the user’s system or
finding the needed software to examine the contents of a file.
Medici 2.0’s scalability/parallelization, flexibility and
robustness as well as its overall performance are improved by
decoupling the preprocessor software from the main server
(with extractors being allowed to run on different machines in a
distributed architecture). A powerful and flexible RDF-enabled
database management system is used for storing datasets.
Dataset files can be uploaded using a variety of interfaces. Use
of the RDF standard makes the datasets’ metadata open and
portable. Medici uses the latest Web technologies to display
large images smoothly (Seadragon, zoom.it [5]), 3D graphics
(X3DOM [6]), video (FFMPEG[7]) and RTI (SpiderGL[8])
along with a special feature added to the RTI web visualization
developed by Bibliotheca Alexandrina to extract 3D models
from RTI files (RTI-to-3D) in order to enhance the possibilities
of the technology.
Different deployments of Medici can be hosted by different
institutions and parts, thus allowing the hosting party better
control of access to resources.
II. FUNCTIONS AND ARCHITECTURE
A diagram describing the dataset upload,preprocessing and
previewing process is displayed in Fig. 1.
The systemis composed of the following parts:
A. Web server
a. Functions and Architecture
Datasets and files are uploaded to the web serverusing one
of an array of methods. One of them is regular HTML forms
(both when creating a new dataset to upload its first file as well
when viewing a dataset to add files to the dataset). However,
files can also be uploaded in other ways, including uploading
individual files that do not belong to a dataset.
The web server is responsible for sending a request to the
database to store the file or dataset and send a message to the
RabbitMQ message broker [9] defined in the server’s
configuration. This message will contain the necessary
information pertaining to the file or dataset that are needed by
the RabbitMQ broker to distribute the file or dataset’s
preprocessing jobs to the available extractors/preprocessors as
well as by the extractors/preprocessors themselves to decide
how to process the file or dataset.Datasets contained as files in
a zipped archive are inspected by the web server. These files
are unzipped to identify the type of dataset whose
preprocessing will be handled by extractors on the RabbitMQ
bus.
The web server can then passively await for any auto-
extracted file or dataset metadata or previews (generated by the
extractors-preprocessors that received a job for the file or
dataset) to be uploaded to the server through a REST API [10].
When that happens, the web server calls the database to save
the previews/metadata and associate them with the file or
dataset.
Extractors can be chained. This means that the resulting file
output from an extractor may be fed as input to another
extractor, thus allowing even greater flexibility regarding
execution of preprocessing subtasks common to the
preprocessing of different file or dataset types. For this to
happen, an extractor that outputs an intermediate result can
upload it as an intermediate result for the file or dataset using
the server’s REST API. The server then calls the database to
save the intermediate result and a new job description is sent to
RabbitMQ, now containing the identifiers of both the original
file or dataset (with which the final result will be associated)
and the intermediate file (which the extractor(s) working on the
new job need to download and process).
The web server is also responsible for selecting and setting
up the previewers that will display the contents of a file or
dataset based on the type (i.e. file format) of a file or dataset.
Other functions include working with custommetadata and
searches,an interface for adding metadata to a dataset by users
based on a mapping dependent on the institution managing the
Medici implementation, searching for datasets having metadata
satisfying a query and text-based search.The server also allows
social annotation of datasets (i.e. tagging, notes, comments,
custommetadata).
b. Technologies used
The web server is built on the Play web application
framework [11][12], written for allowing coding in Scala and
Java. As supported by the framework, the server uses the
model-view-controller (MVC) architectural pattern [13]. It
relies on a number of plugins for communicating with the
RabbitMQ broker, communicating with the database and
authenticating users. It uses dynamic HTML (HTML version
3. 5) webpages generation (views) according to the results of the
processing of the input (Scala-based controllers). The models
(defined in Scala) are closely associated with collections in the
database.
Preprocessors and scripts (i.e. previewers) running on
users’ browsers communicate with the server using a REST
API.
B. Preprocessors / Extractors
Preprocessors are used for both extracting metadata from
datasets and files as well as generating previews for them.
Each extractor listens to a RabbitMQ broker via its own
named job queue, which is bound to particular routing keys. It
is by these keys that RabbitMQ forwards the extraction job
messages to the extractors’ queues. Each job message is
handled separately. Whenever a job is received, the extractor
first downloads the dataset or file having the identifier stated in
the job description from the web server via the server’s REST
API.
What follows depends on the task each extractor is
responsible for. Extractors in general use integrated libraries or
external system calls to third-party software to process dataset
files to generate the previews or extract the metadata. The
third-party software include for example 3D model and video
processing programs that are installed to the extractor’s
environment and are called through Java command-line calls.
Finally, the extractor returns the result back to the web
server:
If the result is an intermediate file for use in extractor
chaining, the file is uploaded to the server as an
intermediate result to be forwarded to other extractors
through RabbitMQ to continue processing.
If the result is a final preview, it is returned to the
server as a preview and then a command is sent to the
serverto associate the preview with the original dataset
or file.
If the result is a set of dataset or file metadata, a
command is sent to the server to associate the
metadata with the original dataset or file.
Many extractors of the same type can be run
simultaneously on the same or different machines and have the
jobs sent by a Medici web server be distributed among themby
RabbitMQ as long as they are registered with the same
RabbitMQ routing exchange. Also, an extractor may receive
messages from more than one Medici web server instance as
long as the web servers are registered with the same RabbitMQ
Fig. 1. Dataset upload, preprocessing and previewing process.
4. routing exchange. This allows for great flexibility, scalability
and robustness through a distributed architecture.
Extractors can be implemented using Java or Python.
C. Previewers
Previewers are used for viewing the contents of a dataset or
a file from within the user’s browser.
On the server side, the server selects which previewer to be
launched for each preview of each file in the dataset when a
dataset page is requested fromthe web server. The selection is
done by comparing each preview’s file type with the file types
accepted by each previewer. The latter information is held as
JavaScript Object Notation (JSON) data in a public file for
each previewer.
Before the previewer scripts are sent to the browser along
with the rest of the dataset or file page to be executed, certain
variables are set dynamically by the server, mostly for the
user’s browser to be able to find the files needed by the
previewer on the server (e.g. the dataset files’ URLs).
The previewers use dynamically-added JavaScript scripts
downloaded from the web server at runtime and added to the
webpage’s DOM (Document Object Model) using jQuery [14].
The previews to be displayed,as well as any other needed files,
are also downloaded at runtime using AJAX calls. After the
preview is ready to be displayed, the preview HTML is
appended to the webpage’s HTML.
The previewers use browser-side JavaScript and jQuery.
D. Database
The NoSQL MongoDB is used as the system’s database
management system (DBMS). Selecting a NoSQL DBMS
offers advantages important for Medici. Flexibility is a key
attribute for Medici, as CH community requirements change
constantly as described in the Introduction. This gives value to
the horizontal scalability of NoSQL. Simplicity of design of
NoSQL databases allows the addition of complex features to
Medici (such as community-generated dataset metadata)
without too much difficulty.
E. Job broker
If the web server is the heart of Medici 2, the job broker is
its veins. The role of the RabbitMQ message broker is to take
preprocessing messages fromthe web server that are sent once
a dataset or file is uploaded and distribute them to the
extractors that can handle the jobs.
Each extractor, once activated registers one or more
delivery queues on RabbitMQ, on which it listens. Each queue
is associated with a particular routing key set, which defines
which routing keys a job can have in order for it to be routed to
that queue (and thus to the extractor listening to the queue).
Each routing key set is defined by a four-field termlike the
following:
localhost.file.model.obj.#
In the above, the first field defines the URL or IP of the
machine on which the Medici web server implementation
sending job messages for the extractors is based. The second
field defines whether the extractor takes jobs pertaining to
preprocessing a file or a whole dataset. The third and
fourth field are for the MIME types of the files (or the types of
datasets) that the extractor accepts, and the fifth the possible
subtype ofthe datasets orfiles. The syntaxof the set definitions
uses the conventions specified in the RabbitMQ tutorials [9].
III. METADATA
Medici can accept, persist and process two kinds of
metadata for each dataset or file:
Automatically-generated. Generated by metadata
extractors and associated with the dataset or file
through a REST API call to the web server by the
extractors. They usually pertain to the more technical
aspects of datasets and files. Their subjects, predicates
and objects can vary from dataset orfile type to dataset
or file type and from Medici implementation to Medici
implementation depending on what metadata are
available for each dataset type and dataset, as well as
the type of the extractor which extracted the metadata.
They cannot be modified by users.
Community-generated. Generated by the users
through completion of an HTML formfor each dataset.
Their allowed subjects, predicates and objects are
defined on an institutional schema that can differ for
each Medici deployment. The schema is formulated by
each deployer or institution at its discretion, however it
is recommended for the schema to be compatible with
the Dublin Core content description model [15] to
facilitate metadata sharing between different
repositories. The CIDOC Conceptual Reference Model
is based on Dublin Core and is specialized for CH
metadata and can also be used to facilitate metadata
sharing among the CH scientific community [16].
The input metadata, either coming from an extractor or an
HTML form, are turned to JSON structures by the extractor or
a browser-side script respectively and sent to the web server,
where they are parsed back.
Metadata is understood by the web server and the
JavaScript script manipulating them on the browser-side (for
community-generated metadata) as a tree structure (for
community-generated metadata) or a list of tree structures (for
auto-generated metadata). This distinction applies because
different extractors may generate different sets of metadata
with no contact points between them for the same dataset or
file and so the reasonable option here is to present them as a
list. On the other hand, community-generated metadata are
generated by users by modifying the whole metadata structure
of a dataset using an HTML form at the same time, meaning
that they can be merged.
Metadata are formulated as tree structures because each
predicate of a metadata subject is either a property of the file or
dataset or a subproperty of another property, all describing the
same file or dataset. An example of a simple metadata tree
(here for community-generated metadata) is presented in Fig.
2. The tree structures (one for each dataset) are stored on
5. the web server’s application layer as Scala maps.
A. Automatically-generated metadata
There are currently metadata extractors for images and
RTIs, with more to be added in the future. For a discussion of
the metadata extractors for each file type, see the sections
discussing Medici’s features for their respective types.
B. Community-generated metadata
The schema for the community-generated metadata is
defined in two comma-separated value files.
The first file stores the name of each metadata node
along with whether the node is a leaf in the metadata
tree or a higher-level node. Leaves consist of string
data which are input by users and serve as the objects
of the description of the dataset, answering the
questions generated by the tree paths leading to them.
Branches are subproperties of higher-level properties
or top-level properties. In the example of Fig. 2, the
leaf node ―Title: Ancient Mari‖ is a leaf answering the
question ―what is a/the Name of a/the Project Location
of a/the Project associated with this dataset?‖.
The second file stores the possible relationships
associating nodes with nodes or leaves, along with the
cardinalities of each relationship.
The above are automatically enforced by the systemwhen
the user adds or modifies a dataset’s community-generated
metadata.
C. Metadata search
Users can search for datasets that contain user-entered
metadata that satisfy formulated queries.
The queries are formulated by filling out an HTML form
similar to the HTML form for adding/ modifying community-
generated metadata. For each level of the metadata tree, the
user can select which properties must be satisfied by selected
nodes at that level. For example, search for datasets having a
Project having Cyprus as a Country and having a Project
Location, with the Project Location having ―Choirokitia‖ as a
Name, and the Project also having Responsible Institution
having a Name of ―Example Institution‖, as in the query
shown in Fig. 3.
Users will be able to use a logical expression-like query
formulation to generate queries. For example, he/she may wish
to search for datasets satisfying the query above but having
Cyprus as a Country OR having a Project Location for which
the rest apply, or having Cyprus as a Country and NOT having
any Project Locations for which the rest apply.
The JSON-formulated query is sent to the web server,
which then parses it as a Scala map, compares it to the
community-generated metadata maps of each existing dataset,
and returns the list of datasets satisfying the query. The
comparison is done using a Depth-First-Search-esque
algorithm [17] that recursively checks for all paths in the query
tree where there is a path in the dataset’s community-generated
metadata tree that satisfies the path. The server formats the
result list as HTML and returns it to the client through AJAX
to be displayed.
IV. IMAGES
The support of multi-format and multi-resolution image
storage and presentation is highly beneficial to CH
communities due to the development of image acquisition
instruments, that satisfy current demands of research analysis
in archaeology. Medici currently supports the JPEG/JPG, PNG,
TIFF, GIF and BMP image formats. Regarding low-resolution
images, the image is displayed by directly embedding it in the
image previewer’s webpage. However, for high-resolution
images, a dedicated gigaimaging view extractor is used.
A. Deep Zoom
Viewing of large images efficiently and without loss of
quality is achieved by constructing an image pyramid using a
technique similar to frustum culling. Specifically, when a large
image is to be viewed in a limited screen space, only a part of it
can be viewed in full quality and the rest of the information in
the image that have no interest for the user is essentially
occluded and thus it is not necessary for it to be transmitted.
Consequently, data transmission size is limited by tiling the
Fig. 2. A simple metadata tree for community-generated metadata.
Fig. 3. Example community-generated metadata search.
6. image at various scales in a preprocessing step executed
server-side so that only the tiles needed for the currently shown
scale and position are transmitted. Deep Zoom, a technology
developed by Microsoft as part of Microsoft Silverlight [18],
works exactly like that- with the user initially seeing a lower-
resolution view of the entire image which they can then zoom
in to see portions of the image in more detail.
B. Zoom.it
Zoomit.js1
is a JavaScript library also used by the Zoom.it
web service [5]. It relies on the technology behind the
Seadragon gigaimaging viewer, which utilizes Deep Zoom.
This library is used by Medici 2 to view large images that have
already been split up into various levels of Deep Zoom tiles
(i.e. the image pyramid) by a dedicated image pyramid
extractor. The pyramids are loaded on the web server together
with an XML file (DZI) containing basic image metadata
needed by zoom.it.
C. Gigaimage processing
When an image is uploaded on the web server, a message is
sent to RabbitMQ for any extractors that can process images. If
there are any active gigaimaging previewers registered with
RabbitMQ, a gigaimaging previewer will receive the job
message for the image.
The job message will contain the image’s size. If the
image’s size is below a threshold input by whoever is
responsible for calling/ activating the extractor program as an
input parameter, the job is rejected and no XML file for
zoom.it or the image pyramid tiles are generated. This means
that when a user requests to view the dataset containing the
image, no zoom.it preview will be found by the web server and
therefore the original image will be shown, embedded with an
HTML tag.
If the size of the image exceeds the threshold, first the
zoom.it XML file is generated and uploaded to the web server
as a preview, it being given an identifier. That identifier will be
used to associate the image tiles with the preview and thus with
the original image. Then the image tiles for all the image zoom
levels are generated, concurrently uploaded back to the web
server and associated with the preview.
Zoomit.js accesses the image’s pyramid tiles via a virtual
directory structure generated on the web server for each
preview. This structure is implemented with Scala functions
that take the parts of the virtual directory path as parameters
and return the requested tile after querying the database where
the tiles are stored. These parameters include the tile’s level
and 2D-coordinates of its position on the level.
An example of a zoom.it preview can be seen in Fig. 4 and
Fig. 5, where it is used for visualizing ancient cylinder seals
and byzantine wall paintings fromCyprus.
D. Image technical metadata
There are currently two different image metadata
extractors, any of which can be used at the discretion of the
deploying institution, each extracting different sets of
1
http://zoom.it/h.js
Fig. 4. A two-dimensional giga-image of a cylinder seal dating to the 13th
century BC taken with a large format camera system and previewed in
Medici’s 2.0 web viewer. Courtesy of the Bank of Cyprus Cultural
Foundation.
metadata. The first uses the ImageMagick image processing
suite [19] and the second uses the Java-basedimage metadata
extractor library developed by Drew Noakes [20]. The
ImageMagick extractor extracts many more details, even
though it needs ImageMagick to be installed on the machine
running the ImageMagick extractor.
V. 3D MODELS
Complete digital documentation in archaeological and
Cultural Heritage research is a multidimensional process. New
opportunities and challenges for the development of web-based
Virtual Reality (VR) applications in these research fields have
been the direct consequence of advances in the field of three-
dimensional representation and Internet-related technology
[21]. High accuracy and multi level of detail (LOD) models
can be obtained by using various methods for 3D model
creation such as laser scanning technology, photogrammetry
and 3D
7. Fig. 5. Giga-image of byzantinewall paintings takeninside the monastery of
St. John Lambadistis in Cyprus andpreviewedin the webviewer. Courtesy of
the Department of Antiquities in Cyprus.
modeling (architectural reconstruction techniques) [22].
Nonetheless, the fusion of these techniques during post-
processing may occur, thus leading to the creation of highly
complex models with a diversity of information encapsulated
within a single file. The development of a web-based
application for user access and interactive exploration of three-
dimensional models has been studied and worked upon since
1995. The WEB 3D consortium composes open standards for
real time 3D data and models exchange over the Internet.
VRML became the first web 3D format [23]. The latest
successor of VRML is X3D, providing a flexible solution for
real time 3D representation and communication for Medici.
The system architecture of the 3D processing pipeline
consists ofopen source tools and free software, thus providing
full transparency on adopted methodology and data processing
methods providing a cost effective solution both for server and
client. The main feature of this web VR system previewer is to
provide the user with a completely new visit experience based
on a free interactive exploration interface of the object (i.e., not
constrained by any predefined pathway) and on the opportunity
to get more detailed information on specific parts of interest.
Medici currently supports 3D file formats that are widely
used in CH such as Wavefront (obj), Polygon File Format
(ply), X3D (x3d) format and 3D models embedded in PDF
(3DPDF). Models in the obj or ply format are converted to x3d
and then translated to their center of rotation. ―Virtual scenes‖
containing multiple objects can also be previewed in 3D fly-
through and walk-through mode (primary mode is ―examine‖
mode). If the model needs optimization for smooth
transmission over the internet, it is decimated in an adequate
LOD, before being converted to the X3DOM format [6] (that
is, HTML5 files) and sent back to the server as a preview to be
added to the DOM of the 3D previewer and displayed upon
request by the user.
There are also previewers for 3DPDF (also used for general
PDF viewing). One of them was written by NCSA and uses
Mozilla’s PDF.js library [24], while the other is a simple
integration of Adobe Acrobat using the associated Acrobat
browser plugin.
A. X3D extraction
The X3D extractor is responsible for converting obj and ply
models to X3D and decimating them if needed through
external calls to MeshLab’s Meshlabserver [25] and setting
job identity flags on each of the models uploaded by the client.
For the X3D extractor to work, MeshLab must be installed
on the extractor’s machine. The extractor accepts jobs from
RabbitMQ for obj, ply and X3D. However, if the 3D model is
not in X3D form, MeshLab’s meshlabserver command-line
interface is called by the extractor to generate the equivalent
X3D model.
On the other hand, if the model’s file size is above a bound
set by the client for calling/activating the extractor programas
an input parameter, a dynamically-generated meshlabserver re-
meshing/simplification script is used. A new call to decimate
the model to the adequate LOD is now generated according to
the client’s preset boundaries in order to reduce its complexity
and thus its size.
Finally, a flag is set on the job received by the extractor
indicating that the first phase of the processing of the model
was completed and the post-processed model is uploaded back
to the web server as an intermediate result together with the
model’s new job flag. The web server then transmits the
intermediate result to the HTML5-X3DOM extractor for
preparation for front-end display.
B. HTML5-X3DOM extraction
Converting X3D models to X3DOM/HTML5 facilitates
and optimizes their integration with the X3D previewer’s
HTML DOM structure. This is done by the X3DOM extractor,
which uses aopt, a command-line tool bundled with the
InstantReality framework [26].
For the HTML5-X3DOM extractor to work, InstantReality
must be installed on the extractor’s machine.
8. C. Models with separate materials/texture files
In special cases and in order to have obj and X3D files
display their geometry with the adequate color information,
models at times have separate material/texture files. Medici
2.0 considers this case. Those files can be still previewed
correctly if they are uploaded contained in a ZIP file with their
corresponding materials/textures. The web server uses a utility
function that unzips the file and uses rules regarding the
existence of files of certain formats in the zipped dataset to
resolve the type of model contained in the ZIP file (whether it
is a zipped obj, a zipped X3D or something different). The type
discovered is sent as part of the routing key to RabbitMQ when
the file’s extraction job is sent.
After the automated post-processing of the model with its
textures from the X3D extractor, the HTML5-X3DOM
extractor completes work on the X3D preview and sends back
the HTML5 X3DOM as the file’s final preview.
D. X3DOM previewer
The X3DOM previewer works by dynamically
downloading the HTML5 file in which the X3D preview of the
model is encapsulated and embedding it into an X3D HTML5
element. Afterwards, events thrown by the new elements
activate functions in the X3DOM JavaScript library,
initializing the model’s preview through an X3D scene.
The preview allows (among others) rotating the model,
zooming, panning, displaying model statistics, changing
rendering geometry mode from default view to wireframe or
vertex view, with or without texture and surface depth map
simulation.
X3dom.js2
accesses any image textures via a virtual
directory structure generated on the web server for each
preview, which for files uploaded as ZIP files simulates the
original ZIP file’s internal directory structure. This structure is
implemented with Scala functions that take the parts of the
virtual directory path as parameters and return the requested
image texture after querying the database where the image
textures are stored. These parameters include the (virtual) path
to the image texture, as it was in the original ZIP file
containing the textures of the original model.
3D model preview examples (with the depth maps and
model statistics also being displayed) are shown in Fig. 6, 7.
VI. RTI
The application of Reflectance Transformation Imaging
Technology (RTI) has offered great possibilities for research as
well as for the documentation and preservation of cultural
heritage objects and works of art. Polynomial Texture Map
(PTM) is a subset of the RTI method and was developed by
Hewlett-Packard Imaging Labs at the beginning of the past
decade [27][28] and enhanced by the West Semitic Project at
University of Southern California (USC) and Cultural Heritage
Imaging (CHI) [4]. PTM addresses the challenges in the
photography of objects’ faded, damaged or badly preserved
surfaces especially when they feature inscription, decorative
2
http://www.x3dom.org/download/dev/x3dom.js
Fig. 6. Hellenistic-Romantheater,300BC-365AD, previewedin the web 3D
viewer. Courtesy Cyprus Dept. of Antiquities-University of Sydney.
Fig. 7. Ancient vessel fromthe Pyrgos area of Cyprus, previewed in the web
viewer. Courtesy of the 3D-coform EU project.
9. Fig. 8. An ancient replicacoin that shows the capabilities of RTI photography.
Self-shadows andinterreflections are such derivatives presentedon the upper-
left cornerin comparison withthe default mode onthe lower-left side that has
no such properties.
Fig. 9. This snapshot is taken from the Inscriptifact desktop RTI viewer
developedby West Semitic Project at USC where the user is able to interact
with the artifact usinga ―virtual torch‖fordynamic illumination and surface
enhancement.
patterns and designs. They surfaces of an array of
archaeological objects and works of art such as stone or clay,
marble, plaques, coins, paintings, mosaics, sculpture, jewellery
and other small objects (see Fig. 8), present optimal study cases
for the utilization of RTI technology.
A. RTI/PTM viewer
The PTM algorithm synthesizes the data from images taken
under varying lighting directions to create a single image that
can be examined on a RTI/PTM viewer with a ―virtual torch‖.
The viewer allows the user to move the light angle intuitively
in real time, so that the combination of light and shadow
representing the relief features of the object’s surface can be
freely altered. RTI also permits the enhancement of the
subject’s surface shape, color and luminance attributes, which
allows one to extract detail out of the surface that cannot be
otherwise derived (see Fig. 9) [1].
Medici currently supports viewing real-time web RTI
through a Java applet developed by Clifford Lyon [29] (see
Fig. 10). There are three stand-alone desktop viewers that can
interpret RTI/PTM files but for the web just the viewer
mentioned above [30][31][32] is used. The current online RTI
viewer has pitfalls concerning incompatibility with current
W3C standards.
In the future, the current viewer will be replaced by
one based on the SpiderGL library [8], which is based on
WebGL. This will allow direct execution of the viewer by the
user’s browser without the need of using a third-party plugin,
improving cohesion between the user’s Graphics Processing
Unit, the viewer and the user’s browser, and also improving
security.
Medici 2.0 will currently avoid the use of this viewer, but
instead will use an innovative method for extracting 3D models
from RTI files for front-end previewing. Based on this the user
can recognize the object and then (down)load the original high-
resolution RTI file to his/her desktop for actual interaction.
B. RTI metadata
In the same manner as traditional web images, RTI
technical metadata is extracted by a standalone preprocessor.
The preprocessor downloads the RTI file from the web server
and then reads its topmost lines for the types of metadata
included in the most common RTI file formats as defined in the
PTM file format specification by Hewlett-Packard [28]. The
metadata is parsed to JSON and returned to the web server to
be associated with the RTI graphic.
Fig. 10. Replica of anancient coinfrom Petra, Jordan, viewed using the
current PTM viewer.
10. VII. EXTRACTING 3D MODELS FROM RTI
The ability to interactively change light direction and apply
various filters makes RTI one of the best techniques for
examining archaeological artifacts. Human perception of
highlights and shadows in a 2D image of a surface helps the
viewer interpret the surface topology. However, this ability to
perceive surface topology varies from person to person. Also,
some surfaces with varying contrast and surface characteristics
might be difficult to perceive even with dynamic lighting.
Further studies on that issue can be found in [33].
RTIs capture both color and geometric properties of the
object which makes it possible to use them to reconstruct 3D
approximations of the original surfaces. Doing this greatly
improves the perception of the artifacts’ surface details. The
approach used is based on an approach similar to that of
photometric stereo [34] and uses three 2D texture maps that are
extracted from the RTI to then reconstruct the final 3D surface
given known lighting directions. Luminance info is used to
estimate a normal map which is integrated to obtain a height
map. The height map is used to build a 3D surface by shifting
vertices of a Delaunay mesh. Color info is used to extract a
uniformly lit diffuse map that is used for surface texture
mapping.
A. Maps used
a. Diffuse Map
Diffuse maps (a.k.a. albedo maps) define the main color of
the surface. A good diffuse map should contain only color
information without any directional light effects, inter-
reflections, specular highlights or self-shadowing. Having any
of above in a diffuse map means that the object will respond in
an incorrect way to virtual incident light; such as casting
shadows in the wrong directions or showing highlights where
no direct incident light hits the object.
PTM allows the extraction of accurate diffuse maps
because the luminance and chromaticity information are stored
separately. In LRGB PTM format the color information exists
out of the box. For RGB PTM format, similar results can be
obtained by casting light perpendicular to surface per texel to
make sure all texels get the same amount of light. This results
in a uniformly lit texture that is ideal for use as a diffuse map.
b. Normal Map
A Normal Map is a texture map containing surface normals
at each texel stored as RGB. Directional Lighting information
for each pixel is already stored in a PTM, which makes it
possible to get a very good estimate of the surface normal at
that pixel. Since Luminance is maximum when incident light is
perpendicular to surface (i.e. incident light vector = surface
normal), the surface normal at each texel can be estimated as
the orientation maximizing light response.
c. Height Map
Height maps are grayscale texture maps in which each
pixel’s white level corresponds to the height value of a vertice
of a 3D grid mesh (usually in the Z direction). Height maps
used here are generated by integrating normal maps obtained as
described in the previous section. Normals at every surface
point are perpendicular to the height map gradient. Integration
is not always guaranteed to yield precise results since it is
based on an estimated normal map. Also, information about
surface discontinuities is lost in normal maps. The 3D models
generated are good approximations of the real surface.
Height map generation is implemented in an iterative
manner in which each iteration improves contrast between low
and high points. Height map pixels are initiated to zero height.
On each new iteration, each pixel’s height is slightly modified
according to the slopes of the surrounding pixels’ normals and
their heights in the current iteration. Contributions from all
eight surrounding pixels will be averaged and added to the
current height. Fig. 11 shows a summary of surrounding pixels
contributions and associated signs, where Nxand Ny are the X
and Y components, respectively, of the surface normal.
-Nx
+Ny
+Ny
+Nx
+Ny
-Nx
Current
Pixel
+Nx
-Nx
-Ny
-Ny
+Nx
-Ny
Fig. 11. Contributions of surroundingpixels tothe current pixel height when
generating the height map from the normal map.
B. Testing of algorithm
In testing done by BA, a 2D Delaunay triangulation was
used to generate rectangular grids that were deformed using
resulting height maps and texture-mapped using diffuse and
normal maps. The quality of the generated models improved
proportionally with number of iterations used. Iterations
beyond 100 000 iterations had no noticeable effect.
Further testing of the algorithm took place at CyI. The
testing showed that for low and mezzo-relief RTI the algorithm
generated 3D models of good quality with no more than 10 000
iterations needed.
An example of testing the algorithmis shown in Fig. 12.
C. Implementation
The algorithm will be implemented as a standalone
command-line program that will take as input the desired
number of height map generation iterations, the desired height
modifier for the generated 3D model and the RTI file and
output the generated 3D model. This program will be called
using a standard extractor taking the above input parameters as
11. Fig. 12. Example ofa PTMfile andits 3D model generatedby the RTI-to-3D
algorithm.
its input. The result of the extraction will be sent to the 3D
extractors for further processing, as it is for all 3D models.
VIII. FUTURE DIRECTIONS
A. Immediate plans
The community-generated metadata search will be
enhanced. The user will be able to use logical expression-like
query formulation to generate queries with AND, OR, NOT
operators. Technical metadata will also be searchable.
Integration of more MeshLab features in Medici will be
attempted (if supported by X3DOM). These may include
shading modes, a measuring caliber, community-generated
annotations on the model display and being able to change the
view’s light source.
Support for more file formats and model types will be
added, including printer-ready stereolithography (STL)
diagrams, Virtual Reality spherical panoramas and videos.
B. Further designs
Medici metadata will be able to be integrated with
Inscriptifact [1] and other Cultural Heritage repositories.
Concerning 2D high-resolution imaging, the IIPImage
framework [35] will serve as a mediator for flexibility on the
presentation of more specialized CH imaging (e.g. able to
support multispectral imaging and real time annotation).
Regarding 3D, technical metadata extractors may be added
and also the capability to generate 3D models from image
datasets produced by photogrammetry similar to ARC 3D web
service [36]. Furthermore, the use of Medici’s interactive
system could therefore be potentially extended to more
complex virtual exploration such as a digitized archaeological
site, to serve not just as a previewer but as an intuitive highly
VR environment for web. An interactive learning environment,
a “virtual world” [37].
A SpiderGL-based one will replace the current java-based
RTI viewer once SpiderGL is developed and made stable by
the digital CH communities.
Medici will also be able to extract semantically-meaningful
features for content-based comparison (e.g. textual data from
handwritten text). This will be made possible by constructing
descriptors of each file according to derivations of
semantically-meaningful features from a file’s
data[38][39][40][41]. The generated descriptors for each file
can be compared with descriptors of the same type generated
from search query data.
IX. CONCLUSION
Though still in development, Medici 2 already supports a
broad range of throughput-intensive research techniques and
community data management. Many file and dataset formats
can be uploaded, analyzed and visualized. These include the
latest formats used by the CH scientific community, i.e. large,
detailed 3D models and RTI, as well as the ever-present large
images and audio/video. Not only can Medici 2 extract
technical metadata from files and datasets, it also allows
searching of metadata and social metadata generation
according to each implementer’s user-input metadata schema.
The above, together with the many important additions
scheduled for the future, make it clear that Medici 2 will be a
CMS more than worth considering for satisfying the CH
scientific community’s dataset management, analysis and
visualization needs.
ACKNOWLEDGMENTS
The authors thank the Cyprus Institute, the LinkSCEEM-2
project and its partners, the National Centre for
Supercomputing Applications (University of Illinois) (NCSA)
and Bibliotheca Alexandrina as organizations for providing
them the means and guidelines enabling their contributions to
the development of Medici. Moreover, they thank every
member of staff in CyI, NCSA and Bibliotheca Alexandrina
who provided them with user requirements and technical
assistance.
SOURCE CODE
The current Medici source code is available from the
NCSA Medici repository
https://opensource.ncsa.illinois.edu/stash/projects/MED
12. REFERENCES
[1] L. Hunt, M. Lundberg, and B. Zuckerman, ―InscriptiFact: A
virtual archive of ancient inscriptions from the Near East‖,
International Journal on Digital Libraries, vol. 5, no. 3, pp.
153-166, May 2005.
[2] L. Marini et al, ―Medici: A Scalable Multimedia Environment
for Research‖, white paper, presented at Microsoft eScience
Workshop 2010, Berkeley, CA, Oct. 2010.
[3] The International Committee for Documentation of Cultural
Heritage (CIPA). http://cipa.icomos.org
[4] M. Mudge et al, “Image-Based Empirical Information
Acquisition, Scientific Reliability, and Long-Term Digital
Preservation for the Natural Sciences and Cultural Heritage”, in
Eurographics 2008, Hersonissos, Greece, 2008.
[5] Zoom.it – About. http://zoom.it/pages/about/
[6] J. Behr, P. Eschler, Y. Jung, and M. Zöllner, ―X3DOM:aDOM-
based HTML5/X3D integration model‖, in Proc.of the 14th Int.
Conf. on 3D Web Technology, Darmstadt, Germany, Jun. 2009,
pp. 129-135.
[7] S. Tomar, ―Converting video formats with FFmpeg‖, Linux J.,
vol.2006, no.146, pp. 10, Jun. 2006.
[8] M. Di Benedetto, F. Ponchio, F. Ganovelli, and R. Scopigno,
―SpiderGL: a JavaScript 3D graphics library for next-generation
WWW‖, in Proc, of the 15th Int. Conf. on Web 3D Technology,
Los Angeles, CA, Jul. 2010, pp. 165-174.
[9] A. Videla and J. J. W. Williams, ―RabbitMQ in action:
distributed messaging for everyone‖, Shelter Island, NY:
Manning, 2012.
[10] R. T. Fielding, ―Architectural Styles and the Design of Network-
based Software Architectures‖, Ph.D. dissertation, Dept.
Comput. Sci., Univ. California Irvine, Irvine, CA, 2000.
[11] P. Hilton, E. Bakker, and F. Canedo, ―Play for Scala (Early
Access Edition)‖, Manning, 2012.
[12] N. Leroux and S. De Kaper, ―Play for Java (Early Access
Edition)‖, Manning, 2012.
[13] A. Leff and J. Watson, ―Web-application development using the
Model/View/Controller design pattern‖, in Proc. of the
Enterprise Distributed Object Comp. Conf. 2001, Seattle, WA,
Sep. 2001, pp. 118-127.
[14] JQuery. http://jquery.com/
[15] S. Weibel, ―The Dublin Core: A Simple Content Description
Model for Electronic Resources‖, Bulletin of the Amer. Soc. for
Inform. Sci. and Technology, vol.24, no.1, pp. 9-11, Oct./Nov.
1997.
[16] M. Doerr, C. E. Ore, and S. Stead, ―The CIDOC conceptual
reference model: a new standard for knowledge sharing‖, in
Tutorials, posters, panels and industrial contributions at the
26th int. conf. on Conceptual modeling, Auckland, New
Zealand, Nov. 2007, pp. 51-56.
[17] R. Tarjan, ―Depth-First Search and Linear Graph Algorithms‖,
SIAM J. Comput.., vol.1, no.2, pp.146-160, Jun. 1972.
[18] Deep Zoom-Features-Microsoft Silverlight.
http://www.microsoft.com/silverlight/deep-zoom/
[19] M. Still, ―The Definitive Guide to ImageMagick‖, New York
City, NY: Apress, 2005.
[20] Drewnoakes.com - jpeg exif / iptc metadata extraction in java.
http://drewnoakes.com/code/exif/
[21] A. Guarnieri, F. Pirotti, and A. Vettore, ―Cultural heritage
interactive 3D models on the Web: An approach using open
source and free software‖, J. of Cultural Heritage, vol. 11, no. 3,
pp. 350-353, Jul.-Sept. 2010.
[22] F. Bernardini and H. Rushmeier, ―The 3D Model Acquisition
Pipeline‖, Comput. Graph. Forum, vol. 21, no. 2, pp. 149-172,
Jun. 2002.
[23] Web3D Consortium: Open Standards for Real-Time 3D
Communication. http://www.web3d.org/realtime-3d/
[24] pdf.js by andreasgal. http://mozilla.github.io/pdf.js/
[25] P. Cignoni et al., ―MeshLab: an Open-Source Mesh Processing
Tool‖, presented at Eurographics Italian Chapter Conf. 2008,
Salerno, Italy, Jul. 2008, pp. 129-136.
[26] J. Behr, U. Bockholt, and D. Fellner, ―Instantreality — A
Framework for Industrial Augmented and Virtual Reality
Applications‖, in Virtual Reality & Augmented Reality in
Industry. Berlin, Germany: Heidelberg, 2011, ch. 5, pp. 91-99.
[27] T. Malzbender, D. Gelb, and H. Wolters, ―Polynomial texture
maps‖, in SIGGRAPH ’01 Proc. of the 28th Annu. Conf. on
Comput.. graph.. and interactive techniques, Los Angeles, CA,
Aug. 2001, pp. 519-528.
[28] T. Malzbender, D. Gelb, and H. Wolters, ―Polynomial Texture
Map (.ptm) FileFormat‖, Client and Media Systems Laboratory,
HP Laboratories, Palo Alto, CA, Tech. Rep. HPL-2001-104,
Apr. 2001.
[29] PTM Web Viewer by Clifford Lyon.
http://materialobjects.com/ptm/
[30] Hewlett-Packard labs PTM Desktop viewer.
http://www.hpl.hp.com/research/ptm/downloads/download.html
[31] Cultural Heritage Imaging (CHI) RTI Desktop viewer.
http://culturalheritageimaging.org/Technologies/RTI/
[32] Inscriptifact Desktop Viewer.
http://www.inscriptifact.com/instructions/
[33] J. T. Todd, ―The visual perception of 3D shape.‖, in Trends in
cognitive sci., vol. 8, no. 3, pp. 115-121, Mar 2004.
[34] R. J. Woodham, ―Photometric method for determining surface
orientation from multiple images.‖, in Optical Eng., vol. 19, no.
1, pp. 139-144, Feb 1980.
[35] IIPImage framework.
http://iipimage.sourceforge.net/
[36] ARC 3D Web service.
http://homes.esat.kuleuven.be/~visit3d/webservice/v2/index.php
[37] B. Harper. J. G. Hedberg and R. Wright, ―Who benefits from
virtuality?‖, Comput. & Educ., vol. 34, no. 3/4, pp. 163-176,
Apr./May 2000.
[38] L. Marini et al, ―Versus: A Framework for General Content-
Based Comparisons‖, in IEEE eScience, Chicago, IL, 2012.
[39] L. Diesendruck et al, ―Digitization and Search: A Non-
Traditional Use of HPC‖, in IEEE eScience Workshop on
Extending High Performance Computing Beyond its Traditional
User Communities, Chicago, IL, 2012.
[40] L. Diesendruck et al, ―A Framework to Access Handwritten
Information within Large Digitized Paper Collections‖, in IEEE
eScience, Chicago, IL, 2012.
[41] L. Diesendruck et al, ―Using Lucene to Index and Search the
Digitized 1940 US Census‖, in Extreme Science and
Engineering Discovery Environment, San Diego, CA, 2013.