Today’s concept of Open Government Data (OGD) for openness, transparency and ease of
access of data owned by government agencies becomes increasingly important. This initiative emerges
from the demand of data usersforthe data belongs to the government agencies. The data services
providing an easy access, cheap, fast, and interoperability are needed by the users and becomes
important indicator performance for respective government agencies. Statistical Data and Metadata
Exchange (SDMX) is a new standard format in the data dissemination activities particularly in the
exchange of statistical data and metadata via Internet. In this respect SDMX support the implementation of
OGD project. This paper is on the technical design, development and implementation of data and
metadata exchange service of statistical data using SDMX format to support interoperability data through
web services. Three results are proposed: (i) framework for standardization of structure of statistical
publications data model with SDMX; (ii) design architecture of data sharing model; and (iii) web service
implementation of data and metadata exchange service using Service Oriented Analysis and Design
(SOAD) method. Implementation at Statistics Indonesia (BPS) is chosen as a case study to prove the
design concept. It is shown through quantitative assessment and black box testing that the design
achieves its objective.
XML Schema Design and Management for e-Government Data Interoperability Thomas Lee
One-stop public services and single window systems are primary goals of many e-government initiatives. How to facilitate the technical and data interoperability among the systems in different government agencies is a key of meeting these goals. While many software standards, such as Web Services and ebXML, have been formulated to address the interoperability between different technical platforms, the data interoperability problem remains to be a big challenge. The data interoperability concerns how different parties agree on what information to exchange, and the definition and representation of such information. To address this problem, the Hong Kong government has released the XML Schema Design and Management Guide as well as the Registry of Data Standards under its e-Government Interoperability Framework initiative. This paper introduces how the data modelling methodology provided by the Guide can be used to develop data interfaces
and standards for e-government systems. We also discuss how the Macao government has formulated their data interoperability policy and has applied the Guide in their situation.
A SERVICE ORIENTED DESIGN APPROACH FOR E-GOVERNANCE SYSTEMSijitcs
The document describes a service-oriented design approach for e-governance systems. It discusses key challenges in developing e-governance systems and proposes addressing these challenges through a service-oriented paradigm. The approach defines concepts like service types (readily available, composable, collaborative), service windows, service composition, and service collaboration. Service types depend on the complexity of processing required - readily available services require minimal processing, composable services may invoke other related services, and collaborative services require coordination across service windows. The approach aims to provide reusable, interoperable services and facilitate integration of existing applications in e-governance systems.
A Generic Model for Student Data Analytic Web Service (SDAWS)Editor IJCATR
Any university management system accumulates a cartload of data and analytics can be applied on it to gather useful
information to aid the academic decision making process. This paper is a novel attempt to demonstrate the significance of a data
analytic web service in the education domain. This can be integrated with the University Management System or any other application
of the university easily. Analytics as a web service offers much benefits over the traditional analysis methods. The web service can be
hosted on a web server and accessed over the internet or on to the private cloud of the campus. The data from various courses from
different departments can be uploaded and analyzed easily. In this paper we design a web service framework to be used in educational
data mining that provide analysis as a service.
This document provides an overview of service oriented architecture (SOA) and directions for SOA. It introduces a conceptual framework for understanding software integration as different layers, from the communication layer to the presentation layer. It then evaluates existing SOA realization approaches like WS-* specifications, ebXML, semantic web services, and RESTful services based on this framework. The document concludes by outlining future directions in SOA to further simplify the problem of integration.
The white paper proposes a concept called "Data Convergence" to provide a unified view of open datasets from different sources and formats. It describes a software application that identifies linked datasets, generates API endpoints to retrieve converged data in JSON/XML formats, and visualizes relationships between datasets. The goal is to make it easier for users to access and analyze related open datasets.
Marketplace affiliates potential analysis using cosine similarity and vision-...journalBEEI
One success factor of an online affiliate is determined by the quality of the content source. Therefore, affiliate marketplaces need to do an objective assessment to retrieve content data that will be used to choose the right product in the appropriate product filter. Usually, the selection is not made using a good and measured system so that the selection of product content is only based on parts that are not in accordance with what is seen or subjective. However, if analyzed using a good and measurable system will produce an objective product content and can have a positive impact on users because the selection is based on factual data. The purpose of this research is to analyze the potential of the affiliate marketplace by combining cosine similarity with vision-based page segmentation. This is a new breakthrough made for optimization to get the best content in accordance with the required criteria. This work will produce a number of product recommendations that are appropriate for publication and then made use of for comparison that matches the required criteria. At the limited evaluation stage, the performance of the proposed model obtained satisfactory results, in which 5 queries tested were all as expected.
eGovernment Services Composition Using Multi-Faceted Metadata Classification...Sotiris Koussouris
The document discusses a methodology for designing an e-government portal using a multi-faceted taxonomy. Key aspects of the methodology include identifying core government services, modeling the services using a Service Description Worksheet to extract metadata, and developing generalized service patterns to facilitate composition of e-government services. Populating the taxonomy and Service Description Database allows for analysis of services, development of service models, and identification of relationships that support service homogenization. Future work includes formalizing document and system exchanges and providing the taxonomy and query outputs in XML format.
This white paper discusses the importance of data quality for successful Master Data Management (MDM) systems. It defines MDM as consolidating all relevant company data from different systems into a single version of truth. High quality meta data and content data are critical for MDM success. The paper describes how data profiling can analyze meta data quality issues across sources. It also discusses challenges in keeping data quality high as MDM systems operate continuously with live data updates and entries. The paper proposes a Data Quality Life Cycle approach including identifying data sources, initial data cleansing, real-time data validation, and ongoing monitoring to help maintain high quality master data.
XML Schema Design and Management for e-Government Data Interoperability Thomas Lee
One-stop public services and single window systems are primary goals of many e-government initiatives. How to facilitate the technical and data interoperability among the systems in different government agencies is a key of meeting these goals. While many software standards, such as Web Services and ebXML, have been formulated to address the interoperability between different technical platforms, the data interoperability problem remains to be a big challenge. The data interoperability concerns how different parties agree on what information to exchange, and the definition and representation of such information. To address this problem, the Hong Kong government has released the XML Schema Design and Management Guide as well as the Registry of Data Standards under its e-Government Interoperability Framework initiative. This paper introduces how the data modelling methodology provided by the Guide can be used to develop data interfaces
and standards for e-government systems. We also discuss how the Macao government has formulated their data interoperability policy and has applied the Guide in their situation.
A SERVICE ORIENTED DESIGN APPROACH FOR E-GOVERNANCE SYSTEMSijitcs
The document describes a service-oriented design approach for e-governance systems. It discusses key challenges in developing e-governance systems and proposes addressing these challenges through a service-oriented paradigm. The approach defines concepts like service types (readily available, composable, collaborative), service windows, service composition, and service collaboration. Service types depend on the complexity of processing required - readily available services require minimal processing, composable services may invoke other related services, and collaborative services require coordination across service windows. The approach aims to provide reusable, interoperable services and facilitate integration of existing applications in e-governance systems.
A Generic Model for Student Data Analytic Web Service (SDAWS)Editor IJCATR
Any university management system accumulates a cartload of data and analytics can be applied on it to gather useful
information to aid the academic decision making process. This paper is a novel attempt to demonstrate the significance of a data
analytic web service in the education domain. This can be integrated with the University Management System or any other application
of the university easily. Analytics as a web service offers much benefits over the traditional analysis methods. The web service can be
hosted on a web server and accessed over the internet or on to the private cloud of the campus. The data from various courses from
different departments can be uploaded and analyzed easily. In this paper we design a web service framework to be used in educational
data mining that provide analysis as a service.
This document provides an overview of service oriented architecture (SOA) and directions for SOA. It introduces a conceptual framework for understanding software integration as different layers, from the communication layer to the presentation layer. It then evaluates existing SOA realization approaches like WS-* specifications, ebXML, semantic web services, and RESTful services based on this framework. The document concludes by outlining future directions in SOA to further simplify the problem of integration.
The white paper proposes a concept called "Data Convergence" to provide a unified view of open datasets from different sources and formats. It describes a software application that identifies linked datasets, generates API endpoints to retrieve converged data in JSON/XML formats, and visualizes relationships between datasets. The goal is to make it easier for users to access and analyze related open datasets.
Marketplace affiliates potential analysis using cosine similarity and vision-...journalBEEI
One success factor of an online affiliate is determined by the quality of the content source. Therefore, affiliate marketplaces need to do an objective assessment to retrieve content data that will be used to choose the right product in the appropriate product filter. Usually, the selection is not made using a good and measured system so that the selection of product content is only based on parts that are not in accordance with what is seen or subjective. However, if analyzed using a good and measurable system will produce an objective product content and can have a positive impact on users because the selection is based on factual data. The purpose of this research is to analyze the potential of the affiliate marketplace by combining cosine similarity with vision-based page segmentation. This is a new breakthrough made for optimization to get the best content in accordance with the required criteria. This work will produce a number of product recommendations that are appropriate for publication and then made use of for comparison that matches the required criteria. At the limited evaluation stage, the performance of the proposed model obtained satisfactory results, in which 5 queries tested were all as expected.
eGovernment Services Composition Using Multi-Faceted Metadata Classification...Sotiris Koussouris
The document discusses a methodology for designing an e-government portal using a multi-faceted taxonomy. Key aspects of the methodology include identifying core government services, modeling the services using a Service Description Worksheet to extract metadata, and developing generalized service patterns to facilitate composition of e-government services. Populating the taxonomy and Service Description Database allows for analysis of services, development of service models, and identification of relationships that support service homogenization. Future work includes formalizing document and system exchanges and providing the taxonomy and query outputs in XML format.
This white paper discusses the importance of data quality for successful Master Data Management (MDM) systems. It defines MDM as consolidating all relevant company data from different systems into a single version of truth. High quality meta data and content data are critical for MDM success. The paper describes how data profiling can analyze meta data quality issues across sources. It also discusses challenges in keeping data quality high as MDM systems operate continuously with live data updates and entries. The paper proposes a Data Quality Life Cycle approach including identifying data sources, initial data cleansing, real-time data validation, and ongoing monitoring to help maintain high quality master data.
Annotation Approach for Document with Recommendation ijmpict
An enormous number of organizations generate and share textual descriptions of their products, facilities, and activities. Such collections of textual data comprise a significant amount of controlled information, which residues buried in the unstructured text. Whereas information extraction systems simplify the extraction of structured associations, they are frequently expensive and incorrect, particularly when working on top of text that does not comprise any examples of the targeted structured data. Projected an alternative methodology that simplifies the structured metadata generation by recognizing documents that are possible to contain information of awareness and this data will be beneficial for querying the database. Moreover, we intend algorithms to extract attribute-value pairs, and similarly devise new mechanisms to map such pairs to manually created schemes. We apply clustering technique to the item content information to complement the user rating information, which improves the correctness of collaborative similarity, and solves the cold start problem.
Data migration involves moving data from legacy systems to a new consolidated system. It requires extracting, transforming, and loading data while addressing issues like duplicate or invalid data. CXAIR can help by allowing business users to explore source system data through a simple interface, understand data quality issues, and provide input to help IT construct accurate migration rules. CXAIR also enables consolidated views of related data across systems and retaining historical data to support validation and analysis during and after migration.
The document discusses different data models including hierarchical, network, relational, object-oriented, and object-relational models. It provides details on each model's structure and advantages and disadvantages. It also discusses using the relational model for a database to manage information for the Fly High Airlines, including passenger, payment, and seat information. The relational model is justified as the best fit due to its ability to efficiently query and join table data while ensuring data integrity.
A Review on Classification of Data Imbalance using BigDataIJMIT JOURNAL
Classification is one among the data mining function that assigns items in a collection to target categories or collection of data to provide more accurate predictions and analysis. Classification using supervised learning method aims to identify the category of the class to which a new data will fall under. With the advancement of technology and increase in the generation of real-time data from various sources like Internet, IoT and Social media it needs more processing and challenging. One such challenge in processing is data imbalance. In the imbalanced dataset, majority classes dominate over minority classes causing the machine learning classifiers to be more biased towards majority classes and also most classification algorithm predicts all the test data with majority classes. In this paper, the author analysis the data imbalance models using big data and classification algorithm.
This white paper proposes a concept called "Data Convergence" to provide a unified view of open government datasets from different sources and formats. The solution would build a software application with an HTTP API to integrate datasets and identify relationships between them based on common attributes. This would allow users to more easily analyze linked datasets and derive useful information. The benefits of this approach include easy access to real-time converged data through standard JSON/XML formats with loose coupling between the underlying data storage and applications.
Management Architecture for Dynamic Federated Identity Management csandit
We present the concept and design of Dynamic Automa
ted Metadata Exchange (DAME) in
Security Assertion Markup Language (SAML) based use
r authentication and authorization
infrastructures. This approach solves the real-worl
d limitations in scalability of pre-exchanged
metadata in SAML-based federations and inter-federa
tions. The user initiates the metadata
exchange on demand, therefore reducing the size of
the exchanged metadata compared to
traditional metadata aggregation. In order to speci
fy and discuss the necessary changes to
identity federation architectures, we apply the Mun
ich Network Management (MNM) service
model to Federated Identity Management via a truste
d third party (TTP); an overview of all
components and interactions is created. Based on th
is model, the management architecture of
the TTP with its basic management functionalities i
s designed. This management architecture
includes further functionality for automated manage
ment of entities and dynamic federations.
Challenges Faced by Novices While Developing and Designing the Visualization ...IRJET Journal
This document discusses challenges faced by novices in developing visualizations using business intelligence (BI) tools. It finds that novices struggle with three main activities: selecting appropriate data attributes based on the question, choosing a suitable visual template, and interpreting complex visualizations. Common problems include high visual complexity, unfamiliar visualization types, inappropriate scaling, and difficulties understanding measurement semantics. Novices also tend to incompletely or improperly specify visualization details. They prefer simple charts like bar, line and pie charts that they are familiar with. The development process faces barriers at each stage of selecting data, creating visual mappings, and interpretation.
An Improved Support Vector Machine Classifier Using AdaBoost and Genetic Algo...Eswar Publications
Predicting the objective of internet users has divergent applications in the areas such as e-commerce, entertainment in online, and several internet-based applications. The critical part of the classifying internet queries based on obtainable features namely contextual information, keywords and their semantic relationships. This research paper presents an improved support vector machine classifier that makes use of ad boost genetic algorithmic approach towards web interaction mining. Around 31 participants are chosen and given topics to
search web contents. Parameters such as precision, recall and F1 score are taken for comparing the proposed classifier with the classical support vector machine. Results proved that the proposed classifier achieves better performance than that of the conventional SVM.
A Survey on Heterogeneous Data Exchange using XmlIRJET Journal
This document summarizes a research paper on heterogeneous data exchange using XML. It discusses how XML has become a standard for data transmission due to its flexibility, extensibility and ability to represent heterogeneous data. The document then reviews related work on XML data exchange and mapping between relational and XML models. It also describes the process of exporting data from a source database to XML, importing XML data by validating, transforming and storing it in the target database, and transmitting data between different servers.
The semantic MDM system is an innovation in the area of consolidation of reference data, unification of services for their processing, consolidation of knowledge in semantic models and standardization of data exchange processes.
This document summarizes a research paper on using clustering approaches to improve the discovery of semantic web services. It begins by defining semantic web services and semantic similarity measures. It then discusses using clustering to eliminate irrelevant services from a collection before applying semantic algorithms. Specifically, it proposes a clustering probabilistic semantic approach (CPLSA) that filters services based on compatibility with a query before clustering the remaining services into semantically related groups using probabilistic latent semantic analysis (PLSA). The document concludes by discussing applications of approximate semantics and challenges in scaling semantic algorithms.
Time-Ordered Collaborative Filtering for News RecommendationIRJET Journal
This document presents a novel news recommendation system called Time-Ordered Collaborative Filtering for News Recommendation. The system filters and recommends news to users based on their choices and preferences using collaborative filtering and content-based filtering. It extracts news data from popular online newspapers and stores it in a database. The recommendation framework then preprocesses the news data and applies algorithms like calculating sparse matrices to rank news and recommend the top news to each user based on their interests.
EVALUATION OF COMPUTABILITY CRITERIONS FOR RUNTIME WEB SERVICE INTEGRATIONijwscjournal
T Today’s competitive environment drives the enterprises to extend their focus and collaborate with their business partners to carry out the necessities. Tight coordination among business partners assists to share and integrate the service logic globally. But integrating service logics across diverse enterprises leads to
exponential problem which stipulates developers to comprehend the whole service and must resolve suitable method to integrate the services. It is complex and time-consuming task. So the present focus is to have a mechanized system to analyze the Business logics and convey the proper mode to integrate them.
There is no standard model to undertake these issues and one such a framework proposed in this paper examines the Business logics individually and suggests proper structure to integrate them. One of the innovative concepts of proposed model is Property Evaluation System which scrutinizes the service logics and generates Business Logic Property Schema (BLPS) for the required services. BLPS holds necessary information to recognize the correct structure for integrating the service logics. At the time of integration, System consumes this BLPS schema and suggests the feasible ways to integrate the service logics. Also if the service logics are attempted to integrate in invalid structure or attempted to violate accessibility levels, system will throw exception with necessary information. This helps developers to ascertain the efficient structure to integrate the services with least effort.
Comparative Study on Graph-based Information Retrieval: the Case of XML DocumentIJAEMSJORNAL
The processing of massive amounts of data has become indispensable especially with the potential proliferation of big data. The volume of information available nowadays makes it difficult for the user to find relevant information in a vast collection of documents. As a result, the exploitation of vast document collections necessitates the implementation of automated technologies that enable appropriate and effective retrieval. In this paper, we will examine the state of the art of IR in XML documents. We will also discuss some works that have used graphs to represent documents in the context of IR. In the same vein, the relationships between the components of a graph are the center of our attention.
P REFIX - BASED L ABELING A NNOTATION FOR E FFECTIVE XML F RAGMENTATIONijcsit
XML is
gradually
emplo
yed as
a standard of data exchange
in
web
environment
since its inception
in the
90s
until
present
.
It
serves
as a data exchange between system
s
and other application
s
.
Meanwhile t
he data
volume has grown substantially
in the web and
thus effective methods
of
storing and retrieving
these
data
is
essential
.
One recommended way is
p
hysically or virtually
fragments
the large chunk of data
and
distributes
the fragments
into different nodes.
F
ragmentation design
of XML document
contains of two
parts: fragmentat
ion operation and fragmentation method. The
three
fragmentation o
peration
s are
Horizontal, Vertical
and Hybrid. It
determines how the XML should be fragmented.
This
paper
aims
to give
an overview on the fragmentation design consideration
and
subsequently,
propose a
fragmentation
technique
using
number addressing
.
2010 07 modeling web evolution amarantidis antoniou vafopoulos finalvafopoulos
This paper analyzes and simulates the KKPS model of web evolution to address open issues. Simulations show that as the number of endorsed documents (b) or topics (k) increases, the power law distribution of document in-degrees becomes less significant. They also found that increasing the number of endorsements (b) can decrease the efficiency of the search algorithm, contrary to expectations. The paper aims to clarify these issues and highlight future directions for more realistic web modeling.
Proceedings of the 2nd International REA Technology Workshop - 2006Demetrius_Gallitzin
Piet Erik Adolf Vandenbossche and J.C. Wortmann:
"Why accounting data models from research are not incorporated in ERP systems",
Mette Jaquet:
"A Property Driven Approach towards Describing Semantics of REA Entities",
Frederik Gailly and Geert Poels:
"Towards an Operational REA Ontology Using Web Ontology Languages",
Fred van Blommestein:
"REA as an e-business ontology",
Anders Hessellund:
"Modeling Issues in REA",
Pavel Hruby and Jesper Kiehn:
"Revised Classification of the Enterprise Information Architecture Elements"
IRJET- Achieving Data Truthfulness and Privacy Preservation in Data MarketsIRJET Journal
This document discusses achieving data truthfulness and privacy preservation in data markets. It proposes a system called TPDM that aims to ensure data honesty and privacy preservation simultaneously. TPDM allows data contributors to honestly submit their own data without impersonating others, and requires the service provider to honestly collect and process data. It also protects both the personally identifiable information and sensitive data of contributors. The document describes how TPDM could be implemented for profile matching and data distribution services, and evaluates its performance on real-world datasets. Ensuring both data honesty and privacy is challenging but important for the healthy development of data markets.
SERVICE ORIENTED QUALITY REQUIREMENT FRAMEWORK FOR CLOUD COMPUTINGijcsit
This research paper introduces a framework to identify the quality requirements of cloud computing services. It considered two dominant sub-layers; functional layer and runtime layer against cloud characteristics. SERVQUAL model attributes and the opinions of the industry experts were used to derive the quality constructs in cloud computing environment. The framework gives proper identification of cloud computing service quality expectations of users. The validity of the framework was evaluated by using
questionnaire based survey. Partial least squares-structural equation modelling (PLS-SEM) technique was
used to evaluate the outcome. The research findings shows that the significance of functional layer is
higher than runtime layer and prioritized quality factors of two layers are Service time, Information and
data security, Recoverability, Service Transparency, and Accessibility.
Comparing and analyzing various method of data integration in big dataIRJET Journal
This document discusses various methods for integrating data from multiple heterogeneous sources in big data systems. It begins by defining data integration and explaining the challenges posed by big data's volume, velocity, variety and veracity. The document then examines different techniques for big data integration including Extract-Transform-Load (ETL), Enterprise Information Integration (EII), data migration, data consolidation, data propagation, data federation and change data capture. It analyzes the advantages and limitations of each technique. In conclusion, the document states that data integration is a major challenge and there is no single best method, as the appropriate solution depends on the specific use case and data characteristics.
This document discusses big data challenges for e-governance systems in distributed systems. It proposes a map reducing architecture model to provide the basics for building interoperable big data infrastructure for e-governance. The model includes stages for data management, access control, and security. It explains how this can be implemented using distributed structures and a provisioning model. Key challenges addressed include the exponential growth of data from various sources, need for data sharing across organizations, and ensuring security, access control, and data integrity in distributed systems.
Semantic Data Integration Approaches for E-Governance dannyijwest
This document summarizes approaches for semantic data integration for e-governance. It discusses how semantic technologies using ontologies can help integrate heterogeneous data sources for improved e-government services. Current data integration approaches like data warehousing have limitations that semantic approaches can address. The document outlines challenges in current workflows and different approaches to semantic data management including using RDF, OWL and mapping relational databases to semantic models. It argues that semantic data integration using ontologies can provide a unified view of data to enable improved cross-agency information sharing and services for both governments and users.
Annotation Approach for Document with Recommendation ijmpict
An enormous number of organizations generate and share textual descriptions of their products, facilities, and activities. Such collections of textual data comprise a significant amount of controlled information, which residues buried in the unstructured text. Whereas information extraction systems simplify the extraction of structured associations, they are frequently expensive and incorrect, particularly when working on top of text that does not comprise any examples of the targeted structured data. Projected an alternative methodology that simplifies the structured metadata generation by recognizing documents that are possible to contain information of awareness and this data will be beneficial for querying the database. Moreover, we intend algorithms to extract attribute-value pairs, and similarly devise new mechanisms to map such pairs to manually created schemes. We apply clustering technique to the item content information to complement the user rating information, which improves the correctness of collaborative similarity, and solves the cold start problem.
Data migration involves moving data from legacy systems to a new consolidated system. It requires extracting, transforming, and loading data while addressing issues like duplicate or invalid data. CXAIR can help by allowing business users to explore source system data through a simple interface, understand data quality issues, and provide input to help IT construct accurate migration rules. CXAIR also enables consolidated views of related data across systems and retaining historical data to support validation and analysis during and after migration.
The document discusses different data models including hierarchical, network, relational, object-oriented, and object-relational models. It provides details on each model's structure and advantages and disadvantages. It also discusses using the relational model for a database to manage information for the Fly High Airlines, including passenger, payment, and seat information. The relational model is justified as the best fit due to its ability to efficiently query and join table data while ensuring data integrity.
A Review on Classification of Data Imbalance using BigDataIJMIT JOURNAL
Classification is one among the data mining function that assigns items in a collection to target categories or collection of data to provide more accurate predictions and analysis. Classification using supervised learning method aims to identify the category of the class to which a new data will fall under. With the advancement of technology and increase in the generation of real-time data from various sources like Internet, IoT and Social media it needs more processing and challenging. One such challenge in processing is data imbalance. In the imbalanced dataset, majority classes dominate over minority classes causing the machine learning classifiers to be more biased towards majority classes and also most classification algorithm predicts all the test data with majority classes. In this paper, the author analysis the data imbalance models using big data and classification algorithm.
This white paper proposes a concept called "Data Convergence" to provide a unified view of open government datasets from different sources and formats. The solution would build a software application with an HTTP API to integrate datasets and identify relationships between them based on common attributes. This would allow users to more easily analyze linked datasets and derive useful information. The benefits of this approach include easy access to real-time converged data through standard JSON/XML formats with loose coupling between the underlying data storage and applications.
Management Architecture for Dynamic Federated Identity Management csandit
We present the concept and design of Dynamic Automa
ted Metadata Exchange (DAME) in
Security Assertion Markup Language (SAML) based use
r authentication and authorization
infrastructures. This approach solves the real-worl
d limitations in scalability of pre-exchanged
metadata in SAML-based federations and inter-federa
tions. The user initiates the metadata
exchange on demand, therefore reducing the size of
the exchanged metadata compared to
traditional metadata aggregation. In order to speci
fy and discuss the necessary changes to
identity federation architectures, we apply the Mun
ich Network Management (MNM) service
model to Federated Identity Management via a truste
d third party (TTP); an overview of all
components and interactions is created. Based on th
is model, the management architecture of
the TTP with its basic management functionalities i
s designed. This management architecture
includes further functionality for automated manage
ment of entities and dynamic federations.
Challenges Faced by Novices While Developing and Designing the Visualization ...IRJET Journal
This document discusses challenges faced by novices in developing visualizations using business intelligence (BI) tools. It finds that novices struggle with three main activities: selecting appropriate data attributes based on the question, choosing a suitable visual template, and interpreting complex visualizations. Common problems include high visual complexity, unfamiliar visualization types, inappropriate scaling, and difficulties understanding measurement semantics. Novices also tend to incompletely or improperly specify visualization details. They prefer simple charts like bar, line and pie charts that they are familiar with. The development process faces barriers at each stage of selecting data, creating visual mappings, and interpretation.
An Improved Support Vector Machine Classifier Using AdaBoost and Genetic Algo...Eswar Publications
Predicting the objective of internet users has divergent applications in the areas such as e-commerce, entertainment in online, and several internet-based applications. The critical part of the classifying internet queries based on obtainable features namely contextual information, keywords and their semantic relationships. This research paper presents an improved support vector machine classifier that makes use of ad boost genetic algorithmic approach towards web interaction mining. Around 31 participants are chosen and given topics to
search web contents. Parameters such as precision, recall and F1 score are taken for comparing the proposed classifier with the classical support vector machine. Results proved that the proposed classifier achieves better performance than that of the conventional SVM.
A Survey on Heterogeneous Data Exchange using XmlIRJET Journal
This document summarizes a research paper on heterogeneous data exchange using XML. It discusses how XML has become a standard for data transmission due to its flexibility, extensibility and ability to represent heterogeneous data. The document then reviews related work on XML data exchange and mapping between relational and XML models. It also describes the process of exporting data from a source database to XML, importing XML data by validating, transforming and storing it in the target database, and transmitting data between different servers.
The semantic MDM system is an innovation in the area of consolidation of reference data, unification of services for their processing, consolidation of knowledge in semantic models and standardization of data exchange processes.
This document summarizes a research paper on using clustering approaches to improve the discovery of semantic web services. It begins by defining semantic web services and semantic similarity measures. It then discusses using clustering to eliminate irrelevant services from a collection before applying semantic algorithms. Specifically, it proposes a clustering probabilistic semantic approach (CPLSA) that filters services based on compatibility with a query before clustering the remaining services into semantically related groups using probabilistic latent semantic analysis (PLSA). The document concludes by discussing applications of approximate semantics and challenges in scaling semantic algorithms.
Time-Ordered Collaborative Filtering for News RecommendationIRJET Journal
This document presents a novel news recommendation system called Time-Ordered Collaborative Filtering for News Recommendation. The system filters and recommends news to users based on their choices and preferences using collaborative filtering and content-based filtering. It extracts news data from popular online newspapers and stores it in a database. The recommendation framework then preprocesses the news data and applies algorithms like calculating sparse matrices to rank news and recommend the top news to each user based on their interests.
EVALUATION OF COMPUTABILITY CRITERIONS FOR RUNTIME WEB SERVICE INTEGRATIONijwscjournal
T Today’s competitive environment drives the enterprises to extend their focus and collaborate with their business partners to carry out the necessities. Tight coordination among business partners assists to share and integrate the service logic globally. But integrating service logics across diverse enterprises leads to
exponential problem which stipulates developers to comprehend the whole service and must resolve suitable method to integrate the services. It is complex and time-consuming task. So the present focus is to have a mechanized system to analyze the Business logics and convey the proper mode to integrate them.
There is no standard model to undertake these issues and one such a framework proposed in this paper examines the Business logics individually and suggests proper structure to integrate them. One of the innovative concepts of proposed model is Property Evaluation System which scrutinizes the service logics and generates Business Logic Property Schema (BLPS) for the required services. BLPS holds necessary information to recognize the correct structure for integrating the service logics. At the time of integration, System consumes this BLPS schema and suggests the feasible ways to integrate the service logics. Also if the service logics are attempted to integrate in invalid structure or attempted to violate accessibility levels, system will throw exception with necessary information. This helps developers to ascertain the efficient structure to integrate the services with least effort.
Comparative Study on Graph-based Information Retrieval: the Case of XML DocumentIJAEMSJORNAL
The processing of massive amounts of data has become indispensable especially with the potential proliferation of big data. The volume of information available nowadays makes it difficult for the user to find relevant information in a vast collection of documents. As a result, the exploitation of vast document collections necessitates the implementation of automated technologies that enable appropriate and effective retrieval. In this paper, we will examine the state of the art of IR in XML documents. We will also discuss some works that have used graphs to represent documents in the context of IR. In the same vein, the relationships between the components of a graph are the center of our attention.
P REFIX - BASED L ABELING A NNOTATION FOR E FFECTIVE XML F RAGMENTATIONijcsit
XML is
gradually
emplo
yed as
a standard of data exchange
in
web
environment
since its inception
in the
90s
until
present
.
It
serves
as a data exchange between system
s
and other application
s
.
Meanwhile t
he data
volume has grown substantially
in the web and
thus effective methods
of
storing and retrieving
these
data
is
essential
.
One recommended way is
p
hysically or virtually
fragments
the large chunk of data
and
distributes
the fragments
into different nodes.
F
ragmentation design
of XML document
contains of two
parts: fragmentat
ion operation and fragmentation method. The
three
fragmentation o
peration
s are
Horizontal, Vertical
and Hybrid. It
determines how the XML should be fragmented.
This
paper
aims
to give
an overview on the fragmentation design consideration
and
subsequently,
propose a
fragmentation
technique
using
number addressing
.
2010 07 modeling web evolution amarantidis antoniou vafopoulos finalvafopoulos
This paper analyzes and simulates the KKPS model of web evolution to address open issues. Simulations show that as the number of endorsed documents (b) or topics (k) increases, the power law distribution of document in-degrees becomes less significant. They also found that increasing the number of endorsements (b) can decrease the efficiency of the search algorithm, contrary to expectations. The paper aims to clarify these issues and highlight future directions for more realistic web modeling.
Proceedings of the 2nd International REA Technology Workshop - 2006Demetrius_Gallitzin
Piet Erik Adolf Vandenbossche and J.C. Wortmann:
"Why accounting data models from research are not incorporated in ERP systems",
Mette Jaquet:
"A Property Driven Approach towards Describing Semantics of REA Entities",
Frederik Gailly and Geert Poels:
"Towards an Operational REA Ontology Using Web Ontology Languages",
Fred van Blommestein:
"REA as an e-business ontology",
Anders Hessellund:
"Modeling Issues in REA",
Pavel Hruby and Jesper Kiehn:
"Revised Classification of the Enterprise Information Architecture Elements"
IRJET- Achieving Data Truthfulness and Privacy Preservation in Data MarketsIRJET Journal
This document discusses achieving data truthfulness and privacy preservation in data markets. It proposes a system called TPDM that aims to ensure data honesty and privacy preservation simultaneously. TPDM allows data contributors to honestly submit their own data without impersonating others, and requires the service provider to honestly collect and process data. It also protects both the personally identifiable information and sensitive data of contributors. The document describes how TPDM could be implemented for profile matching and data distribution services, and evaluates its performance on real-world datasets. Ensuring both data honesty and privacy is challenging but important for the healthy development of data markets.
SERVICE ORIENTED QUALITY REQUIREMENT FRAMEWORK FOR CLOUD COMPUTINGijcsit
This research paper introduces a framework to identify the quality requirements of cloud computing services. It considered two dominant sub-layers; functional layer and runtime layer against cloud characteristics. SERVQUAL model attributes and the opinions of the industry experts were used to derive the quality constructs in cloud computing environment. The framework gives proper identification of cloud computing service quality expectations of users. The validity of the framework was evaluated by using
questionnaire based survey. Partial least squares-structural equation modelling (PLS-SEM) technique was
used to evaluate the outcome. The research findings shows that the significance of functional layer is
higher than runtime layer and prioritized quality factors of two layers are Service time, Information and
data security, Recoverability, Service Transparency, and Accessibility.
Comparing and analyzing various method of data integration in big dataIRJET Journal
This document discusses various methods for integrating data from multiple heterogeneous sources in big data systems. It begins by defining data integration and explaining the challenges posed by big data's volume, velocity, variety and veracity. The document then examines different techniques for big data integration including Extract-Transform-Load (ETL), Enterprise Information Integration (EII), data migration, data consolidation, data propagation, data federation and change data capture. It analyzes the advantages and limitations of each technique. In conclusion, the document states that data integration is a major challenge and there is no single best method, as the appropriate solution depends on the specific use case and data characteristics.
This document discusses big data challenges for e-governance systems in distributed systems. It proposes a map reducing architecture model to provide the basics for building interoperable big data infrastructure for e-governance. The model includes stages for data management, access control, and security. It explains how this can be implemented using distributed structures and a provisioning model. Key challenges addressed include the exponential growth of data from various sources, need for data sharing across organizations, and ensuring security, access control, and data integrity in distributed systems.
Semantic Data Integration Approaches for E-Governance dannyijwest
This document summarizes approaches for semantic data integration for e-governance. It discusses how semantic technologies using ontologies can help integrate heterogeneous data sources for improved e-government services. Current data integration approaches like data warehousing have limitations that semantic approaches can address. The document outlines challenges in current workflows and different approaches to semantic data management including using RDF, OWL and mapping relational databases to semantic models. It argues that semantic data integration using ontologies can provide a unified view of data to enable improved cross-agency information sharing and services for both governments and users.
Efficient Cost Minimization for Big Data ProcessingIRJET Journal
This document discusses efficient cost minimization techniques for big data processing. It characterizes big data processing using a two-dimensional Markov chain model to evaluate expected completion time. The problem is formulated as a mixed non-linear programming problem to optimize data assignment, placement, and migration across distributed data centers. A weighted bloom filter approach is presented to reduce communication costs through distributed incomplete pattern matching.
This document discusses big data and its challenges related to the Internet of Things (IoT). It first defines big data and explains how the aggregation of data from many IoT systems can lead to big data. It then discusses some key challenges of big data, including issues with data volume, velocity, variety, and veracity. Specific challenges for big data from IoT systems are also reviewed, such as authentication, security, and uncertainty of data. Finally, the document outlines some potential solutions to big data challenges, such as using MapReduce for heterogeneous data, data cleaning techniques for inconsistencies, and cloud-based security platforms for IoT devices.
A Combinational Approach of GIS and SOA for Performance Improvement of Organi...dannyijwest
A GIS (Geographic Information System), is a system designed to capture, store, manipulate, analyze, manage, and present all types of geographically referenced data. In other words, GIS is the merging of statistical analysis, database technology and cartography that can be integrated into any enterprise information system framework. But despite all these advantages, geographic information systems that used
in organizational structures, is faced with some problems such as lack of interoperability, business
alignment, and agility.
Abstract In early days information contain in increasingly corporate area, now IT organization help to right module to store, manage ,retrieve and transfer information in the more reliable and powerful manner. As part of an Information Lifecycle Management (ILM) best-practices strategy, organizations require solutions for migrating data between in heterogeneous environments and system storage. In early days information contain in increasingly corporate area, today IT organization help to right module to store, manage ,retrieve and transfer information in the more reliable and powerful manner. This paper helps to planned to design powerful modules that high-performances data migration of storage area with less time complexity. This project contain unique information of data migration in dynamic IT nature and business advantage that design to provide new tool used for data migration. Keywords— Heterogeneous Environment, data migration, data mapping
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A NOVEL SERVICE ORIENTED ARCHITECTURE (SOA) TO SECURE SERVICES IN E-CITYijsptm
Many cities in the world have moved toward being e-city using IT and there are some who have
implemented it and or seeking out to make it operational. However, experience shows that implementation
of e-city faces Challenges which the effectiveness of e-city, improvement of provided services and security
are the most important part of these challenges. Today, for realization of a perfect e-city, to overcome
faced challenges Such as security issues is an urgent need. Considering that e-city consists of multiple
information systems, it’s the most important challenge to create integration between these systems and
provide its security which service oriented architecture as a computational model and an approach for data
integration could largely overcome these problems. In this paper through studying challenges of
information systems in e-city layers and with concentrating on advantages of service oriented architecture,
a new architecture to improve the security of e-city’s systems and their provided services and also
overcoming the challenges of information systems has been proposed
Recommendation System for Information Services Adapted, Over Terrestrial Digi...CSEIJJournal
The development of digital television in Colombia has grown in last year’s, specially the digital terrestrial
television (DTT), which is an essential part to the projects of National Minister of ICT, thanks to the big
distribution and use of the television network and Internet in the country. This article explains how joining
different technologies like social networks, information adaptation and DTT, to get an application that
offers information services to users, based on their data, preferences, inclinations, use and interaction with
others users and groups inside the network.
Big data service architecture: a surveyssuser0191d4
This document discusses big data service architecture. It begins with an introduction to big data services and their economic benefits. It then describes the key components of big data service architecture, including data collection and storage, data processing, and applications. For data collection and storage, it covers Extract-Transform-Load tools, distributed file systems, and NoSQL databases. For data processing, it discusses batch, stream, and hybrid processing frameworks like MapReduce, Storm, and Spark. It concludes by noting big data applications in various fields and cloud computing services for big data.
SUITABILITY OF SERVICE ORIENTED ARCHITECTURE FOR SOLVING GIS PROBLEMSijait
Nowadays spatial data is becoming as a key element for effective planning and decision making in all aspects of societies. Spatial data are those data which are related to the features on the ground. In this way, a Geographic Information System (GIS) is a system that captures, analyzes, and manages any spatially referenced data. This paper analyzes the architecture and main features of Geographic Information Systems and aims at discussing some important problems emerged in the research of applying GIS in the organizations. It focuses on some of them such as lack of interoperability, agility and business alignment. We explain that SOA as a service oriented software architecture model can support the transformation of geographic information software from "system and function" to "service and application" and as the best practice of the architectural concepts can increase business alignment in the enterprise applications.
Service Level Comparison for Online Shopping using Data MiningIIRindia
The term knowledge discovery in databases (KDD) is the analysis step of data mining. The data mining goal is to extract the knowledge and patterns from large data sets, not the data extraction itself. Big-Data Computing is a critical challenge for the ICT industry. Engineers and researchers are dealing with the cloud computing paradigm of petabyte data sets. Thus the demand for building a service stack to distribute, manage and process massive data sets has risen drastically. We investigate the problem for a single source node to broadcast the big chunk of data sets to a set of nodes to minimize the maximum completion time. These nodes may locate in the same datacenter or across geo-distributed data centers. The Big-data broadcasting problem is modeled into a LockStep Broadcast Tree (LSBT) problem. And the main idea of the LSBT is defining a basic unit of upload bandwidth, r, a node with capacity c broadcasts data to a set of [c=r] children at the rate r. Note that r is a parameter to be optimized as part of the LSBT problem. The broadcast data are further divided into m chunks. In a pipeline manner, these m chunks can then be broadcast down the LSBT. In a homogeneous network environment in which each node has the same upload capacity c, the optimal uplink rate r, of LSBT is either c=2 or 3, whichever gives the smaller maximum completion time. For heterogeneous environments, an O(nlog2n) algorithm is presented to select an optimal uplink rate r, and to construct an optimal LSBT. With lower computational complexity and low maximum completion time, the numerical results shows better performance.The methodology includes Various Web applications Building and Broadcasting followed by the Gateway Application and Batch Processing over the TSV Data after which the Web Crawling for Resources and MapReduce process takes place and finally Picking Products from Recommendations and Purchasing it.
Design and implementation of the web (extract, transform, load) process in da...IAESIJAI
Owing to the recent increased size and complexity of data in addition to management issues, data storage requires extensive attention so that it can be employed in realistic applications. Hence, the requirement for designing and implementing a data ware-house system has become an urgent necessity. Data extraction, transformation and loading (ETL) is a vital part of the data warehouse architecture. This study designs a data warehouse application that contains a web ETL process that divides the work between the input device and server. This system is proposed to solve the lack of work partitioning between the input device and server. The designed system can be used in many branches and disciplines because of its high performance in adding data, analyzing data by using a web server and building queries. Analysis of the results proves that the designed system is fast in cleaning and transferring data between the remote parts of the system connected to the internet. ETL without missing any data consumes 0.00582 seconds.
Implementing Proof of Retriavaibility for Multiple Replica of Data File using...IRJET Journal
1. The document proposes a protocol for implementing proof of retrievability for multiple replicas of data files stored on cloud servers using a NoSQL database. It aims to verify the integrity of data when both the cloud storage server and third party auditor cannot fully be trusted.
2. The proposed system replaces a relational database with a NoSQL database to improve data operation performance and scaling for large datasets. It designs a protocol where the third party auditor generates signatures for dataset blocks and integrity proofs, which are then verified by the user to check the trustworthiness of the third party auditor.
3. Experimental results show that the time required for operations like tag generation and challenge-proof-verify are lower than previous solutions,
IRJET - A Framework for Tourist Identification and Analytics using Transport ...IRJET Journal
This document presents a framework for identifying and analyzing tourists using transport data. Big data technologies are used to monitor tourist movement and evaluate travel behavior in scenic areas. Transport data is isolated using Hadoop tools like HDFS, MapReduce, Sqoop, Hive and Pig. This allows processing large transport data sets without data loss issues. The data is analyzed to represent tourist hotspots, locations and preferences. Visualization tools like R are then used to provide insights into the analytics results. The framework aims to provide better information and perspectives to stakeholders like tour companies and transport operators using transport data.
A HEALTH RESEARCH COLLABORATION CLOUD ARCHITECTUREijccsa
Cloud computing platforms used for research collaborations within public health domains are limited in terms of how service components are organized and provisioned. This challenge is intensified by platform level challenges of transparency, confidentiality, privacy and trust. Addressing these collaboration issues will necessitate that components are reorganized. There is a need for secure and efficient approaches of reorganizing the service components, with trust to support collaboration related requirements. Through iterative design, a reliable and trust-aware re-organization of cloud components – the Collaboration cloud architecture is achieved. We utilize SOA, Privacy-By-Design principles and insights from blockchain to enforce trust. We illustrate its potential with multi-layer security process flow based on server-to-client tokens, Role-based Access Control and the traditional authentication – username and password with assurance for privacy and trust. The architecture shows promise towards data governance and the overall management of internal or external data flows.
A HEALTH RESEARCH COLLABORATION CLOUD ARCHITECTUREijccsa
Cloud computing platforms used for research collaborations within public health domains are limited in terms of how service components are organized and provisioned. This challenge is intensified by platform level challenges of transparency, confidentiality, privacy and trust. Addressing these collaboration issues will necessitate that components are reorganized. There is a need for secure and efficient approaches of reorganizing the service components, with trust to support collaboration related requirements. Through iterative design, a reliable and trust-aware re-organization of cloud components – the Collaboration cloud architecture is achieved. We utilize SOA, Privacy-By-Design principles and insights from blockchain to enforce trust. We illustrate its potential with multi-layer security process flow based on server-to-client tokens, Role-based Access Control and the traditional authentication – username and password with assurance for privacy and trust. The architecture shows promise towards data governance and the overall management of internal or external data flows.
The document discusses privacy-preserving location query protocols (PLQP). It proposes a novel PLQP that allows fine-grained location query and access control on encrypted location information. The protocol allows different levels of location query for different users based on access policies defined by the location publisher, such as revealing the city a user is in to friends or exact distance between two users who attended the same university. This achieves privacy preservation while providing location-based utility.
A HEALTH RESEARCH COLLABORATION CLOUD ARCHITECTUREijccsa
Cloud computing platforms used for research collaborations within public health domains are limited in terms of how service components are organized and provisioned. This challenge is intensified by platform level challenges of transparency, confidentiality, privacy and trust. Addressing these collaboration issues will necessitate that components are reorganized. There is a need for secure and efficient approaches of reorganizing the service components, with trust to support collaboration related requirements. Through
iterative design, a reliable and trust-aware re-organization of cloud components – the Collaboration cloud architecture is achieved. We utilize SOA, Privacy-By-Design principles and insights from blockchain to
enforce trust. We illustrate its potential with multi-layer security process flow based on server-to-client tokens, Role-based Access Control and the traditional authentication – username and password with assurance for privacy and trust. The architecture shows promise towards data governance and the overall management of internal or external data flows.
Similar to Data Exchange Design with SDMX Format for Interoperability Statistical Data (20)
Technology Content Analysis with Technometric Theory Approach to Improve Perf...Nooria Sukmaningtyas
Radiologic installation be facilitated with medical equipment for supporting of health services in
investigation of disease. There are 3 (three) criteria technology of equipment: investigation with
sophisticated equipment (CT scans single slice), investigation with a medium-sized enterprises equipment
(general x-ray 300 mA / 125 KV) and simple equipment (Portable Dental x-ray 8mA /70 KV). In view of
contribution to the Hospital, Radiologic installation from 2008 until 2012 has decreased, the data as
follows: The year 2008 was 6.8 percent; in 2009 was 4.3 percent; in 2010 was 2.5 percent; in 2011 was
2.3 percent; and in 2012 was 2.6 percent.By using approach of technometric theory, the study measures
the significant contribution of each component of technology that consists of aspects: Technoware,
Humanware, Inforware, Orgaware in Radiologic Installation, and also want to know about the
sophistication of the technology which is used in indicators Technology contribution coefficien (TCC), so
factors that affect experienced performance result can be known, where TCC are: TCC High Technology
TCC_tt =0,490 if T_tt= 0,387 H= 0,519 I= 0,538 O=0,534 TCC Middle technology TCC_tm=0,443 if
T_tm=0,258 H=0,519 I=0,538 O=0,534TCC Simple technology TCC_ts =0,398 if T_ts=0,168 H=0,519
I=0,538 O=0,534If the value of component technology (T,H,I,O) is less than TCC, its means that
Radiologic Installation Unit is in decreasing phase, a condition that cannot be left, need the directors’s
action immediately to formulate the right and fast policies to protect it from lost. The final result of study is a
gap almost everywhere from the three technology component Humanware = 0,519, Inforware = 0,538,
Orgaware = 0,534, but the gap between most components in technology aspects Technoware (0,387,
0,258, 0,168), that means that development strategy of Radiologic Installation unit be prioritized on
increasing aspects Technoware (rejuvenation medical equipment).
The knowledge management is a model involving the information system in the knowledge
processing. “Tim Penyelesaian Kerugian Negara” (TPKN) is one of sources of information related to the
state loss settlement, so it needs the development of knowledge management system on the state loss
settlement to ease the users when looking for the references of knowledge as completely as possible,
accurately, and quickly. This research aims to develop the system of knowledge management on the state
loss in LAPAN (SIMAPKLA). The used research methodology is Knowledge Management System Life
Cycle (KMSLC). The tacit, explicit knowledge is taken from the experts and it is stored in the Knowledge
Base (KB). The design model uses the approach with the orientation of object and implementation with Yii
Framework and blackbox testing. The menu on this system includes home, about us, dictionary, news,
meeting schedulles, knowledge about state loss, e-document, progress, forum, and contact us. Based on a
series of tests, in the aspect of functionality, this system is suitable and useful to share knowledge and
know the development of state loss settlement.
Automatic Extraction of Diaphragm Motion and Respiratory Pattern from Time-se...Nooria Sukmaningtyas
The document appears to be short, containing only the letters "TEL". It may refer to the word "telephone" or be an abbreviation. Without more context it is difficult to provide a more detailed summary in 3 sentences or less about the essential information or high level meaning of this brief document.
A dual-frequency microstrip patch antennas has been presented and used for 802.11WLAN
applications. The antennas had been designed, simulated and parametrically studied in CST Microwave
studio. By introducing u-slot, dual-band operation with its operating mode centered at frequency 2.4GHz,
3.65GHz and 5.2GHz had been obtained. The gain and directivity had been improved by adjusting the
parameters of the antennas. The gain of the proposed designs was 6.019dBi, 4.04dBi and 6.22dBi and
directivity was 6.02dBi, 4.05dBi and 6.22dBi at resonant frequencies 2.4GHz, 3.6GHz and 5.2GHz
respectively. The patch antennas had been proposed to be used in portable devices that require
miniaturized constituent parts.
The Detection of Straight and Slant Wood Fiber through Slop Angle Fiber FeatureNooria Sukmaningtyas
Quality control is one of important process that can not be avoided in industry. Image processing
technique is required to distinguish the quality of wood. If it can be done automatically by the computer, it
will be very helpful. This paper discusses the detection of straight and slant wood fiber to distinguish its
quality. This paper proposes an algorithm by using only two features i.e. mean (average value of slop
angle fiber) and maximumangle (the maximum value of slop angle fiber). Then the classification method is
used by tresholding. The result shows the performance is achieved on accuracy 79.2%
Active Infrared Night Vision System of Agricultural VehiclesNooria Sukmaningtyas
Active infrared night vision system was significant for night driving and it has been greatly used on
limousine car. Design active infrared night vision system for agricultural vehicles greatly improved the night
vision of them and it was an inevitable trend. Comparing parameters of various night vision systems and
designing active infrared night vision system of agricultural vehicles was significant for improving active
security of agricultural vehicles working at nighttime. By analyzing the infrared night vision system basic
parameters determined the structure form and basic parameters, calculated the infrared light wave width
and emission power to choose each components, designed active infrared night vision system’s structure
and determined parameters of agricultural vehicles.
Robot Three Dimensional Space Path-planning Applying the Improved Ant Colony ...Nooria Sukmaningtyas
This document describes an improved ant colony optimization (ACO) algorithm for 3D robot path planning to avoid obstacles. It proposes modifying the pheromone updating and transition probability functions of the ACO algorithm. Specifically, it introduces a distance factor to the transition probability function to encourage paths closer to a direct line between start and end points. It also uses fuzzy control to vary the pheromone amount based on iteration count and path length, rather than a fixed value. Simulation results show the improved ACO finds better paths with fewer iterations than the conventional ACO algorithm.
Research on a Kind of PLC Based Fuzzy-PID Controller with Adjustable FactorNooria Sukmaningtyas
A kind of fuzzy-PID controller with adjustable factor is designed in this paper. Scale factor’s selfadjust
will come true. Fuzzy control algorithm is finished in STEP7 software, and then downloaded in S7-
300 PLC. WinCC software will be used to control the change-trend in real time. Data communication
between S7-300 PLC and WinCC is achieved by MPI. The research shows that this fuzzy-PID controller
has better robust capability and stability. It’s an effective method in controlling complex long time-varying
delay systems.
This paper proposed a nonlinear robust control for spacecraft attitude based on passivity and
disturbance suppression vector. The spacecraft model was described using quaternion. The control law
introduced the suppression vector of external disturbances and had no information related to the system
parameters. The desired performance of spacecraft attitude control could be achieved using the designed
control law. And stability conditions of the nonlinear robust control for spacecraft attitude were given. The
stability could be proved by applying Lyapunov approach. The verification of the proposed attitude control
method was performed through a series of simulations. The numerical results showed the effectiveness of
the proposed control method in controlling the spacecraft attitude in the presence of external disturbances.
The main benefit of the proposed attitude control method does not need angular velocity measurement
and has its robustness against model uncertainties and external disturbances.
Remote Monitoring System for Communication Base Based on Short MessageNooria Sukmaningtyas
The automatic monitoring system of communication base which is an important means to realize
modernization of mobile communication base station management. In this paper, we implement a
monitoring system for communication base with three essential functions which are telemetry, remote
control and communication. In this system, data acquisition unit, data transmit unit and monitoring centre
unit are combined to form this monitoring system. The system can check the communication base status
anytime through GSM SMS (short message service), and can send predefined command to perform
remote data collection and monitoring in the special conditions. It is suitable especially for the alarm of
unusual situation, the monitoring of environmental information and entrance guard information. The paper,
firstly, proposes the architecture of the monitoring system; secondly, proposes the terminal of monitoring
system. The data collection terminal is studied and designed, including hardware design based on
embedded system and software design. Finally, presents implmentation and results. The monitoring
system can improve the integrity, reliability, flexibility and intellectuality of monitoring system. The system
with modular structure, which is low-cost, fitter and easier to move and operate, can be expanded
according to practical need and is reliable and effective through field test.
Tele-Robotic Assisted Dental Implant Surgery with Virtual Force FeedbackNooria Sukmaningtyas
The dental implant surgical applications full of risk because of the complex anatomical
architecture of craio-maxillofacial area. Therefore, the surgeons move towards computer-aided planning
for surgeries and then implementation using robotic assisted tele-operated techniques. This study divided
into four main parts. The first part is developed by computer-aided surgical planning by image modalities.
The second part is based on Virtual Surgical Environment through virtual force feedback haptic device.
The third part is implemented the experimental surgery by integrating the prototype surgical manipulator
with the haptic device poses using inverse kinematics method. The fourth part based on monitoring the
robotic manipulator pose by using image guided navigation system to calculate the position error of the
surgical manipulator. Thus, this tele-robotic system is able to comprehend the sense of complete practice,
improve skills and gain experience of the surgeon during the surgery. Finally, the experimental outcomes
show in satisfactory boundaries.
This paper proposes an adaptation mechanism based on adaptation planning graph for servicebased
business processes. First, a three-layer representation model of service-based business process is
introduced. Second, control-flow patterns of tasks, goal, logic model of service-based business process
and adaptation planning graph are introduced to enforce reliability of composite web services at run-time.
Finally, a simulation example of adaptation in service-based business processes is given. Simulations
prove that this approach can efficiently guarantee the reliability of composite services at run-time.
Review on Islanding Detection Methods for Photovoltaic InverterNooria Sukmaningtyas
This document summarizes and reviews various methods for detecting islanding effects that can occur when a photovoltaic inverter is connected to the electric grid. It divides detection methods into two categories: grid-side detection, which involves communication between the PV system and the grid; and local detection, which is further divided into passive methods that detect changes in frequency or voltage, and active methods that intentionally disturb the frequency, current, or phase to detect islanding. The document analyzes several representative methods in each category and discusses their advantages and disadvantages, proposing that improving detection of rate of change in power or frequency could help address limitations of passive methods. It focuses on the widely used active frequency drift method and improvements developed to reduce its non
Stabilizing Planar Inverted Pendulum System Based on Fuzzy Nine-point ControllerNooria Sukmaningtyas
In order to stabilize planar inverted pendulum, after analyzing the physical characteristics of the
planar inverted pendulum system, a pendulum nine-point controller and a car nine-point controller for Xaxis
and Y-axis were designed respectively. Then a fuzzy coordinator was designed using the fuzzy
control theory based on the priority of those two controllers, and the priority level of the pendulum is higher
than the car. Thus, the control tasks of each controller in each axis were harmonized successfully. The
designed control strategy did not depend on mathematical model of the system; it depended on the control
experience of people or the control experts. The compared experiments showed that the control strategy
was easy and effective, what was’s more; it had a very good robust feature.
Gross Error Elimination Based on the Polynomial Least Square Method in Integr...Nooria Sukmaningtyas
The measurement data of parameter in the electrical equipment contains many noises in subway
integrated monitoring system. To eliminate the impact of gross error in the measurement data, a
polynomial least square curve fitting algorithm is used in this paper. Based on the Rajda criterion, the
algorithm gives the variance estimation of the noises, and then uses dynamic threshold to detect and
replace the measurement data with gross error by statistical estimation. Finally, a data processing
procedure has been presented to deal with the gross error. The practical application indicates that the
proposed algorithm can effectively eliminate the gross error in many types of measurement signals so as
to ensure the reliability of the monitoring system.
Design on the Time-domain Airborne Electromagnetic Weak Signal Data Acquisiti...Nooria Sukmaningtyas
According to principle of transient electromagnetic method as well as its signal characteristics,
this paper designed and implemented a time-domain airborne electromagnetic weak signal data
acquisition system. With the use of the floating-point amplification technology, the system amplifies the
weak transient electromagnetic signal dynamically. CPLD and DSP were used as the decoding control
circuit and the main controller for processing the sampled data, respectively. The transient electromagnetic
signal acquisition system, which was designed with a dynamic range up to 144dB and a sampling rate up
to 100 kHz, meets the requirements of the high sampling rate with high precision and it has been applied in
the time-domain fixed-wing airborne electromagnetic mineral exploration.
INS/GPS integrated navigation system is studied in this paper for the hypersonic UAV in order to
satisfy the precise guidance requirements of hypersonic UAV and in response to the defects while the
inertial navigation system (INS) and the global positioning system (GPS) are being applied separately. The
information of UAV including position, velocity and attitude can be obtained by using INS and GPS
respectively after generating a reference trajectory. The corresponding errors of two navigation systems
can be obtained through comparing the navigation information of the above two guidance systems.
Kalman filter is designed to estimate the navigation errors and then the navigation information of INS are
corrected. The non-equivalence relationship between the platform misalignment angle and attitude error
angle are considered so that the navigation accuracy is further improved. The Simulink simulation results
show that INS/GPS integrated navigation system can help to achieve higher accuracy and better antiinterference
ability than INS navigation system and this system can also satisfy the navigation accuracy
requirements of hypersonic UAV.
Research on Space Target Recognition Algorithm Based on Empirical Mode Decomp...Nooria Sukmaningtyas
The space target recognition algorithm, which is based on the time series of radar cross section
(RCS), is proposed in this paper to solve the problems of space target recognition in the active radar
system. In the algorithm, EMD method is applied for the first time to extract the eigen of RCS time series.
The normalized instantaneous frequencies of high-frequency intrinsic mode functions obtained by EMD are
used as the eigen values for the recognition, and an effective target recognition criterion is established.
The effectiveness and the stability of the algorithm are verified by both simulation data and real data. In
addition, the algorithm could reduce the estimation bias of RCS caused by inaccurate evaluation, and it is
of great significance in promoting the target recognition ability of narrow-band radar in practice.
With the expanding of database of the watch list of anti-money laundering, improving the speed in
matching between the watch list and the database of account holders and clients’ transaction is especially
important. This paper proposes an improved AC-BM Algorithm, a matching algorithm of subsection, to
improve the speed of matching. Experiment results show the time performance of the improved algorithm
is better than traditional BM algorithm, AC algorithm and the AC-BM algorithm. It can improve the
efficiency of on-line monitoring of anti-money laundering.
Design of a Subsequent Water Detection System Controlled by the MicrocontrollerNooria Sukmaningtyas
This article instructs the subsequent water detection system design and implementation of sanitary
ware. This system used C8051f040 Microcontroller as the main control module and communication
module which controlled the operation of the entire system, touch screen as a position machine,and
made use of metal probing technique and weighing sensor technology to realize data test, collection,
display, storage and export. At last, the experimental results showed that this system meets the expected
requirement and can measure the subsequent water more accurately.
We have compiled the most important slides from each speaker's presentation. This year’s compilation, available for free, captures the key insights and contributions shared during the DfMAy 2024 conference.
KuberTENes Birthday Bash Guadalajara - K8sGPT first impressionsVictor Morales
K8sGPT is a tool that analyzes and diagnoses Kubernetes clusters. This presentation was used to share the requirements and dependencies to deploy K8sGPT in a local environment.
International Conference on NLP, Artificial Intelligence, Machine Learning an...gerogepatton
International Conference on NLP, Artificial Intelligence, Machine Learning and Applications (NLAIM 2024) offers a premier global platform for exchanging insights and findings in the theory, methodology, and applications of NLP, Artificial Intelligence, Machine Learning, and their applications. The conference seeks substantial contributions across all key domains of NLP, Artificial Intelligence, Machine Learning, and their practical applications, aiming to foster both theoretical advancements and real-world implementations. With a focus on facilitating collaboration between researchers and practitioners from academia and industry, the conference serves as a nexus for sharing the latest developments in the field.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELgerogepatton
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
Embedded machine learning-based road conditions and driving behavior monitoringIJECEIAES
Car accident rates have increased in recent years, resulting in losses in human lives, properties, and other financial costs. An embedded machine learning-based system is developed to address this critical issue. The system can monitor road conditions, detect driving patterns, and identify aggressive driving behaviors. The system is based on neural networks trained on a comprehensive dataset of driving events, driving styles, and road conditions. The system effectively detects potential risks and helps mitigate the frequency and impact of accidents. The primary goal is to ensure the safety of drivers and vehicles. Collecting data involved gathering information on three key road events: normal street and normal drive, speed bumps, circular yellow speed bumps, and three aggressive driving actions: sudden start, sudden stop, and sudden entry. The gathered data is processed and analyzed using a machine learning system designed for limited power and memory devices. The developed system resulted in 91.9% accuracy, 93.6% precision, and 92% recall. The achieved inference time on an Arduino Nano 33 BLE Sense with a 32-bit CPU running at 64 MHz is 34 ms and requires 2.6 kB peak RAM and 139.9 kB program flash memory, making it suitable for resource-constrained embedded systems.
Using recycled concrete aggregates (RCA) for pavements is crucial to achieving sustainability. Implementing RCA for new pavement can minimize carbon footprint, conserve natural resources, reduce harmful emissions, and lower life cycle costs. Compared to natural aggregate (NA), RCA pavement has fewer comprehensive studies and sustainability assessments.
A SYSTEMATIC RISK ASSESSMENT APPROACH FOR SECURING THE SMART IRRIGATION SYSTEMSIJNSA Journal
The smart irrigation system represents an innovative approach to optimize water usage in agricultural and landscaping practices. The integration of cutting-edge technologies, including sensors, actuators, and data analysis, empowers this system to provide accurate monitoring and control of irrigation processes by leveraging real-time environmental conditions. The main objective of a smart irrigation system is to optimize water efficiency, minimize expenses, and foster the adoption of sustainable water management methods. This paper conducts a systematic risk assessment by exploring the key components/assets and their functionalities in the smart irrigation system. The crucial role of sensors in gathering data on soil moisture, weather patterns, and plant well-being is emphasized in this system. These sensors enable intelligent decision-making in irrigation scheduling and water distribution, leading to enhanced water efficiency and sustainable water management practices. Actuators enable automated control of irrigation devices, ensuring precise and targeted water delivery to plants. Additionally, the paper addresses the potential threat and vulnerabilities associated with smart irrigation systems. It discusses limitations of the system, such as power constraints and computational capabilities, and calculates the potential security risks. The paper suggests possible risk treatment methods for effective secure system operation. In conclusion, the paper emphasizes the significant benefits of implementing smart irrigation systems, including improved water conservation, increased crop yield, and reduced environmental impact. Additionally, based on the security analysis conducted, the paper recommends the implementation of countermeasures and security approaches to address vulnerabilities and ensure the integrity and reliability of the system. By incorporating these measures, smart irrigation technology can revolutionize water management practices in agriculture, promoting sustainability, resource efficiency, and safeguarding against potential security threats.
2. ISSN: 2302-4046
TELKOMNIKA Vol. 14, No. 2, May 2015 : 343 – 352
344
implementation of the design using web services. This is a natural choice since the web service
is designed to support interoperability and interactions between system on a network [9]. In this
paper we propose three main results: (i) standardization framework of statistical
publication/aggregate data model structure, (ii) design architecture of data sharing service
model, and (iii) implementation of design architecture of data exchangeusing web service. We
will also present the validationresult of the proposed design with both user/expert judgment and
black box testing.
2. Research Method
2.1. Standardization Framework of Statistical Data Structure with SDMX
This section will explain our first result on the standardization framework of statistical
publication data structure. For case illustration we will use the business process of Data
Dissemination Directorate (DDS) of BPS. The business process of the mentioned directorateis a
process flow of activities to disseminate the result of statistical information in the form of
services to the public users. The objective of our proposed scheme is to improve the value of
public organization, in this case BPS, through increasing the openess of data, ease of access,
and data interoperability. Openess of the data can be measured by availability of both data
structure and content that can be accessed and used, reused and redistributed by all data
users. Ease of access can be measured by data servicesmedium thatcan beaccessed
easilythrough thedevicegadgets, mobile phones, PCandlaptop, etc. Meanwhile interoperability
can be measured by capability service to be operated on different platforms.In order to achieve
thesevalues and to show the conceptproposed in this paper, we adopt General Statistic
Business Process Model (GSBPM) for business process flow [10]. By adopting GSBPM we can
derive the business process as shown in Figure 1 whereas the business process of data
disseminationcan be elaborated further in Figure 2. In this business process model, the data to
be published will be taken from the dissemination data warehouse, not directly from subject
matters. Based on the GSBPM, we propose the new business process architecture for data
dissemination as shown in Figure 3. In this architecture, the published data is taken from a data
warehouse after being validated and analyzed by Analysis and Development Statistics
Directorate. Before publication, the data will be validated and mapped to the model
structures/data scheme predetermined in the form of DSD/MSD. This process is conducted as a
methodto standardize format and model structure of the published data, and to avoid the
inconsistency of data. The data is ready to be published/released to the web service provider
only after it is validated. We focus on the dissemination of statistics information through creating
services of data and metadata exchange where for this purpose we createa new service for
data and metadata exchange. This service provides easy access for the users.
Figure 1. Proposed business process activities at Data Dissemination Directorate of Statistics
Indonesia after adopting the GSBPM. Note that difference from existing condition, the data to be
published will be taken from the dissemination data warehouse, not directly from subject matters
3. TELKOMNIKA ISSN: 2302-4046
Data Exchange Design with SDMX Format for Interoperability Statistical Data (Jaka Sembiring)
345
Figure 2. Detail business process of data dissemination derived from GSBPM. Note that this
decomposition provides us with candidate services such as dissemination service
Figure 3. Proposed business process architecture of statistics data dissemination. This
architecture has been aligned with the future contextual plan of BPS
The framework of data and metadata model structure is designed with SDMX-
Information Model (SDMX-IM) methodwhich incudes several components: DSD, MSD, Data set,
Metadata set and SDMX-ML message. Based on the analysis on the data to be published and
SDMX-IM components, we can define the Level 1 standardization structure model as in Figure 4
where we have three main activities in this process as follows:
a) delivery of data to be published to the Data Dissemination Repository,
b) mapping local data with Data Structure Definition (DSD) and Metadata Structure
Definition (MSD), and
c) publish/release of data publication.
Figure 4. Standardization process data structure Level 1
4. ISSN: 2302-4046
TELKOMNIKA Vol. 14, No. 2, May 2015 : 343 – 352
346
We proceed to the Level 2 standardization of the model structure for detail process as
in Figure 5.
Figure 5. Standardization process data structure Level 2
The general process in Level 2 can be explained in detail as follows.
1) Determine the Concept and Concept Scheme. Concept has a very important role in the
SDMX-IM because it is used to describe the multi-dimensional table structure or metadata
structurere port. Concept structure consists of Dimensions, Measures, and Attributes. The
sample result of concept and concept scheme in this paper can be seen in Table 3.
2) Creating Code List. Code list in the SDMX-IM is a list containing codes that represent
concepts (dimension or attribute) in either DSD or MSD [11]. Sample of code list created in
this paper is shown in Table 3.
3) Data Structure Definition (DSD). DSD is basically a description for concepts that have been
identified and established. It describes whether the concept is a dimension or an attribute.
DSD can be used in Time Series (TS), Cross Sectional (CS) and multidimensional data
tables. In our case to create DSD, first we need to differentiate between TS data and CS
data. Then we identify the concept, determine the measure and make an association with
the Code Lists. The sample result for our case can be seen in Table 3.
4) Metadata Structure Definition (MSD). As in DSD, MSD is a structure for metadata. Metadata
is used as a description or reference for the object to be exchanged. The purpose of
metadata structure is to facilitate indexing, searching, processing and reporting of statistical
activities. BPS devide metadata into three types: descriptive metadata, structural metadata
and administrative metadata. These metadata are associated with reference metadata in
SDMX.
5) Mapping Data. The main problem of data exchange on the data sharing systems is that
there are various database structure integrated in one system, so that itis necessary to
perform data mapping amongst related database to improve the convergence [12]. In our
case we employ five steps for mapping between local databases and the predefined DSD
concept as follows.
a) List theconceptsthat existin thelocal data.
b) Pairingalllocal dataconceptwiththe DSD concept.
c) Pairinglocal codeswithDSDcodes.
d) TranslatingSDMXQuery.
e) Translate intodataset.
6) SDMX-ML message. SDMX-ML is created based on XML; so that SDMX complieswiththe
concept and the rules of XML.The most important component in the implement SDMX XML
are an XML namespace and XML Schema (XSD).
5. TELKOMNIKA ISSN: 2302-4046
Data Exchange Design with SDMX Format for Interoperability Statistical Data (Jaka Sembiring)
347
2.2. Architecture of Data Sharing Model
The general modelof dataandmetadataexchange processproposedin this paper isa
simple model where dataproviderpublish/disseminatethe datatothe consumer using web
services for data sharing implementation. The system diagram can be seen in Figure 6.
Figure 6. Data sharing model using web service
In our case, two models of datasharing are involved, i.e.: (i) datasharing from data
provider to the publisher system, and (ii) data sharing from the publisher system to consumer.
Data provider including subject matters and government institutions; sends the data to the
system using push method, in other words the data provider push/send the data to the Data
Dissemination publication repository. Whereas thedata transmission from system to consumers
is using pull method where consumers ends a data request message and will pull/retrieve
dataset fromweb service system. The architecture design of data dissemination that supports
the above mentioned data sharing can be seen in Figure 7.
Figure 7. Detail of data sharing system design
The dissemination data warehouse saves all the local data publication comes from the
subject matters and other data sources. Based on GSBPM, this warehouse will be managed by
the Statisticals Information Systems Division and Statistical Dissemination Divisionso that the
data can be validated before publication.In addition to dissemination repository, we have a
mapping store to save the data concept. Statistical Dissemination Division who also performs
the data mapping maintains the mapping store. With this architecture, all data will have the
same concept. Both data storage are designed to support data exchange services. User is
allowed to query or request the data via the web client, and this web client will send a message
to the web service. This service will retrieve the data from dissemination data warehouse and
mapping store located in the BPS servers.
Figure 8. Data dissemination web service architecture
6. ISSN: 2302-4046
TELKOMNIKA Vol. 14, No. 2, May 2015 : 343 – 352
348
The web service architecture for data exchange services is shown in Figure 8. Users
who send request data in SDMX Query messages, which will be translated by SDMX Query
Parser, initiate the process of data dissemination in the web service. The SQLQuery retrieve the
data structures and datasets from the store Mapping and data dissemination repository, and the
system will generate data in SDMX-ML format which will be sent to the webclient.
2.2. Design of Data Exchange Service with SOAD
In this section we will discuss the design process of data and metadata exchange
services. We use SOAD methodology as an approach of SOA [8]. Following the SOAD
methodology for analysis and design, first step is the Conceptual View (CV) where we obtain
the business process and sub business process for data and metadata service as shown in
Figure 9.
Figure 9. Business process and sub-business process analysis
Based the proposed system depicted in Figure 3, there are two systemswe need to
create, i.e.(i) Web Service Provider, which is used by data publisher to provide, report and
disseminate statistical data, and (ii) Web Application Client,which is usedbythe usersto invoke
data.The general function of the system is described in Table 1.
Table 1. General function of the proposed system
Integrated Data Publication and Data Sharing Management
System Service
Description
Web Service Provider
Providing publication data and
metadata
Facilitatethe
dataprovidertoprepareandtransmit
data/metadata for the users
Reporting and dissemination
publicationdata
Facilitatetheinstrumentforpublication and
reportingthe data for
publicationsafterbeingvalidated
Web Application Client
Requesting publication data
Facilitatetheuserstorequestthe databased on
the published categoryof data
Displaying publication data
Facilitate the users by displayingdata
fromthe userqueryaccording to the provided
format
7. TELKOMNIKA ISSN: 2302-4046
Data Exchange Design with SDMX Format for Interoperability Statistical Data (Jaka Sembiring)
349
Second step, the Logical View (LG) consists of three layers: business layer, service
layer and components layer which is derived from CV. The result of analysis can be seen in
Table 2.
Table 2. Proposed services at Logical View
System FunctionalDomain
Coverageof Functional
system
Description
Providing publication data
and metadata
Registing data provider Servicesfor registeringthe publication data
Sending data publications
Servicestouploadthe publication datatothe
database of dissemination/publication
Mapping local data to DSD
and MSD SDMX
Services used by datapublishertoverify,
mapandvalidatethe publicationdatawiththe pre
defined DSDandMSD
Publishing the publications
data
Servicesused by dataprovidersto approvethe
publication datato be published to the public
Reporting and dissemination
publication data
Reporting and dissemination
publication data
Servicefor the users to view the published data
Requesting publication data
Selecting data category
Servicesfor the usersto selectappropriatedata
categories
Selecting data flow Servicesfor theusersto choosedataflow
Sending request
Servicesforthe userto choose andinvokethe
datarequest
Displaying publication data
Displaying data
Servicefordisplaying datawithSDMX-ML,
PDFandXLS data format,andfor
visualizationwithtabulationandgraphs
Dowloading data Service for downloading datawith specificformat
Finally, the Physical View (PV) consists of four layers web service layer, presentation
layer, appplication layer and data model layer. Web service layer is divided into two parts: web
service design and service sequence diagram. Presentation layer is an interface design for
Integrated Data Publication and Data Sharing Management System where in this paper we
provide seven interfaces including: Home Page, Data Category, Data and Metadata Set Upload,
Data Mapping, Login/Register, Web Service Data Published, and Dataset Presentation.
Application Service layer is the interfaces for business logic implementation constist of eight
interfaces including: IdataCategory AppS, IdataMetadataSetUploadAppS, IdataMappingAppS,
IloginAppS, ImemberRegisterAppS, IdataFlowAppS, IdataPresentationAppS, and
IDSDMSDAppS. Data model layer consists of two Data Transfer Object (DTO). The first is the
Data Dissemination DTO consists of LogQuery, DataSet, MetadataSet, Log Download and
Publisher. This componentis used to store local publications data i.e. dataset and metadataset.
The secondis Mapping Store consists of DSD, MSD, Mapping, Master Data Provider, CodeList,
Category Data and Data Flow. This component is used to store data check, validation and
mapping so that the data is readyto be disseminated.
3. Results and Analysis
In this section we will show the implementation result of the design involving public
domain data on population/demography of Statistics Indonesia. Aggregate data for sample case
is(i) the population of Indonesia by province 1971, 1980, 1990, 1995, 2000, and 2010, and (ii)
total and percentage of poverty and poverty line by province 2007, 2008, 2009, 2010, 2011,
2012. Both data can be viewed in two categories, which are Time Series (TS) and Cross
Sectional (CS). Using the method described in the previous section, first we determine the
concept where in this simple case is denoted with POP_CONCEPT since we deal with data on
population/demographic. Based on the concept, we can derive the code list such as in column
five of Table 3. We can create the DSD/MSD by first separating the TS and CS, then using the
method described in the previous section we can obtain the DSD/MSD where the sample result
can be seen in Table 3.
8. ISSN: 2302-4046
TELKOMNIKA Vol. 14, No. 2, May 2015 : 343 – 352
350
Table 3. Format of Message Implementation Guide (MIG) DSD
DIMENSIONS
Position in
key
Concept Scheme Codelist
ID Name ID Scheme ID CodeList
1 FREQ Frequency POP_CONCEPTS CL_FREQ
2 PROVINCE Population country POP_CONCEPTS CL_PROVINCE
3 INDIC_POP Population Indicator POP_CONCEPTS CL_POP_INDICAT
4 POOR_POP POORNESS Measure POP_CONCEPTS CL_POP_POOR
TIME TIME_PERIOD Time Period POP_CONCEPTS
MEASURES
TYPE
Concept Scheme Codelist
ID Name ID Scheme ID CodeList
Primary OBS_VALUE Observation value POP_CONCEPTS
CS ID-AC Aceh CL_PROVINCE
ID-BA Bali
ID-BB Bangka-Belitung
ID-BE Bengkulu
ATTRIBUTES
Attachment
Level
Concept Scheme Codelist
ID Name ID Scheme ID CodeList
Observation OBS_STATUS Status of the observation POP_CONCEPTS CL_OBS_STATUS
Series UNIT Unit POP_CONCEPTS CL_UNIT
Series TIME_FORMAT Time Format POP_CONCEPTS CL_TIME_FORMAT
We continue presenting the result with the implementation of webservice provider which
can be seen in Figure 10.
Figure 10. Implementation example of web service provider
We verify the standardization and architecture design quantitatively using simple Likert
scale by presenting the result to the user/expert, which in this case is the relevant personnel at
Directorate of Statistical Dissemination, Statistics Indonesia (BPS).Forthe web service
implementation we perform both quantitative evaluation through Likert scale and black box
testing using soapUI software. We obtained that for quantitative assessment the result is 22.3 in
9. TELKOMNIKA ISSN: 2302-4046
Data Exchange Design with SDMX Format for Interoperability Statistical Data (Jaka Sembiring)
351
Likert scale, whichmeansvery good. Meanwhile for black box testing resultusing soapUI can be
seen in Figure 11. We conclude that all of the designed service perform as intended.
Figure 11. Result of black box testing using soapUI
4. Conclusion
In this paper we report on the design of data and meta data exchange using SDMX
format for statistical data at Statistics Indonesia (BPS). We implement the design concept with
web services. Implementation at Statistics Indonesia (BPS) is chosen as a case study to prove
the design concept. We describe three main results. The first result is on the framework for
standardization of structure of statistical publications data model with SDMX. The simple case
we presented has been able to demonstrate the advantageous of the effort for interoperability of
data and metadata. Second result is on the design architecture of data sharing model. The
design is based on the user needs where we able to align the design with the long term master
plan of BPS. Finally, the third result is onthe implementation of data and metadata exchange
service using Service Oriented Analysis and Design (SOAD) method. We present the prototype
system implementation with web services.It is shown through quantitative Likert assessment
and black box testing that the overall standard, architecture design and implementation have
achievedtheirrespective objectives.
References
[1] Hoxha J, Brahaj A. Open Goverment Data on the Web: A Semantic Approach. International
Conference on Emerging Intelligent Data and Web Technologies. Tirana, Albania. 2011: 107-113.
[2] Bizer C, Health T, Berners Lee T. Linked Data-The Story So Far. International Journal on Semantic
Web and Information Systems (IJSWIS). 2009; 5(3): 1-22.
[3] Zakaria Z. Key Performance Indicators (KPIs) in the Public Sector: A Study in Malaysia. Canadian
Center of Science and Education. 2011; 7: 102-107.
[4] Statistical Data and Metadata eXchange. 2012. sdmx.org/page_id=10.
[5] SDMX Standards Section 1 -Framework. 2013. www.sdmx.org.
[6] SDMX Standards Section 1 -SDMX-Information Model. 2013. www.sdmx.org.
[7] Erl T. Service Oriented Architecture: Concept, Tehnology and Design. New York: Prentice Hall PTR.
2005.
10. ISSN: 2302-4046
TELKOMNIKA Vol. 14, No. 2, May 2015 : 343 – 352
352
[8] Sarno R, Herdiyanti A. A Service Portfolio for an Enterprise Resource Planning. International Journal
of Computer Science and Network Security. 2010: 144-156.
[9] Web Services Architecture. 2004. http://www.w3.org/TR/ws-arch/.
[10] Vale S. GSBPM. http://www1.unece.org.
[11] Cross-Domain Code List. 2009. www.sdmx.org.
[12] Ning L, Bin X, Xin Z. Database Conversion Based on Relationship Schema Mapping. International
Conference on Internet Technology and Applications (iTAP). Wuhan. 2011: 16-18.