Nowadays, the database field has gotten much more diverse, and as a result, a variety of non-relational (NoSQL) databases have been created, including JSON-document databases and key-value stores, as well as extensible markup language (XML) and graph databases. Due to the emergence of a new generation of data services, some of the problems associated with big data have been resolved. In addition, in the haste to address the challenges of big data, NoSQL abandoned several core databases features that make them extremely efficient and functional, for instance the global view, which enables users to access data regardless of how it is logically structured or physically stored in its sources. In this article, we propose a method that allows us to query non-relational databases based on the ontology-based access data (OBDA) framework by delegating SPARQL protocol and resource description framework (RDF) query language (SPARQL) queries from ontology to the NoSQL database. We applied the method on a popular database called Couchbase and we discussed the result obtained.
Towards a new hybrid approach for building documentoriented data warehIJECEIAES
Schemaless databases offer a large storage capacity while guaranteeing high performance in data processing. Unlike relational databases, which are rigid and have shown their limitations in managing large amounts of data. However, the absence of a well-defined schema and structure in not only SQL (NoSQL) databases makes the use of data for decision analysis purposes even more complex and difficult. In this paper, we propose an original approach to build a document-oriented data warehouse from unstructured data. The new approach follows a hybrid paradigm that combines data analysis and user requirements analysis. The first data-driven step exploits the fast and distributed processing of the spark engine to generate a general schema for each collection in the database. The second requirement-driven step consists of analyzing the semantics of the decisional requirements expressed in natural language and mapping them to the schemas of the collections. At the end of the process, a decisional schema is generated in JavaScript object notation (JSON) format and the data loading with the necessary transformations is performed.
Study on potential capabilities of a nodb systemijitjournal
There is a need of optimal data to query processing technique to handle the increasing database size,
complexity, diversity of use. With the introduction of commercial website, social network, expectations are
that the high scalability, more flexible database will replace the RDBMS. Complex application and Big
Table require highly optimized queries. Users are facing the increasing bottlenecks in their data analysis. A
growing part of the database community recognizes the need for significant and fundamental changes to
database design. A new philosophy for creating database systems called noDB aims at minimizing the datato-
query time, most prominently by removing the need to load data before launching queries. That will
process queries without any data preparation or loading steps. There may not need to store data. User can
pipe raw data from websites, DBs, excel sheets into two promise sample inputs without storing anything.
This study is based on PostgreSQL systems. A series of the baseline experiment are executed to evaluate the
Performance of this system as per -a. Data loading cost, b-Query processing timing, c-Avoidance of
Collision and Deadlock, d-Enabling the Big data storage and e-Optimize query processing etc. The study
found significant possible capabilities of noDB system over the traditional database management system.
Towards a new hybrid approach for building documentoriented data warehIJECEIAES
Schemaless databases offer a large storage capacity while guaranteeing high performance in data processing. Unlike relational databases, which are rigid and have shown their limitations in managing large amounts of data. However, the absence of a well-defined schema and structure in not only SQL (NoSQL) databases makes the use of data for decision analysis purposes even more complex and difficult. In this paper, we propose an original approach to build a document-oriented data warehouse from unstructured data. The new approach follows a hybrid paradigm that combines data analysis and user requirements analysis. The first data-driven step exploits the fast and distributed processing of the spark engine to generate a general schema for each collection in the database. The second requirement-driven step consists of analyzing the semantics of the decisional requirements expressed in natural language and mapping them to the schemas of the collections. At the end of the process, a decisional schema is generated in JavaScript object notation (JSON) format and the data loading with the necessary transformations is performed.
Study on potential capabilities of a nodb systemijitjournal
There is a need of optimal data to query processing technique to handle the increasing database size,
complexity, diversity of use. With the introduction of commercial website, social network, expectations are
that the high scalability, more flexible database will replace the RDBMS. Complex application and Big
Table require highly optimized queries. Users are facing the increasing bottlenecks in their data analysis. A
growing part of the database community recognizes the need for significant and fundamental changes to
database design. A new philosophy for creating database systems called noDB aims at minimizing the datato-
query time, most prominently by removing the need to load data before launching queries. That will
process queries without any data preparation or loading steps. There may not need to store data. User can
pipe raw data from websites, DBs, excel sheets into two promise sample inputs without storing anything.
This study is based on PostgreSQL systems. A series of the baseline experiment are executed to evaluate the
Performance of this system as per -a. Data loading cost, b-Query processing timing, c-Avoidance of
Collision and Deadlock, d-Enabling the Big data storage and e-Optimize query processing etc. The study
found significant possible capabilities of noDB system over the traditional database management system.
Development of a Web based Shopping Cart using the Mongo DB Database for Huma...AI Publications
The databases in use today are of SQL-type. This has its drawbacks such as unnecessary complex queries, rigid schema, non-asynchronous persistence and they are definitely not object oriented. Moreover, SQL-shopping cart is expensive by requiring more programs to function. Therefore, the development of a modern shopping cart using MongoDB will eradicate these set backs. The main aim of this study is to design and execute a modern e-commerce shopping cart using MongoDB database. The method used here is the agile development methodology. Stages involved here include: Brainstorm, Design, development stage, Quality Assurance, deployment and Cycle. The User interface is written with HTML, CSS and JavaScript. The HTML (Hyper Text markup language) is used to create the web pages involved, including the forms through which the user supplies input to the system. Each item in the web page is well labeled to optimize user friendliness. The CSS (cascading Style Sheet) is used to create a mobile-friendly, responsive interface to enable mobile devices to seamlessly use the system.The developed shopping cart will save time and effort for programmers rather than using SQL tools with all the labors with it.
OUTCOME ANALYSIS IN ACADEMIC INSTITUTIONS USING NEO4Jijcsity
Databases are an integral part of a computing system and users heavily rely on the services they provide.When interact with a computing system, we expect that data be stored for future use, that the data is able to be looked up fastly, and we can perform complex queries against the data stored in the database. Many
different emerging database types available for use such as relational databases, object databases, keyvalue databases, graph databases, and RDF databases. Each type of database provides unique qualities that have applications in certain domains. Our work aims to investigate and compare the performance and
scalability of relational databases to graph databases in terms of handling multilevel queries such as finding the impact of a particular subject with the working area of pass out students. MySQL was chosen as the relational database, Neo4j as the graph database.
DATABASE SYSTEMS PERFORMANCE EVALUATION FOR IOT APPLICATIONSijdms
ABSTRACT
The amount of data stored in IoT databases increases as the IoT applications extend throughout smart city appliances, industry and agriculture. Contemporary database systems must process huge amounts of sensory and actuator data in real-time or interactively. Facing this first wave of IoT revolution, database vendors struggle day-by-day in order to gain more market share, develop new capabilities and attempt to overcome the disadvantages of previous releases, while providing features for the IoT.
There are two popular database types: The Relational Database Management Systems and NoSQL databases, with NoSQL gaining ground on IoT data storage. In the context of this paper these two types are examined. Focusing on open source databases, the authors experiment on IoT data sets and pose an answer to the question which one performs better than the other. It is a comparative study on the performance of the commonly market used open source databases, presenting results for the NoSQL MongoDB database and SQL databases of MySQL and PostgreSQL
The growth of data and its effi cient handling is becoming more popular trend in recent years bringing
new challenges to explore new avenues. Data analytics can be done more effi ciently with the availability of
distributed architecture of “Not Only SQL” NoSQL databases.
Analysis and evaluation of riak kv cluster environment using basho benchStevenChike
Many institutions and companies with technological development have been producing large size of structured and unstructured data. Therefore, we need special databases to deal with these data and thus emerged NoSQL databases. They are widely used in the cloud databases and the distributed systems. In the era of big data, those databases provide a scalable high availability solution. So we need new architectures to try to meet the need to store more and more different kinds of different data. In order to arrive at a good structure of large and diverse data, this structure must be tested and analyzed in depth with the use of different benchmark tools. In this paper, we experiment the Riak key-value database to measure their performance in terms of throughput and latency, where huge amounts of data are stored and retrieved in different sizes in a distributed database environment. Throughput and latency of the NoSQL database over different types of experiments and different sizes of data are compared and then results were discussed.
EVALUATING CASSANDRA, MONGO DB LIKE NOSQL DATASETS USING HADOOP STREAMINGijiert bestjournal
An unstructured data poses challenges to storing da ta. Experts estimate that 80 to 90 percent of the d ata in any organization is unstructured. And the amount of uns tructured data in enterprises is growing significan tly� often many times faster than structured databases are gro wing. As structured data is existing in table forma t i,e having proper scheme but unstructured data is schema less database So it�s directly signifying the importance of NoSQL storage Model and Map Reduce platform. For processi ng unstructured data,where in existing it is given to Cassandra dataset. Here in present system along wit h Cassandra dataset,Mongo DB is to be implemented. As Mongo DB provide flexible data model and large amou nt of options for querying unstructured data. Where as Cassandra model their data in such a way as to mini mize the total number of queries through more caref ul planning and renormalizations. It offers basic secondary ind exes but for the best performance it�s recommended to model our data as to use them infrequently. So to process
MONGODB VS MYSQL: A COMPARATIVE STUDY OF PERFORMANCE IN SUPER MARKET MANAGEME...ijcsity
A database is information collection that is organized in tables so that it can easily be accessed, managed, and updated. It is the collection of tables, schemas, queries, reports, views and other objects. The data are typically organized to model in a way that supports processes requiring information, such as modelling to find a hotel with availability of rooms, thus the people can easily locate the hotels with vacancies. There are many databases commonly, relational and non relational databases. Relational databases usually work with structured data and non relational databases are work with semi structured data. In this paper, the performance evaluation of MySQL and MongoDB is performed where MySQL is an example of relational database and MongoDB is an example of non relational databases. A relational database is a data structure that allows you to connect information from different 'tables', or different types of data buckets. Non-relational database stores data without explicit and structured mechanisms to link data from different buckets to one another. This paper discuss about the performance of MongoDB and MySQL in the field of Super Market Management System. A supermarket is a large form of the traditional grocery store also a self-service shop offering a wide variety of food and household products, organized in systematic manner. It is larger and has a open selection than a traditional grocery store.
MONGODB VS MYSQL: A COMPARATIVE STUDY OF PERFORMANCE IN SUPER MARKET MANAGEME...ijcsity
A database is information collection that is organized in tables so that it can easily be accessed, managed,
and updated. It is the collection of tables, schemas, queries, reports, views and other objects. The data are
typically organized to model in a way that supports processes requiring information, such as modelling to
find a hotel with availability of rooms, thus the people can easily locate the hotels with vacancies. There
are many databases commonly, relational and non relational databases. Relational databases usually work
with structured data and non relational databases are work with semi structured data. In this paper, the
performance evaluation of MySQL and MongoDB is performed where MySQL is an example of relational
database and MongoDB is an example of non relational databases. A relational database is a data
structure that allows you to connect information from different 'tables', or different types of data buckets.
Non-relational database stores data without explicit and structured mechanisms to link data from different
buckets to one another. This paper discuss about the performance of MongoDB and MySQL in the field of
Super Market Management System. A supermarket is a large form of the traditional grocery store also a
self-service shop offering a wide variety of food and household products, organized in systematic manner.
It is larger and has a open selection than a traditional grocery store.
MONGODB VS MYSQL: A COMPARATIVE STUDY OF PERFORMANCE IN SUPER MARKET MANAGEME...ijcsity
A database is information collection that is organized in tables so that it can easily be accessed, managed,
and updated. It is the collection of tables, schemas, queries, reports, views and other objects. The data are
typically organized to model in a way that supports processes requiring information, such as modelling to
find a hotel with availability of rooms, thus the people can easily locate the hotels with vacancies. There
are many databases commonly, relational and non relational databases. Relational databases usually work
with structured data and non relational databases are work with semi structured data. In this paper, the
performance evaluation of MySQL and MongoDB is performed where MySQL is an example of relational
database and MongoDB is an example of non relational databases. A relational database is a data
structure that allows you to connect information from different 'tables', or different types of data buckets.
Non-relational database stores data without explicit and structured mechanisms to link data from different
buckets to one another. This paper discuss about the performance of MongoDB and MySQL in the field of
Super Market Management System. A supermarket is a large form of the traditional grocery store also a
self-service shop offering a wide variety of food and household products, organized in systematic manner.
It is larger and has a open selection than a traditional grocery store.
Linked Data Generation for the University Data From Legacy Database dannyijwest
Web was developed to share information among the users through internet as some hyperlinked documents.
If someone wants to collect some data from the web he has to search and crawl through the documents to
fulfil his needs. Concept of Linked Data creates a breakthrough at this stage by enabling the links within
data. So, besides the web of connected documents a new web developed both for humans and machines, i.e.,
the web of connected data, simply known as Linked Data Web. Since it is a very new domain, still a very
few works has been done, specially the publication of legacy data within a University domain as Linked
Data.
The aim of this paper is to evaluate, through indexing techniques, the performance of Neo4j and
OrientDB, both graph databases technologies and to come up with strength and weaknesses os each
technology as a candidate for a storage mechanism of a graph structure. An index is a data structure that
makes the searching faster for a specific node in concern of graph databases. The referred data structure
is habitually a B-tree, however, can be a hash table or some other logic structure as well. The pivotal
point of having an index is to speed up search queries, primarily by reducing the number of nodes in a
graph or table to be examined. Graphs and graph databases are more commonly associated with social
networking or “graph search” style recommendations. Thus, these technologies remarkably are a core
technology platform for some Internet giants like Hi5, Facebook, Google, Badoo, Twitter and LinkedIn.
The key to understanding graph database systems, in the social networking context, is they give equal
prominence to storing both the data (users, favorites) and the relationships between them (who liked
what, who ‘follows’ whom, which post was liked the most, what is the shortest path to ‘reach’ who). By a
suitable application case study, in case a Twitter social networking of almost 5,000 nodes imported in
local servers (Neo4j and Orient-DB), one queried to retrieval the node with the searched data, first
without index (full scan), and second with index, aiming at comparing the response time (statement query
time) of the aforementioned graph databases and find out which of them has a better performance (the
speed of data or information retrieval) and in which case. Thereof, the main results are presented in the
section 6.
Growth of relational model: Interdependence and complementary to big data IJECEIAES
A database management system is a constant application of science that provides a platform for the creation, movement, and use of voluminous data. The area has witnessed a series of developments and technological advancements from its conventional structured database to the recent buzzword, bigdata. This paper aims to provide a complete model of a relational database that is still being widely used because of its well known ACID properties namely, atomicity, consistency, integrity and durability. Specifically, the objective of this paper is to highlight the adoption of relational model approaches by bigdata techniques. Towards addressing the reason for this in corporation, this paper qualitatively studied the advancements done over a while on the relational data model. First, the variations in the data storage layout are illustrated based on the needs of the application. Second, quick data retrieval techniques like indexing, query processing and concurrency control methods are revealed. The paper provides vital insights to appraise the efficiency of the structured database in the unstructured environment, particularly when both consistency and scalability become an issue in the working of the hybrid transactional and analytical database management system.
Representing Non-Relational Databases with Darwinian NetworksIJERA Editor
The Darwinian networks (DNs) are first introduced by Dr Butz [1] to simplify and clarify how to work with Bayesian networks (BNs). DNs can unify modeling and reasoning tasks into a single platform using the graphical manipulation of the probability tables that takes on a biological feel. From this view of the DNs, we propose a graphical library to represent and depict non-relational databases using DNs. Because of the growing of this kind of databases, we need even more tools to help in the management work, and the DNs can help with these tasks.
An elastic , effective, activety or intelligent ,graceful networking architecture layout be desired to make processing massive data. next to that ,existent network architectures be considerably incapable for
cleatting the huge data. massive data thrusts network exchequers into border it consequence with in network overcrowding ,needy achievement, then permicious employer exprtises. this offered the current state-of-the-art research affronts ,potential solutions into huge data networking notion. More specifically, present the state of networking problems into massive data connected intrequirements,capacity,running ,
data manipulating also will introduce the architectures of MapReduce , Hadoop paradigm within research
requirements, fabric networks and software defined networks which utilizized into making today’s idly growing digital world and compare and contrast into identify relevant drawbacks and solutions.
Bibliometric analysis highlighting the role of women in addressing climate ch...IJECEIAES
Fossil fuel consumption increased quickly, contributing to climate change
that is evident in unusual flooding and draughts, and global warming. Over
the past ten years, women's involvement in society has grown dramatically,
and they succeeded in playing a noticeable role in reducing climate change.
A bibliometric analysis of data from the last ten years has been carried out to
examine the role of women in addressing the climate change. The analysis's
findings discussed the relevant to the sustainable development goals (SDGs),
particularly SDG 7 and SDG 13. The results considered contributions made
by women in the various sectors while taking geographic dispersion into
account. The bibliometric analysis delves into topics including women's
leadership in environmental groups, their involvement in policymaking, their
contributions to sustainable development projects, and the influence of
gender diversity on attempts to mitigate climate change. This study's results
highlight how women have influenced policies and actions related to climate
change, point out areas of research deficiency and recommendations on how
to increase role of the women in addressing the climate change and
achieving sustainability. To achieve more successful results, this initiative
aims to highlight the significance of gender equality and encourage
inclusivity in climate change decision-making processes.
Voltage and frequency control of microgrid in presence of micro-turbine inter...IJECEIAES
The active and reactive load changes have a significant impact on voltage
and frequency. In this paper, in order to stabilize the microgrid (MG) against
load variations in islanding mode, the active and reactive power of all
distributed generators (DGs), including energy storage (battery), diesel
generator, and micro-turbine, are controlled. The micro-turbine generator is
connected to MG through a three-phase to three-phase matrix converter, and
the droop control method is applied for controlling the voltage and
frequency of MG. In addition, a method is introduced for voltage and
frequency control of micro-turbines in the transition state from gridconnected mode to islanding mode. A novel switching strategy of the matrix
converter is used for converting the high-frequency output voltage of the
micro-turbine to the grid-side frequency of the utility system. Moreover,
using the switching strategy, the low-order harmonics in the output current
and voltage are not produced, and consequently, the size of the output filter
would be reduced. In fact, the suggested control strategy is load-independent
and has no frequency conversion restrictions. The proposed approach for
voltage and frequency regulation demonstrates exceptional performance and
favorable response across various load alteration scenarios. The suggested
strategy is examined in several scenarios in the MG test systems, and the
simulation results are addressed.
More Related Content
Similar to Bridging the gap between the semantic web and big data: answering SPARQL queries over NoSQL databases
Development of a Web based Shopping Cart using the Mongo DB Database for Huma...AI Publications
The databases in use today are of SQL-type. This has its drawbacks such as unnecessary complex queries, rigid schema, non-asynchronous persistence and they are definitely not object oriented. Moreover, SQL-shopping cart is expensive by requiring more programs to function. Therefore, the development of a modern shopping cart using MongoDB will eradicate these set backs. The main aim of this study is to design and execute a modern e-commerce shopping cart using MongoDB database. The method used here is the agile development methodology. Stages involved here include: Brainstorm, Design, development stage, Quality Assurance, deployment and Cycle. The User interface is written with HTML, CSS and JavaScript. The HTML (Hyper Text markup language) is used to create the web pages involved, including the forms through which the user supplies input to the system. Each item in the web page is well labeled to optimize user friendliness. The CSS (cascading Style Sheet) is used to create a mobile-friendly, responsive interface to enable mobile devices to seamlessly use the system.The developed shopping cart will save time and effort for programmers rather than using SQL tools with all the labors with it.
OUTCOME ANALYSIS IN ACADEMIC INSTITUTIONS USING NEO4Jijcsity
Databases are an integral part of a computing system and users heavily rely on the services they provide.When interact with a computing system, we expect that data be stored for future use, that the data is able to be looked up fastly, and we can perform complex queries against the data stored in the database. Many
different emerging database types available for use such as relational databases, object databases, keyvalue databases, graph databases, and RDF databases. Each type of database provides unique qualities that have applications in certain domains. Our work aims to investigate and compare the performance and
scalability of relational databases to graph databases in terms of handling multilevel queries such as finding the impact of a particular subject with the working area of pass out students. MySQL was chosen as the relational database, Neo4j as the graph database.
DATABASE SYSTEMS PERFORMANCE EVALUATION FOR IOT APPLICATIONSijdms
ABSTRACT
The amount of data stored in IoT databases increases as the IoT applications extend throughout smart city appliances, industry and agriculture. Contemporary database systems must process huge amounts of sensory and actuator data in real-time or interactively. Facing this first wave of IoT revolution, database vendors struggle day-by-day in order to gain more market share, develop new capabilities and attempt to overcome the disadvantages of previous releases, while providing features for the IoT.
There are two popular database types: The Relational Database Management Systems and NoSQL databases, with NoSQL gaining ground on IoT data storage. In the context of this paper these two types are examined. Focusing on open source databases, the authors experiment on IoT data sets and pose an answer to the question which one performs better than the other. It is a comparative study on the performance of the commonly market used open source databases, presenting results for the NoSQL MongoDB database and SQL databases of MySQL and PostgreSQL
The growth of data and its effi cient handling is becoming more popular trend in recent years bringing
new challenges to explore new avenues. Data analytics can be done more effi ciently with the availability of
distributed architecture of “Not Only SQL” NoSQL databases.
Analysis and evaluation of riak kv cluster environment using basho benchStevenChike
Many institutions and companies with technological development have been producing large size of structured and unstructured data. Therefore, we need special databases to deal with these data and thus emerged NoSQL databases. They are widely used in the cloud databases and the distributed systems. In the era of big data, those databases provide a scalable high availability solution. So we need new architectures to try to meet the need to store more and more different kinds of different data. In order to arrive at a good structure of large and diverse data, this structure must be tested and analyzed in depth with the use of different benchmark tools. In this paper, we experiment the Riak key-value database to measure their performance in terms of throughput and latency, where huge amounts of data are stored and retrieved in different sizes in a distributed database environment. Throughput and latency of the NoSQL database over different types of experiments and different sizes of data are compared and then results were discussed.
EVALUATING CASSANDRA, MONGO DB LIKE NOSQL DATASETS USING HADOOP STREAMINGijiert bestjournal
An unstructured data poses challenges to storing da ta. Experts estimate that 80 to 90 percent of the d ata in any organization is unstructured. And the amount of uns tructured data in enterprises is growing significan tly� often many times faster than structured databases are gro wing. As structured data is existing in table forma t i,e having proper scheme but unstructured data is schema less database So it�s directly signifying the importance of NoSQL storage Model and Map Reduce platform. For processi ng unstructured data,where in existing it is given to Cassandra dataset. Here in present system along wit h Cassandra dataset,Mongo DB is to be implemented. As Mongo DB provide flexible data model and large amou nt of options for querying unstructured data. Where as Cassandra model their data in such a way as to mini mize the total number of queries through more caref ul planning and renormalizations. It offers basic secondary ind exes but for the best performance it�s recommended to model our data as to use them infrequently. So to process
MONGODB VS MYSQL: A COMPARATIVE STUDY OF PERFORMANCE IN SUPER MARKET MANAGEME...ijcsity
A database is information collection that is organized in tables so that it can easily be accessed, managed, and updated. It is the collection of tables, schemas, queries, reports, views and other objects. The data are typically organized to model in a way that supports processes requiring information, such as modelling to find a hotel with availability of rooms, thus the people can easily locate the hotels with vacancies. There are many databases commonly, relational and non relational databases. Relational databases usually work with structured data and non relational databases are work with semi structured data. In this paper, the performance evaluation of MySQL and MongoDB is performed where MySQL is an example of relational database and MongoDB is an example of non relational databases. A relational database is a data structure that allows you to connect information from different 'tables', or different types of data buckets. Non-relational database stores data without explicit and structured mechanisms to link data from different buckets to one another. This paper discuss about the performance of MongoDB and MySQL in the field of Super Market Management System. A supermarket is a large form of the traditional grocery store also a self-service shop offering a wide variety of food and household products, organized in systematic manner. It is larger and has a open selection than a traditional grocery store.
MONGODB VS MYSQL: A COMPARATIVE STUDY OF PERFORMANCE IN SUPER MARKET MANAGEME...ijcsity
A database is information collection that is organized in tables so that it can easily be accessed, managed,
and updated. It is the collection of tables, schemas, queries, reports, views and other objects. The data are
typically organized to model in a way that supports processes requiring information, such as modelling to
find a hotel with availability of rooms, thus the people can easily locate the hotels with vacancies. There
are many databases commonly, relational and non relational databases. Relational databases usually work
with structured data and non relational databases are work with semi structured data. In this paper, the
performance evaluation of MySQL and MongoDB is performed where MySQL is an example of relational
database and MongoDB is an example of non relational databases. A relational database is a data
structure that allows you to connect information from different 'tables', or different types of data buckets.
Non-relational database stores data without explicit and structured mechanisms to link data from different
buckets to one another. This paper discuss about the performance of MongoDB and MySQL in the field of
Super Market Management System. A supermarket is a large form of the traditional grocery store also a
self-service shop offering a wide variety of food and household products, organized in systematic manner.
It is larger and has a open selection than a traditional grocery store.
MONGODB VS MYSQL: A COMPARATIVE STUDY OF PERFORMANCE IN SUPER MARKET MANAGEME...ijcsity
A database is information collection that is organized in tables so that it can easily be accessed, managed,
and updated. It is the collection of tables, schemas, queries, reports, views and other objects. The data are
typically organized to model in a way that supports processes requiring information, such as modelling to
find a hotel with availability of rooms, thus the people can easily locate the hotels with vacancies. There
are many databases commonly, relational and non relational databases. Relational databases usually work
with structured data and non relational databases are work with semi structured data. In this paper, the
performance evaluation of MySQL and MongoDB is performed where MySQL is an example of relational
database and MongoDB is an example of non relational databases. A relational database is a data
structure that allows you to connect information from different 'tables', or different types of data buckets.
Non-relational database stores data without explicit and structured mechanisms to link data from different
buckets to one another. This paper discuss about the performance of MongoDB and MySQL in the field of
Super Market Management System. A supermarket is a large form of the traditional grocery store also a
self-service shop offering a wide variety of food and household products, organized in systematic manner.
It is larger and has a open selection than a traditional grocery store.
Linked Data Generation for the University Data From Legacy Database dannyijwest
Web was developed to share information among the users through internet as some hyperlinked documents.
If someone wants to collect some data from the web he has to search and crawl through the documents to
fulfil his needs. Concept of Linked Data creates a breakthrough at this stage by enabling the links within
data. So, besides the web of connected documents a new web developed both for humans and machines, i.e.,
the web of connected data, simply known as Linked Data Web. Since it is a very new domain, still a very
few works has been done, specially the publication of legacy data within a University domain as Linked
Data.
The aim of this paper is to evaluate, through indexing techniques, the performance of Neo4j and
OrientDB, both graph databases technologies and to come up with strength and weaknesses os each
technology as a candidate for a storage mechanism of a graph structure. An index is a data structure that
makes the searching faster for a specific node in concern of graph databases. The referred data structure
is habitually a B-tree, however, can be a hash table or some other logic structure as well. The pivotal
point of having an index is to speed up search queries, primarily by reducing the number of nodes in a
graph or table to be examined. Graphs and graph databases are more commonly associated with social
networking or “graph search” style recommendations. Thus, these technologies remarkably are a core
technology platform for some Internet giants like Hi5, Facebook, Google, Badoo, Twitter and LinkedIn.
The key to understanding graph database systems, in the social networking context, is they give equal
prominence to storing both the data (users, favorites) and the relationships between them (who liked
what, who ‘follows’ whom, which post was liked the most, what is the shortest path to ‘reach’ who). By a
suitable application case study, in case a Twitter social networking of almost 5,000 nodes imported in
local servers (Neo4j and Orient-DB), one queried to retrieval the node with the searched data, first
without index (full scan), and second with index, aiming at comparing the response time (statement query
time) of the aforementioned graph databases and find out which of them has a better performance (the
speed of data or information retrieval) and in which case. Thereof, the main results are presented in the
section 6.
Growth of relational model: Interdependence and complementary to big data IJECEIAES
A database management system is a constant application of science that provides a platform for the creation, movement, and use of voluminous data. The area has witnessed a series of developments and technological advancements from its conventional structured database to the recent buzzword, bigdata. This paper aims to provide a complete model of a relational database that is still being widely used because of its well known ACID properties namely, atomicity, consistency, integrity and durability. Specifically, the objective of this paper is to highlight the adoption of relational model approaches by bigdata techniques. Towards addressing the reason for this in corporation, this paper qualitatively studied the advancements done over a while on the relational data model. First, the variations in the data storage layout are illustrated based on the needs of the application. Second, quick data retrieval techniques like indexing, query processing and concurrency control methods are revealed. The paper provides vital insights to appraise the efficiency of the structured database in the unstructured environment, particularly when both consistency and scalability become an issue in the working of the hybrid transactional and analytical database management system.
Representing Non-Relational Databases with Darwinian NetworksIJERA Editor
The Darwinian networks (DNs) are first introduced by Dr Butz [1] to simplify and clarify how to work with Bayesian networks (BNs). DNs can unify modeling and reasoning tasks into a single platform using the graphical manipulation of the probability tables that takes on a biological feel. From this view of the DNs, we propose a graphical library to represent and depict non-relational databases using DNs. Because of the growing of this kind of databases, we need even more tools to help in the management work, and the DNs can help with these tasks.
An elastic , effective, activety or intelligent ,graceful networking architecture layout be desired to make processing massive data. next to that ,existent network architectures be considerably incapable for
cleatting the huge data. massive data thrusts network exchequers into border it consequence with in network overcrowding ,needy achievement, then permicious employer exprtises. this offered the current state-of-the-art research affronts ,potential solutions into huge data networking notion. More specifically, present the state of networking problems into massive data connected intrequirements,capacity,running ,
data manipulating also will introduce the architectures of MapReduce , Hadoop paradigm within research
requirements, fabric networks and software defined networks which utilizized into making today’s idly growing digital world and compare and contrast into identify relevant drawbacks and solutions.
Bibliometric analysis highlighting the role of women in addressing climate ch...IJECEIAES
Fossil fuel consumption increased quickly, contributing to climate change
that is evident in unusual flooding and draughts, and global warming. Over
the past ten years, women's involvement in society has grown dramatically,
and they succeeded in playing a noticeable role in reducing climate change.
A bibliometric analysis of data from the last ten years has been carried out to
examine the role of women in addressing the climate change. The analysis's
findings discussed the relevant to the sustainable development goals (SDGs),
particularly SDG 7 and SDG 13. The results considered contributions made
by women in the various sectors while taking geographic dispersion into
account. The bibliometric analysis delves into topics including women's
leadership in environmental groups, their involvement in policymaking, their
contributions to sustainable development projects, and the influence of
gender diversity on attempts to mitigate climate change. This study's results
highlight how women have influenced policies and actions related to climate
change, point out areas of research deficiency and recommendations on how
to increase role of the women in addressing the climate change and
achieving sustainability. To achieve more successful results, this initiative
aims to highlight the significance of gender equality and encourage
inclusivity in climate change decision-making processes.
Voltage and frequency control of microgrid in presence of micro-turbine inter...IJECEIAES
The active and reactive load changes have a significant impact on voltage
and frequency. In this paper, in order to stabilize the microgrid (MG) against
load variations in islanding mode, the active and reactive power of all
distributed generators (DGs), including energy storage (battery), diesel
generator, and micro-turbine, are controlled. The micro-turbine generator is
connected to MG through a three-phase to three-phase matrix converter, and
the droop control method is applied for controlling the voltage and
frequency of MG. In addition, a method is introduced for voltage and
frequency control of micro-turbines in the transition state from gridconnected mode to islanding mode. A novel switching strategy of the matrix
converter is used for converting the high-frequency output voltage of the
micro-turbine to the grid-side frequency of the utility system. Moreover,
using the switching strategy, the low-order harmonics in the output current
and voltage are not produced, and consequently, the size of the output filter
would be reduced. In fact, the suggested control strategy is load-independent
and has no frequency conversion restrictions. The proposed approach for
voltage and frequency regulation demonstrates exceptional performance and
favorable response across various load alteration scenarios. The suggested
strategy is examined in several scenarios in the MG test systems, and the
simulation results are addressed.
Enhancing battery system identification: nonlinear autoregressive modeling fo...IJECEIAES
Precisely characterizing Li-ion batteries is essential for optimizing their
performance, enhancing safety, and prolonging their lifespan across various
applications, such as electric vehicles and renewable energy systems. This
article introduces an innovative nonlinear methodology for system
identification of a Li-ion battery, employing a nonlinear autoregressive with
exogenous inputs (NARX) model. The proposed approach integrates the
benefits of nonlinear modeling with the adaptability of the NARX structure,
facilitating a more comprehensive representation of the intricate
electrochemical processes within the battery. Experimental data collected
from a Li-ion battery operating under diverse scenarios are employed to
validate the effectiveness of the proposed methodology. The identified
NARX model exhibits superior accuracy in predicting the battery's behavior
compared to traditional linear models. This study underscores the
importance of accounting for nonlinearities in battery modeling, providing
insights into the intricate relationships between state-of-charge, voltage, and
current under dynamic conditions.
Smart grid deployment: from a bibliometric analysis to a surveyIJECEIAES
Smart grids are one of the last decades' innovations in electrical energy.
They bring relevant advantages compared to the traditional grid and
significant interest from the research community. Assessing the field's
evolution is essential to propose guidelines for facing new and future smart
grid challenges. In addition, knowing the main technologies involved in the
deployment of smart grids (SGs) is important to highlight possible
shortcomings that can be mitigated by developing new tools. This paper
contributes to the research trends mentioned above by focusing on two
objectives. First, a bibliometric analysis is presented to give an overview of
the current research level about smart grid deployment. Second, a survey of
the main technological approaches used for smart grid implementation and
their contributions are highlighted. To that effect, we searched the Web of
Science (WoS), and the Scopus databases. We obtained 5,663 documents
from WoS and 7,215 from Scopus on smart grid implementation or
deployment. With the extraction limitation in the Scopus database, 5,872 of
the 7,215 documents were extracted using a multi-step process. These two
datasets have been analyzed using a bibliometric tool called bibliometrix.
The main outputs are presented with some recommendations for future
research.
Use of analytical hierarchy process for selecting and prioritizing islanding ...IJECEIAES
One of the problems that are associated to power systems is islanding
condition, which must be rapidly and properly detected to prevent any
negative consequences on the system's protection, stability, and security.
This paper offers a thorough overview of several islanding detection
strategies, which are divided into two categories: classic approaches,
including local and remote approaches, and modern techniques, including
techniques based on signal processing and computational intelligence.
Additionally, each approach is compared and assessed based on several
factors, including implementation costs, non-detected zones, declining
power quality, and response times using the analytical hierarchy process
(AHP). The multi-criteria decision-making analysis shows that the overall
weight of passive methods (24.7%), active methods (7.8%), hybrid methods
(5.6%), remote methods (14.5%), signal processing-based methods (26.6%),
and computational intelligent-based methods (20.8%) based on the
comparison of all criteria together. Thus, it can be seen from the total weight
that hybrid approaches are the least suitable to be chosen, while signal
processing-based methods are the most appropriate islanding detection
method to be selected and implemented in power system with respect to the
aforementioned factors. Using Expert Choice software, the proposed
hierarchy model is studied and examined.
Enhancing of single-stage grid-connected photovoltaic system using fuzzy logi...IJECEIAES
The power generated by photovoltaic (PV) systems is influenced by
environmental factors. This variability hampers the control and utilization of
solar cells' peak output. In this study, a single-stage grid-connected PV
system is designed to enhance power quality. Our approach employs fuzzy
logic in the direct power control (DPC) of a three-phase voltage source
inverter (VSI), enabling seamless integration of the PV connected to the
grid. Additionally, a fuzzy logic-based maximum power point tracking
(MPPT) controller is adopted, which outperforms traditional methods like
incremental conductance (INC) in enhancing solar cell efficiency and
minimizing the response time. Moreover, the inverter's real-time active and
reactive power is directly managed to achieve a unity power factor (UPF).
The system's performance is assessed through MATLAB/Simulink
implementation, showing marked improvement over conventional methods,
particularly in steady-state and varying weather conditions. For solar
irradiances of 500 and 1,000 W/m2
, the results show that the proposed
method reduces the total harmonic distortion (THD) of the injected current
to the grid by approximately 46% and 38% compared to conventional
methods, respectively. Furthermore, we compare the simulation results with
IEEE standards to evaluate the system's grid compatibility.
Enhancing photovoltaic system maximum power point tracking with fuzzy logic-b...IJECEIAES
Photovoltaic systems have emerged as a promising energy resource that
caters to the future needs of society, owing to their renewable, inexhaustible,
and cost-free nature. The power output of these systems relies on solar cell
radiation and temperature. In order to mitigate the dependence on
atmospheric conditions and enhance power tracking, a conventional
approach has been improved by integrating various methods. To optimize
the generation of electricity from solar systems, the maximum power point
tracking (MPPT) technique is employed. To overcome limitations such as
steady-state voltage oscillations and improve transient response, two
traditional MPPT methods, namely fuzzy logic controller (FLC) and perturb
and observe (P&O), have been modified. This research paper aims to
simulate and validate the step size of the proposed modified P&O and FLC
techniques within the MPPT algorithm using MATLAB/Simulink for
efficient power tracking in photovoltaic systems.
Adaptive synchronous sliding control for a robot manipulator based on neural ...IJECEIAES
Robot manipulators have become important equipment in production lines, medical fields, and transportation. Improving the quality of trajectory tracking for
robot hands is always an attractive topic in the research community. This is a
challenging problem because robot manipulators are complex nonlinear systems
and are often subject to fluctuations in loads and external disturbances. This
article proposes an adaptive synchronous sliding control scheme to improve trajectory tracking performance for a robot manipulator. The proposed controller
ensures that the positions of the joints track the desired trajectory, synchronize
the errors, and significantly reduces chattering. First, the synchronous tracking
errors and synchronous sliding surfaces are presented. Second, the synchronous
tracking error dynamics are determined. Third, a robust adaptive control law is
designed,the unknown components of the model are estimated online by the neural network, and the parameters of the switching elements are selected by fuzzy
logic. The built algorithm ensures that the tracking and approximation errors
are ultimately uniformly bounded (UUB). Finally, the effectiveness of the constructed algorithm is demonstrated through simulation and experimental results.
Simulation and experimental results show that the proposed controller is effective with small synchronous tracking errors, and the chattering phenomenon is
significantly reduced.
Remote field-programmable gate array laboratory for signal acquisition and de...IJECEIAES
A remote laboratory utilizing field-programmable gate array (FPGA) technologies enhances students’ learning experience anywhere and anytime in embedded system design. Existing remote laboratories prioritize hardware access and visual feedback for observing board behavior after programming, neglecting comprehensive debugging tools to resolve errors that require internal signal acquisition. This paper proposes a novel remote embeddedsystem design approach targeting FPGA technologies that are fully interactive via a web-based platform. Our solution provides FPGA board access and debugging capabilities beyond the visual feedback provided by existing remote laboratories. We implemented a lab module that allows users to seamlessly incorporate into their FPGA design. The module minimizes hardware resource utilization while enabling the acquisition of a large number of data samples from the signal during the experiments by adaptively compressing the signal prior to data transmission. The results demonstrate an average compression ratio of 2.90 across three benchmark signals, indicating efficient signal acquisition and effective debugging and analysis. This method allows users to acquire more data samples than conventional methods. The proposed lab allows students to remotely test and debug their designs, bridging the gap between theory and practice in embedded system design.
Detecting and resolving feature envy through automated machine learning and m...IJECEIAES
Efficiently identifying and resolving code smells enhances software project quality. This paper presents a novel solution, utilizing automated machine learning (AutoML) techniques, to detect code smells and apply move method refactoring. By evaluating code metrics before and after refactoring, we assessed its impact on coupling, complexity, and cohesion. Key contributions of this research include a unique dataset for code smell classification and the development of models using AutoGluon for optimal performance. Furthermore, the study identifies the top 20 influential features in classifying feature envy, a well-known code smell, stemming from excessive reliance on external classes. We also explored how move method refactoring addresses feature envy, revealing reduced coupling and complexity, and improved cohesion, ultimately enhancing code quality. In summary, this research offers an empirical, data-driven approach, integrating AutoML and move method refactoring to optimize software project quality. Insights gained shed light on the benefits of refactoring on code quality and the significance of specific features in detecting feature envy. Future research can expand to explore additional refactoring techniques and a broader range of code metrics, advancing software engineering practices and standards.
Smart monitoring technique for solar cell systems using internet of things ba...IJECEIAES
Rapidly and remotely monitoring and receiving the solar cell systems status parameters, solar irradiance, temperature, and humidity, are critical issues in enhancement their efficiency. Hence, in the present article an improved smart prototype of internet of things (IoT) technique based on embedded system through NodeMCU ESP8266 (ESP-12E) was carried out experimentally. Three different regions at Egypt; Luxor, Cairo, and El-Beheira cities were chosen to study their solar irradiance profile, temperature, and humidity by the proposed IoT system. The monitoring data of solar irradiance, temperature, and humidity were live visualized directly by Ubidots through hypertext transfer protocol (HTTP) protocol. The measured solar power radiation in Luxor, Cairo, and El-Beheira ranged between 216-1000, 245-958, and 187-692 W/m 2 respectively during the solar day. The accuracy and rapidity of obtaining monitoring results using the proposed IoT system made it a strong candidate for application in monitoring solar cell systems. On the other hand, the obtained solar power radiation results of the three considered regions strongly candidate Luxor and Cairo as suitable places to build up a solar cells system station rather than El-Beheira.
An efficient security framework for intrusion detection and prevention in int...IJECEIAES
Over the past few years, the internet of things (IoT) has advanced to connect billions of smart devices to improve quality of life. However, anomalies or malicious intrusions pose several security loopholes, leading to performance degradation and threat to data security in IoT operations. Thereby, IoT security systems must keep an eye on and restrict unwanted events from occurring in the IoT network. Recently, various technical solutions based on machine learning (ML) models have been derived towards identifying and restricting unwanted events in IoT. However, most ML-based approaches are prone to miss-classification due to inappropriate feature selection. Additionally, most ML approaches applied to intrusion detection and prevention consider supervised learning, which requires a large amount of labeled data to be trained. Consequently, such complex datasets are impossible to source in a large network like IoT. To address this problem, this proposed study introduces an efficient learning mechanism to strengthen the IoT security aspects. The proposed algorithm incorporates supervised and unsupervised approaches to improve the learning models for intrusion detection and mitigation. Compared with the related works, the experimental outcome shows that the model performs well in a benchmark dataset. It accomplishes an improved detection accuracy of approximately 99.21%.
Developing a smart system for infant incubators using the internet of things ...IJECEIAES
This research is developing an incubator system that integrates the internet of things and artificial intelligence to improve care for premature babies. The system workflow starts with sensors that collect data from the incubator. Then, the data is sent in real-time to the internet of things (IoT) broker eclipse mosquito using the message queue telemetry transport (MQTT) protocol version 5.0. After that, the data is stored in a database for analysis using the long short-term memory network (LSTM) method and displayed in a web application using an application programming interface (API) service. Furthermore, the experimental results produce as many as 2,880 rows of data stored in the database. The correlation coefficient between the target attribute and other attributes ranges from 0.23 to 0.48. Next, several experiments were conducted to evaluate the model-predicted value on the test data. The best results are obtained using a two-layer LSTM configuration model, each with 60 neurons and a lookback setting 6. This model produces an R 2 value of 0.934, with a root mean square error (RMSE) value of 0.015 and a mean absolute error (MAE) of 0.008. In addition, the R 2 value was also evaluated for each attribute used as input, with a result of values between 0.590 and 0.845.
A review on internet of things-based stingless bee's honey production with im...IJECEIAES
Honey is produced exclusively by honeybees and stingless bees which both are well adapted to tropical and subtropical regions such as Malaysia. Stingless bees are known for producing small amounts of honey and are known for having a unique flavor profile. Problem identified that many stingless bees collapsed due to weather, temperature and environment. It is critical to understand the relationship between the production of stingless bee honey and environmental conditions to improve honey production. Thus, this paper presents a review on stingless bee's honey production and prediction modeling. About 54 previous research has been analyzed and compared in identifying the research gaps. A framework on modeling the prediction of stingless bee honey is derived. The result presents the comparison and analysis on the internet of things (IoT) monitoring systems, honey production estimation, convolution neural networks (CNNs), and automatic identification methods on bee species. It is identified based on image detection method the top best three efficiency presents CNN is at 98.67%, densely connected convolutional networks with YOLO v3 is 97.7%, and DenseNet201 convolutional networks 99.81%. This study is significant to assist the researcher in developing a model for predicting stingless honey produced by bee's output, which is important for a stable economy and food security.
A trust based secure access control using authentication mechanism for intero...IJECEIAES
The internet of things (IoT) is a revolutionary innovation in many aspects of our society including interactions, financial activity, and global security such as the military and battlefield internet. Due to the limited energy and processing capacity of network devices, security, energy consumption, compatibility, and device heterogeneity are the long-term IoT problems. As a result, energy and security are critical for data transmission across edge and IoT networks. Existing IoT interoperability techniques need more computation time, have unreliable authentication mechanisms that break easily, lose data easily, and have low confidentiality. In this paper, a key agreement protocol-based authentication mechanism for IoT devices is offered as a solution to this issue. This system makes use of information exchange, which must be secured to prevent access by unauthorized users. Using a compact contiki/cooja simulator, the performance and design of the suggested framework are validated. The simulation findings are evaluated based on detection of malicious nodes after 60 minutes of simulation. The suggested trust method, which is based on privacy access control, reduced packet loss ratio to 0.32%, consumed 0.39% power, and had the greatest average residual energy of 0.99 mJoules at 10 nodes.
Fuzzy linear programming with the intuitionistic polygonal fuzzy numbersIJECEIAES
In real world applications, data are subject to ambiguity due to several factors; fuzzy sets and fuzzy numbers propose a great tool to model such ambiguity. In case of hesitation, the complement of a membership value in fuzzy numbers can be different from the non-membership value, in which case we can model using intuitionistic fuzzy numbers as they provide flexibility by defining both a membership and a non-membership functions. In this article, we consider the intuitionistic fuzzy linear programming problem with intuitionistic polygonal fuzzy numbers, which is a generalization of the previous polygonal fuzzy numbers found in the literature. We present a modification of the simplex method that can be used to solve any general intuitionistic fuzzy linear programming problem after approximating the problem by an intuitionistic polygonal fuzzy number with n edges. This method is given in a simple tableau formulation, and then applied on numerical examples for clarity.
The performance of artificial intelligence in prostate magnetic resonance im...IJECEIAES
Prostate cancer is the predominant form of cancer observed in men worldwide. The application of magnetic resonance imaging (MRI) as a guidance tool for conducting biopsies has been established as a reliable and well-established approach in the diagnosis of prostate cancer. The diagnostic performance of MRI-guided prostate cancer diagnosis exhibits significant heterogeneity due to the intricate and multi-step nature of the diagnostic pathway. The development of artificial intelligence (AI) models, specifically through the utilization of machine learning techniques such as deep learning, is assuming an increasingly significant role in the field of radiology. In the realm of prostate MRI, a considerable body of literature has been dedicated to the development of various AI algorithms. These algorithms have been specifically designed for tasks such as prostate segmentation, lesion identification, and classification. The overarching objective of these endeavors is to enhance diagnostic performance and foster greater agreement among different observers within MRI scans for the prostate. This review article aims to provide a concise overview of the application of AI in the field of radiology, with a specific focus on its utilization in prostate MRI.
Seizure stage detection of epileptic seizure using convolutional neural networksIJECEIAES
According to the World Health Organization (WHO), seventy million individuals worldwide suffer from epilepsy, a neurological disorder. While electroencephalography (EEG) is crucial for diagnosing epilepsy and monitoring the brain activity of epilepsy patients, it requires a specialist to examine all EEG recordings to find epileptic behavior. This procedure needs an experienced doctor, and a precise epilepsy diagnosis is crucial for appropriate treatment. To identify epileptic seizures, this study employed a convolutional neural network (CNN) based on raw scalp EEG signals to discriminate between preictal, ictal, postictal, and interictal segments. The possibility of these characteristics is explored by examining how well timedomain signals work in the detection of epileptic signals using intracranial Freiburg Hospital (FH), scalp Children's Hospital Boston-Massachusetts Institute of Technology (CHB-MIT) databases, and Temple University Hospital (TUH) EEG. To test the viability of this approach, two types of experiments were carried out. Firstly, binary class classification (preictal, ictal, postictal each versus interictal) and four-class classification (interictal versus preictal versus ictal versus postictal). The average accuracy for stage detection using CHB-MIT database was 84.4%, while the Freiburg database's time-domain signals had an accuracy of 79.7% and the highest accuracy of 94.02% for classification in the TUH EEG database when comparing interictal stage to preictal stage.
Analysis of driving style using self-organizing maps to analyze driver behaviorIJECEIAES
Modern life is strongly associated with the use of cars, but the increase in acceleration speeds and their maneuverability leads to a dangerous driving style for some drivers. In these conditions, the development of a method that allows you to track the behavior of the driver is relevant. The article provides an overview of existing methods and models for assessing the functioning of motor vehicles and driver behavior. Based on this, a combined algorithm for recognizing driving style is proposed. To do this, a set of input data was formed, including 20 descriptive features: About the environment, the driver's behavior and the characteristics of the functioning of the car, collected using OBD II. The generated data set is sent to the Kohonen network, where clustering is performed according to driving style and degree of danger. Getting the driving characteristics into a particular cluster allows you to switch to the private indicators of an individual driver and considering individual driving characteristics. The application of the method allows you to identify potentially dangerous driving styles that can prevent accidents.
Hyperspectral object classification using hybrid spectral-spatial fusion and ...IJECEIAES
Because of its spectral-spatial and temporal resolution of greater areas, hyperspectral imaging (HSI) has found widespread application in the field of object classification. The HSI is typically used to accurately determine an object's physical characteristics as well as to locate related objects with appropriate spectral fingerprints. As a result, the HSI has been extensively applied to object identification in several fields, including surveillance, agricultural monitoring, environmental research, and precision agriculture. However, because of their enormous size, objects require a lot of time to classify; for this reason, both spectral and spatial feature fusion have been completed. The existing classification strategy leads to increased misclassification, and the feature fusion method is unable to preserve semantic object inherent features; This study addresses the research difficulties by introducing a hybrid spectral-spatial fusion (HSSF) technique to minimize feature size while maintaining object intrinsic qualities; Lastly, a soft-margins kernel is proposed for multi-layer deep support vector machine (MLDSVM) to reduce misclassification. The standard Indian pines dataset is used for the experiment, and the outcome demonstrates that the HSSF-MLDSVM model performs substantially better in terms of accuracy and Kappa coefficient.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
Event Management System Vb Net Project Report.pdfKamal Acharya
In present era, the scopes of information technology growing with a very fast .We do not see any are untouched from this industry. The scope of information technology has become wider includes: Business and industry. Household Business, Communication, Education, Entertainment, Science, Medicine, Engineering, Distance Learning, Weather Forecasting. Carrier Searching and so on.
My project named “Event Management System” is software that store and maintained all events coordinated in college. It also helpful to print related reports. My project will help to record the events coordinated by faculties with their Name, Event subject, date & details in an efficient & effective ways.
In my system we have to make a system by which a user can record all events coordinated by a particular faculty. In our proposed system some more featured are added which differs it from the existing system such as security.
TECHNICAL TRAINING MANUAL GENERAL FAMILIARIZATION COURSEDuvanRamosGarzon1
AIRCRAFT GENERAL
The Single Aisle is the most advanced family aircraft in service today, with fly-by-wire flight controls.
The A318, A319, A320 and A321 are twin-engine subsonic medium range aircraft.
The family offers a choice of engines
Overview of the fundamental roles in Hydropower generation and the components involved in wider Electrical Engineering.
This paper presents the design and construction of hydroelectric dams from the hydrologist’s survey of the valley before construction, all aspects and involved disciplines, fluid dynamics, structural engineering, generation and mains frequency regulation to the very transmission of power through the network in the United Kingdom.
Author: Robbie Edward Sayers
Collaborators and co editors: Charlie Sims and Connor Healey.
(C) 2024 Robbie E. Sayers
Water scarcity is the lack of fresh water resources to meet the standard water demand. There are two type of water scarcity. One is physical. The other is economic water scarcity.
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
About
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Technical Specifications
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
Key Features
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface
• Compatible with MAFI CCR system
• Copatiable with IDM8000 CCR
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
Application
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Democratizing Fuzzing at Scale by Abhishek Aryaabh.arya
Presented at NUS: Fuzzing and Software Security Summer School 2024
This keynote talks about the democratization of fuzzing at scale, highlighting the collaboration between open source communities, academia, and industry to advance the field of fuzzing. It delves into the history of fuzzing, the development of scalable fuzzing platforms, and the empowerment of community-driven research. The talk will further discuss recent advancements leveraging AI/ML and offer insights into the future evolution of the fuzzing landscape.
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Bridging the gap between the semantic web and big data: answering SPARQL queries over NoSQL databases
1. International Journal of Electrical and Computer Engineering (IJECE)
Vol. 12, No. 6, December 2022, pp. 6829~6835
ISSN: 2088-8708, DOI: 10.11591/ijece.v12i6.pp6829-6835 6829
Journal homepage: http://ijece.iaescore.com
Bridging the gap between the semantic web and big data:
answering SPARQL queries over NoSQL databases
Hakim El Massari, Sajida Mhammedi, Noreddine Gherabi
Lasti Laboratory, National School of Applied Sciences, Sultan Moulay Slimane University, Khouribga, Morocco
Article Info ABSTRACT
Article history:
Received Aug 24, 2021
Revised Jun 17, 2022
Accepted Jul 10, 2022
Nowadays, the database field has gotten much more diverse, and as a result,
a variety of non-relational (NoSQL) databases have been created, including
JSON-document databases and key-value stores, as well as extensible
markup language (XML) and graph databases. Due to the emergence of a
new generation of data services, some of the problems associated with big
data have been resolved. In addition, in the haste to address the challenges of
big data, NoSQL abandoned several core databases features that make them
extremely efficient and functional, for instance the global view, which
enables users to access data regardless of how it is logically structured or
physically stored in its sources. In this article, we propose a method that
allows us to query non-relational databases based on the ontology-based
access data (OBDA) framework by delegating SPARQL protocol and
resource description framework (RDF) query language (SPARQL) queries
from ontology to the NoSQL database. We applied the method on a popular
database called Couchbase and we discussed the result obtained.
Keywords:
Big data
NoSQL
Ontology
Semantic web
SPARQL
This is an open access article under the CC BY-SA license.
Corresponding Author:
Hakim El Massari
Lasti Laboratory, National School of Applied Sciences, Sultan Moulay Slimane University
Khouribga, Morocco
Email: h.elmassari@usms.ma
1. INTRODUCTION
Every business needs an efficient and reliable method to manage data accurately. Databases remain
one of the most extensively used methods for keeping client information, inventory, or any other kind of
corporate data. For decades, the greatly utilized paradigm for storing and managing structured data has been
relational data management.
For many reasons and constraints, companies are used to storing data on various database systems
(each subsidiary/organization uses its preferred system, which is often different from others). This makes
accessing data across these various and heterogenous databases a complicated task. Some works have been
performed to make this task easier, we will discover them in detail in the next section. These works fall into
two categories; the first one consists in unifying the needed databases into a single database, this results in
querying only a single database rather than many. However, this database is simply a snapshot of the used
databases and once their states change, this snapshot becomes obsolete and needs to be generated again. The
second approach is different in that it creates a logical layer upon the needed databases and allows using
high-level queries. This approach is efficient and sustainable since it gathers data in real-time from the lower
levels (databases) and allows these databases to evolve without having to create the logical layer again. Our
work focuses on this second approach that allows data access, integration, quality checking, and governance
through an ontology and its application in the semantic web world.
2. ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 12, No. 6, December 2022: 6829-6835
6830
Ontology-based access data (OBDA) [1] is one of the paradigms used to implement the second
approach, a powerful solution for high-level data access, [2] it allows users to create high-level queries that
OBDA automatically converts into low-level queries suited for the used database (DB) engines [3]. One of
the most known implementations of OBDA was Ontop, a project introduced by the Free University of
Bozen-Bolzano released under the Apache license. Ontop [4] is a project that exposes the content of arbitrary
relational databases as knowledge graphs [5]. These graphs are virtual, which means that the data remains in
their original sources rather than being moved to a new database. By simplifying the method, Ontop helps in
such circumstances by converting ontology inquiries into queries that can be effectively executed over
traditional databases. Among the limitations of Ontop is that it only supports relational databases. And we all
know that the emergence of increasingly large-scale applications, exposed the drawback of relational data
management in managing the storing and querying of big data efficiently and horizontally. Thus, that is why
in our work we have added a layer, to enable us to convert the queries to be suitable for the DB engine of not
only SQL (NoSQL) databases.
In this article we proposed a method to enable integrating the semantic web with NoSQL databases,
we are based on the OBDA Ontop framework to generate adequate queries to query a NoSQL database. We
implement our method on Couchbase server a popular JSON documents database. The structure of this paper
is as follows. In section 2, we present the literature review from previous research. The method used in
section 3. The evaluation and environment in section 4. We discuss the results and conclude the paper in
sections 5 and 6 respectively.
2. RELATED WORK
This section gives an overview of many methodologies linked to our work in this article. Other
studies have been carried out to give approaches and tools allowing integration of the semantic web and big
data. Transforming legacy data from various forms into resource description framework (RDF) is an
important first step in enabling RDF-based data integration [6]–[8]. RDF is more and more used as a hinge
format for combining disparate data sources. It offers a data model that enables the expansion of a wide
number of current terminologies and domain ontologies while still utilizing the Semantic Web's reasoning
capacity. RDF data is rapidly being published on the web, particularly in accordance with the linked data
standards [9], [10]. This data is frequently supplied from heterogeneous sources that are inaccessible to data
integration tools and search engines.
Since the 2000s, a lot of research has been devoted to the conversion of prevalent databases and data
formats to RDF. The research has been focused on relational databases [11]. There have been several efforts
to use OBDA with NoSQL. To better handle the analysis of this sort of data, Ravat et al. [12] relies on pre-
aggregation procedures. In particular, they construct a conceptual model to represent the original RDF data in
a multidimensional structure with aggregations.
On the other side, ontologies are the outcome of common knowledge that has been arranged to be
machine-readable and captures a certain view of the universe that has been clearly specified. Ontologies are
utilized in a variety of domains, including software engineering, information extraction, semantic search,
knowledge management, recommender systems, and so on. Many studies have been published about the
creating of ontology from NoSQL databases [13] and compute ontology similarity to determine the semantic
similarity of initial retrieval concepts and execute extended queries [14]–[16].
Many kinds of research have focused on the response to SPARQL protocol and RDF query
language (SPARQL) queries on NoSQL databases, we cite a few. Massari et al. [17] rely on the OBDA
mechanism Ontop to answer SPARQL queries over Couchbase. Michel et al. [18] propose a two-step
technique for converting a SPARQL query into an abstract pivot query using MongoDB to RDF mappings
expressed in xR2RML, then converting the pivot query into a concrete MongoDB query. On top of key-value
storage, Mugnier et al. [19] investigate the topic of responding to ontology-mediated queries. They define
these systems' data models and fundamental queries and provide a rule language for expressing lightweight
ontologies on top of data.
A few research papers were published related to our topic which quering NoSQL databases via
virtual knowledge graph (VKG) using ontology and SPARQL language to delegate queries and get result
from NoSQL databases. To make such databases more accessible and to enable data integration from non-
relational data sources, [20] the authors applied the well-known OBDA framework to enable for querying
arbitrary databases via a mediating ontology; the implementation was on the MongoDB database. Using the
same DB engine, Araujo et al. [21] describe a unique OBDA technique based on document-oriented NoSQL
databases. The technique employs an access interface with an expandable and adaptable intermediary
conceptual layer capable of giving access to many types of database management systems.
3. Int J Elec & Comp Eng ISSN: 2088-8708
Bridging the gap between the semantic web and big data: answering SPARQL … (Hakim El Massari)
6831
3. METHOD
3.1. Background
Since past decades, OBDA has been a widely used strategy for resolving the challenge of accessing
current data sources through scalable, effective, and efficient techniques [22]–[25]. An ontology, in the form
of a conceptual layer, provides a common language, builds the domain, covers the data source structure, and
increases the context data of unintelligible information in OBDA. Thus, Queries may be done utilizing a
high-level conceptual viewpoint since users do not need to know anything about the data sources, linkages or
encoding of the data. Connecting ontology and data sources is done by a declarative specification that
describes mappings between the ontology and the data views. R2RML, a World Wide Web Consortium
(W3C) standard for specifying mappings in an OBDA environment, was developed with this objective in
mind. Ontology and mappings provide a virtual RDF graph that can be searched using SPARQL.
In the semantic web environment, query answering is critical because it enables users and
applications to interact with ontologies and data. As a result, many query languages, such as SPARQL have
been created. The W3C defined the SPARQL query language in 2008, and it is currently supported by the
majority of RDF triple stores, which is why we chose it.
In this paper, we used the Ontop OBDA system which is used in a variety of applications. All W3C
OBDA recommendations, including OWL2 QL, SWRL, R2RML, and SPARQL, are supported by Ontop, as
support for all current relational databases. Ontop is accessible via a Protégé plugin, and a Java library that
supports the OWL API. As ontology languages, Ontop supports RDFS and OWL2QL. OWL2QL is based on
the DL-Lite family of compact description logics, ensuring that ontology inquiries may be converted into
database queries in an equivalent manner.
3.2. Onto-Couchbase architecture
To illustrate the many thoughts and concepts mentioned in this article, we propose using the OBDA
model, which consists of ontology and mappings as well as an intermediary conceptual layer, to access the
contents of a relational database primarily, but we have added a layer that enables us to access the documents
of NoSQL database. We demonstrate the onto-Couchbase system, which implements the query translation
technique based on the Ontop system, allowing us to query a NoSQL database, in this instance Couchbase,
and create a collection of JSON documents as a consequence. The following are the major components of the
Onto-Couchbase system, as shown in Figure 1. The user part consists of OWL Ontology with SPARQL
queries, then access interface consists of classes plus mapping and query adjustment, last part is a NoSQL
database Choubase server in our case. More details about each part are in the sub sub-sections below.
Figure 1. Architecture of onto-Couchbase
3.2.1. User
In this part, we create our ontology shown in Figure 2 using Protégé software [26]. Protégé it is an
open-source platform that offers a suite of tools to a growing user community for building domain models
and knowledge-based applications with ontologies. The ontology consists of all information from the
database which is about a university and all staff and courses.
3.2.2. Access interface
This part is the core of our system, where we made an interface that is a component able to convert
SPARQL queries and retrieving JSON documents from the Couchbase server. With the help of Java API of
Ontop and Couchbase server, we were able to achieve this work and make this interface a window between the
ontology and NoSQL database. The classes in our Java application, along with the mappings in Figure 3(a),
expose a virtual RDF graph, which will be interrogated using SPARQL Figure 3(b) by transforming
4. ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 12, No. 6, December 2022: 6829-6835
6832
SPARQL queries into SQL queries Figure 3(c) and get as a result N1QL query Figure 3(d) executed by
Couchbase server.
Figure 2. The ontology graphs
Figure 3. Process of querying adjustments (a) mappings, (b) SPARQL query, (c) SQL query, and (d) N1QL
The produced SQL queries are often not optimal and cannot be performed directly by Couchbase
server. As a result, we must modify the SQL syntax by including the adjustment query phase in order to
construct an N1QL query, keeping in mind that N1QL is called SQL for JSON since it appears very similar
to a SQL query. It is intended to deal with both structured and semi-structured data, and it is built on the
original SQL with enhancements to allow it to function with JSON document databases by loosening its data
model requirements. As a result, the query language preserves the benefits of SQL, such as its high-level
(declarative) nature, while also allowing it to handle the more flexible structures inherent in the semi-
structured world. On the basis of and because our DB engine does not allow slightly produced SQL dialect,
we must modify the SQL syntax. For example, the operator for string concatenation in Couchbase is ||, and
the CONCAT function in other relational databases; we utilized backticks instead of double quotation marks
in Couchbase, and because Couchbase does not support the CAST function, we removed it. Finally, the
modified SQL query is conducted over the Couchbase database and resulting in the retrieval of a JSON
document.
3.2.3. NoSQL database
We chose Couchbase server a document-oriented NoSQL database for our Onto-Couchbase system.
There are two universities in the database, labeled “uni1” and “uni2”. The University data was created at
random using a java method based on an exciting relational structure consisting of 8 tables (Student,
academic, courses, and so on). We created 2 million JSON documents, which were distributed among both
“uni1” and “uni2” 1 million for each one. The reason why we chose Couchbase server is the only NoSQL
database that supports SQL dialect which makes the process of translating SPARQL query to SQL query then
to N1QL easy. In addition, is a distributed database that combines the characteristics of relational databases,
such as SQL and ACID transactions, with the JSON flexibility and scale that characterizes NoSQL.
(a)
(b)
(c)
(d)
5. Int J Elec & Comp Eng ISSN: 2088-8708
Bridging the gap between the semantic web and big data: answering SPARQL … (Hakim El Massari)
6833
4. EVALUATION AND ENVIRONMENT
In this section, we will share the evaluation and environment chosen to make the experimental and
draw if Onto-Couchbase is an efficient solution to cover and resolve one of the big data problems which is
variety. As we describe in the previous section our system is composed of a SPARQL query and an access
interface and a NoSQL database (Couchbase). We provide the database and all files with the documentation
online, so that the experiment may be recreated.
To implement the Onto-Couchbase system we import the database to a cluster in Couchbase server,
to do so we develop a java program to fill the database with real data. Then we used 5 distinct SPARQL
queries to test the operation of the system, the queries used are (details of the queries can be found online):
i) Q1: returns all permanent professor; ii) Q2: returns all faculty members; iii) Q3: returns all person; iv) Q4:
returns all teachers; and v) Q5: returns information of students taking a course. The queries were tested on
HPE ProLiant ML350 Gen9 Server with characteristic of Processor Intel(R) Xeon(R) CPU E5-2620 v3 @
2.40GHz, 2397 MHz, 6 Core(s), 12 Logical Processor(s), 48 GB of memory, and 1TB SAS 12 Gbps Hard
drive. Our Onto-Couchbase system is developed using Java programming language based on Couchbase
server and Ontop API's.
5. RESULTS AND DISCUSSION
In this section, we present and discuss the results of the Onto-Couchbase system. Table 1
summarizes the results, and Figures 4 and 5 demonstrate the influence of the number of documents on the
execution time for the five Onto-Couchbase queries. Table 1 shows the execution time for our system based
on the number of documents returned.
Table 1. Query answering times vs. the number of documents returned in Couchbase
Q1 Q2 Q3 Q4 Q5
ETa
21572 41498 49105 50480 51030
NDRb
473 1400730 3000000 644230 502675
a: Query execution time (ET), b: Number of documents returned (NDR).
Figure 4. Execution time of all queries in milliseconds
Figure 5. Number of documents retrieved
0
10000
20000
30000
40000
50000
60000
Q1 Q2 Q3 Q4 Q5
Queries
Execution time (ms)
Execution
time
0
1000000
2000000
3000000
4000000
Q1 Q2 Q3 Q4 Q5
Queries
Number of documents returned
Number
of
documents
returned
6. ISSN: 2088-8708
Int J Elec & Comp Eng, Vol. 12, No. 6, December 2022: 6829-6835
6834
We are presently concentrating on query evaluation times. Observing the results of our experimental
research, there are many factors to consider that can be made, but the most essential one is that there is no
relationship between the number of documents returned and query execution time, concluded from Q3 and
Q5 we got exactly the same execution time even if we retrieve a different number of documents returned.
Although execution time is proportional to the complexity of the SQL query, Couchbase server supports a
variety of data processing features, including filtering, deep traversal of nested documents, querying via
relationships via JOINs or subqueries, grouping, combining result sets using operators, sorting, and
aggregating. The major cause of this is that the Ontop system produces rewrites containing complex
subqueries, consisting of unions of multiple select-project-join queries, and these kinds of queries are not
efficiently evaluated. As a result, it affects the query execution time.
Our solution illustrates that OBDA is capable of integrating some of the most prevalent NoSQL
database capabilities. The characteristics discussed in this study extend the well-known ontology-based data
access system for effective data management by allowing high-level conceptual integration. However, we
have demonstrated that using OBDA provides capabilities that goes beyond the expectations of most NoSQL
developers and consumers, such as querying NoSQL databases through ontologies. The technique used in
this study provides benefits for the research community, primarily because it employs Ontop to answer
SPARQL queries by rewriting them into SQL queries and sending them to the database. We made this
procedure easy for OBDA Systems by developing the entire project in JAVA and using APIs.
6. CONCLUSION
In this paper we tried to query heterogeneous data based on the OBDA Ontop system approach, this
is achieved by adding a layer that converts a SPARQL query to an N1QL query known by Couchbase server
which is used as a NoSQL database in our case. We built a system called Onto-Couchbase, based on the
architecture of the OBDA Ontop system and we evaluate the implementation on a real case of data using a
popular NoSQL Couchbase. The study we conducted demonstrates that Onto-Couchbase can generate
queries and get JSON documents as a consequence. We conclude that the use of semantic web and ontology
to query NoSQL databases will help to connect the semantic web with big data. In future work, we intend to
integrate our system into Protégé software as a plugin that will help end-user to exploit it.
REFERENCES
[1] K. Bereta, G. Papadakis, and M. Koubarakis, “OBDA for the web: Creating virtual RDF graphs on top of web data sources,”
arXiv:2005.11264, May 2020.
[2] E. Botoeva, D. Calvanese, B. Cogrel, J. Corman, and G. Xiao, “Ontology-based data access – beyond relational sources,”
Intelligenza Artificiale, vol. 13, no. 1, pp. 21–36, Aug. 2019, doi: 10.3233/IA-190023.
[3] T. Bagosi et al., “The ontop framework for ontology based data access,” in The Semantic Web and Web Science, Berlin,
Heidelberg, 2014, pp. 67–77. doi: 10.1007/978-3-662-45495-4_6.
[4] G. Xiao et al., “The virtual knowledge graph system ontop,” in The Semantic Web – ISWC 2020, 2020, pp. 259–277. doi:
10.1007/978-3-030-62466-8_17.
[5] D. Calvanese et al., “Ontop: Answering SPARQL queries over relational databases,” Semantic Web, vol. 8, no. 3, pp. 471–487,
Dec. 2016, doi: 10.3233/SW-160217.
[6] K. Saleh Aloufi, “Generating RDF resources from web open data portals,” Indonesian Journal of Electrical Engineering and
Computer Science (IJEECS), vol. 16, no. 3, pp. 1521–1529, Dec. 2019, doi: 10.11591/ijeecs.v16.i3.pp1521-1529.
[7] M. Buron, F. Goasdoué, I. Manolescu, and M.-L. Mugnier, “Obi-Wan: ontology-based RDF integration of heterogeneous data,”
Proceedings of the VLDB Endowment, vol. 13, no. 12, pp. 2933–2936, Aug. 2020, doi: 10.14778/3415478.3415512.
[8] J. F. Sequeda, S. H. Tirmizi, O. Corcho, and D. P. Miranker, “Survey of directly mapping SQL databases to the semantic Web,”
The Knowledge Engineering Review, vol. 26, no. 4, pp. 445–486, Dec. 2011, doi: 10.1017/S0269888911000208.
[9] L. Po, N. Bikakis, F. Desimoni, and G. Papastefanatos, “Linked data visualization: Techniques, tools, and big data,” Synthesis
Lectures on the Semantic Web: Theory and Technology, vol. 10, no. 1, pp. 1–157, Mar. 2020, doi:
10.2200/S00967ED1V01Y201911WBE019.
[10] P. Zangeneh and B. McCabe, “Ontology-based knowledge representation for industrial megaprojects analytics using linked data
and the semantic web,” Advanced Engineering Informatics, vol. 46, p. 101164, Oct. 2020, doi: 10.1016/j.aei.2020.101164.
[11] B. Lakzaei and M. Shamsfard, “Ontology learning from relational databases,” Information Sciences, vol. 577, pp. 280–297, Oct.
2021, doi: 10.1016/j.ins.2021.06.074.
[12] F. Ravat, J. Song, O. Teste, and C. Trojahn, “Efficient querying of multidimensional RDF data with aggregates: Comparing
NoSQL, RDF and relational data stores,” International Journal of Information Management, vol. 54, Oct. 2020, doi:
10.1016/j.ijinfomgt.2020.102089.
[13] S. Mhammedi, H. El Massari, and N. Gherabi, “Composition of large modular ontologies based on structure,” in Advances in
Information, Communication and Cybersecurity, 2022, pp. 144–154. doi: 10.1007/978-3-030-91738-8_14.
[14] V. Devarajan and R. Subramanian, “Analyzing semantic similarity amongst textual documents to suggest near duplicates,”
Indonesian Journal of Electrical Engineering and Computer Science (IJEECS), vol. 25, no. 3, pp. 1703–1711, Mar. 2022, doi:
10.11591/ijeecs.v25.i3.pp1703-1711.
[15] H. El Massari, N. Gherabi, S. Mhammedi, Z. Sabouri, and H. Ghandi, “Ontology-based decision tree model for prediction of
cardiovascular disease,” Indian Journal of Computer Science and Engineering, vol. 13, no. 3, pp. 851–859, Jun. 2022, doi:
10.21817/indjcse/2022/v13i3/221303143.
7. Int J Elec & Comp Eng ISSN: 2088-8708
Bridging the gap between the semantic web and big data: answering SPARQL … (Hakim El Massari)
6835
[16] M. Fariss, N. El Allali, H. Asaidi, and M. Bellouki, “A semantic web services discovery approach integrating multiple similarity
measures and k-means clustering,” Indonesian Journal of Electrical Engineering and Computer Science (IJEECS), vol. 24, no. 2,
pp. 1228–1237, Nov. 2021, doi: 10.11591/ijeecs.v24.i2.pp1228-1237.
[17] H. El Massari, S. Mhammedi, N. Gherabi, and M. Nasri, “Virtual OBDA mechanism ontop for answering SPARQL queries over
couchbase,” in Advanced Technologies for Humanity, 2022, pp. 193–205. doi: 10.1007/978-3-030-94188-8_19.
[18] F. Michel, C. Faron-Zucker, and J. Montagnat, “A mapping-based method to query mongoDB documents with SPARQL,” in
Database and Expert Systems Applications, 2016, pp. 52–67. doi: 10.1007/978-3-319-44406-2_6.
[19] M.-L. Mugnier, M.-C. Rousset, and F. Ulliana, “Ontology-mediated queries for NOSQL databases,” in AAI Conference on
Artificial Intelligence, 2016, pp. 1051–1057.
[20] E. Botoeva, D. Calvanese, B. Cogrel, M. Rezk, and G. Xiao, “OBDA beyond relational DBs: A study for mongoDB,” 2016.
[21] T. H. D. Araujo, B. T. Agena, K. R. Braghetto, and R. Wassermann, “OntoMongo - ontology-based data access for NoSQL,” in
Proceedings of the X Seminar on Ontology Research, 2017, pp. 55–66.
[22] H. El Massari, Z. Sabouri, S. Mhammedi, and N. Gherabi, “Diabetes prediction using machine learning algorithms and ontology,”
Journal of ICT Standardization, pp. 319–338, May 2022, doi: 10.13052/jicts2245-800X.10212.
[23] M. Hajji, M. Qbadou, and K. Mansouri, “Onto2DB: towards an eclipse plugin for automated database design from an ontology,”
International Journal of Electrical and Computer Engineering (IJECE), vol. 9, no. 4, pp. 3298–3306, Aug. 2019, doi:
10.11591/ijece.v9i4.pp3298-3306.
[24] R. Y. Alsalhee and A. M. Abdullah, “Building Quranic stories ontology using MappingMaster domain-specific language,”
International Journal of Electrical and Computer Engineering (IJECE), vol. 12, no. 1, pp. 684–693, Feb. 2022, doi:
10.11591/ijece.v12i1.pp684-693.
[25] H. El Massari, N. Gherabi, S. Mhammedi, H. Ghandi, F. Qanouni, and M. Bahaj, “An ontological model based on machine
learning for predicting breast cancer,” International Journal of Advanced Computer Science and Applications, vol. 13, no. 7,
2022, doi: 10.14569/IJACSA.2022.0130715.
[26] M. A. Musen, “The protégé project: A look back and a look forward,” AI Matters, vol. 1, no. 4, pp. 4–12, Jun. 2015, doi:
10.1145/2757001.2757003.
BIOGRAPHIES OF AUTHORS
Hakim El Massari received his master degree from Normal Superior School of
Abdelmalek Essaadi University, Tétouan, Morocco, in 2014. Currently, he is preparing his
Ph.D. in computer science at the National School of Applied Sciences, Sultan Moulay Slimane
University, Khouribga, Morocco. His research areas include machine learning, deep learning,
big data, semantic web, and ontology. He can be contacted at email: h.elmassari@usms.ma.
Sajida Mhammedi received her Ms Degree in Computer Engineering from
Faculty of Science and Technology, Beni Mellal Morocco, she worked as a visiting researcher
at the Sultane Moulay Slimane University, her research interests include machine learning,
semantic web, recommendation systems, ontology, and big data. She can be contacted at
email: sajida.mhammedi@usms.ac.ma.
Noreddine Gherabi is a professor of computer science with industrial and
academic experience. He holds a doctorate degree in computer science. In 2013, he worked as
a professor of computer science at Mohamed Ben Abdellah University and since 2015 has
worked as a research professor at Sultan Moulay Slimane University, Morocco. Member of the
International Association of Engineers (IAENG). Professor Gherabi having several
contributions in information systems namely: big data, semantic web, pattern recognition,
and intelligent systems. He has papers (book chapters, international journals, and
conferences/workshops), and edited books. He has served on executive and technical program
committees and as a reviewer of numerous international conferences and journals, he convened
and chaired more than 30 conferences and workshops. He is member of the editorial board of
several other renowned international journals: Co-editor in chief (Editorial Board) in the
journal “The International Journal of sports science and engineering for children” (IJSSEC).
Associate Editor in the journal International Journal of Engineering Research and Sports
Science. Reviewer in several journals/Conferences. Excellence Award, the best innovation in
science and technology 2009. His research areas include machine learning, deep learning, big
data, semantic web, and ontology. He can be contacted at email: n.gherabi@usms.ma.