Nowadays, grid systems are vital technology for programs running with high performance and problems solving with largescale
in scientific, engineering and business. In grid systems, heterogeneous computational resources and data should be shared
between independent organizations that are scatter geographically. A data grid is a kind of grid types that make relations computational
and storage resources. Data replication is an efficient way in data grid to obtain high performance and high availability by saving
numerous replicas in different locations e.g. grid sites. In this research, we propose a new architecture for dynamic Group data
replication. In our architecture, we added two components to OptorSim architecture: Group Replication Management component
(GRM) and Management of Popular Files Group component (MPFG). OptorSim developed by European Data Grid projects for
evaluate replication algorithm. By using this architecture, popular files group will be replicated in grid sites at the end of each
predefined time interval.
This document provides a survey of file replication techniques used in grid systems. It begins with an introduction to grid systems and discusses their use of replication to improve response times and reduce bandwidth consumption. It then categorizes replication techniques as static or dynamic and describes challenges of replication including maintaining consistency and overhead. The document surveys various replication strategies for different grid topologies like peer-to-peer, tree and hybrid. It evaluates strategies based on factors like access latency, bandwidth consumption and fault tolerance. Specific replication techniques are discussed for peer-to-peer architectures aimed at availability, placement strategies and balancing workloads.
The Impact of Data Replication on Job Scheduling Performance in Hierarchical ...graphhoc
In data-intensive applications data transfer is a primary cause of job execution delay. Data access time depends on bandwidth. The major bottleneck to supporting fast data access in Grids is the high latencies of Wide Area Networks and Internet. Effective scheduling can reduce the amount of data transferred across the internet by dispatching a job to where the needed data are present. Another solution is to use a data replication mechanism. Objective of dynamic replica strategies is reducing file access time which leads to reducing job runtime. In this paper we develop a job scheduling policy and a dynamic data replication strategy, called HRS (Hierarchical Replication Strategy), to improve the data access efficiencies. We study our approach and evaluate it through simulation. The results show that our algorithm has improved 12% over the current strategies
Survey on Synchronizing File Operations Along with Storage Scalable MechanismIRJET Journal
The document summarizes research on efficient file operations and storage scalability mechanisms. It discusses how data is divided into chunks and distributed to nodes for transmission in peer-to-peer networks. The proposed system aims to provide efficient load balancing, eliminate single points of failure, and ensure synchronization and security during data transmission. It uses synchronization algorithms and a hybrid distribution model combining features of peer-to-peer and client-server networks. The system is designed to securely handle insertions, deletions, splits, and concatenations of file chunks in a distributed storage system.
CONTENT BASED DATA TRANSFER MECHANISM FOR EFFICIENT BULK DATA TRANSFER IN GRI...ijgca
A new class of Data Grid infrastructure is needed to support management, transport, distributed access, and analysis of terabyte and peta byte of data collections by thousands of users. Even though some of the existing data management systems (DMS) of Grid computing infrastructures provides methodologies to handle bulk data transfer. These technologies are not usable in addressing some kind of simultaneous data
access requirements. Often, in most of the scientific computing environments, a common data will be needed to access from different locations. Further, most of such computing entities will wait for a common scientific data (such as a data belonging to an astronomical phenomenon) which will be published only
when it is available. These kinds of data access needs were not yet addressed in the design of data component Grid Access to Secondary Storage (GASS) or GridFTP. In this paper, we address an application layer content based data transfer scheme for grid computing environments. By using the
proposed scheme in a grid computing environment, we can simultaneously move bulk data in an efficient way using simple subscribe and publish mechanism.
AN ENTROPIC OPTIMIZATION TECHNIQUE IN HETEROGENEOUS GRID COMPUTING USING BION...ijcsit
This document summarizes a research paper that proposes a new method for improving both fault tolerance and load balancing in grid computing networks. The method converts the tree structure of grid computing nodes into a distributed R-tree index structure and then applies an entropy estimation technique. This entropy estimation helps discard nodes with high entropy from the tree, reducing complexity. The method then uses thresholding and control algorithms to select optimal route paths based on load balance and fault tolerance. Various optimization techniques like genetic algorithms, ant colony optimization, and particle swarm optimization are also applied to reach better solutions. Experimental results showed the proposed method improved performance over other existing methods.
MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...acijjournal
Apriori is one of the key algorithms to generate frequent itemsets. Analysing frequent itemset is a crucial
step in analysing structured data and in finding association relationship between items. This stands as an
elementary foundation to supervised learning, which encompasses classifier and feature extraction
methods. Applying this algorithm is crucial to understand the behaviour of structured data. Most of the
structured data in scientific domain are voluminous. Processing such kind of data requires state of the art
computing machines. Setting up such an infrastructure is expensive. Hence a distributed environment
such as a clustered setup is employed for tackling such scenarios. Apache Hadoop distribution is one of
the cluster frameworks in distributed environment that helps by distributing voluminous data across a
number of nodes in the framework. This paper focuses on map/reduce design and implementation of
Apriori algorithm for structured data analysis.
ANALYSIS STUDY ON CACHING AND REPLICA PLACEMENT ALGORITHM FOR CONTENT DISTRIB...ijp2p
Recently there has been significant research focus on distributed computing network massively caching
and replica placement problems for content distribution in globally. Peer-to-Peer (P2P) network provides
dynamically decentralized, self organized, scalable objects in distributed computing system. However such
networks suffer from high latency, network traffic and cache update problems. The existing caching and
replica placement techniques for placing objects across peer-to-peer network have no complete solution to
these problems. This paper presents an overview of the current challenges present in P2P overlay
networks, followed by describes briefly the analysis study of the existing algorithms and their merits and
demerits. And also suggest a new popularity based QoS-aware(Quality of Service) smart replica
placement algorithm for content distribution in peer- to-peer overlay networks which overcomes the access
latency, fault tolerance, network traffic and redundancy problems with low cost. The new algorithm
suggested is based on the outcome of the analysis study
A novel cache resolution technique for cooperative caching in wireless mobile...csandit
Cooperative caching is used in mobile ad hoc networks to reduce the latency perceived by the
mobile clients while retrieving data and to reduce the traffic load in the network. Caching also
increases the availability of data due to server disconnections. The implementation of a
cooperative caching technique essentially involves four major design considerations (i) cache
placement and resolution, which decides where to place and how to locate the cached data (ii)
Cache admission control which decides the data to be cached (iii) Cache replacement which
makes the replacement decision when the cache is full and (iv) consistency maintenance, i.e.
maintaining consistency between the data in server and cache. In this paper we propose an
effective cache resolution technique, which reduces the number of messages flooded in to the
network to find the requested data. The experimental results gives a promising result based on
the metrics of studies.
This document provides a survey of file replication techniques used in grid systems. It begins with an introduction to grid systems and discusses their use of replication to improve response times and reduce bandwidth consumption. It then categorizes replication techniques as static or dynamic and describes challenges of replication including maintaining consistency and overhead. The document surveys various replication strategies for different grid topologies like peer-to-peer, tree and hybrid. It evaluates strategies based on factors like access latency, bandwidth consumption and fault tolerance. Specific replication techniques are discussed for peer-to-peer architectures aimed at availability, placement strategies and balancing workloads.
The Impact of Data Replication on Job Scheduling Performance in Hierarchical ...graphhoc
In data-intensive applications data transfer is a primary cause of job execution delay. Data access time depends on bandwidth. The major bottleneck to supporting fast data access in Grids is the high latencies of Wide Area Networks and Internet. Effective scheduling can reduce the amount of data transferred across the internet by dispatching a job to where the needed data are present. Another solution is to use a data replication mechanism. Objective of dynamic replica strategies is reducing file access time which leads to reducing job runtime. In this paper we develop a job scheduling policy and a dynamic data replication strategy, called HRS (Hierarchical Replication Strategy), to improve the data access efficiencies. We study our approach and evaluate it through simulation. The results show that our algorithm has improved 12% over the current strategies
Survey on Synchronizing File Operations Along with Storage Scalable MechanismIRJET Journal
The document summarizes research on efficient file operations and storage scalability mechanisms. It discusses how data is divided into chunks and distributed to nodes for transmission in peer-to-peer networks. The proposed system aims to provide efficient load balancing, eliminate single points of failure, and ensure synchronization and security during data transmission. It uses synchronization algorithms and a hybrid distribution model combining features of peer-to-peer and client-server networks. The system is designed to securely handle insertions, deletions, splits, and concatenations of file chunks in a distributed storage system.
CONTENT BASED DATA TRANSFER MECHANISM FOR EFFICIENT BULK DATA TRANSFER IN GRI...ijgca
A new class of Data Grid infrastructure is needed to support management, transport, distributed access, and analysis of terabyte and peta byte of data collections by thousands of users. Even though some of the existing data management systems (DMS) of Grid computing infrastructures provides methodologies to handle bulk data transfer. These technologies are not usable in addressing some kind of simultaneous data
access requirements. Often, in most of the scientific computing environments, a common data will be needed to access from different locations. Further, most of such computing entities will wait for a common scientific data (such as a data belonging to an astronomical phenomenon) which will be published only
when it is available. These kinds of data access needs were not yet addressed in the design of data component Grid Access to Secondary Storage (GASS) or GridFTP. In this paper, we address an application layer content based data transfer scheme for grid computing environments. By using the
proposed scheme in a grid computing environment, we can simultaneously move bulk data in an efficient way using simple subscribe and publish mechanism.
AN ENTROPIC OPTIMIZATION TECHNIQUE IN HETEROGENEOUS GRID COMPUTING USING BION...ijcsit
This document summarizes a research paper that proposes a new method for improving both fault tolerance and load balancing in grid computing networks. The method converts the tree structure of grid computing nodes into a distributed R-tree index structure and then applies an entropy estimation technique. This entropy estimation helps discard nodes with high entropy from the tree, reducing complexity. The method then uses thresholding and control algorithms to select optimal route paths based on load balance and fault tolerance. Various optimization techniques like genetic algorithms, ant colony optimization, and particle swarm optimization are also applied to reach better solutions. Experimental results showed the proposed method improved performance over other existing methods.
MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...acijjournal
Apriori is one of the key algorithms to generate frequent itemsets. Analysing frequent itemset is a crucial
step in analysing structured data and in finding association relationship between items. This stands as an
elementary foundation to supervised learning, which encompasses classifier and feature extraction
methods. Applying this algorithm is crucial to understand the behaviour of structured data. Most of the
structured data in scientific domain are voluminous. Processing such kind of data requires state of the art
computing machines. Setting up such an infrastructure is expensive. Hence a distributed environment
such as a clustered setup is employed for tackling such scenarios. Apache Hadoop distribution is one of
the cluster frameworks in distributed environment that helps by distributing voluminous data across a
number of nodes in the framework. This paper focuses on map/reduce design and implementation of
Apriori algorithm for structured data analysis.
ANALYSIS STUDY ON CACHING AND REPLICA PLACEMENT ALGORITHM FOR CONTENT DISTRIB...ijp2p
Recently there has been significant research focus on distributed computing network massively caching
and replica placement problems for content distribution in globally. Peer-to-Peer (P2P) network provides
dynamically decentralized, self organized, scalable objects in distributed computing system. However such
networks suffer from high latency, network traffic and cache update problems. The existing caching and
replica placement techniques for placing objects across peer-to-peer network have no complete solution to
these problems. This paper presents an overview of the current challenges present in P2P overlay
networks, followed by describes briefly the analysis study of the existing algorithms and their merits and
demerits. And also suggest a new popularity based QoS-aware(Quality of Service) smart replica
placement algorithm for content distribution in peer- to-peer overlay networks which overcomes the access
latency, fault tolerance, network traffic and redundancy problems with low cost. The new algorithm
suggested is based on the outcome of the analysis study
A novel cache resolution technique for cooperative caching in wireless mobile...csandit
Cooperative caching is used in mobile ad hoc networks to reduce the latency perceived by the
mobile clients while retrieving data and to reduce the traffic load in the network. Caching also
increases the availability of data due to server disconnections. The implementation of a
cooperative caching technique essentially involves four major design considerations (i) cache
placement and resolution, which decides where to place and how to locate the cached data (ii)
Cache admission control which decides the data to be cached (iii) Cache replacement which
makes the replacement decision when the cache is full and (iv) consistency maintenance, i.e.
maintaining consistency between the data in server and cache. In this paper we propose an
effective cache resolution technique, which reduces the number of messages flooded in to the
network to find the requested data. The experimental results gives a promising result based on
the metrics of studies.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
Peer to peer cache resolution mechanism for mobile ad hoc networksijwmn
In this paper we investigate the problem of cache resolution in a mobile peer to peer ad hoc network. In our
vision cache resolution should satisfy the following requirements: (i) it should result in low message
overhead and (ii) the information should be retrieved with minimum delay. In this paper, we show that
these goals can be achieved by splitting the one hop neighbours in to two sets based on the transmission
range. The proposed approach reduces the number of messages flooded in to the network to find the
requested data. This scheme is fully distributed and comes at very low cost in terms of cache overhead. The
experimental results gives a promising result based on the metrics of studies
THRESHOLD BASED VM PLACEMENT TECHNIQUE FOR LOAD BALANCED RESOURCE PROVISIONIN...IJCNCJournal
The unbalancing load issue is a multi-variation, multi-imperative issue that corrupts the execution and productivity of processing assets. Workload adjusting methods give solutions of load unbalancing circumstances for two bothersome aspects over-burdening and under-stacking. Cloud computing utilizes planning and workload balancing for a virtualized environment, resource partaking in cloud foundation. These two factors must be handled in an improved way in cloud computing to accomplish ideal resource sharing. Henceforth, there requires productive resource, asset reservation for guaranteeing load advancement in the cloud. This work aims to present an incorporated resource, asset reservation, and workload adjusting calculation for effective cloud provisioning. The strategy develops a Priority-based Resource Scheduling Model to acquire the resource, asset reservation with threshold-based load balancing for improving the proficiency in cloud framework. Extending utilization of Virtual Machines through the suitable and sensible outstanding task at hand modifying is then practiced by intensely picking a job from submitting jobs using Priority-based Resource Scheduling Model to acquire resource asset reservation. Experimental evaluations represent, the proposed scheme gives better results by reducing execution time, with minimum resource cost and improved resource utilization in dynamic resource provisioning conditions.
This document provides a comparative analysis of two main hierarchical distributed hash table (DHT) systems - the homogenous design and the superpeer design. It presents an analytical framework and cost model to evaluate these designs. The analysis reveals that contrary to initial expectations, the costs incurred by the hierarchical superpeer design are not necessarily minimized. Key aspects of the two designs like load balancing, fault tolerance, and advantages/disadvantages are discussed. The document aims to help identify the better hierarchical DHT design for a given workload or application.
Data Warehouses store integrated and consistent data in a subject-oriented data repository dedicated
especially to support business intelligence processes. However, keeping these repositories updated usually
involves complex and time-consuming processes, commonly denominated as Extract-Transform-Load tasks.
These data intensive tasks normally execute in a limited time window and their computational requirements
tend to grow in time as more data is dealt with. Therefore, we believe that a grid environment could suit
rather well as support for the backbone of the technical infrastructure with the clear financial advantage of
using already acquired desktop computers normally present in the organization. This article proposes a
different approach to deal with the distribution of ETL processes in a grid environment, taking into account
not only the processing performance of its nodes but also the existing bandwidth to estimate the grid
availability in a near future and therefore optimize workflow distribution.
Java Abs Peer To Peer Design & Implementation Of A Tuple Spacencct
Final Year Projects, IEEE Projects, Final Year Projects in Chennai, Final Year IEEE Projects, final year projects, college projects, student projects, java projects, asp.net projects, software projects, software ieee projects, ieee 2009 projects, 2009 ieee projects, embedded projects, final year software projects, final year embedded projects, ieee embedded projects, matlab projects, microcontroller projects, vlsi projects, dsp projects, free projects, project review, project report, project presentation, free source code, free project report, Final Year Projects, IEEE Projects, Final Year Projects in Chennai, Final Year IEEE Projects, final year projects, college projects, student projects, java projects, asp.net projects, software projects, software ieee projects, ieee 2009 projects, 2009 ieee projects, embedded projects, final year software projects, final year embedded projects, ieee embedded projects, matlab projects, final year java projects, final year asp.net projects, final year vb.net projects, vb.net projects, c# projects, final year c# projects, electrical projects, power electronics projects, motors and drives projects, robotics projects, ieee electrical projects, ieee power electronics projects, ieee robotics projects, power system projects, power system ieee projects, engineering projects, ieee engineering projects, engineering students projects, be projects, mca projects, mtech projects, btech projects, me projects, mtech projects, college projects, polytechnic projects, real time projects, ieee projects, non ieee projects, project presentation, project ppt, project pdf, project source code, project review, final year project, final year projects
International Journal of Computational Engineering Research(IJCER) ijceronline
nternational Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
1. The document proposes using FPGA for data offloading in cloud datacenter networks to improve quality of service. It discusses how current high-performance processors are not keeping pace with required QoS and how FPGA can help balance workloads.
2. It reviews related works in cloud datacenter networking that have not considered using FPGA for acceleration and quality of service improvement. Most have focused on topology, architecture, flow scheduling, and virtualization but not FPGA-based acceleration.
3. The document argues that an FPGA-based spine-leaf model could be an alternative to traditional network models for enterprise applications like EETACP, offering benefits like lower latency, offloading, scalability, and
Cloud Based Datacenter Network Acceleration Using FPGA for Data-Offloading Onyebuchi nosiri
Currently, the high-performance processors in Spine-Leaf, Mesh, and Router layer-3 (SLMR-3) backend server domain have multiple cores, but data offloading from processor to the peripheral is not keeping pace with the required Quality of Service (QoS) needed to balance the workload on a Warehouse Scaled Computer (WSC) running a developed Enterprise Energy Tracking Analytic Cloud Portal (EETACP) data center network. High speed with low latency interconnects between the processors and Field Programmable Gate Array (FPGA) is critical for achieving performance benefits in EETACP deployment. Most of the servers in WSC architectures are running at average utilization rates and perform well under peak processing power. These servers are good candidates for FPGA processors in cloud-based data centers owing to its acceleration coherency. This paper made a strong case for cloudbased support for EETACP. An FPGA-based Spine-Leaf model is proposed to be an alternative to traditional network models for EETACP provisioning. The paper analyzed reconfigurable FPGAs, characterized a simplified process model for hyperscale FPGA cloud design description. To validate the performance, comparisons was made with two similar networks, namely DCell and BCube for enterprise application deployments. It was concluded that FPGA-based DCN acceleration for EETACP offers acceptable QoS expectations
Abstract: Energy consumption is one of the constraints in Wireless Sensor Networks (WSNs). The routing protocols are the hot areas to address quality-of-service (QoS) related issues viz. Energy consumption, network lifetime, network scalability and packet overhead. In existing system a hybrid optimization based PEGASIS-DSR optimized routing protocol (PDORP) is presented which used cache and directional transmission concept of both proactive and reactive routing protocols. The performance of PDORP has been evaluated and the results indicated that it performs better in most significant parameters. The performance of the existing method is checked when it is evaluated and validated with the nodes which are highly dynamic in nature based on the application requirement. The current system finds the trusted nodes in the case of only static environment. To overcome the issue the proposed system is applied for dynamic WSN’s with the location frequently being changed. The PDORP-LC is applied with local caching (LC) to acquire the location information so that the path learning can be dynamic without depending on the fixed location. The proposed work is performing in dynamic environment with the dynamic derivation of trusted nodes.
Keywords: local caching (LC), Wireless Sensor Networks (WSNs), PEGASIS-DSR optimized routing protocol (PDORP).
Title: Energy Efficient Optimal Paths Using PDORP-LC
Author: ADARSH KUMAR B, BIBIN CHRISTOPHER, ISSAC SAJAN, AJ DEEPA
ISSN 2350-1022
International Journal of Recent Research in Mathematics Computer Science and Information Technology
Paper Publications
International Refereed Journal of Engineering and Science (IRJES) is a peer reviewed online journal for professionals and researchers in the field of computer science. The main aim is to resolve emerging and outstanding problems revealed by recent social and technological change. IJRES provides the platform for the researchers to present and evaluate their work from both theoretical and technical aspects and to share their views.
An Efficient PDP Scheme for Distributed Cloud StorageIJMER
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
International Journal of Modern Engineering Research (IJMER) covers all the fields of engineering and science: Electrical Engineering, Mechanical Engineering, Civil Engineering, Chemical Engineering, Computer Engineering, Agricultural Engineering, Aerospace Engineering, Thermodynamics, Structural Engineering, Control Engineering, Robotics, Mechatronics, Fluid Mechanics, Nanotechnology, Simulators, Web-based Learning, Remote Laboratories, Engineering Design Methods, Education Research, Students' Satisfaction and Motivation, Global Projects, and Assessment…. And many more.
A New Efficient Cache Replacement Strategy for Named Data NetworkingIJCNCJournal
The Information-Centric Network (ICN) is a future internet architecture with efficient content retrieval and distribution. Named Data Networking (NDN) is one of the proposed architectures for ICN. NDN’s innetwork caching improves data availability, reduce retrieval delays, network load, alleviate producer load, and limit data traffic. Despite the existence of several caching decision algorithms, the fetching and distribution of contents with minimum resource utilization remains a great challenge. In this paper, we introduce a new cache replacement strategy called Enhanced Time and Frequency Cache Replacement strategy (ETFCR) where both cache hit frequency and cache retrieval time are used to select evicted data chunks. ETFCR adds time cycles between the last two requests to adjust data chunk’s popularity and cache hits. We conducted extensive simulations using the ccnSim simulator to evaluate the performance of ETFCR and compare it to that of some well-known cache replacement strategies. Simulations results show that ETFCR outperforms the other cache replacement strategies in terms of cache hit ratio, and lower content retrieval delay.
This document discusses the challenges of collecting, storing, and analyzing large volumes of internet measurement data. It examines issues such as distributed and resilient data collection, handling multi-timescale and heterogeneous data from various sources, and developing standardized tools and formats. The paper proposes the "datapository" - an internet data repository designed to address these challenges through a collaborative framework for data sharing, storage, and analysis tools. The goal is to help both network operators and researchers more effectively harness the wealth of data available.
DWDM-RAM: a data intensive Grid service architecture enabled by dynamic optic...Tal Lavian Ph.D.
The DWDM-RAM project develops an architecture for data-intensive grid services enabled by dynamic optical networks. It encapsulates optical network resources like wavelengths and lightpaths as grid services. This allows applications to schedule large data transfers using these network resources. The architecture consists of application and resource middleware layers. It is being implemented on the OMNInet photonic testbed to demonstrate on-demand and scheduled data retrieval using a dynamically switched DWDM network. In summary, DWDM-RAM schedules high-bandwidth network resources through dynamic lightpath provisioning and makes large-scale data services accessible via a grid service interface.
A presentation about how to make MySQL highly available, presented at the San Francisco MySQL Meetup (http://www.sfmysql.org/events/15760472/) on January 26th, 2011.
A video recording of this presentation is available from Ustream: http://ustre.am/fyLk
A tutorial on MySQL High Availability on the Pacemaker stack. Covers both MySQL on DRBD, and MySQL with MySQL replication.
Presented by Florian Haas and Yves Trudeau at the Percona Live MySQL Conference & Expo, 2012
Software developers love tools for coding, debugging, testing, and configuration management. The more these tools improve the How of coding, the more we see that we're behind the curve on improving the What, Why, and When. If you've been on a project that seemed vague, adrift, and endless, this talk can help. Make your projects run SMART.
The nightmare of locking, blocking and isolation levels!Boris Hristov
am sure you all know that troubleshooting problems related to locking and blocking (hey, sometimes there are deadlocks too) can be a real nightmare! In this session, you will be able to see and understand why and how locking actually works, what problems it causes and how can we use isolation levels and various other techniques to resolve them!
Just about anyone can write a basic SQL query for a table. Not everyone can write a good query though - that takes practice and knowing how to understand what the optimizer is doing with the query. Learn the basics of query optimization so you keep your application engaging the user rather then showing the progress bar as they wait on the database.
Everything You Need to Know About MySQL Group ReplicationNuno Carvalho
MySQL Group Replication is a new plugin that implements an exciting extension to the proven and long standing MySQL Replication technology. It leverages advanced distributed protocols to ultimately provide to the end user features such as data replication, high availability, split brain protection and automation.
It can be deployed in single-primary mode (default), in which primary fail-over is handled gracefully and automatically, or in multi-master mode, in which row level conflicts are detected and handled automatically as well. Regardless of the deployment mode, the end result is that this new addition provides a consistent and dependable replicated state machine, thus effectively enabling a fault-tolerant MySQL database service.
At the end of the presentation, you will be able to understand how it works, the use cases it address, its limitations and also its roadmap ahead. Moreover, you will get to know how it fits in the overall high availability roadmap at MySQL.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
Peer to peer cache resolution mechanism for mobile ad hoc networksijwmn
In this paper we investigate the problem of cache resolution in a mobile peer to peer ad hoc network. In our
vision cache resolution should satisfy the following requirements: (i) it should result in low message
overhead and (ii) the information should be retrieved with minimum delay. In this paper, we show that
these goals can be achieved by splitting the one hop neighbours in to two sets based on the transmission
range. The proposed approach reduces the number of messages flooded in to the network to find the
requested data. This scheme is fully distributed and comes at very low cost in terms of cache overhead. The
experimental results gives a promising result based on the metrics of studies
THRESHOLD BASED VM PLACEMENT TECHNIQUE FOR LOAD BALANCED RESOURCE PROVISIONIN...IJCNCJournal
The unbalancing load issue is a multi-variation, multi-imperative issue that corrupts the execution and productivity of processing assets. Workload adjusting methods give solutions of load unbalancing circumstances for two bothersome aspects over-burdening and under-stacking. Cloud computing utilizes planning and workload balancing for a virtualized environment, resource partaking in cloud foundation. These two factors must be handled in an improved way in cloud computing to accomplish ideal resource sharing. Henceforth, there requires productive resource, asset reservation for guaranteeing load advancement in the cloud. This work aims to present an incorporated resource, asset reservation, and workload adjusting calculation for effective cloud provisioning. The strategy develops a Priority-based Resource Scheduling Model to acquire the resource, asset reservation with threshold-based load balancing for improving the proficiency in cloud framework. Extending utilization of Virtual Machines through the suitable and sensible outstanding task at hand modifying is then practiced by intensely picking a job from submitting jobs using Priority-based Resource Scheduling Model to acquire resource asset reservation. Experimental evaluations represent, the proposed scheme gives better results by reducing execution time, with minimum resource cost and improved resource utilization in dynamic resource provisioning conditions.
This document provides a comparative analysis of two main hierarchical distributed hash table (DHT) systems - the homogenous design and the superpeer design. It presents an analytical framework and cost model to evaluate these designs. The analysis reveals that contrary to initial expectations, the costs incurred by the hierarchical superpeer design are not necessarily minimized. Key aspects of the two designs like load balancing, fault tolerance, and advantages/disadvantages are discussed. The document aims to help identify the better hierarchical DHT design for a given workload or application.
Data Warehouses store integrated and consistent data in a subject-oriented data repository dedicated
especially to support business intelligence processes. However, keeping these repositories updated usually
involves complex and time-consuming processes, commonly denominated as Extract-Transform-Load tasks.
These data intensive tasks normally execute in a limited time window and their computational requirements
tend to grow in time as more data is dealt with. Therefore, we believe that a grid environment could suit
rather well as support for the backbone of the technical infrastructure with the clear financial advantage of
using already acquired desktop computers normally present in the organization. This article proposes a
different approach to deal with the distribution of ETL processes in a grid environment, taking into account
not only the processing performance of its nodes but also the existing bandwidth to estimate the grid
availability in a near future and therefore optimize workflow distribution.
Java Abs Peer To Peer Design & Implementation Of A Tuple Spacencct
Final Year Projects, IEEE Projects, Final Year Projects in Chennai, Final Year IEEE Projects, final year projects, college projects, student projects, java projects, asp.net projects, software projects, software ieee projects, ieee 2009 projects, 2009 ieee projects, embedded projects, final year software projects, final year embedded projects, ieee embedded projects, matlab projects, microcontroller projects, vlsi projects, dsp projects, free projects, project review, project report, project presentation, free source code, free project report, Final Year Projects, IEEE Projects, Final Year Projects in Chennai, Final Year IEEE Projects, final year projects, college projects, student projects, java projects, asp.net projects, software projects, software ieee projects, ieee 2009 projects, 2009 ieee projects, embedded projects, final year software projects, final year embedded projects, ieee embedded projects, matlab projects, final year java projects, final year asp.net projects, final year vb.net projects, vb.net projects, c# projects, final year c# projects, electrical projects, power electronics projects, motors and drives projects, robotics projects, ieee electrical projects, ieee power electronics projects, ieee robotics projects, power system projects, power system ieee projects, engineering projects, ieee engineering projects, engineering students projects, be projects, mca projects, mtech projects, btech projects, me projects, mtech projects, college projects, polytechnic projects, real time projects, ieee projects, non ieee projects, project presentation, project ppt, project pdf, project source code, project review, final year project, final year projects
International Journal of Computational Engineering Research(IJCER) ijceronline
nternational Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
1. The document proposes using FPGA for data offloading in cloud datacenter networks to improve quality of service. It discusses how current high-performance processors are not keeping pace with required QoS and how FPGA can help balance workloads.
2. It reviews related works in cloud datacenter networking that have not considered using FPGA for acceleration and quality of service improvement. Most have focused on topology, architecture, flow scheduling, and virtualization but not FPGA-based acceleration.
3. The document argues that an FPGA-based spine-leaf model could be an alternative to traditional network models for enterprise applications like EETACP, offering benefits like lower latency, offloading, scalability, and
Cloud Based Datacenter Network Acceleration Using FPGA for Data-Offloading Onyebuchi nosiri
Currently, the high-performance processors in Spine-Leaf, Mesh, and Router layer-3 (SLMR-3) backend server domain have multiple cores, but data offloading from processor to the peripheral is not keeping pace with the required Quality of Service (QoS) needed to balance the workload on a Warehouse Scaled Computer (WSC) running a developed Enterprise Energy Tracking Analytic Cloud Portal (EETACP) data center network. High speed with low latency interconnects between the processors and Field Programmable Gate Array (FPGA) is critical for achieving performance benefits in EETACP deployment. Most of the servers in WSC architectures are running at average utilization rates and perform well under peak processing power. These servers are good candidates for FPGA processors in cloud-based data centers owing to its acceleration coherency. This paper made a strong case for cloudbased support for EETACP. An FPGA-based Spine-Leaf model is proposed to be an alternative to traditional network models for EETACP provisioning. The paper analyzed reconfigurable FPGAs, characterized a simplified process model for hyperscale FPGA cloud design description. To validate the performance, comparisons was made with two similar networks, namely DCell and BCube for enterprise application deployments. It was concluded that FPGA-based DCN acceleration for EETACP offers acceptable QoS expectations
Abstract: Energy consumption is one of the constraints in Wireless Sensor Networks (WSNs). The routing protocols are the hot areas to address quality-of-service (QoS) related issues viz. Energy consumption, network lifetime, network scalability and packet overhead. In existing system a hybrid optimization based PEGASIS-DSR optimized routing protocol (PDORP) is presented which used cache and directional transmission concept of both proactive and reactive routing protocols. The performance of PDORP has been evaluated and the results indicated that it performs better in most significant parameters. The performance of the existing method is checked when it is evaluated and validated with the nodes which are highly dynamic in nature based on the application requirement. The current system finds the trusted nodes in the case of only static environment. To overcome the issue the proposed system is applied for dynamic WSN’s with the location frequently being changed. The PDORP-LC is applied with local caching (LC) to acquire the location information so that the path learning can be dynamic without depending on the fixed location. The proposed work is performing in dynamic environment with the dynamic derivation of trusted nodes.
Keywords: local caching (LC), Wireless Sensor Networks (WSNs), PEGASIS-DSR optimized routing protocol (PDORP).
Title: Energy Efficient Optimal Paths Using PDORP-LC
Author: ADARSH KUMAR B, BIBIN CHRISTOPHER, ISSAC SAJAN, AJ DEEPA
ISSN 2350-1022
International Journal of Recent Research in Mathematics Computer Science and Information Technology
Paper Publications
International Refereed Journal of Engineering and Science (IRJES) is a peer reviewed online journal for professionals and researchers in the field of computer science. The main aim is to resolve emerging and outstanding problems revealed by recent social and technological change. IJRES provides the platform for the researchers to present and evaluate their work from both theoretical and technical aspects and to share their views.
An Efficient PDP Scheme for Distributed Cloud StorageIJMER
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
International Journal of Modern Engineering Research (IJMER) covers all the fields of engineering and science: Electrical Engineering, Mechanical Engineering, Civil Engineering, Chemical Engineering, Computer Engineering, Agricultural Engineering, Aerospace Engineering, Thermodynamics, Structural Engineering, Control Engineering, Robotics, Mechatronics, Fluid Mechanics, Nanotechnology, Simulators, Web-based Learning, Remote Laboratories, Engineering Design Methods, Education Research, Students' Satisfaction and Motivation, Global Projects, and Assessment…. And many more.
A New Efficient Cache Replacement Strategy for Named Data NetworkingIJCNCJournal
The Information-Centric Network (ICN) is a future internet architecture with efficient content retrieval and distribution. Named Data Networking (NDN) is one of the proposed architectures for ICN. NDN’s innetwork caching improves data availability, reduce retrieval delays, network load, alleviate producer load, and limit data traffic. Despite the existence of several caching decision algorithms, the fetching and distribution of contents with minimum resource utilization remains a great challenge. In this paper, we introduce a new cache replacement strategy called Enhanced Time and Frequency Cache Replacement strategy (ETFCR) where both cache hit frequency and cache retrieval time are used to select evicted data chunks. ETFCR adds time cycles between the last two requests to adjust data chunk’s popularity and cache hits. We conducted extensive simulations using the ccnSim simulator to evaluate the performance of ETFCR and compare it to that of some well-known cache replacement strategies. Simulations results show that ETFCR outperforms the other cache replacement strategies in terms of cache hit ratio, and lower content retrieval delay.
This document discusses the challenges of collecting, storing, and analyzing large volumes of internet measurement data. It examines issues such as distributed and resilient data collection, handling multi-timescale and heterogeneous data from various sources, and developing standardized tools and formats. The paper proposes the "datapository" - an internet data repository designed to address these challenges through a collaborative framework for data sharing, storage, and analysis tools. The goal is to help both network operators and researchers more effectively harness the wealth of data available.
DWDM-RAM: a data intensive Grid service architecture enabled by dynamic optic...Tal Lavian Ph.D.
The DWDM-RAM project develops an architecture for data-intensive grid services enabled by dynamic optical networks. It encapsulates optical network resources like wavelengths and lightpaths as grid services. This allows applications to schedule large data transfers using these network resources. The architecture consists of application and resource middleware layers. It is being implemented on the OMNInet photonic testbed to demonstrate on-demand and scheduled data retrieval using a dynamically switched DWDM network. In summary, DWDM-RAM schedules high-bandwidth network resources through dynamic lightpath provisioning and makes large-scale data services accessible via a grid service interface.
A presentation about how to make MySQL highly available, presented at the San Francisco MySQL Meetup (http://www.sfmysql.org/events/15760472/) on January 26th, 2011.
A video recording of this presentation is available from Ustream: http://ustre.am/fyLk
A tutorial on MySQL High Availability on the Pacemaker stack. Covers both MySQL on DRBD, and MySQL with MySQL replication.
Presented by Florian Haas and Yves Trudeau at the Percona Live MySQL Conference & Expo, 2012
Software developers love tools for coding, debugging, testing, and configuration management. The more these tools improve the How of coding, the more we see that we're behind the curve on improving the What, Why, and When. If you've been on a project that seemed vague, adrift, and endless, this talk can help. Make your projects run SMART.
The nightmare of locking, blocking and isolation levels!Boris Hristov
am sure you all know that troubleshooting problems related to locking and blocking (hey, sometimes there are deadlocks too) can be a real nightmare! In this session, you will be able to see and understand why and how locking actually works, what problems it causes and how can we use isolation levels and various other techniques to resolve them!
Just about anyone can write a basic SQL query for a table. Not everyone can write a good query though - that takes practice and knowing how to understand what the optimizer is doing with the query. Learn the basics of query optimization so you keep your application engaging the user rather then showing the progress bar as they wait on the database.
Everything You Need to Know About MySQL Group ReplicationNuno Carvalho
MySQL Group Replication is a new plugin that implements an exciting extension to the proven and long standing MySQL Replication technology. It leverages advanced distributed protocols to ultimately provide to the end user features such as data replication, high availability, split brain protection and automation.
It can be deployed in single-primary mode (default), in which primary fail-over is handled gracefully and automatically, or in multi-master mode, in which row level conflicts are detected and handled automatically as well. Regardless of the deployment mode, the end result is that this new addition provides a consistent and dependable replicated state machine, thus effectively enabling a fault-tolerant MySQL database service.
At the end of the presentation, you will be able to understand how it works, the use cases it address, its limitations and also its roadmap ahead. Moreover, you will get to know how it fits in the overall high availability roadmap at MySQL.
Java MySQL Connector & Connection Pool Features & OptimizationKenny Gryp
This talk will give an overview of the different available Java MySQL connectors (Connector/J, MariaDB Java Connector) and connection pools (Commons-DBCP, C3P0, ...).
One of the things with the default configuration of these solutions is that they are very chatty. This can have a noticeable impact on application performance and database load. I've seen many environments where over 50% of the total amount of queries are caused by such behavior. This behavior will be explained through examples seen on production systems as well as recommendations on optimization will be made.
Load balancing and transparent failover solutions will be described for both MySQL's traditional asynchronous replication and Galera based replication (Percona XtraDB Cluster or MariaDB Galera Cluster).
The document discusses testing HBase source code. It describes the required JAR files for a simple test, inserts test data into HBase tables, and measures the insertion performance. It also summarizes the entry points for HMaster and HRegionServer source code, and describes some common HBase operations like building tables, handling DDL exceptions, flushing, compacting, and splitting regions.
Mix ‘n’ Match Async and Group Replication for Advanced Replication SetupsPedro Gomes
The document discusses mixing asynchronous replication and group replication for advanced MySQL replication setups. It provides an overview of asynchronous replication, semi-synchronous replication, multi-source replication, and group replication. It then discusses some basic scenarios for mixing these technologies, such as using asynchronous replication for read scaling beyond the group's 9 member limit or aggregating data from multiple groups. The document also covers migrating from asynchronous to group replication and migrating disjoint servers into a group.
Many questions on database newsgroups and forums can be answered with uses of outer joins. Outer joins are part of the standard SQL language and supported by all RDBMS brands. Many programmers are expected to use SQL in their work, but few know how to use outer joins effectively.
Learn to use this powerful feature of SQL, increase your employability, and amaze your friends!
Karwin will explain outer joins, show examples, and demonstrate a Sudoku puzzle solver implemented in a single SQL query.
Every website wants to become successful. Few websites however undertake the basic and fundamental steps to build a rock solid foundation to ensure a scalable
MySQL Storage Engines - which do you use? TokuDB? MyRocks? InnoDB?Sveta Smirnova
"MySQL Storage Engines - which do you use? TokuDB? MyRocks? InnoDB?" session at https://www.percona.com/live/17/sessions/mysql-storage-engines-which-do-you-use-tokudb-myrocks-innodb
This document summarizes a presentation about MySQL Group Replication. The presentation discusses how Group Replication provides enhanced high availability for MySQL databases by allowing multiple MySQL servers to act as equal masters that can handle writes and remain available even if one server fails. It covers the theory behind Group Replication, how to configure and use it, and management of Group Replication deployments.
High Availability Using MySQL Group ReplicationOSSCube
MySQL Group Replication is a recent MySQL plugin that brings together group communication techniques and database replication, providing both a high availability and a multi-master update everywhere replication solution.
The PPT provide provide a broad overview of MySQL Group Replication plugin, what it can achieve and how it helps keep your MySQL databases highly available and your business up and running, without fail.
Group Replication: A Journey to the Group Communication CoreAlfranio Júnior
This document discusses the journey of MySQL Group Replication from using a third-party group communication system (Corosync) to its own built-in group communication engine (XCOM) based on Paxos algorithms. It describes the key components of MySQL Group Replication including the group communication interface, XCOM engine, and optimizations in XCOM like pipelining and batching. It also provides performance results of MySQL Group Replication with multiple writers.
Upgrading MySQL databases do not come without risk. There is no guarantee that no problems will happen if you move to a new major MySQL version.
Should we just upgrade and rollback immediately if problems occur? But what if these problems only happen a few days after migrating to this new version?
You might have a database environment that is risk-adverse, where you really have to be sure that this new MySQL version will handle the workload properly.
Examples:
- Both MySQL 5.6 and 5.7 have a lot of changes in the MySQL Optimizer. It is expected that this improves performance of my queries, but is it really the case? What if there is a performance regression? How will this affect my database performance?
- Also, there are a lot of incompatible changes which are documented in the release notes, how do I know if I'm affected by this in my workload? It's a lot to read..
- Can I go immediately from MySQL 5.5 to 5.7 and skip MySQL 5.6 even though the MySQL documentation states that this is not supported?
- Many companies have staging environments, but is there a QA team and do they really test all functionality, under a similar workload?
This presentation will show you a process, using open source tools, of these types of migrations with a focus on assessing risk and fixing any problems you might run into prior to the migration.
This process can then be used for various changes:
- MySQL upgrades for major version upgrades
- Switching storage engines
- Changing hardware architecture
Additionally, we will describe ways to do the actual migration and rollback with the least amount of downtime.
MHA (MySQL High Availability): Getting started & moving past quirksColin Charles
MHA (Master High Availability Manager) is a tool for automating MySQL master failover and slave promotion with minimal downtime. It consists of Perl scripts that monitor a replication topology and automatically switch a slave to a master if the current master fails. When a failure is detected, MHA identifies the most up-to-date slave and applies any missing binary logs to make it the new master. It then changes the slaves to replicate from the new master. MHA requires installing node packages on all servers and a manager package to coordinate monitoring and failover.
Multi Source Replication With MySQL 5.7 @ VerisureKenny Gryp
Verisure migrated their data warehouse from using Tungsten Replicator to native multi-source replication in MySQL 5.7 to simplify operations. They loaded data from production shards into the new data warehouse setup using XtraBackup backups and improved replication capacity with MySQL's parallel replication features. Some issues were encountered with replication lag reporting and crashes during the upgrade but most were resolved. Monitoring and management tools also required updates to support the new multi-source replication configuration.
We all have tasks from time to time for bulk-loading external data into MySQL. What's the best way of doing this? That's the task I faced recently when I was asked to help benchmark a multi-terrabyte database. We had to find the most efficient method to reload test data repeatedly without taking days to do it each time. In my presentation, I'll show you several alternative methods for bulk data loading, and describe the practical steps to use them efficiently. I'll cover SQL scripts, the mysqlimport tool, MySQL Workbench import, the CSV storage engine, and the Memcached API. I'll also give MySQL tuning tips for data loading, and how to use multi-threaded clients.
A Survey of File Replication Techniques In Grid SystemsEditor IJCATR
Grid is a type of parallel and distributed systems that is designed to provide reliable access to data
and computational resources in wide area networks. These resources are distributed in different geographical
locations. Efficient data sharing in global networks is complicated by erratic node failure, unreliable network
connectivity and limited bandwidth. Replication is a technique used in grid systems to improve the
applications’ response time and to reduce the bandwidth consumption. In this paper, we present a survey on
basic and new replication techniques that have been proposed by other researchers. After that, we have a full
comparative study on these replication strategies.
A Survey of File Replication Techniques In Grid SystemsEditor IJCATR
Grid is a type of parallel and distributed systems that is designed to provide reliable access to data
and computational resources in wide area networks. These resources are distributed in different geographical
locations. Efficient data sharing in global networks is complicated by erratic node failure, unreliable network
connectivity and limited bandwidth. Replication is a technique used in grid systems to improve the
applications’ response time and to reduce the bandwidth consumption. In this paper, we present a survey on
basic and new replication techniques that have been proposed by other researchers. After that, we have a full
comparative study on these replication strategies
Analyse the performance of mobile peer to Peer network using ant colony optim...IJCI JOURNAL
The document describes analyzing the performance of a mobile peer-to-peer network using ant colony optimization. It proposes using a distributed spanning tree (DST) structure to improve efficiency by reducing the large number of messages. The DST is optimized using ant colony optimization to give an optimal solution. Simulation results show the approach reduces the number of messages, average delay, and increases packet delivery ratio in the network.
ANALYSE THE PERFORMANCE OF MOBILE PEER TO PEER NETWORK USING ANT COLONY OPTIM...ijcsity
A mobile peer-to-peer computer network is the one in which each computer in the network can act as a
client or server for the other computers in the network. The communication process among the nodes in the
mobile peer to peer network requires more no of messages. Due to this large number of messages passing,
propose an interconnection structure called distributed Spanning Tree (DST) and it improves the efficiency
of the mobile peer to peer network. The proposed method improves the data availability and consistency
across the entire network and also reduces the data latency and the required number of message passes for
any specific application in the network. Further to enhance the effectiveness of the proposed system, the
DST network is optimized with the Ant Colony Optimization method. It gives the optimal solution of the
DST method and increased availability, enhanced consistency and scalability of the network. The
simulation results shows that reduces the number of message sent for any specific application and average
delay and increases the packet delivery ratio in the network.
This document summarizes an article from the International Journal of Computer Engineering and Technology (IJCET) that proposes an algorithm called Replica Placement in Graph Topology Grid (RPGTG) to optimally place data replicas in a graph-based data grid while ensuring quality of service (QoS). The algorithm aims to minimize data access time, balance load among replica servers, and avoid unnecessary replications, while restricting QoS in terms of number of hops and deadline to complete requests. The article describes how the algorithm converts the graph structure of the data grid to a hierarchical structure to better manage replica servers and proposes services to facilitate dynamic replication, including a replica catalog to track replica locations and a replica manager to perform replication
An asynchronous replication model to improve data available into a heterogene...Alexander Decker
This document summarizes a research paper that proposes an asynchronous replication model to improve data availability in heterogeneous systems. The proposed model uses a loosely coupled architecture between main and replication servers to reduce dependencies. It also supports heterogeneous systems, allowing different parts of an application to run on different systems for better performance. This makes it a cost-effective solution for data replication across different system types.
ANALYSIS STUDY ON CACHING AND REPLICA PLACEMENT ALGORITHM FOR CONTENT DISTRIB...ijp2p
Recently there has been significant research focus on distributed computing network massively caching
and replica placement problems for content distribution in globally. Peer-to-Peer (P2P) network provides
dynamically decentralized, self organized, scalable objects in distributed computing system. However such
networks suffer from high latency, network traffic and cache update problems. The existing caching and
replica placement techniques for placing objects across peer-to-peer network have no complete solution to
these problems. This paper presents an overview of the current challenges present in P2P overlay
networks, followed by describes briefly the analysis study of the existing algorithms and their merits and
demerits. And also suggest a new popularity based QoS-aware(Quality of Service) smart replica
placement algorithm for content distribution in peer- to-peer overlay networks which overcomes the access
latency, fault tolerance, network traffic and redundancy problems with low cost. The new algorithm
suggested is based on the outcome of the analysis study.
This document summarizes an article from the International Journal of Computer Engineering and Technology (IJCET) that proposes a new dynamic data replication and job scheduling strategy for data grids. The strategy aims to improve data access time and reduce bandwidth consumption by replicating data based on file popularity, storage limitations at nodes, and data category. It replicates more popular files that are in the same category as frequently accessed data to nodes close to where jobs are run. This is intended to optimize performance by locating data and jobs close together. The document provides context on related work and outlines the proposed system architecture and replication/scheduling approach.
This document discusses load balancing strategies for grid computing. It proposes a dynamic tree-based model to represent grid architecture in a hierarchical way that supports heterogeneity and scalability. It then develops a hierarchical load balancing strategy and algorithms based on neighborhood properties to decrease communication overhead. Conventional scheduling algorithms like Min-Min, Max-Min, and Sufferage are discussed but determined to ignore dynamic network status, which is important for load balancing. Genetic algorithms are also mentioned as a potential solution.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
Propose a Method to Improve Performance in Grid Environment, Using Multi-Crit...Editor IJCATR
The most important purpose of grid networks is resource subscription in a dynamic and heterogeneous environment.
They are accessible through using various methods. Subscription has mainly computational, scientific and other implications. In
order to reach grid purposes and to use available resources in grid environment, subtasks are distributed among resources and are
scheduled by considering the quality of service. It has been tried to distribute subtasks between resources in a way that maximum
QOS can be obtained. In this study, a method has been presented. In this method, three parameters; namely, sent and transferred
time between RMS and resource, process time of subtask by the resource, and the load of available tasks in resources row, have
been taken into account. In this way, multi-criteria decision is made by using TOPSIS method and this priority of the resources
are determined to assign them to subtasks. Finally, time response, as an efficient parameter, has been improved and optimized by
optimal assignment of the resources to subtasks.
Hadoop is an open source implementation of the MapReduce Framework in the realm of distributed processing.
A Hadoop cluster is a unique type of computational cluster designed for storing and analyzing large datasets
across cluster of workstations. To handle massive scale data, Hadoop exploits the Hadoop Distributed File
System termed as HDFS. The HDFS similar to most distributed file systems share a familiar problem on data
sharing and availability among compute nodes, often which leads to decrease in performance. This paper is an
experimental evaluation of Hadoop's computing performance which is made by designing a rack aware cluster
that utilizes the Hadoop’s default block placement policy to improve data availability. Additionally, an adaptive
data replication scheme that relies on access count prediction using Langrange’s interpolation is adapted to fit
the scenario. To prove, experiments were conducted on a rack aware cluster setup which significantly reduced
the task completion time, but once the volume of the data being processed increases there is a considerable
cutback in computational speeds due to update cost. Further the threshold level for balance between the update
cost and replication factor is identified and presented graphically.
Scalable and adaptive data replica placement for geo distributed cloud storagesVenkat Projects
This document discusses scalable and adaptive data replica placement schemes for geo-distributed cloud storage systems. It proposes using overlapping community discovery and adjustment to determine where to place data replicas in a distributed way. An offline scheme uses average read/write rates to find an initial replica placement solution. An online scheme then adaptively handles bursty requests without completely overriding the existing placement. Performance evaluations using real-world data show the approaches reduce inter-node traffic and system overhead compared to traditional centralized placement schemes.
This document summarizes a research paper that proposes using a Google File System (GFS) configuration with MapReduce to improve CPU and storage utilization in a cloud computing system. It discusses how GFS with MapReduce can split large files into chunks and distribute the processing of those chunks across idle cloud nodes to make better use of resources. The document also addresses using encryption to improve security of data in the cloud.
Comparative Analysis, Security Aspects & Optimization of Workload in Gfs Base...IOSR Journals
This document summarizes a research paper that proposes using Google File System (GFS) and MapReduce in a cloud computing environment to improve resource utilization and processing of large datasets. The paper discusses GFS architecture with a master node and chunk servers, and how MapReduce can split large files into chunks and process them in parallel across idle cloud nodes. It also proposes encrypting data for security and using a third party to audit client files. The goal is to provide fault tolerance, optimize workload processing time, and maximize utilization of cloud resources for data-intensive applications.
T AXONOMY OF O PTIMIZATION A PPROACHES OF R ESOURCE B ROKERS IN D ATA G RIDSijcsit
A novel taxonomy of replica selection techniques is proposed. We studied some data grid approaches
where the selection strategies of data management is different. The aim of the study is to determine the
common concepts and observe their performance and to compare their performance with our strategy
REPLICATION STRATEGY BASED ON DATA RELATIONSHIP IN GRID COMPUTINGcsandit
This study discusses the utilization of three types of relationships in performing data replication. As grid
computing offers the ability of sharing huge amount of resources, resource availability is an important issue
to be addressed. The undertaken approach combines the viewpoint of user, system and the grid itself in
ensuring resource availability. The realization of the proposed strategy is demonstrated via OptorSim and
evaluation is made based on execution time, storage usage, network bandwidth and computing element
usage. Results suggested that the proposed strategy produces a better outcome than an existing method even
though various job workload is introduced.
Welcome to International Journal of Engineering Research and Development (IJERD)IJERD Editor
The document summarizes research on vertical fragmentation, allocation, and re-fragmentation in distributed object relational database systems. It proposes an algorithm for vertical fragmentation and allocation that considers the usage of attributes and methods by queries at different sites. The algorithm forms usage matrices, calculates affinity between methods, clusters methods, and partitions the data into fragments that are allocated to sites where they see the most demand. It also describes handling update queries by redirecting them to a server for processing and then propagating the updates to relevant fragments.
This document proposes a framework for managing and querying multidimensional historical data distributed across an unstructured peer-to-peer network. When data is published, it is summarized into a synopsis containing an index and sub-synopses storing compressed data portions. The index and sub-synopses are distributed across the network with suitable replication to provide coverage for queries. The framework supports data publication and querying while ensuring peer autonomy and handling volatile peers through prompt reaction to disconnections to prevent broken index references.
In a Mobile Ad hoc Network (MANET), due to mobility, limited battery power and poor features of nodes,
network partitioning and nodes disconnecting occur frequently. To improve data availability, database
systems create multiple copies of each data object and allocate them on different nodes. This paper
proposes Automated Re-allocator of Replicas Over MANET (ARROM), that addresses these issues.
ARROM reduces the average response time of requests between clients and database servers by
reallocating replicas frequently. In addition, ARROM increases the average throughput in the network. Our
performance study indicates that ARROM improves average response time and average network
throughput in MANET as compared to resent existing scheme.
Similar to A New Architecture for Group Replication in Data Grid (20)
Text Mining in Digital Libraries using OKAPI BM25 ModelEditor IJCATR
The emergence of the internet has made vast amounts of information available and easily accessible online. As a result, most libraries have digitized their content in order to remain relevant to their users and to keep pace with the advancement of the internet. However, these digital libraries have been criticized for using inefficient information retrieval models that do not perform relevance ranking to the retrieved results. This paper proposed the use of OKAPI BM25 model in text mining so as means of improving relevance ranking of digital libraries. Okapi BM25 model was selected because it is a probability-based relevance ranking algorithm. A case study research was conducted and the model design was based on information retrieval processes. The performance of Boolean, vector space, and Okapi BM25 models was compared for data retrieval. Relevant ranked documents were retrieved and displayed at the OPAC framework search page. The results revealed that Okapi BM 25 outperformed Boolean model and Vector Space model. Therefore, this paper proposes the use of Okapi BM25 model to reward terms according to their relative frequencies in a document so as to improve the performance of text mining in digital libraries.
Green Computing, eco trends, climate change, e-waste and eco-friendlyEditor IJCATR
This document discusses green computing practices and sustainable IT services. It provides an overview of factors driving adoption of green computing to reduce costs and environmental impact of data centers, such as rising energy costs and density. Green strategies discussed include improving infrastructure efficiency, power management, thermal management, efficient product design, and virtualization to optimize resource utilization. The document examines how green computing aims to lower costs and environmental footprint, and how sustainable IT services take a broader approach considering economic, environmental and social impacts.
Policies for Green Computing and E-Waste in NigeriaEditor IJCATR
Computers today are an integral part of individuals’ lives all around the world, but unfortunately these devices are toxic to the environment given the materials used, their limited battery life and technological obsolescence. Individuals are concerned about the hazardous materials ever present in computers, even if the importance of various attributes differs, and that a more environment -friendly attitude can be obtained through exposure to educational materials. In this paper, we aim to delineate the problem of e-waste in Nigeria and highlight a series of measures and the advantage they herald for our country and propose a series of action steps to develop in these areas further. It is possible for Nigeria to have an immediate economic stimulus and job creation while moving quickly to abide by the requirements of climate change legislation and energy efficiency directives. The costs of implementing energy efficiency and renewable energy measures are minimal as they are not cash expenditures but rather investments paid back by future, continuous energy savings.
Performance Evaluation of VANETs for Evaluating Node Stability in Dynamic Sce...Editor IJCATR
Vehicular ad hoc networks (VANETs) are a favorable area of exploration which empowers the interconnection amid the movable vehicles and between transportable units (vehicles) and road side units (RSU). In Vehicular Ad Hoc Networks (VANETs), mobile vehicles can be organized into assemblage to promote interconnection links. The assemblage arrangement according to dimensions and geographical extend has serious influence on attribute of interaction .Vehicular ad hoc networks (VANETs) are subclass of mobile Ad-hoc network involving more complex mobility patterns. Because of mobility the topology changes very frequently. This raises a number of technical challenges including the stability of the network .There is a need for assemblage configuration leading to more stable realistic network. The paper provides investigation of various simulation scenarios in which cluster using k-means algorithm are generated and their numbers are varied to find the more stable configuration in real scenario of road.
Optimum Location of DG Units Considering Operation ConditionsEditor IJCATR
The optimal sizing and placement of Distributed Generation units (DG) are becoming very attractive to researchers these days. In this paper a two stage approach has been used for allocation and sizing of DGs in distribution system with time varying load model. The strategic placement of DGs can help in reducing energy losses and improving voltage profile. The proposed work discusses time varying loads that can be useful for selecting the location and optimizing DG operation. The method has the potential to be used for integrating the available DGs by identifying the best locations in a power system. The proposed method has been demonstrated on 9-bus test system.
Analysis of Comparison of Fuzzy Knn, C4.5 Algorithm, and Naïve Bayes Classifi...Editor IJCATR
Early detection of diabetes mellitus (DM) can prevent or inhibit complication. There are several laboratory test that must be done to detect DM. The result of this laboratory test then converted into data training. Data training used in this study generated from UCI Pima Database with 6 attributes that were used to classify positive or negative diabetes. There are various classification methods that are commonly used, and in this study three of them were compared, which were fuzzy KNN, C4.5 algorithm and Naïve Bayes Classifier (NBC) with one identical case. The objective of this study was to create software to classify DM using tested methods and compared the three methods based on accuracy, precision, and recall. The results showed that the best method was Fuzzy KNN with average and maximum accuracy reached 96% and 98%, respectively. In second place, NBC method had respective average and maximum accuracy of 87.5% and 90%. Lastly, C4.5 algorithm had average and maximum accuracy of 79.5% and 86%, respectively.
Web Scraping for Estimating new Record from Source SiteEditor IJCATR
Study in the Competitive field of Intelligent, and studies in the field of Web Scraping, have a symbiotic relationship mutualism. In the information age today, the website serves as a main source. The research focus is on how to get data from websites and how to slow down the intensity of the download. The problem that arises is the website sources are autonomous so that vulnerable changes the structure of the content at any time. The next problem is the system intrusion detection snort installed on the server to detect bot crawler. So the researchers propose the use of the methods of Mining Data Records and the method of Exponential Smoothing so that adaptive to changes in the structure of the content and do a browse or fetch automatically follow the pattern of the occurrences of the news. The results of the tests, with the threshold 0.3 for MDR and similarity threshold score 0.65 for STM, using recall and precision values produce f-measure average 92.6%. While the results of the tests of the exponential estimation smoothing using ? = 0.5 produces MAE 18.2 datarecord duplicate. It slowed down to 3.6 datarecord from 21.8 datarecord results schedule download/fetch fix in an average time of occurrence news.
Evaluating Semantic Similarity between Biomedical Concepts/Classes through S...Editor IJCATR
Most of the existing semantic similarity measures that use ontology structure as their primary source can measure semantic similarity between concepts/classes using single ontology. The ontology-based semantic similarity techniques such as structure-based semantic similarity techniques (Path Length Measure, Wu and Palmer’s Measure, and Leacock and Chodorow’s measure), information content-based similarity techniques (Resnik’s measure, Lin’s measure), and biomedical domain ontology techniques (Al-Mubaid and Nguyen’s measure (SimDist)) were evaluated relative to human experts’ ratings, and compared on sets of concepts using the ICD-10 “V1.0” terminology within the UMLS. The experimental results validate the efficiency of the SemDist technique in single ontology, and demonstrate that SemDist semantic similarity techniques, compared with the existing techniques, gives the best overall results of correlation with experts’ ratings.
Semantic Similarity Measures between Terms in the Biomedical Domain within f...Editor IJCATR
The techniques and tests are tools used to define how measure the goodness of ontology or its resources. The similarity between biomedical classes/concepts is an important task for the biomedical information extraction and knowledge discovery. However, most of the semantic similarity techniques can be adopted to be used in the biomedical domain (UMLS). Many experiments have been conducted to check the applicability of these measures. In this paper, we investigate to measure semantic similarity between two terms within single ontology or multiple ontologies in ICD-10 “V1.0” as primary source, and compare my results to human experts score by correlation coefficient.
A Strategy for Improving the Performance of Small Files in Openstack Swift Editor IJCATR
This is an effective way to improve the storage access performance of small files in Openstack Swift by adding an aggregate storage module. Because Swift will lead to too much disk operation when querying metadata, the transfer performance of plenty of small files is low. In this paper, we propose an aggregated storage strategy (ASS), and implement it in Swift. ASS comprises two parts which include merge storage and index storage. At the first stage, ASS arranges the write request queue in chronological order, and then stores objects in volumes. These volumes are large files that are stored in Swift actually. During the short encounter time, the object-to-volume mapping information is stored in Key-Value store at the second stage. The experimental results show that the ASS can effectively improve Swift's small file transfer performance.
Integrated System for Vehicle Clearance and RegistrationEditor IJCATR
Efficient management and control of government's cash resources rely on government banking arrangements. Nigeria, like many low income countries, employed fragmented systems in handling government receipts and payments. Later in 2016, Nigeria implemented a unified structure as recommended by the IMF, where all government funds are collected in one account would reduce borrowing costs, extend credit and improve government's fiscal policy among other benefits to government. This situation motivated us to embark on this research to design and implement an integrated system for vehicle clearance and registration. This system complies with the new Treasury Single Account policy to enable proper interaction and collaboration among five different level agencies (NCS, FRSC, SBIR, VIO and NPF) saddled with vehicular administration and activities in Nigeria. Since the system is web based, Object Oriented Hypermedia Design Methodology (OOHDM) is used. Tools such as Php, JavaScript, css, html, AJAX and other web development technologies were used. The result is a web based system that gives proper information about a vehicle starting from the exact date of importation to registration and renewal of licensing. Vehicle owner information, custom duty information, plate number registration details, etc. will also be efficiently retrieved from the system by any of the agencies without contacting the other agency at any point in time. Also number plate will no longer be the only means of vehicle identification as it is presently the case in Nigeria, because the unified system will automatically generate and assigned a Unique Vehicle Identification Pin Number (UVIPN) on payment of duty in the system to the vehicle and the UVIPN will be linked to the various agencies in the management information system.
Assessment of the Efficiency of Customer Order Management System: A Case Stu...Editor IJCATR
The Supermarket Management System deals with the automation of buying and selling of good and services. It includes both sales and purchase of items. The project Supermarket Management System is to be developed with the objective of making the system reliable, easier, fast, and more informative.
Energy-Aware Routing in Wireless Sensor Network Using Modified Bi-Directional A*Editor IJCATR
Energy is a key component in the Wireless Sensor Network (WSN)[1]. The system will not be able to run according to its function without the availability of adequate power units. One of the characteristics of wireless sensor network is Limitation energy[2]. A lot of research has been done to develop strategies to overcome this problem. One of them is clustering technique. The popular clustering technique is Low Energy Adaptive Clustering Hierarchy (LEACH)[3]. In LEACH, clustering techniques are used to determine Cluster Head (CH), which will then be assigned to forward packets to Base Station (BS). In this research, we propose other clustering techniques, which utilize the Social Network Analysis approach theory of Betweeness Centrality (BC) which will then be implemented in the Setup phase. While in the Steady-State phase, one of the heuristic searching algorithms, Modified Bi-Directional A* (MBDA *) is implemented. The experiment was performed deploy 100 nodes statically in the 100x100 area, with one Base Station at coordinates (50,50). To find out the reliability of the system, the experiment to do in 5000 rounds. The performance of the designed routing protocol strategy will be tested based on network lifetime, throughput, and residual energy. The results show that BC-MBDA * is better than LEACH. This is influenced by the ways of working LEACH in determining the CH that is dynamic, which is always changing in every data transmission process. This will result in the use of energy, because they always doing any computation to determine CH in every transmission process. In contrast to BC-MBDA *, CH is statically determined, so it can decrease energy usage.
Security in Software Defined Networks (SDN): Challenges and Research Opportun...Editor IJCATR
In networks, the rapidly changing traffic patterns of search engines, Internet of Things (IoT) devices, Big Data and data centers has thrown up new challenges for legacy; existing networks; and prompted the need for a more intelligent and innovative way to dynamically manage traffic and allocate limited network resources. Software Defined Network (SDN) which decouples the control plane from the data plane through network vitalizations aims to address these challenges. This paper has explored the SDN architecture and its implementation with the OpenFlow protocol. It has also assessed some of its benefits over traditional network architectures, security concerns and how it can be addressed in future research and related works in emerging economies such as Nigeria.
Measure the Similarity of Complaint Document Using Cosine Similarity Based on...Editor IJCATR
Report handling on "LAPOR!" (Laporan, Aspirasi dan Pengaduan Online Rakyat) system depending on the system administrator who manually reads every incoming report [3]. Read manually can lead to errors in handling complaints [4] if the data flow is huge and grows rapidly, it needs at least three days to prepare a confirmation and it sensitive to inconsistencies [3]. In this study, the authors propose a model that can measure the identities of the Query (Incoming) with Document (Archive). The authors employed Class-Based Indexing term weighting scheme, and Cosine Similarities to analyse document similarities. CoSimTFIDF, CoSimTFICF and CoSimTFIDFICF values used in classification as feature for K-Nearest Neighbour (K-NN) classifier. The optimum result evaluation is pre-processing employ 75% of training data ratio and 25% of test data with CoSimTFIDF feature. It deliver a high accuracy 84%. The k = 5 value obtain high accuracy 84.12%
Hangul Recognition Using Support Vector MachineEditor IJCATR
The recognition of Hangul Image is more difficult compared with that of Latin. It could be recognized from the structural arrangement. Hangul is arranged from two dimensions while Latin is only from the left to the right. The current research creates a system to convert Hangul image into Latin text in order to use it as a learning material on reading Hangul. In general, image recognition system is divided into three steps. The first step is preprocessing, which includes binarization, segmentation through connected component-labeling method, and thinning with Zhang Suen to decrease some pattern information. The second is receiving the feature from every single image, whose identification process is done through chain code method. The third is recognizing the process using Support Vector Machine (SVM) with some kernels. It works through letter image and Hangul word recognition. It consists of 34 letters, each of which has 15 different patterns. The whole patterns are 510, divided into 3 data scenarios. The highest result achieved is 94,7% using SVM kernel polynomial and radial basis function. The level of recognition result is influenced by many trained data. Whilst the recognition process of Hangul word applies to the type 2 Hangul word with 6 different patterns. The difference of these patterns appears from the change of the font type. The chosen fonts for data training are such as Batang, Dotum, Gaeul, Gulim, Malgun Gothic. Arial Unicode MS is used to test the data. The lowest accuracy is achieved through the use of SVM kernel radial basis function, which is 69%. The same result, 72 %, is given by the SVM kernel linear and polynomial.
Application of 3D Printing in EducationEditor IJCATR
This paper provides a review of literature concerning the application of 3D printing in the education system. The review identifies that 3D Printing is being applied across the Educational levels [1] as well as in Libraries, Laboratories, and Distance education systems. The review also finds that 3D Printing is being used to teach both students and trainers about 3D Printing and to develop 3D Printing skills.
Survey on Energy-Efficient Routing Algorithms for Underwater Wireless Sensor ...Editor IJCATR
In underwater environment, for retrieval of information the routing mechanism is used. In routing mechanism there are three to four types of nodes are used, one is sink node which is deployed on the water surface and can collect the information, courier/super/AUV or dolphin powerful nodes are deployed in the middle of the water for forwarding the packets, ordinary nodes are also forwarder nodes which can be deployed from bottom to surface of the water and source nodes are deployed at the seabed which can extract the valuable information from the bottom of the sea. In underwater environment the battery power of the nodes is limited and that power can be enhanced through better selection of the routing algorithm. This paper focuses the energy-efficient routing algorithms for their routing mechanisms to prolong the battery power of the nodes. This paper also focuses the performance analysis of the energy-efficient algorithms under which we can examine the better performance of the route selection mechanism which can prolong the battery power of the node
Comparative analysis on Void Node Removal Routing algorithms for Underwater W...Editor IJCATR
The designing of routing algorithms faces many challenges in underwater environment like: propagation delay, acoustic channel behaviour, limited bandwidth, high bit error rate, limited battery power, underwater pressure, node mobility, localization 3D deployment, and underwater obstacles (voids). This paper focuses the underwater voids which affects the overall performance of the entire network. The majority of the researchers have used the better approaches for removal of voids through alternate path selection mechanism but still research needs improvement. This paper also focuses the architecture and its operation through merits and demerits of the existing algorithms. This research article further focuses the analytical method of the performance analysis of existing algorithms through which we found the better approach for removal of voids
Decay Property for Solutions to Plate Type Equations with Variable CoefficientsEditor IJCATR
In this paper we consider the initial value problem for a plate type equation with variable coefficients and memory in
1 n R n ), which is of regularity-loss property. By using spectrally resolution, we study the pointwise estimates in the spectral
space of the fundamental solution to the corresponding linear problem. Appealing to this pointwise estimates, we obtain the global
existence and the decay estimates of solutions to the semilinear problem by employing the fixed point theorem
Main Java[All of the Base Concepts}.docxadhitya5119
This is part 1 of my Java Learning Journey. This Contains Custom methods, classes, constructors, packages, multithreading , try- catch block, finally block and more.
How to Add Chatter in the odoo 17 ERP ModuleCeline George
In Odoo, the chatter is like a chat tool that helps you work together on records. You can leave notes and track things, making it easier to talk with your team and partners. Inside chatter, all communication history, activity, and changes will be displayed.
How to Setup Warehouse & Location in Odoo 17 InventoryCeline George
In this slide, we'll explore how to set up warehouses and locations in Odoo 17 Inventory. This will help us manage our stock effectively, track inventory levels, and streamline warehouse operations.
Walmart Business+ and Spark Good for Nonprofits.pdfTechSoup
"Learn about all the ways Walmart supports nonprofit organizations.
You will hear from Liz Willett, the Head of Nonprofits, and hear about what Walmart is doing to help nonprofits, including Walmart Business and Spark Good. Walmart Business+ is a new offer for nonprofits that offers discounts and also streamlines nonprofits order and expense tracking, saving time and money.
The webinar may also give some examples on how nonprofits can best leverage Walmart Business+.
The event will cover the following::
Walmart Business + (https://business.walmart.com/plus) is a new shopping experience for nonprofits, schools, and local business customers that connects an exclusive online shopping experience to stores. Benefits include free delivery and shipping, a 'Spend Analytics” feature, special discounts, deals and tax-exempt shopping.
Special TechSoup offer for a free 180 days membership, and up to $150 in discounts on eligible orders.
Spark Good (walmart.com/sparkgood) is a charitable platform that enables nonprofits to receive donations directly from customers and associates.
Answers about how you can do more with Walmart!"
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
This slide is special for master students (MIBS & MIFB) in UUM. Also useful for readers who are interested in the topic of contemporary Islamic banking.
How to Fix the Import Error in the Odoo 17Celine George
An import error occurs when a program fails to import a module or library, disrupting its execution. In languages like Python, this issue arises when the specified module cannot be found or accessed, hindering the program's functionality. Resolving import errors is crucial for maintaining smooth software operation and uninterrupted development processes.
How to Manage Your Lost Opportunities in Odoo 17 CRMCeline George
Odoo 17 CRM allows us to track why we lose sales opportunities with "Lost Reasons." This helps analyze our sales process and identify areas for improvement. Here's how to configure lost reasons in Odoo 17 CRM
This presentation includes basic of PCOS their pathology and treatment and also Ayurveda correlation of PCOS and Ayurvedic line of treatment mentioned in classics.
Strategies for Effective Upskilling is a presentation by Chinwendu Peace in a Your Skill Boost Masterclass organisation by the Excellence Foundation for South Sudan on 08th and 09th June 2024 from 1 PM to 3 PM on each day.
Your Skill Boost Masterclass: Strategies for Effective Upskilling
A New Architecture for Group Replication in Data Grid
1. International Journal of Computer Applications Technology and Research
Volume 2– Issue 2, 147 - 151, 2013
www.ijcat.com 147
A New Architecture for Group Replication in Data Grid
Leila Azari
Department of Computer,
Science and Research Branch,
Islamic Azad University,
Khouzestan, Iran
Mashalla Abbasi Dezfouli
Department of Computer,
Science and Research Branch,
Islamic Azad University,
Khouzestan, Iran
Abstract: Nowadays, grid systems are vital technology for programs running with high performance and problems solving with large-
scale in scientific, engineering and business. In grid systems, heterogeneous computational resources and data should be shared
between independent organizations that are scatter geographically. A data grid is a kind of grid types that make relations computational
and storage resources. Data replication is an efficient way in data grid to obtain high performance and high availability by saving
numerous replicas in different locations e.g. grid sites. In this research, we propose a new architecture for dynamic Group data
replication. In our architecture, we added two components to OptorSim architecture: Group Replication Management component
(GRM) and Management of Popular Files Group component (MPFG). OptorSim developed by European Data Grid projects for
evaluate replication algorithm. By using this architecture, popular files group will be replicated in grid sites at the end of each
predefined time interval.
Keywords: data Grid; data replication; group replication; architecture; replica.
1. INTRODUCTION
In grid systems, heterogeneous computational resources and
data should be shared between independent organizations that
are scatter geographically [1]. On the other hand, the data
Grid is a kind of grid services that provides services and
infrastructure for distributed data applications with large
volumes of data. These basic services which are provided by
the data grid architecture are: storage systems, data access and
meta-data services [2]. Data replication is one of the main
services to manage volume data in data grid. This service save
replica in different locations. Therefore, when each user needs
those files, it will be available locally and causes reducing
access latency time, response time and bandwidth
consumption. So far various strategies have been presented in
this field. Grid architecture has an efficient factor in
replication technique. There are different architectures for
data grid that all of them are based on hierarchal data model
[3]. Data grids can be classified into two types [4]: multi-tier
data grids that first introduced by the MONARC project [5],
and cluster data grids that initially proposed by Chang et al.
[6]. Multi-tier data grid architecture contains two models, tree
model and graph (peer to peer) model. In tree model, Each
node can communicate only to the parent node and there is
only one path from a leaf to the root. Because of architecture
failure we can not simply add nodes to the grid. For example,
at GriPhyN project has been used tree model. There are five
tiers in this model. In this multi-tier data grid architecture, the
root site holds all files that are produced initially in the data
grid. Next tiers are national center, regional center, work
groups and leafs node represents desktop. In graph model,
grid sites connect to each others as a peer to peer architecture.
In this model, some of the limitations of the tree type are
improved. In 2001, six replication techniques were presented
by Ranganathan and foster [7] for a multi-tier data grid. In
2002 Ranganathan [8] presented a replication technique based
on peer to peer communities. A cluster represents an
organization unit which is a group of sites that are
geographically close to each others. A cluster data grid
consists of n clusters connected by the Internet [9]. Recent
researchers have used hierarchal models as a basis and
proposed other architectures for replication techniques. The
most above mentioned techniques are the based on a single
replication. If a replica does not exist locally in a requester
grid site replication techniques will replicate it to that grid
site. In this paper, we propose a new architecture for dynamic
group data replication which has two added components to
OptorSim architecture: Group Replication Management
component (GRM) and Management of Popular Files Group
component (MPFG). This paper is organized in the following
way: Section 2 gives a brief introduction of related work on
architecture for data grid. In section 3, the architecture of
OptorSim is presented. Section 4 describes proposed
architecture and finally concludes our presents some future
work.
2. RELATED WORK
Chang et al. [6] presented a hierarchical architecture with
concept of clustering. Cluster grid is a simple hierarchical
form of a grid system. There are two kinds of communications
between grid sites in a cluster grid. Intra-communication is the
communication between grid sites within the same cluster and
inter-communication is the communication between grid sites
across clusters. Network bandwidth between grid sites within
a cluster will be larger than across clusters. In this
architecture, clusters are connected via internet. A job
scheduling policy called HCS is presented based on this
architecture that considers not only computational capability
and data location but also cluster information. Also a dynamic
replica optimization strategy (HRS) is suggested where the
nearby data has a higher priority to access than to generate
new replicas. The simulation results showed that HCS
successfully reduced data access time and the amount of inter-
cluster-communications in comparison with LRU (Least
Recently Used), LFU (Least Frequently Used), BHR
(Bandwidth Hierarchy-based Replication) [10]. Since the file
transmission time is the most important factor to influence the
job execution time for data-intensive jobs in data grids, HCS
with HRS can reduce the file transmission time effectively by
virtue of valid scheduling and proper data replication.
Sepahvand et al. [11] presented three-level hierarchical
architecture. This architecture represents a real structure in
most academic data centers, cities and countries. First level is
grid sites that contain personal computers. These sites are
connected by a high bandwidth network. Second level
contains LAN. Compared with first level, LANs are
connected by a low bandwidth network. Finally, there are
2. International Journal of Computer Applications Technology and Research
Volume 2– Issue 2, 147 - 151, 2013
www.ijcat.com 148
several LANs in each region which are connected via Internet
having a low bandwidth network. If there is not space for
replica, only those file will be deleted that have a low cost of
transfer i.e. considering the bandwidth between source and
destination. So it deletes those files that are available in local
LAN. In comparison to BHR algorithm which considers 2-
level, the 3 level has performed better and it is more realistic.
The simulation results with OptorSim have showed better
performance comparing to LRU and BHR.
Sashi and Thanamani [12] presented a replication algorithm
which is called Modified BHR with multi-tier architecture
based on region. Region comprise group of sites that are
geographically located close together. Each region has a
header. All files were produced in a master site and were
distributed in each region headers. The Modified BHR Region
increases the data availability by replicating files within the
region to the region header and storing them in the site where
the file has been accessed frequently. Instead of storing files
in many sites, Modified BHR stores them in a particular site
so that the storage usage can be reduced. In comparison to No
Replication, (LFU), (LRU) and BHR algorithms, Modified
BHR reduces mean job execution time and the network traffic
too.
Chang et al. [9] presented a replication algorithm which is
called Latest Access Largest Weight (LALW) with multi-tier
architecture based on a centralized data replication
management which has a Dynamic Replication Policymaker
(Policymaker) that responsible for replica management. This
architecture contains grid sites within a cluster. There is a
header used to maintain the site’s information in a cluster. In
the time interval, Policymaker sends a request for collecting
site’s information. Each header sends the information of
accessed files from all sites to Policymaker. LALW selects a
popular file for replication and calculates a suitable number of
copies and grid sites for replication. According to access
frequencies for all files that have been requested, a popular
file is found and replicated to grid sites. In LALW, the data
access records in the nearer past have higher weigh and higher
value of references. The simulation results showed that the
average job execution time of LALW is similar to LFU
optimizer, but exceeded in terms of effective network usage.
Saadat and rahmani [13] presented a replication algorithm
named PDDRA hierarchical architecture supporting their
dynamic replication technique. Grid sites is first level several
grid sites constitute a virtual organization (VO). Next level is
Local Server for each VO. Regional Server (RS) comprise the
final level. Each RS consists of one or more VOs. Due to the
far distance between the RS, they are connected via internet
which has low bandwidth. Sites within a VO have similar
interests. PDDRA predicts future needs of grid sites and pre-
fetches a sequence of files to the requester grid site, so the
next time that this site needs a file, it will be locally available.
The simulation results showed that PDDRA has better
performance in comparison with No Replication, LRU, LFU,
EcoModel, EcoModel Zipf-like distribution and PRA [14] in
terms of job execution time, effective network usage, total
number of replications, hit ratio and percentage of storage
filled.
3. THE ARCHITECTURE OF
OPTORSIM
OptorSim is used as the simulator tool written in Java to
evaluate the performance of replication strategies. It was
developed by the European Data Grid projects [15]. It was
developed to study the effectiveness of replica optimization
algorithms within a data grid environment [16]. The
architecture used in OptorSim is the CMS testbed architecture
[17]. Figure 1, describes the Grid topology; that is, the
resource available and the network connections to other sites.
Sites are represented as green nodes and routers as red nodes.
In this architecture, there are twenty sites, two of which only
have a storage element and act as the master node. CERN and
FNAL are considered the master sites where data is produced
initially. The master site has the most capacity, which allows
it to hold all the master files at the beginning of the
simulation. The storage capacity of the master site is 100 GB,
and the storage capacity of all other sites is 50 GB. Each data
file to be accessed is 1 GB. Jobs are processed in the
remaining sites that have computing elements. There are eight
routers that are used to forward requests to other sites.
Figure 1. CMS topology
The internal architecture [18] of OptorSim and the content of
each site is illustrated in Figure 2, Each site may provide
computational and data storage resources called the
Computing Element (CE) and the Storage Element (SE). CEs
run jobs by processing data files, which are stored in the SEs.
A Resource Broker (RB) controls the scheduling of jobs to
Grid Sites, and schedules jobs to CEs according to scheduling
algorithm. Each site handles its file content with Replica
Manager (RM), within which a Replica Optimizer (RO)
contains the replication algorithm which drives automatic
creation and deletion of replicas [18]. Jobs are submitted to
the grid over a period of time via the RB. The RB schedules
each job to the CE with the goal to improve the overall
throughput of the grid. RM at each site manages the data flow
between sites. The RO inside the RM is responsible for the
selection and dynamic creation and deletion of file replicas.
4. PROPOSED ARCHITECTURE
Our proposed architecture contains Local Servers (LS) and
grid sites. We have used virtual organization concept in our
3. International Journal of Computer Applications Technology and Research
Volume 2– Issue 2, 147 - 151, 2013
www.ijcat.com 149
Figure 2. OptorSim architecture
architecture. Collection of grid sites constitute a VO. Grid
sites within a VO have similar interests and request similar
files. There is also a LS within each VO. LSs are connected
via internet which has low bandwidth. Therefore, speed of
data access within VO is larger than across VOs. OptorSim
[19] assumes that each grid sites contains zero or more
Computing Elements (CE) to run jobs and zero or more
Storage Elements (SE) to store files, or a combination of both.
In OptorSim, there is a Resource Broker (RB) that controls
job scheduling between different CEs. Figure 3 shows the
proposed architecture. First of all, end users submit their jobs
to RB. Then RB schedules jobs between LS by scheduling
optimization and are appropriate jobs to computing sites
which have CE for execution. Computing sites request some
resources for running job that they can be files. If these files
are not available locally then they will be accessed remotely.
Figure 3. group replication architecture
4.1 The architecture of LS and grid sites
In this section, the LS components and components of grid
sites and communications between them are presented. Figure
4 shows messages passing between LS components and grid
sites.
4.1.1 LS Components
LS is composed of two components as follows:
4.1.1.1 Replica Catalogue (RC)
this component maps between physical address and logical
name of a replica.
4.1.1.2 Replica Manager (RM)
after getting information from the Replica Catalogue, the RM
checks the locations of the physical file and selects the best
replica of file. It is important that the file must be fetched
from a low cost grid site. OptorSim finds the best replica
according to minimum cost of replication and then RM selects
replicas that have lowest instance of CE.
4.1.2 Components of grid sites
Each grid site is composed of two components as follows:
4.1.2.1 Management of Popular Files Group
component (MPFG)
This component is composed of two parts as follows:
• Access history database: Each job must have access to a
set of requested files and send them to a grid site.
Sequence of the requested files in each grid site is stored
in a local database.
• Produce engine: This component is responsible for
processing on database and finding the most popular
files group (PFG).
4.1.2.2 Group Replication Management
component (GRM)
In a predefined time interval, this component sends a request
for MPFG component. Replacement Management (RM) is
part of Replication Management. If there is not enough space
for replicating of the PFG to a requester grid site, then RM
will replace the PFG with old replica according to the
replacement algorithm.
4.2 Communication between grid sites and LS
Internal components of proposed architecture and messages
passing between them are shown in figure 3. It consists of two
new components which are shown by dashed lines in figure 3.
At the beginning of time interval GRM sends a replication
request to MPFG component. MPFG processes on the access
history database and finds PFG by using existing replication
algorithm. Then it sends the requested replica of PFG to LS of
grid sites. Also, it sends the name of requester site to LS with
each replication request. Then, Replica Catalogue component
finds physical addresses of each replica of PFG and transfers
them to the Replica Manager component. RM checks the
locations of the physical files and determines the best replica
by minimum cost of replication. Finally, it will begin the
group replication process for the requester sites.
4. International Journal of Computer Applications Technology and Research
Volume 2– Issue 2, 147 - 151, 2013
www.ijcat.com 150
5. CONCLUSION AND FUTURE WORK
In grid systems, heterogeneous computational resources and
data should be shared between independent organizations that
are scatter geographically. On the other hand, the data Grid is
a kind of grid services that provides services and
infrastructure for distributed data applications with large
volumes of data. Data replication is one of the main services
for managing volume data in data grid. This service saves
replica in different locations. Grid architecture has an efficient
factor in replication technique. In this paper, we proposed a
new architecture for group replication in data grid. We added
two components in our architecture, (MPFG) for Management
of Popular Files Group and (GRM) for Group Replication
Management. By using this architecture, popular files group
will be replicated in grid sites at the end of predefined time
interval. For future works, we plan to implement this
architecture and to combine it with group replication and
replacement technique.
6. REFERENCES
[1] Foster, I., Kesselman, C., Tuecke, S., “The anatomy of
the grid: Enabling scalable virtual organizations”,
International Journal of Supercomputer Applications,
pp. 200-222, 2001.
[2] Jacob, B., Brown, M., Fukui, K., Trivedi, N.,
”Introduction to Grid Computing ”,IBM Red books , pp.
3-6, 2005.
[3] Amjad, T., Sher, M., Daud, A., “A survey of dynamic
replication strategies for improving data availability in
data grids”, Future Generation Computer Systems 28,
pp.337–349, 2012.
[4] Ming-Chang, L., Fang-Yie, L., Ying-ping, Ch., “PFRF:
An adaptive data replication algorithm based on star-
topology data grids”, Future Generation Computer
Systems 28, pp.1045–1057, 2012.
[5] The MONARC project.
http://monarc.web.cern.ch/MONARC/, [Accessed 15
Jan 2012].
[6] Chang, R-S., Chang, J-S., Lin, S-Y., “Job scheduling
and data replication on data grids”, Future Generation
Computer Systems 23, pp. 846-860, 2007.
[7] Ranganathan, K., Foster, I., “Design and evaluation of
dynamic replication strategies for a high performance
data grid”, in: International Conference on Computing
in High Energy an Nuclear Physics, vol. 2001, 2001.
[8] Ranganathan, K., Iamnitchi, A., Foster, I., Improving
data availability through dynamic model driven
replication in large peer-to-peer communities, in:
CCGrid, p. 376, 2002.
[9] Chang, R-S., Chang, H-P., Wang, Y-T., “A Dynamic
Weighted Data Replication Stratrgy in Data Grids”, The
Journal of Supercomputing, 45(3), pp. 277-295, 2008.
[10] Park, S.-M., Kim, J.-H., Go, Y.-B., Yoon, W.-S.,
“Dynamic grid replication strategy based on internet
hierarchy, in: International Workshop on Grid and
Cooperative Computing”, in: Lecture Note in Computer
Science, vol. 1001, pp. 1324–1331, 2003.
[11] Sepahvand, R., Horri, A., Dastghaibyfard, Gh., “A New
3-layer Replication and Scheduling Strategy in Data
Grid”, International Symposium on
Telecommunications, pp. 464-469, 2008.
[12] Sashi, K., Thanamani, A.S., “Dynamic replication in a
data grid using a Modified BHR Region Based
Algorithm”, Future Generation Computer Systems 27,
pp. 202-210, 2011.
[13] Saadat, N., rahmani, A-M., “PDDRA: A new pre-
fetching based dynamic data replication algorithm in
Figure 4. Messages passing between grid sites and LS.
5. International Journal of Computer Applications Technology and Research
Volume 2– Issue 2, 147 - 151, 2013
www.ijcat.com 151
data grids”, Future Generation Computer Systems 28,
pp. 666–681, 2012.
[14] Tian, T., Luo, J., Wu, Z., Song, A., “A pre-fetching-
based replication algorithm in data grid, in: 3th
International Conference on Pervasive Computing and
Applications”, pp. 526–531, 2008.
[15] The European DataGrid Project, http://www.edg.org,
[Accessed 15 Jan 2012].
[16] Bell, W.H., Cameron, D.G., Capozza, L., Millar, A.P.,
Stockinger, K., Zini, F., “Simulation of dynamic grid
replication strategies in OptorSim”, International
Journal of High performance Computing Applications
17 (4), 2003.
[17] CMS Data Challenge, 2004:
http://www.uscms.org/s&c/dc04/, [Accessed 15 Jan
2012].
[18] Cameron, D.G., Millar, A.P., Nicholson, C., “OptorSim:
a simulation tool for scheduling and replica optimization
in data grids”, Proceedings of Computing in High
Energy and Nuclear Physics (CHEP) , 2004.
[19] OptorSim – A Replica Optimiser Simulation, http://grid-
data-management.web.cern.ch/grid-data-
management/optimization/optor, [Accessed 15 Jan
2012].