In data-intensive applications data transfer is a primary cause of job execution delay. Data access time depends on bandwidth. The major bottleneck to supporting fast data access in Grids is the high latencies of Wide Area Networks and Internet. Effective scheduling can reduce the amount of data transferred across the internet by dispatching a job to where the needed data are present. Another solution is to use a data replication mechanism. Objective of dynamic replica strategies is reducing file access time which leads to reducing job runtime. In this paper we develop a job scheduling policy and a dynamic data replication strategy, called HRS (Hierarchical Replication Strategy), to improve the data access efficiencies. We study our approach and evaluate it through simulation. The results show that our algorithm has improved 12% over the current strategies
A New Architecture for Group Replication in Data GridEditor IJCATR
Nowadays, grid systems are vital technology for programs running with high performance and problems solving with largescale
in scientific, engineering and business. In grid systems, heterogeneous computational resources and data should be shared
between independent organizations that are scatter geographically. A data grid is a kind of grid types that make relations computational
and storage resources. Data replication is an efficient way in data grid to obtain high performance and high availability by saving
numerous replicas in different locations e.g. grid sites. In this research, we propose a new architecture for dynamic Group data
replication. In our architecture, we added two components to OptorSim architecture: Group Replication Management component
(GRM) and Management of Popular Files Group component (MPFG). OptorSim developed by European Data Grid projects for
evaluate replication algorithm. By using this architecture, popular files group will be replicated in grid sites at the end of each
predefined time interval.
TASK-DECOMPOSITION BASED ANOMALY DETECTION OF MASSIVE AND HIGH-VOLATILITY SES...ijdpsjournal
The Science Information Network (SINET) is a Japanese academic backbone network for more than 800 universities and research institutions. The characteristic of SINET traffic is that it is enormous and highly variable. In this paper, we present a task-decomposition based anomaly detection of massive and highvolatility session data of SINET. Three main features are discussed: Tash scheduling, Traffic discrimination, and Histogramming. We adopt a task-decomposition based dynamic scheduling method to handle the massive session data stream of SINET. In the experiment, we have analysed SINET traffic from 2/27 to 3/8 and detect some anomalies by LSTM based time-series data processing.
Propose a Method to Improve Performance in Grid Environment, Using Multi-Crit...Editor IJCATR
The most important purpose of grid networks is resource subscription in a dynamic and heterogeneous environment.
They are accessible through using various methods. Subscription has mainly computational, scientific and other implications. In
order to reach grid purposes and to use available resources in grid environment, subtasks are distributed among resources and are
scheduled by considering the quality of service. It has been tried to distribute subtasks between resources in a way that maximum
QOS can be obtained. In this study, a method has been presented. In this method, three parameters; namely, sent and transferred
time between RMS and resource, process time of subtask by the resource, and the load of available tasks in resources row, have
been taken into account. In this way, multi-criteria decision is made by using TOPSIS method and this priority of the resources
are determined to assign them to subtasks. Finally, time response, as an efficient parameter, has been improved and optimized by
optimal assignment of the resources to subtasks.
Research Inventy : International Journal of Engineering and Scienceinventy
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
Grid is a type of parallel and distributed systems that is designed to provide reliable access to data
and computational resources in wide area networks. These resources are distributed in different geographical
locations. Efficient data sharing in global networks is complicated by erratic node failure, unreliable network
connectivity and limited bandwidth. Replication is a technique used in grid systems to improve the
applications’ response time and to reduce the bandwidth consumption. In this paper, we present a survey on
basic and new replication techniques that have been proposed by other researchers. After that, we have a full
comparative study on these replication strategies
A New Architecture for Group Replication in Data GridEditor IJCATR
Nowadays, grid systems are vital technology for programs running with high performance and problems solving with largescale
in scientific, engineering and business. In grid systems, heterogeneous computational resources and data should be shared
between independent organizations that are scatter geographically. A data grid is a kind of grid types that make relations computational
and storage resources. Data replication is an efficient way in data grid to obtain high performance and high availability by saving
numerous replicas in different locations e.g. grid sites. In this research, we propose a new architecture for dynamic Group data
replication. In our architecture, we added two components to OptorSim architecture: Group Replication Management component
(GRM) and Management of Popular Files Group component (MPFG). OptorSim developed by European Data Grid projects for
evaluate replication algorithm. By using this architecture, popular files group will be replicated in grid sites at the end of each
predefined time interval.
TASK-DECOMPOSITION BASED ANOMALY DETECTION OF MASSIVE AND HIGH-VOLATILITY SES...ijdpsjournal
The Science Information Network (SINET) is a Japanese academic backbone network for more than 800 universities and research institutions. The characteristic of SINET traffic is that it is enormous and highly variable. In this paper, we present a task-decomposition based anomaly detection of massive and highvolatility session data of SINET. Three main features are discussed: Tash scheduling, Traffic discrimination, and Histogramming. We adopt a task-decomposition based dynamic scheduling method to handle the massive session data stream of SINET. In the experiment, we have analysed SINET traffic from 2/27 to 3/8 and detect some anomalies by LSTM based time-series data processing.
Propose a Method to Improve Performance in Grid Environment, Using Multi-Crit...Editor IJCATR
The most important purpose of grid networks is resource subscription in a dynamic and heterogeneous environment.
They are accessible through using various methods. Subscription has mainly computational, scientific and other implications. In
order to reach grid purposes and to use available resources in grid environment, subtasks are distributed among resources and are
scheduled by considering the quality of service. It has been tried to distribute subtasks between resources in a way that maximum
QOS can be obtained. In this study, a method has been presented. In this method, three parameters; namely, sent and transferred
time between RMS and resource, process time of subtask by the resource, and the load of available tasks in resources row, have
been taken into account. In this way, multi-criteria decision is made by using TOPSIS method and this priority of the resources
are determined to assign them to subtasks. Finally, time response, as an efficient parameter, has been improved and optimized by
optimal assignment of the resources to subtasks.
Research Inventy : International Journal of Engineering and Scienceinventy
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
Grid is a type of parallel and distributed systems that is designed to provide reliable access to data
and computational resources in wide area networks. These resources are distributed in different geographical
locations. Efficient data sharing in global networks is complicated by erratic node failure, unreliable network
connectivity and limited bandwidth. Replication is a technique used in grid systems to improve the
applications’ response time and to reduce the bandwidth consumption. In this paper, we present a survey on
basic and new replication techniques that have been proposed by other researchers. After that, we have a full
comparative study on these replication strategies
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
THRESHOLD BASED VM PLACEMENT TECHNIQUE FOR LOAD BALANCED RESOURCE PROVISIONIN...IJCNCJournal
The unbalancing load issue is a multi-variation, multi-imperative issue that corrupts the execution and productivity of processing assets. Workload adjusting methods give solutions of load unbalancing circumstances for two bothersome aspects over-burdening and under-stacking. Cloud computing utilizes planning and workload balancing for a virtualized environment, resource partaking in cloud foundation. These two factors must be handled in an improved way in cloud computing to accomplish ideal resource sharing. Henceforth, there requires productive resource, asset reservation for guaranteeing load advancement in the cloud. This work aims to present an incorporated resource, asset reservation, and workload adjusting calculation for effective cloud provisioning. The strategy develops a Priority-based Resource Scheduling Model to acquire the resource, asset reservation with threshold-based load balancing for improving the proficiency in cloud framework. Extending utilization of Virtual Machines through the suitable and sensible outstanding task at hand modifying is then practiced by intensely picking a job from submitting jobs using Priority-based Resource Scheduling Model to acquire resource asset reservation. Experimental evaluations represent, the proposed scheme gives better results by reducing execution time, with minimum resource cost and improved resource utilization in dynamic resource provisioning conditions.
Dynamic selection of cluster head in in networks for energy managementeSAT Journals
Abstract In this project, we presented Multipath Region Routing (MRR) protocol for energy conservation in Wireless Sensor Networks (WSNs). Large scale dense WSNs are used in different types of applications for accurate monitoring. Energy conservation is an important issue in WSNs. In order to save energy, Multipath Region Routing protocol is used which provides balance in energy consumption and sustains the network life-span. By using this method, we can reduce the number of energy dissipation because the cluster head will collect data directly from other nodes. Hence, the energy can be preserved and network life time is extended to reasonable time. Keywords: Clustering; Wireless Sensor Networks; Security; Multipath Region Routing;
Dynamic selection of cluster head in in networks for energy managementeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A survey of various scheduling algorithm in cloud computing environmenteSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
An efficient approach on spatial big data related to wireless networks and it...eSAT Journals
Abstract
Spatial big data acts as a important key role in wireless networks applications. In that spatial and spatio temporal problems contains the distinct role in big data and it’s compared to common relational problems. If we are solving those problems means describing the three applications for spatial big data. In each applications imposing the specific design and we are developing our work on highly scalable parallel processing for spatial big data in Hadoop frameworks by using map reduce computational model. Our results show that enables highly scalable implementations of algorithms using Hadoop for the purpose of spatial data processing problems. Inspite of developing these implementations requires specialized knowledge and user friendly.
Keywords: Spatial Big Data, Hadoop, Wireless Networks, Map reduce
An adaptive algorithm for task scheduling for computational grideSAT Journals
Abstract
Grid Computing is a collection of computing and storage resources that are collected from multiple administrative domains. Grid resources can be applied to reach a common goal. Since computational grids enable the sharing and aggregation of a wide variety of geographically distributed computational resources, an effective task scheduling is vital for managing the tasks. Efficient scheduling algorithms are the need of the hour to achieve efficient utilization of the unused CPU cycles distributed geographically in various locations. The existing job scheduling algorithms in grid computing are mainly concentrated on the system’s performance rather than the user satisfaction. This research work presents a new algorithm that mainly focuses on better meeting the deadlines of the statically available jobs as expected by the users. This algorithm also concentrates on the better utilization of the available heterogeneous resources.
Keywords: Task Scheduling, Computational Grid, Adaptive Scheduling and User Deadline.
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENTIJCNCJournal
Cloud computing has an indispensable role in the modern digital scenario. The fundamental challenge of cloud systems is to accommodate user requirements which keep on varying. This dynamic cloud environment demands the necessity of complex algorithms to resolve the trouble of task allotment. The overall performance of cloud systems is rooted in the efficiency of task scheduling algorithms. The dynamic property of cloud systems makes it challenging to find an optimal solution satisfying all the evaluation metrics. The new approach is formulated on the Round Robin and the Shortest Job First algorithms. The Round Robin method reduces starvation, and the Shortest Job First decreases the average waiting time. In this work, the advantages of both algorithms are incorporated to improve the makespan of user tasks.
ORCHESTRATING BULK DATA TRANSFERS ACROSS GEO-DISTRIBUTED DATACENTERSNexgen Technology
bulk ieee projects in pondicherry,ieee projects in pondicherry,final year ieee projects in pondicherry
Nexgen Technology Address:
Nexgen Technology
No :66,4th cross,Venkata nagar,
Near SBI ATM,
Puducherry.
Email Id: praveen@nexgenproject.com.
www.nexgenproject.com
Mobile: 9751442511,9791938249
Telephone: 0413-2211159.
NEXGEN TECHNOLOGY as an efficient Software Training Center located at Pondicherry with IT Training on IEEE Projects in Android,IEEE IT B.Tech Student Projects, Android Projects Training with Placements Pondicherry, IEEE projects in pondicherry, final IEEE Projects in Pondicherry , MCA, BTech, BCA Projects in Pondicherry, Bulk IEEE PROJECTS IN Pondicherry.So far we have reached almost all engineering colleges located in Pondicherry and around 90km
In the era of big data, even though we have large infrastructure, storage data varies in size,
formats, variety, volume and several platforms such as hadoop, cloud since we have problem associated
with an application how to process the data which is varying in size and format. Data varying in
application and resources available during run time is called dynamic workflow. Using large
infrastructure and huge amount of resources for the analysis of data is time consuming and waste of
resources, it’s better to use scheduling algorithm to analyse the given data set, for efficient execution of
data set without time consuming and evaluate which scheduling algorithm is best and suitable for the
given data set. We evaluate with different data set understand which is the most suitable algorithm for
analysis of data being efficient execution of data set and store the data after analysis
NETWORK-AWARE DATA PREFETCHING OPTIMIZATION OF COMPUTATIONS IN A HETEROGENEOU...IJCNCJournal
Rapid development of diverse computer architectures and hardware accelerators caused that designing parallel systems faces new problems resulting from their heterogeneity. Our implementation of a parallel
system called KernelHive allows to efficiently run applications in a heterogeneous environment consisting
of multiple collections of nodes with different types of computing devices. The execution engine of the
system is open for optimizer implementations, focusing on various criteria. In this paper, we propose a new
optimizer for KernelHive, that utilizes distributed databases and performs data prefetching to optimize the
execution time of applications, which process large input data. Employing a versatile data management
scheme, which allows combining various distributed data providers, we propose using NoSQL databases
for our purposes. We support our solution with results of experiments with real executions of our OpenCL
implementation of a regular expression matching application in various hardware configurations.
Additionally, we propose a network-aware scheduling scheme for selecting hardware for the proposed
optimizer and present simulations that demonstrate its advantages.
An enhanced adaptive scoring job scheduling algorithm with replication strate...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A Survey of File Replication Techniques In Grid SystemsEditor IJCATR
Grid is a type of parallel and distributed systems that is designed to provide reliable access to data
and computational resources in wide area networks. These resources are distributed in different geographical
locations. Efficient data sharing in global networks is complicated by erratic node failure, unreliable network
connectivity and limited bandwidth. Replication is a technique used in grid systems to improve the
applications’ response time and to reduce the bandwidth consumption. In this paper, we present a survey on
basic and new replication techniques that have been proposed by other researchers. After that, we have a full
comparative study on these replication strategies.
A Survey of File Replication Techniques In Grid SystemsEditor IJCATR
Grid is a type of parallel and distributed systems that is designed to provide reliable access to data
and computational resources in wide area networks. These resources are distributed in different geographical
locations. Efficient data sharing in global networks is complicated by erratic node failure, unreliable network
connectivity and limited bandwidth. Replication is a technique used in grid systems to improve the
applications’ response time and to reduce the bandwidth consumption. In this paper, we present a survey on
basic and new replication techniques that have been proposed by other researchers. After that, we have a full
comparative study on these replication strategies
A Survey of Job Scheduling Algorithms Whit Hierarchical Structure to Load Ba...Editor IJCATR
Due to the advances in human civilization, problems in science and engineering are becoming more complicated than ever
before. To solve these complicated problems, grid computing becomes a popular tool. a grid environment collects, integrates, and uses
heterogeneous or homogeneous resources scattered around the globe by a high-speed network. Scheduling problems are at the heart of
any Grid-like computational system. a good scheduling algorithm can assign jobs to resources efficiently and can balance the system
load. in this paper we survey three algorithms for grid scheduling and compare benefit and disadvantages of their based on makespan.
GROUPING BASED JOB SCHEDULING ALGORITHM USING PRIORITY QUEUE AND HYBRID ALGOR...ijgca
Grid computing enlarge with computing platform which is collection of heterogeneous computing resources connected by a network across dynamic and geographically dispersed organization to form a distributed high performance computing infrastructure. Grid computing solves the complex computing
problems amongst multiple machines. Grid computing solves the large scale computational demands in a high performance computing environment. The main emphasis in the grid computing is given to the resource management and the job scheduler .The goal of the job scheduler is to maximize the resource utilization and minimize the processing time of the jobs. Existing approaches of Grid scheduling doesn’t give much emphasis on the performance of a Grid scheduler in processing time parameter. Schedulers allocate resources to the jobs to be executed using the First come First serve algorithm. In this paper, we have provided an optimize algorithm to queue of the scheduler using various scheduling methods like Shortest Job First, First in First out, Round robin. The job scheduling system is responsible to select best suitable machines in a grid for user jobs. The management and scheduling system generates job schedules for each machine in the grid by taking static restrictions and dynamic parameters of jobs and machines
into consideration. The main purpose of this paper is to develop an efficient job scheduling algorithm to maximize the resource utilization and minimize processing time of the jobs. Queues can be optimized by using various scheduling algorithms depending upon the performance criteria to be improved e.g. response
time, throughput. The work has been done in MATLAB using the parallel computing toolbox.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
THRESHOLD BASED VM PLACEMENT TECHNIQUE FOR LOAD BALANCED RESOURCE PROVISIONIN...IJCNCJournal
The unbalancing load issue is a multi-variation, multi-imperative issue that corrupts the execution and productivity of processing assets. Workload adjusting methods give solutions of load unbalancing circumstances for two bothersome aspects over-burdening and under-stacking. Cloud computing utilizes planning and workload balancing for a virtualized environment, resource partaking in cloud foundation. These two factors must be handled in an improved way in cloud computing to accomplish ideal resource sharing. Henceforth, there requires productive resource, asset reservation for guaranteeing load advancement in the cloud. This work aims to present an incorporated resource, asset reservation, and workload adjusting calculation for effective cloud provisioning. The strategy develops a Priority-based Resource Scheduling Model to acquire the resource, asset reservation with threshold-based load balancing for improving the proficiency in cloud framework. Extending utilization of Virtual Machines through the suitable and sensible outstanding task at hand modifying is then practiced by intensely picking a job from submitting jobs using Priority-based Resource Scheduling Model to acquire resource asset reservation. Experimental evaluations represent, the proposed scheme gives better results by reducing execution time, with minimum resource cost and improved resource utilization in dynamic resource provisioning conditions.
Dynamic selection of cluster head in in networks for energy managementeSAT Journals
Abstract In this project, we presented Multipath Region Routing (MRR) protocol for energy conservation in Wireless Sensor Networks (WSNs). Large scale dense WSNs are used in different types of applications for accurate monitoring. Energy conservation is an important issue in WSNs. In order to save energy, Multipath Region Routing protocol is used which provides balance in energy consumption and sustains the network life-span. By using this method, we can reduce the number of energy dissipation because the cluster head will collect data directly from other nodes. Hence, the energy can be preserved and network life time is extended to reasonable time. Keywords: Clustering; Wireless Sensor Networks; Security; Multipath Region Routing;
Dynamic selection of cluster head in in networks for energy managementeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A survey of various scheduling algorithm in cloud computing environmenteSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
An efficient approach on spatial big data related to wireless networks and it...eSAT Journals
Abstract
Spatial big data acts as a important key role in wireless networks applications. In that spatial and spatio temporal problems contains the distinct role in big data and it’s compared to common relational problems. If we are solving those problems means describing the three applications for spatial big data. In each applications imposing the specific design and we are developing our work on highly scalable parallel processing for spatial big data in Hadoop frameworks by using map reduce computational model. Our results show that enables highly scalable implementations of algorithms using Hadoop for the purpose of spatial data processing problems. Inspite of developing these implementations requires specialized knowledge and user friendly.
Keywords: Spatial Big Data, Hadoop, Wireless Networks, Map reduce
An adaptive algorithm for task scheduling for computational grideSAT Journals
Abstract
Grid Computing is a collection of computing and storage resources that are collected from multiple administrative domains. Grid resources can be applied to reach a common goal. Since computational grids enable the sharing and aggregation of a wide variety of geographically distributed computational resources, an effective task scheduling is vital for managing the tasks. Efficient scheduling algorithms are the need of the hour to achieve efficient utilization of the unused CPU cycles distributed geographically in various locations. The existing job scheduling algorithms in grid computing are mainly concentrated on the system’s performance rather than the user satisfaction. This research work presents a new algorithm that mainly focuses on better meeting the deadlines of the statically available jobs as expected by the users. This algorithm also concentrates on the better utilization of the available heterogeneous resources.
Keywords: Task Scheduling, Computational Grid, Adaptive Scheduling and User Deadline.
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENTIJCNCJournal
Cloud computing has an indispensable role in the modern digital scenario. The fundamental challenge of cloud systems is to accommodate user requirements which keep on varying. This dynamic cloud environment demands the necessity of complex algorithms to resolve the trouble of task allotment. The overall performance of cloud systems is rooted in the efficiency of task scheduling algorithms. The dynamic property of cloud systems makes it challenging to find an optimal solution satisfying all the evaluation metrics. The new approach is formulated on the Round Robin and the Shortest Job First algorithms. The Round Robin method reduces starvation, and the Shortest Job First decreases the average waiting time. In this work, the advantages of both algorithms are incorporated to improve the makespan of user tasks.
ORCHESTRATING BULK DATA TRANSFERS ACROSS GEO-DISTRIBUTED DATACENTERSNexgen Technology
bulk ieee projects in pondicherry,ieee projects in pondicherry,final year ieee projects in pondicherry
Nexgen Technology Address:
Nexgen Technology
No :66,4th cross,Venkata nagar,
Near SBI ATM,
Puducherry.
Email Id: praveen@nexgenproject.com.
www.nexgenproject.com
Mobile: 9751442511,9791938249
Telephone: 0413-2211159.
NEXGEN TECHNOLOGY as an efficient Software Training Center located at Pondicherry with IT Training on IEEE Projects in Android,IEEE IT B.Tech Student Projects, Android Projects Training with Placements Pondicherry, IEEE projects in pondicherry, final IEEE Projects in Pondicherry , MCA, BTech, BCA Projects in Pondicherry, Bulk IEEE PROJECTS IN Pondicherry.So far we have reached almost all engineering colleges located in Pondicherry and around 90km
In the era of big data, even though we have large infrastructure, storage data varies in size,
formats, variety, volume and several platforms such as hadoop, cloud since we have problem associated
with an application how to process the data which is varying in size and format. Data varying in
application and resources available during run time is called dynamic workflow. Using large
infrastructure and huge amount of resources for the analysis of data is time consuming and waste of
resources, it’s better to use scheduling algorithm to analyse the given data set, for efficient execution of
data set without time consuming and evaluate which scheduling algorithm is best and suitable for the
given data set. We evaluate with different data set understand which is the most suitable algorithm for
analysis of data being efficient execution of data set and store the data after analysis
NETWORK-AWARE DATA PREFETCHING OPTIMIZATION OF COMPUTATIONS IN A HETEROGENEOU...IJCNCJournal
Rapid development of diverse computer architectures and hardware accelerators caused that designing parallel systems faces new problems resulting from their heterogeneity. Our implementation of a parallel
system called KernelHive allows to efficiently run applications in a heterogeneous environment consisting
of multiple collections of nodes with different types of computing devices. The execution engine of the
system is open for optimizer implementations, focusing on various criteria. In this paper, we propose a new
optimizer for KernelHive, that utilizes distributed databases and performs data prefetching to optimize the
execution time of applications, which process large input data. Employing a versatile data management
scheme, which allows combining various distributed data providers, we propose using NoSQL databases
for our purposes. We support our solution with results of experiments with real executions of our OpenCL
implementation of a regular expression matching application in various hardware configurations.
Additionally, we propose a network-aware scheduling scheme for selecting hardware for the proposed
optimizer and present simulations that demonstrate its advantages.
An enhanced adaptive scoring job scheduling algorithm with replication strate...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A Survey of File Replication Techniques In Grid SystemsEditor IJCATR
Grid is a type of parallel and distributed systems that is designed to provide reliable access to data
and computational resources in wide area networks. These resources are distributed in different geographical
locations. Efficient data sharing in global networks is complicated by erratic node failure, unreliable network
connectivity and limited bandwidth. Replication is a technique used in grid systems to improve the
applications’ response time and to reduce the bandwidth consumption. In this paper, we present a survey on
basic and new replication techniques that have been proposed by other researchers. After that, we have a full
comparative study on these replication strategies.
A Survey of File Replication Techniques In Grid SystemsEditor IJCATR
Grid is a type of parallel and distributed systems that is designed to provide reliable access to data
and computational resources in wide area networks. These resources are distributed in different geographical
locations. Efficient data sharing in global networks is complicated by erratic node failure, unreliable network
connectivity and limited bandwidth. Replication is a technique used in grid systems to improve the
applications’ response time and to reduce the bandwidth consumption. In this paper, we present a survey on
basic and new replication techniques that have been proposed by other researchers. After that, we have a full
comparative study on these replication strategies
A Survey of Job Scheduling Algorithms Whit Hierarchical Structure to Load Ba...Editor IJCATR
Due to the advances in human civilization, problems in science and engineering are becoming more complicated than ever
before. To solve these complicated problems, grid computing becomes a popular tool. a grid environment collects, integrates, and uses
heterogeneous or homogeneous resources scattered around the globe by a high-speed network. Scheduling problems are at the heart of
any Grid-like computational system. a good scheduling algorithm can assign jobs to resources efficiently and can balance the system
load. in this paper we survey three algorithms for grid scheduling and compare benefit and disadvantages of their based on makespan.
GROUPING BASED JOB SCHEDULING ALGORITHM USING PRIORITY QUEUE AND HYBRID ALGOR...ijgca
Grid computing enlarge with computing platform which is collection of heterogeneous computing resources connected by a network across dynamic and geographically dispersed organization to form a distributed high performance computing infrastructure. Grid computing solves the complex computing
problems amongst multiple machines. Grid computing solves the large scale computational demands in a high performance computing environment. The main emphasis in the grid computing is given to the resource management and the job scheduler .The goal of the job scheduler is to maximize the resource utilization and minimize the processing time of the jobs. Existing approaches of Grid scheduling doesn’t give much emphasis on the performance of a Grid scheduler in processing time parameter. Schedulers allocate resources to the jobs to be executed using the First come First serve algorithm. In this paper, we have provided an optimize algorithm to queue of the scheduler using various scheduling methods like Shortest Job First, First in First out, Round robin. The job scheduling system is responsible to select best suitable machines in a grid for user jobs. The management and scheduling system generates job schedules for each machine in the grid by taking static restrictions and dynamic parameters of jobs and machines
into consideration. The main purpose of this paper is to develop an efficient job scheduling algorithm to maximize the resource utilization and minimize processing time of the jobs. Queues can be optimized by using various scheduling algorithms depending upon the performance criteria to be improved e.g. response
time, throughput. The work has been done in MATLAB using the parallel computing toolbox.
GROUPING BASED JOB SCHEDULING ALGORITHM USING PRIORITY QUEUE AND HYBRID ALGOR...ijgca
Grid computing enlarge with computing platform which is collection of heterogeneous computing resources connected by a network across dynamic and geographically dispersed organization to form a distributed high performance computing infrastructure. Grid computing solves the complex computing problems amongst multiple machines. Grid computing solves the large scale computational demands in a high performance computing environment. The main emphasis in the grid computing is given to the resource management and the job scheduler .The goal of the job scheduler is to maximize the resource utilization and minimize the processing time of the jobs. Existing approaches of Grid scheduling doesn’t give much emphasis on the performance of a Grid scheduler in processing time parameter. Schedulers allocate resources to the jobs to be executed using the First come First serve algorithm. In this paper, we have provided an optimize algorithm to queue of the scheduler using various scheduling methods like Shortest Job First, First in First out, Round robin. The job scheduling system is responsible to select best suitable machines in a grid for user jobs. The management and scheduling system generates job schedules for each machine in the grid by taking static restrictions and dynamic parameters of jobs and machines into consideration. The main purpose of this paper is to develop an efficient job scheduling algorithm to maximize the resource utilization and minimize processing time of the jobs. Queues can be optimized by using various scheduling algorithms depending upon the performance criteria to be improved e.g. response time, throughput. The work has been done in MATLAB using the parallel computing toolbox.
Efficient Resource Management Mechanism with Fault Tolerant Model for Computa...Editor IJCATR
Grid computing provides a framework and deployment environment that enables resource
sharing, accessing, aggregation and management. It allows resource and coordinated use of various
resources in dynamic, distributed virtual organization. The grid scheduling is responsible for resource
discovery, resource selection and job assignment over a decentralized heterogeneous system. In the
existing system, primary-backup approach is used for fault tolerance in a single environment. In this
approach, each task has a primary copy and backup copy on two different processors. For dependent
tasks, precedence constraint among tasks must be considered when scheduling backup copies and
overloading backups. Then, two algorithms have been developed to schedule backups of dependent and
independent tasks. The proposed work is to manage the resource failures in grid job scheduling. In this
method, data source and resource are integrated from different geographical environment. Faulttolerant
scheduling with primary backup approach is used to handle job failures in grid environment.
Impact of communication protocols is considered. Communication protocols such as Transmission
Control Protocol (TCP), User Datagram Protocol (UDP) which are used to distribute the message of
each task to grid resources.
Grid computing can involve lot of computational tasks which requires trustworthy computational nodes. Load balancing in grid computing is a technique which overall optimizes the whole process of assigning computational tasks to processing nodes. Grid computing is a form of distributed computing but different from conventional distributed computing in a manner that it tends to be heterogeneous, more loosely coupled and dispersed geographically. Optimization of this process must contains the overall maximization of resources utilization with balance load on each processing unit and also by decreasing the overall time or output. Evolutionary algorithms like genetic algorithms have studied so far for the implementation of load balancing across the grid networks. But problem with these genetic algorithm is that they are quite slow in cases where large number of tasks needs to be processed. In this paper we give a novel approach of parallel genetic algorithms for enhancing the overall performance and optimization of managing the whole process of load balancing across the grid nodes.
A survey of various scheduling algorithm in cloud computing environmenteSAT Journals
Abstract Cloud computing is known as a provider of dynamic services using very large scalable and virtualized resources over the Internet. Due to novelty of cloud computing field, there is no many standard task scheduling algorithm used in cloud environment. Especially that in cloud, there is a high communication cost that prevents well known task schedulers to be applied in large scale distributed environment. Today, researchers attempt to build job scheduling algorithms that are compatible and applicable in Cloud Computing environment Job scheduling is most important task in cloud computing environment because user have to pay for resources used based upon time. Hence efficient utilization of resources must be important and for that scheduling plays a vital role to get maximum benefit from the resources. In this paper we are studying various scheduling algorithm and issues related to them in cloud computing. Index Terms: cloud computing, scheduling, algorithm
T AXONOMY OF O PTIMIZATION A PPROACHES OF R ESOURCE B ROKERS IN D ATA G RIDSijcsit
A novel taxonomy of replica selection techniques is proposed. We studied some data grid approaches
where the selection strategies of data management is different. The aim of the study is to determine the
common concepts and observe their performance and to compare their performance with our strategy
A Platform for Large-Scale Grid Data Service on Dynamic High-Performance Netw...Tal Lavian Ph.D.
Data intensive Grid applications often deal with multiple terabytes and even petabytes of data. For them to be effectively deployed over distances, it is crucial that Grid infrastructures learn how to best exploit high-performance networks
(such as agile optical networks). The network footprint of these Grid applications show pronounced peaks and valleys in utilization, prompting for a radical overhaul of traditional network provisioning styles such as peak-provisioning, point-and-click or operator-assisted provisioning. A Grid stack must become capable to dynamically orchestrate a complex set of variables related to application requirements, data services, and network provisioning services, all within a rapidly and continually changing environment. Presented here is a platform that addresses some of these issues. This service platform closely integrates a set of large-scale data services with those for dynamic bandwidth allocation, through a network resource middleware service, using an OGSA-compliant interface allowing direct access by external applications. Recently, this platform has been implemented as an experimental research prototype on a unique wide area optical networking testbed incorporating state-of-the-art photonic
components. The paper, which presents initial results of research conducted on this prototype, indicates that these methods have the potential to address multiple major challenges related to data intensive applications. Given the complexities of this topic, especially where scheduling is required, only selected aspects of this platform are considered in this paper.
AN ENTROPIC OPTIMIZATION TECHNIQUE IN HETEROGENEOUS GRID COMPUTING USING BION...ijcsit
The wide usage of the Internet and the availability of powerful computers and high-speed networks as low cost
commodity components have a deep impact on the way we use computers today, in such a way that
these technologies facilitated the usage of multi-owner and geographically distributed resources to address
large-scale problems in many areas such as science, engineering, and commerce. The new paradigm of
Grid computing has evolved from these researches on these topics. Performance and utilization of the grid
depends on a complex and excessively dynamic procedure of optimally balancing the load among the
available nodes. In this paper, we suggest a novel two-dimensional figure of merit that depict the network
effects on load balance and fault tolerance estimation to improve the performance of the network
utilizations. The enhancement of fault tolerance is obtained by adaptively decrease replication time and
message cost. On the other hand, load balance is improved by adaptively decrease mean job response time.
Finally, analysis of Genetic Algorithm, Ant Colony Optimization, and Particle Swarm Optimization is
conducted with regards to their solutions, issues and improvements concerning load balancing in
computational grid. Consequently, a significant system utilization improvement was attained. Experimental
results eventually demonstrate that the proposed method's performance surpasses other methods.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Cache mechanism to avoid dulpication of same thing in hadoop system to speed ...eSAT Journals
Abstract Cloud computing provides a proper platform for hosting large-scale data-intensive applications. MapReduce is a programming model as well as a framework that supports the model. The main idea of the MapReduce model is to hide details of parallel execution and allow users to focus only on data processing strategies. Hadoop is an open-source implementation for MapReduce. For storage and analysis of online or streaming data which is big in size. Most organization are moving toward Apaches Hadoop HDFS. Applications like log processors, search engines etc. ueses hadoop Map reduce for computing and HDFS for storage. Hadoop is popular for analysis, storage and processing of very large data but require to make changes in hadoop system. There is no mechanism to identify duplicate computations which increase processing time and unnecessary data transmission .To co-locate related files by considering content and using locality sensitive hashing algorithm. By storing related files in same cluster using cache mechanism which improve data locality mechanism and avoids repeated execution of task, both helps to speed up execution of hadoop. Keywords-Distributed file system, Datanode, Locality Sensitive Hashing
MAP/REDUCE DESIGN AND IMPLEMENTATION OF APRIORIALGORITHM FOR HANDLING VOLUMIN...acijjournal
Apriori is one of the key algorithms to generate frequent itemsets. Analysing frequent itemset is a crucial
step in analysing structured data and in finding association relationship between items. This stands as an
elementary foundation to supervised learning, which encompasses classifier and feature extraction
methods. Applying this algorithm is crucial to understand the behaviour of structured data. Most of the
structured data in scientific domain are voluminous. Processing such kind of data requires state of the art
computing machines. Setting up such an infrastructure is expensive. Hence a distributed environment
such as a clustered setup is employed for tackling such scenarios. Apache Hadoop distribution is one of
the cluster frameworks in distributed environment that helps by distributing voluminous data across a
number of nodes in the framework. This paper focuses on map/reduce design and implementation of
Apriori algorithm for structured data analysis.
ANALYSE THE PERFORMANCE OF MOBILE PEER TO PEER NETWORK USING ANT COLONY OPTIM...ijcsity
A mobile peer-to-peer computer network is the one in which each computer in the network can act as a
client or server for the other computers in the network. The communication process among the nodes in the
mobile peer to peer network requires more no of messages. Due to this large number of messages passing,
propose an interconnection structure called distributed Spanning Tree (DST) and it improves the efficiency
of the mobile peer to peer network. The proposed method improves the data availability and consistency
across the entire network and also reduces the data latency and the required number of message passes for
any specific application in the network. Further to enhance the effectiveness of the proposed system, the
DST network is optimized with the Ant Colony Optimization method. It gives the optimal solution of the
DST method and increased availability, enhanced consistency and scalability of the network. The
simulation results shows that reduces the number of message sent for any specific application and average
delay and increases the packet delivery ratio in the network.
Similar to The Impact of Data Replication on Job Scheduling Performance in Hierarchical Data Grid (20)
ON THE PROBABILITY OF K-CONNECTIVITY IN WIRELESS AD HOC NETWORKS UNDER DIFFER...graphhoc
We compare the probability of k-Connectivity of an ad hoc network under Random Way Point (RWP),City Section and Manhattan mobility models. A Network is said to be k Connected if there exists at least k edge disjoint paths between any pair of nodes in that network at any given time and velocity. Initially, for each of the three mobility models, the movement of the each node in the ad hoc network at a given velocity and time are captured and stored in the Node Movement Database (NMDB). Using the movements in the NMDB, the location of the node at a given time is computed and stored in the Node
Location Database (NLDB).
DISTANCE TWO LABELING FOR MULTI-STOREY GRAPHSgraphhoc
An L (2, 1)-labeling of a graph G (also called distance two labeling) is a function f from the vertex set V (G) to the non negative integers {0,1,…, k }such that |f(x)-f(y)| ≥2 if d(x, y) =1 and | f(x)- f(y)| ≥1 if d(x, y) =2. The L (2, 1)-labeling number λ (G) or span of G is the smallest k such that there is a f with
max {f (v) : vє V(G)}= k. In this paper we introduce a new type of graph called multi-storey graph. The distance two labeling of multi-storey of path, cycle, Star graph, Grid, Planar graph with maximal edges and its span value is determined. Further maximum upper bound span value for Multi-storey of simple
graph are discussed.
Impact of Mobility for Qos Based Secure Manet graphhoc
Secure multicast communication in Mobile Adhoc Networks (MANETs) is challenging due to its inherent characteristics of infrastructure-less architecture with lack of central authority, limited resources such as bandwidth, energy and power. Several group oriented applications over MANETs create new challenges to routing protocols in terms of QOS requirements. In many multicast interactions, due to its frequent node mobility, new member can join and current members can leave at a time. It is necessary to choose a routing protocol which establishes true connectivity between the mobile nodes. The pattern of movement of members is classified into different mobility models and each one has its own distinct features. It is a crucial part in the performance of MANET. Hence key management is the fundamental challenge in achieving secure communication using multicast key distribution for mobile adhoc networks. This paper describes the impact of mobility models for the performance of a new cluster-based multicast tree algorithm with destination sequenced distance vector routing protocol in terms of QOS requirements such as end to end delay, energy consumption and key delivery ratio. For simulation purposes, three mobility models are considered. Simulation results illustrate the performance of routing protocol with different mobility models and different mobility speed under varying network conditions.
A Transmission Range Based Clustering Algorithm for Topology Control Manetgraphhoc
This paper presents a novel algorithm for clustering of nodes by transmission range based clustering (TRBC).This algorithm does topology management by the usage of coverage area of each node and power management based on mean transmission power within the context of wireless ad-hoc networks. By reducing the transmission range of the nodes, energy consumed by each node is decreased and topology is formed. A new algorithm is formulated that helps in reducing the system power consumption and prolonging the battery life of mobile nodes. Formation of cluster and selection of optimal cluster head and thus forming the optimal cluster taking weighted metrics like battery life, distance, position and mobility is done based on the factors such as node density, coverage area, contention index, required and current node degree of the nodes in the clusters
A Battery Power Scheduling Policy with Hardware Support In Mobile Devices graphhoc
A major issue in the ad hoc networks with energy constraints is to find ways that increase their lifetime. The use of multihop radio relaying requires a sufficient number of relaying nodes to maintainnetwork connectivity. Hence, battery power is a precious resource that must be used efficiently in order to avoid early termination of any node. In this paper, a new battery power scheduling policy based on dynamic programming is proposed for mobile devices.This policy makes use of the state information of each cell provided by the smart battery package and uses the strategy of dynamic programming to optimally satisfy a request for power. Using extensive simulation it is proved that dynamic programming based schedulingpolicyimproves the lifetime of the mobile nodes.Also a hardware support is proposed to succeeds in distinguishing between real-time and non-real-time traffic and provides the appropriate grade of service, to meet the time constraints associated with real time traffic.
A Review of the Energy Efficient and Secure Multicast Routing Protocols for ...graphhoc
This paper presents a thorough survey of recent work addressing energy efficient multicast routing protocols and secure multicast routing protocols in Mobile Ad hoc Networks (MANETs). There are so many issues and solutions which witness the need of energy management and security in ad hoc wireless networks. The objective of a multicast routing protocol for MANETs is to support the propagation of data from a sender to all the receivers of a multicast group while trying to use the available bandwidth efficiently in the presence of frequent topology changes. Multicasting can improve the efficiency of the wireless link when sending multiple copies of messages by exploiting the inherent broadcast property of wireless transmission. Secure multicast routing plays a significant role in MANETs. However, offering energy efficient and secure multicast routing is a difficult and challenging task. In recent years, various multicast routing protocols have been proposed for MANETs. These protocols have distinguishing features and use different mechanisms.
Case Study On Social Engineering Techniques for Persuasion Full Text graphhoc
There are plenty of security software in market; each claiming the best, still we daily face problem of viruses and other malicious activities. If we know the basic working principal of such malware then we can very easily prevent most of them even without security software. Hackers and crackers are experts in psychology to manipulate people into giving them access or the information necessary to get access. This paper discusses the inner working of such attacks. Case study of Spyware is provided. In this case study, we got 100% success using social engineering techniques for deception on Linux operating system, which is considered as the most secure operating system. Few basic principal of defend, for the individual as well as for the organization, are discussed here, which will prevent most of such attack if followed.
Breaking the Legend: Maxmin Fairness notion is no longer effective graphhoc
In this paper we analytically propose an alternative approach to achieve better fairness in scheduling mechanisms which could provide better quality of service particularly for real time application. Our proposal oppose the allocation of the bandwidth which adopted by all previous scheduling mechanism. It rather adopt the opposition approach be proposing the notion of Maxmin-charge which fairly distribute the congestion. Furthermore, analytical proposition of novel mechanism named as Just Queueing is been demonstrated
I-Min: An Intelligent Fermat Point Based Energy Efficient Geographic Packet F...graphhoc
Energy consumption and delay incurred in packet delivery are the two important metrics for measuring the performance of geographic routing protocols for Wireless Adhoc and Sensor Networks (WASN). A protocol capable of ensuring both lesser energy consumption and experiencing lesser delay in packet delivery is thus suitable for networks which are delay sensitive and energy hungry at the same time. Thus a smart packet forwarding technique addressing both the issues is thus the one looked for by any geographic routing protocol. In the present paper we have proposed a Fermat point based forwarding technique which reduces the delay experienced during packet delivery as well as the energy consumed for transmission and reception of data packets.
Fault tolerant wireless sensor mac protocol for efficient collision avoidancegraphhoc
In sensor networks communication by broadcast methods involves many hazards, especially collision. Several MAC layer protocols have been proposed to resolve the problem of collision namely ARBP, where the best achieved success rate is 90%. We hereby propose a MAC protocol which achieves a greater success rate (Success rate is defined as the percentage of delivered packets at the source reaching the destination successfully) by reducing the number of collisions, but by trading off the average propagation delay of transmission. Our proposed protocols are also shown to be more energy efficient in terms of energy dissipation per message delivery, compared to the currently existing protocol.
Enhancing qo s and qoe in ims enabled next generation networksgraphhoc
Managing network complexity, accommodating greater numbers of subscribers, improving coverage to support data services (e.g. email, video, and music downloads), keeping up to speed with fast-changing technology, and driving maximum value from existing networks – all while reducing CapEX and OpEX and ensuring Quality of Service (QoS) for the network and Quality of Experience (QoE) for the user. These are just some of the pressing business issues faced by mobileservice providers, summarized by the demand to “achieve more, for less.” The ultimate goal of optimization techniques at the network and application layer is to ensure End-user perceived QoS. The next generation networks (NGN), a composite environment of proven telecommunications and Internet-oriented mechanisms have become generally recognized as the telecommunications environment of the future. However, the nature of the NGN environment presents several complex issues regarding quality assurance that have not existed in the legacy environments (e.g., multi-network, multi-vendor, and multi-operator IP-based telecommunications environment, distributed intelligence, third-party provisioning, fixed-wireless and mobile access, etc.). In this Research Paper, a service aware policy-based approach to NGN quality assurance is presented, taking into account both perceptual quality of experience and technologydependant quality of service issues. The respective procedures, entities, mechanisms, and profiles are discussed. The purpose of the presented approach is in research, development, and discussion of pursuing the end-to-end controllability of the quality of the multimedia NGN-based communications in an environment that is best effort in its nature and promotes end user’s access agnosticism, service agility, and global mobility
Simulated annealing for location area planning in cellular networksgraphhoc
LA planning in cellular network is useful for minimizing location management cost in GSM network. In fact, size of LA can be optimized to create a balance between the LA update rate and expected paging rate within LA. To get optimal result for LA planning in cellular network simulated annealing algorithm is used. Simulated annealing give optimal results in acceptable run-time
Secure key exchange and encryption mechanism for group communication in wirel...graphhoc
Secured communication in ad hoc wireless networks is primarily important, because the communication signals are openly available as they propagate through air and are more susceptible to attacks ranging from passive eavesdropping to active interfering. The lack of any central coordination and shared wireless medium makes them more vulnerable to attacks than wired networks. Nodes act both as hosts and routers and are interconnected by Multi- hop communication path for forwarding and receiving packets to/from other nodes. The objective of this paper is to propose a key exchange and encryption mechanism that aims to use the MAC address as an additional parameter as the message specific key[to encrypt]and forward data among the nodes. The nodes are organized in spanning tree fashion, as they avoid forming cycles and exchange of key occurs only with authenticated neighbors in ad hoc networks, where nodes join or leave the network dynamically.
Simulation to track 3 d location in gsm through ns2 and real lifegraphhoc
In recent times the cost of mobile communication has dropped significantly leading to a dramatic increase in mobile phone usage. The widespread usage has led mobiles to emerge as a strong alternative for other applications one of which is tracking. This has enabled law-enforcing agencies to detect overspeeding vehicles and organizations to keep track its employees. The 3 major ways of tracking being employed presently are (a) via GPS [1] (b) signal attenuation property of a packet [3] and (c) using GSM Network [2]. The initial cost of GPS is very high resulting in low usage whereas (b) needs a very high precision measuring device. The paper presents a GSM-based tracking technique which eliminates the above mentioned overheads, implements it in NS2 and shows the limitations of the real life simulation. An accuracy of 97% was achieved during NS2 simulation which is comparable to the above mentioned alternate methods of tracking.
Performance Analysis of Ultra Wideband Receivers for High Data Rate Wireless ...graphhoc
For high data rate ultra wideband communication system, performance comparison of Rake, MMSE and Rake-MMSE receivers is attempted in this paper. Further a detail study on Rake-MMSE time domain equalizers is carried out taking into account all the important parameters such as the effect of the number of Rake fingers and equalizer taps on the error rate performance. This receiver combats inter-symbol interference by taking advantages of both the Rake and equalizer structure. The bit error rate performances are investigated using MATLAB simulation on IEEE 802.15.3a defined UWB channel models. Simulation results show that the bit error rate probability of Rake-MMSE receiver is much better than Rake receiver and MMSE equalizer. Study on non-line of sight indoor channel models illustrates that bit error rate performance of Rake-MMSE (both LE and DFE) improves for CM3 model with smaller spread compared to CM4 channel model. It is indicated that for a MMSE equalizer operating at low to medium SNR values, the number of Rake fingers is the dominant factor to improve system performance, while at high SNR values the number of equalizer taps plays a more significant role in reducing the error rate.
Coverage and Connectivity Aware Neural Network Based Energy Efficient Routing...graphhoc
There are many challenges when designing and deploying wireless sensor networks (WSNs). One of the key challenges is how to make full use of the limited energy to prolong the lifetime of the network, because energy is a valuable resource in WSNs. The status of energy consumption should be continuously monitored after network deployment. In this paper, we propose coverage and connectivity aware neural network based energy efficient routing in WSN with the objective of maximizing the network lifetime. In the proposed scheme, the problem is formulated as linear programming (LP) with coverage and connectivity aware constraints. Cluster head selection is proposed using adaptive learning in neural networks followed by coverage and connectivity aware routing with data transmission. The proposed scheme is compared with existing schemes with respect to the parameters such as number of alive nodes, packet delivery fraction, and node residual energy. The simulation results show that the proposed scheme can be used in wide area of applications in WSNs.
An Overview of Mobile Ad Hoc Networks for the Existing Protocols and Applicat...graphhoc
Mobile Ad Hoc Network (MANET) is a collection of two or more devices or nodes or terminals with
wireless communications and networking capability that communicate with each other without the aid of
any centralized administrator also the wireless nodes that can dynamically form a network to exchange
information without using any existing fixed network infrastructure. And it’s an autonomous system in
which mobile hosts connected by wireless links are free to be dynamically and some time act as routers at
the same time, and we discuss in this paper the distinct characteristics of traditional wired networks,
including network configuration may change at any time , there is no direction or limit the movement and
so on, and thus needed a new optional path Agreement (Routing Protocol) to identify nodes for these
actions communicate with each other path, An ideal choice way the agreement should not only be able to
find the right path, and the Ad Hoc Network must be able to adapt to changing network of this type at any
time. and we talk in details in this paper all the information of Mobile Ad Hoc Network which include the
History of ad hoc, wireless ad hoc, wireless mobile approaches and types of mobile ad Hoc networks, and
then we present more than 13 types of the routing Ad Hoc Networks protocols have been proposed. In this
paper, the more representative of routing protocols, analysis of individual characteristics and advantages
and disadvantages to collate and compare, and present the all applications or the Possible Service of Ad
Hoc Networks
An Algorithm for Odd Graceful Labeling of the Union of Paths and Cycles graphhoc
In 1991, Gnanajothi [4] proved that the path graph n
P with n vertex and n −1edge is odd graceful, and
the cycle graph Cm with m vertex and m edges is odd graceful if and only if m even, she proved the
cycle graph is not graceful if m odd. In this paper, firstly, we studied the graphCm∪Pn when m = 4, 6,8,10
and then we proved that the graphCm∪Pn
is odd graceful if m is even. Finally, we described an
algorithm to label the vertices and the edges of the vertex set ( ) m n
V C ∪P and the edge set ( ) m n
E C ∪P .
ACTOR GARBAGE COLLECTION IN DISTRIBUTED SYSTEMS USING GRAPH TRANSFORMATIONgraphhoc
A lot of research work has been done in the area of Garbage collection for both uniprocessor and
distributed systems. Actors are associated with activity (thread) and hence usual garbage collection
algorithms cannot be applied for them. Hence a separate algorithm should be used to collect them. If we
transform the active reference graph into a graph which captures all the features of actors and looks like
passive reference graph then any passive reference graph algorithm can be applied for it. But the cost of
transformation and optimization are the core issues. An attempt has been made to walk through these
issues.
A Proposal Analytical Model and Simulation of the Attacks in Routing Protocol...graphhoc
In this work we have devoted to some proposed analytical methods to simulate these attacks, and node mobility in MANET. The model used to simulate the malicious nodes mobility attacks is based on graphical theory, which is a tool for analyzing the behavior of nodes. The model used to simulate the Blackhole cooperative, Blackmail, Bandwidth Saturation and Overflow attacks is based on malicious nodes and the number of hops. We conducted a simulation of the attacks with a C implementation of the proposed mathematical models.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex ProofsAlex Pruden
This paper presents Reef, a system for generating publicly verifiable succinct non-interactive zero-knowledge proofs that a committed document matches or does not match a regular expression. We describe applications such as proving the strength of passwords, the provenance of email despite redactions, the validity of oblivious DNS queries, and the existence of mutations in DNA. Reef supports the Perl Compatible Regular Expression syntax, including wildcards, alternation, ranges, capture groups, Kleene star, negations, and lookarounds. Reef introduces a new type of automata, Skipping Alternating Finite Automata (SAFA), that skips irrelevant parts of a document when producing proofs without undermining soundness, and instantiates SAFA with a lookup argument. Our experimental evaluation confirms that Reef can generate proofs for documents with 32M characters; the proofs are small and cheap to verify (under a second).
Paper: https://eprint.iacr.org/2023/1886
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...
The Impact of Data Replication on Job Scheduling Performance in Hierarchical Data Grid
1. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
DOI : 10.5121/jgraphoc.2010.2302 15
THE IMPACT OF DATA REPLICATION ON JOB
SCHEDULING PERFORMANCE IN HIERARCHICAL
DATA GRID
Somayeh Abdi1
, Somayeh Mohamadi2
1
Department of Computer Engineering, Eslamabad Gharb branch Islamic Azad
University, Eslamabad Gharb,Kermanshah, Iran
Somayeh.abdi@gmail.com
1
Department of Computer Engineering, Ghasre Shirin branch Islamic Azad University,
Ghasre Shirin, Kermanshah, Iran
s.mohamadi@gmail.com
ABSTRACT
In data-intensive applications data transfer is a primary cause of job execution delay. Data access time
depends on bandwidth. The major bottleneck to supporting fast data access in Grids is the high latencies
of Wide Area Networks and Internet. Effective scheduling can reduce the amount of data transferred
across the internet by dispatching a job to where the needed data are present. Another solution is to use a
data replication mechanism. Objective of dynamic replica strategies is reducing file access time which
leads to reducing job runtime. In this paper we develop a job scheduling policy and a dynamic data
replication strategy, called HRS (Hierarchical Replication Strategy), to improve the data access
efficiencies. We study our approach and evaluate it through simulation. The results show that our
algorithm has improved 12% over the current strategies.
KEYWORDS
Grid, Data Grid, Job Scheduling, Data Replication, Simulation
1. INTRODUCTION
In the increasing demand of scientific and large-scale business application, a large amount of
data are generated and spread for using by users around the world. Many good examples can be
listed such as High Energy Physics, meteorology, computational genomics which processes and
results large amount of data. Such data cannot be stored centralized in any site but distributed
among centre around the world [1]. Data grid tries to store this data in decentralize sites and
then for each application retrieves it from these sites. Data grid as an important branch of grid
computing focuses on supporting an efficient management mechanism for controlled sharing
and large amounts of distributed data. In Data Grid, for each incoming job, the Grid scheduler
decides where to run the job based on the job requirements and the system status. Scheduling
jobs to suitable grid sites is necessary because data movement between different grid sites is
time consuming. If a job is scheduled to a site where the required data are present, the job can
process data in this site without any transmission delay for getting data from a remote site.
2. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
16
Data replication is an important optimization step to manage large data by replicating data in
geographically distributed data stores. When users’ jobs access a large amount of data from
remote sites, dynamic replica optimizer running in the site tries to store replicas on local storage
for future possible repeated requests.
We proposed a new job scheduling algorithm and data replication policy that reduce job
execution time by reducing job data access time. Our new scheduling policy considers the
locations of required data and the job queue length of a computing node. We develop a
replication strategy, called HRS (Hierarchical Replication Strategy). It takes into account
bandwidth as an important factor for replica selection and replica placement. It also increases
the chances of accessing data at a nearby node.
The rest of the paper is organized as follows, in section 2 we present a summary of existing and
related work. In section 3, a hierarchical structure is proposed for data grid based on
classification of networks, along with a scheduling algorithm and replication algorithm for this
structure, Section 4 describes our experiments and the results achieved followed by conclusion
in section 5.
2. RELATED WORK
There are some recent works that address the problem of scheduling and/ or replication in Data
Grid as well as the combination between them.
In [3], it considers two centralized and decentralized replication algorithms. In centralized
method, replica master uses a table that ranks each file access in descending order. If a file
access is less than the average, it will be removed from the table. Then it pop files from top and
replicates using a response-time oriented replica placement algorithm. In the decentralized
method, every site records file access in its table and exchange this table with neighbours. Since
every domain knows average number of access for each file and then deletes those files whose
access is less than the average, and replicates other files in its local storage.
In [5], an algorithm for a two-level hierarchical structure based on internet hierarchy (BHR) has
been introduced which only considers dynamic replication and does not consider scheduling.
Nodes in the first level are connected to each other with high speed networks and in the second
level via internet. The algorithm replicates the file to the site if there is enough space. Next it,
accesses the file remotely if the file is available in the sites that are in the same region.
Otherwise it tries to make available space by deleting files using LRU (Least Recently Used)
method, and replicates the file. It assumes that master site always has a safe copy of file before
deleting.
In [6], a structure with few networks connected via internet has been presented and an algorithm
similar to [5], along with scheduling is proposed. For replicating a file, first computes the total
transfer time, then it selects the best node with shortest transfer time.
In [7] authors introduce dynamic replication placement (RP) that categorizes the data based on
their property. This category is used for job scheduling and replication. Then a job is allocated
to a site which has the file in the required category, this leads to reduce the cost for file transfer.
In [8] a Genetic Algorithm based co-scheduling of data and jobs of independent nature was
proposed; the GA is executed to converge to a schedule by looking at the jobs in the scheduler
queue as well as the replicated data objects at once. A performance overhead might be incurred
by the system as a result of delaying the replication of the data until the scheduling time. The
authors also use an objective function that assumes infinite availability of storage for all data
objects which is infeasible in a realistic grid setting.
3. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
17
3. THE PROPOSED METHOD
In this section, we present hierarchical network structure and then we proposed two strategies
for job scheduling and data replication by considering the hierarchical network structure.
3.1. NETWORK STRUCTURE
The proposed Data Grid structure is shown in Figure 1. This structure has two levels. Regions
comprise the first level and sites that are located in a region comprise the second level. A region
represents an organization unit which is a group of sites that are geographically close to each
other, each region comprises the computers which are connected by a high bandwidth. We
define two kinds of communications between sites, inter-communication and intra-
communication. Intra-communication is the communication between sites within the same
region and inter-communication is the communication between sites across regions. Network
bandwidth between sites within a region will be larger than across regions.
In communication networks, the performance of system is underlying available network
bandwidth and data access latency, especially in networks that hierarchy of bandwidth appears.
Therefore, to reduce access latency and to avoid WAN bandwidth bottleneck, it is important to
reduce the number of inter-communications.
Data Grid Information Service (DGIS) providing resource registration services and keeping
track of a list of resources available in the Data Grid. The Grid Scheduler can query this for
resource contact, configuration, and status information; Resource discovery identify resources
that can be used with their capability through Data Grid Information Service.
In this structure we apply distributed dynamic data replication infrastructure. Replica Manager
at each site manages the data movement between sites. When a job is assigned to Local
Scheduler of site, those required data that do not exist in local site will be transfer to it. Data
request by the job is generated as soon as job is scheduled into Local Scheduler queue and
Replica Manager controls data movement at each site.
In proposed structure there is a centralized Replica Catalogue. It is responsible for indexing
available files on the resources and handles queries from users and resources about location of
replicas.
4. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
18
Figure 1. Data Grid Architecture.
3.2. SCHEDULING ALGORITHM
Our new scheduling policy considers the locations of required data, computing capacity and
number of jobs that allocate on computing nodes. For efficient scheduling of any job, the
algorithm determines the best site and then submit job to Local scheduler. A best site is a site
that holds most of the requested files (from size point of view). This will significantly reduce
total transfer time, and reduce the job execution time.
Each job requires some replicas and the replicas is needed to execute job j are represented as: Rj
= {LFN1, LFN2,…., LFNn}. We assume that jS
is total size of the requested files available in
site j, Ci is computing capacity (in MIPS1) of site i and RelativeLoadi is relative load of site i.
1
. Million Instruction Per Second
5. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
19
∑=
SsiteinLFNiavailableallfor
iS LFNS (1)
RelativeLoadi= SizeofJobsi / Ci (2)
SizeofJobsi is the lengths of queued jobs (in MIPS) on site i and iLFN present size of file
LFNi .
The scheduling algorithm can be summarized as follow:
1- Compute SC for each site from (1)
2- Select the best site i
q
i
SMAXS
1
max
=
= ; where q is the number of site (i.e. site with largest
available requested data files)
3- If there are several sites with most available data, select the site with minimum relative
load from (2).
When a job is assigned to local scheduler, the Replica Manager transfers all the requested files
that are not exist in local site. The objective of file replicating is to transferring required data to
local site before job execution. Therefore, data replication improves job scheduling performance
by reducing job execution time.
The proposed scheduling algorithm takes into consideration the requested data files.
Additionally, where there are the same data requirements in several sites (from size point of
view); it takes into consideration the computing capacity of sites with a view to reducing the
queue waiting time and relative load of sites.
3.3. HIERARCHICAL REPLICATION STRATEGY
After a job is scheduled to site Sj , the requested data will be transferred to Sj to become
replicas. HRS (hierarchical replication Strategy) determines wich replica will be transferred to
Sj and how to handle this new replica. HRS considers the bandwidth between the regions as the
main factor for replica selection / deletion.
When each site stores a new replica, replica manager sends a file register request to RC and then
RC adds this site to the list of sites that holds the replica. Replica Manager controls data
transferring in each site and provides a mechanism for accessing the Replica Catalogue. Data
Grid Architecture for Replica Catalogue and Replica Manager is shwon in in Figure 2.
6. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
20
Figure 2. Data Grid Architecture for replica cataloge and replica manager.
For each required file of the job, Replica Manager controls the existence of the file in local site.
If file doesn’t exist in the local site HRS first searches the file in local region. If the file
duplicated in the same region, then it will creates a list of candidate replicas and selects a replica
with the maximum bandwidth available for transferring it. If there is enough space for new
replica, then stores it in the local site, otherwise it is only stored in the temporary buffer and will
be deleted after the job completes.
If the file doesn’t exit in the same region, then HRS creates a list of replicas in other regions and
selects the replica with the maximum bandwidth available for transferring it. If there is enough
space for new replica, then stores it in the local site, otherwise occupided space will be released
to make engough room for this new replica. First, it removes the replicas that already exist in
other sites in the same region based on LRU(least recently used) replacement algorithm. After
all these replicas are deleted, if the space is still insufficient, the HRS use LRU replacement
algorithm to delete replica in local storage which is duplicated in other regions , till it has
enough room for the new replica.
4. SIMULATIONS
Gridsim is used as the simulation tool to evaluate the performance of the proposed replication
and scheduling algorithms. The Java-based GridSim discrete event simulation toolkit provides
Java classes that represent entities essential for application, resource modelling, scheduling of
jobs to resources, and their execution along with management [9]. Its Java-based design makes
it portable and available on all computational platforms. The components of Gridsim are as
follow and also depicted in Figure 3.
7. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
21
Figure 3. Simulator architecture
Resource Broker: It receives jobs from user and sends them to the best node according to
proposed algorithm; Storage Element (SE): Storage resource in grid; Computing Element (CE):
Computing resource in grid; Replica Manager:controls data transfering at each site; Replica
Catalogue: stores list of sites that holds replicas.
Based on the scheduling algorithm the broker sends jobs to a node. Each job needs a list of files
to run. Reducing file access time is the final objective of optimization algorithms.
4.1. Simulation Environment
There are four regions in our configuration and each region has an average of 13 sites, which all
have CE with associated SE. Table1 specifies the simulation parameters used in our study.
There are 5 job types; each job type requires 12 files to execute. While running, jobs were
randomly picked from 5 job types, then submitted to the Resource Broker. Files are accessed
sequentially within a job without any access pattern. To simplify the requirements, data
replication approaches in Data Grid environments commonly assume that the data is read-only.
8. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
22
4.2. Simulation results and discussion
HRS will be compared with LRU (Least Recently Used) and BHR (Bandwidth Hierarchy based
Replication). The LRU algorithm always replicates and then deletes those files that have been
used least recently. Figure 4 shows the Average job time based on changing number of jobs for
3 algorithms. Figure 5 shows the Average job time for 1000 jobs by 3 mentioned algorithms.
HRS replication strategy uses the concept of “network locality” as BHR [5].
0
5000
10000
15000
20000
25000
100 200 300 400 500
Number of Jobs
AverageJobTime(Sec)
HRS
BHR
LRU
Figure 4. Average job Time based on varying number of jobs.
TABLE 1. Simulation parameters
valueTopology Parameters
4Number of region
13Number of sites in each region
10 GBStorage space at each site
10 MbpsConnectivity bandwidth(WAN)
1000MbpsConnectivity bandwidth(LAN)
valueJob parameters
500Number of jobs
5Number of job types
12Number of file accessed per job
500 MBSize of single file
50GBTotal size of files
9. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
23
The difference between HRS and BHR is that required replica within the same region is
always the top priority used in HRS, while BHR searches all sites to find the best replica and
has no distinction between intra-region communication and inter-region comuuniation. It could
be anticipated that HRS will avoid inter-region communications and be stable in hierarchical
network architecture with variable bandwidth. Our method takes benefit from network level
locality of BHR. Thus, total job execution time is about 12% faster using HRS optimizer than
BHR.
0
10000
20000
30000
40000
50000
MeanJobsTime(sec)
1000
Number of Jobs
LRU
BHR
HRS
Figure 5. Average Job Time for 1000 jobs.
The job execution time is the Max{file transmission time, queue time} plus job processing time.
Since the file transmission time is the most important factor to influence the job execution time
for data-intensive jobs in data grids, the proposed scheduling algorithm with HRS can reduce
the file transmission time effectively by virtue of valid scheduling and proper data replication,
as can be seen from the experiments.
The average number of inter-communications for a job execution is illustrated in Figure 6. By
selecting the best site based on location of required data by the job, the proposed scheduling
algorithm with HRS can decrease the number of inter-communications effectively. Overall the
simulation results with Gridsim show better performance (over 12%) comparing to current
algorithms.
10. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
24
0
1
2
3
4
5
6
7
8
NumberofAverage
inter-communication
100 300 500
Number of jobs
BHR
HRS
Figure 6. Average number of inter-communications
Figures 7 shows the average job time for 500 jobs. We compare HRS, BHR and LRU
algorithms for varying inter-communication bandwidth. As inter-communication bandwidth
increase 3 mentioned algorithms will converge. We can conclude that HRS strategy can be
effectively utilized when hierarchy of bandwidth appears because BHR has no distinction
between intra-region and inter-region. It could be anticipated that HRS will avoid inter-region
communications and be stable in hierarchical network architecture with variable bandwidth.
0
5000
10000
15000
20000
25000
10 50 100 300 600 1000
inter-communication Bandwidth
AverageJobTime(Sec)
HRS
BHR
LRU
Figure 7. Average Jobs Time with varying inter-communication bandwidth for 500 jobs
Overall the simulation results with Gridsim show better performance (over 12%) comparing to
current algorithms.
11. International journal on applications of graph theory in wireless ad hoc networks and sensor networks
(GRAPH-HOC) Vol.2, No.3, September 2010
25
5. CONCLUSION AND FUTURE WORK
In this paper a hierarchical structure for dynamic replicating file and scheduling in data grids
was proposed. To achieve good network bandwidth utilization and reduce data access time, we
propose a job scheduling policy that considers computational capability, job type and data
location in job placement decision. We study and evaluate the performance of various replica
strategies. The simulation results show, first of all, that proposed scheduling algorithm and HRS
both get better performances. Second, we can achieve particularly good performance with
scheduling algorithm where jobs are always scheduled to site with most of the data needed, and
a separate HRS process at each site for replication management. Experimental data show
proposed scheduling algorithm with HRS replica strategy outperforms others combinations in
total job execution time.
REFERENCES
[1] K. Ranganathan and I. Foster, “Identifying Dynamic Replication Strategies for a High
Performance Data Grid”. In Proceedings of the International Grid Computing Workshop,
Denver, Colorado, USA, 2001.
[2] Jianhua Jiang, Huifang Ji, “scheduling algorithm with potential behaviors”, Journal of
Computers, VOL. 3 , NO. 12 , December 2008.
[3] Ming Tang, Bu-Sung Lee, Xueyan Tang, Chai-Kiat Yeo .” The Impact of Data Replication on
Job Scheduling Performance in the Data Grid”, Future Generation Computer Systems, Volume
22, Issue 3, February 2006, Pages 254-268
[4] Ali Elghirani, Riky Subrata, Albert Y. Zomaya, and Ali Al Mazari., “Performance Enhancement
through Hybrid Replication and Genetic Algorithm Co-Scheduling in Data Grids”, Advanced
Networks Research Group, School of Information Technologies, University of Sydney, NSW
2006 Australia.
[5] Sang-Min Park, Jai-Hoon Kim, Young-Bae Ko: “Dynamic Grid Replication Strategy based on
Internet Hierarchy”, Book Series Lecture Notes in Computer Science, Grid and Cooperative
omputing book,Publisher Springer, August 2005, Volume 3033/2004, Pages 838-846
[6] Ruay-Shiung Chang, Jih-Sheng Chang, Shin-Yi Lin, “Job scheduling and data replication on
data grids”, Future Generation Computer Systems, Volume 23, Issue 7, August 2007, Pages 846-
860
[7] Nhan Nguyen Dang, Sang Boem Lim2: “Combination of Replication and Scheduling in Data
Grids”, IJCSNS International Journal of Computer Science and Network Security, VOL.7 No.3,
March 2007.
[8] T. Phan, K. Ranganathan, and R. Sion, “Evolving toward the perfect schedule: Co-scheduling
job assignments and data replication in wide-area systems using a genetic algorithm”, Job
scheduling strategies for parallel processing (11th international workshop), JSSPP 2005,
Cambridge MA, 2005.
[9] Klaus Krauter, Rajkumar Buyya and Muthucumaru Maheswaran, “A taxonomy and survey of
grid resource management systems for distributed computing”, SOFTWARE—PRACTICE
AND EXPERIENCE Softw. Pract. Exper. 2002; 32:135–164 (DOI: 10.1002/spe.432).
[10] Klaus Krauter and Manzur Murshed, GridSim a toolkit for the modelling and simulation of
distributed resource management and scheduling for Grid computing, CONCURRENCY AND
COMPUTATION: PRACTICE AND EXPERIENCE Concurrency Computat.: Pract. Exper.
2002; 14:1175–1220 (DOI: 10.1002/cpe.710).
[11] The Data Grid Project. http://www.eu-datagrid.org
[12] Parallel workload Project. http://www.parrallelworkload.org.