Many computational solutions can be expressed as Directed Acyclic Graph (DAG), in which
nodes represent tasks to be executed and edges represent precedence constraints among tasks.
A Cluster of processors is a shared resource among several users and hence the need for a
scheduler which deals with multi-user jobs presented as DAGs. The scheduler must find the
number of processors to be allotted for each DAG and schedule tasks on allotted processors. In
this work, a new method to find optimal and maximum number of processors that can be allotted
for a DAG is proposed. Regression analysis is used to find the best possible way to share
available processors, among suitable number of submitted DAGs. An instance of a scheduler
for each DAG, schedules tasks on the allotted processors. Towards this end, a new framework
to receive online submission of DAGs, allot processors to each DAG and schedule tasks, is
proposed and experimented using a simulator. This space-sharing of processors among multiple
DAGs shows better performance than the other methods found in literature. Because of spacesharing,
an online scheduler can be used for each DAG within the allotted processors. The use
of online scheduler overcomes the drawbacks of static scheduling which relies on inaccurate
estimated computation and communication costs. Thus the proposed framework is a promising
solution to perform online scheduling of tasks using static information of DAG, a kind of hybrid
scheduling.
Sharing of cluster resources among multiple Workflow Applicationsijcsit
Many computational solutions can be expressed as workflows. A Cluster of processors is a shared
resource among several users and hence the need for a scheduler which deals with multi-user jobs
presented as workflows. The scheduler must find the number of processors to be allotted for each workflow
and schedule tasks on allotted processors. In this work, a new method to find optimal and maximum
number of processors that can be allotted for a workflow is proposed. Regression analysis is used to find
the best possible way to share available processors, among suitable number of submitted workflows. An
instance of a scheduler is created for each workflow, which schedules tasks on the allotted processors.
Towards this end, a new framework to receive online submission of workflows, to allot processors to each
workflow and schedule tasks, is proposed and experimented using a discrete-event based simulator. This
space-sharing of processors among multiple workflows shows better performance than the other methods
found in literature. Because of space-sharing, an instance of a scheduler must be used for each workflow
within the allotted processors. Since the number of processors for each workflow is known only during
runtime, a static schedule can not be used. Hence a hybrid scheduler which tries to combine the advantages
of static and dynamic scheduler is proposed. Thus the proposed framework is a promising solution to
multiple workflows scheduling on cluster.
Scheduling Algorithm Based Simulator for Resource Allocation Task in Cloud Co...IRJET Journal
This document proposes a scheduling algorithm for allocating resources in cloud computing based on the Project Evaluation and Review Technique (PERT). It aims to address issues like starvation of lower priority tasks. The algorithm models task allocation as a directed acyclic graph and uses PERT to schedule critical and non-critical tasks, prioritizing higher priority tasks. The algorithm is evaluated against other scheduling methods and shows improvements in reducing completion time and optimizing resource allocation for all tasks.
This document proposes a genetic algorithm called Workflow Scheduling for Public Cloud Using Genetic Algorithm (WSGA) to optimize the cost of executing workflows in the public cloud. It discusses how genetic algorithms can be applied to the workflow scheduling problem to generate optimal schedules. The WSGA represents potential scheduling solutions as chromosomes, uses a fitness function to evaluate scheduling costs, and applies genetic operators like selection, crossover and mutation to evolve new schedules over multiple iterations. The goal is to minimize total execution cost while meeting workflow dependencies and deadline constraints. An experimental setup is described and the WSGA approach is claimed to reduce costs more than other heuristic scheduling algorithms for communication-intensive workflows.
A novel scheduling algorithm for cloud computing environmentSouvik Pal
The document describes a proposed genetic algorithm-based scheduling approach for cloud computing environments. It aims to minimize waiting time and queue length. The algorithm first permutes task burst times and finds minimum waiting times using FCFS and genetic algorithms. It then applies a queuing model to the sequences with minimum waiting time from each approach. Experimental results on 4 sample tasks show the genetic algorithm reduces waiting time compared to FCFS. The genetic operators of selection, crossover and mutation are applied to evolve optimal task scheduling sequences.
Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing IJORCS
This paper shows the importance of fair scheduling in grid environment such that all the tasks get equal amount of time for their execution such that it will not lead to starvation. The load balancing of the available resources in the computational grid is another important factor. This paper considers uniform load to be given to the resources. In order to achieve this, load balancing is applied after scheduling the jobs. It also considers the Execution Cost and Bandwidth Cost for the algorithms used here because in a grid environment, the resources are geographically distributed. The implementation of this approach the proposed algorithm reaches optimal solution and minimizes the make span as well as the execution cost and bandwidth cost.
This document presents a scheduling strategy that performs dynamic job grouping at runtime to optimize the execution of applications with many fine-grained tasks on global grids. The strategy groups individual jobs into larger "job groups" based on the processing requirements of each job, the capabilities of available grid resources, and a defined granularity size. It aims to minimize overall job execution time and cost while maximizing resource utilization. The strategy is evaluated through simulations using the GridSim toolkit, which models grid resources and application scheduling.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
An enhanced adaptive scoring job scheduling algorithm with replication strate...eSAT Publishing House
This document describes an enhanced adaptive scoring job scheduling algorithm with replication strategy for grid environments. The algorithm aims to improve upon an existing adaptive scoring job scheduling algorithm by identifying whether jobs are data-intensive or computation-intensive. It then divides large jobs into subtasks, replicates the subtasks, and allocates the replicas to clusters based on a computed cluster score in order to improve resource utilization and job completion times. The algorithm is evaluated through simulation using the GridSim toolkit.
Sharing of cluster resources among multiple Workflow Applicationsijcsit
Many computational solutions can be expressed as workflows. A Cluster of processors is a shared
resource among several users and hence the need for a scheduler which deals with multi-user jobs
presented as workflows. The scheduler must find the number of processors to be allotted for each workflow
and schedule tasks on allotted processors. In this work, a new method to find optimal and maximum
number of processors that can be allotted for a workflow is proposed. Regression analysis is used to find
the best possible way to share available processors, among suitable number of submitted workflows. An
instance of a scheduler is created for each workflow, which schedules tasks on the allotted processors.
Towards this end, a new framework to receive online submission of workflows, to allot processors to each
workflow and schedule tasks, is proposed and experimented using a discrete-event based simulator. This
space-sharing of processors among multiple workflows shows better performance than the other methods
found in literature. Because of space-sharing, an instance of a scheduler must be used for each workflow
within the allotted processors. Since the number of processors for each workflow is known only during
runtime, a static schedule can not be used. Hence a hybrid scheduler which tries to combine the advantages
of static and dynamic scheduler is proposed. Thus the proposed framework is a promising solution to
multiple workflows scheduling on cluster.
Scheduling Algorithm Based Simulator for Resource Allocation Task in Cloud Co...IRJET Journal
This document proposes a scheduling algorithm for allocating resources in cloud computing based on the Project Evaluation and Review Technique (PERT). It aims to address issues like starvation of lower priority tasks. The algorithm models task allocation as a directed acyclic graph and uses PERT to schedule critical and non-critical tasks, prioritizing higher priority tasks. The algorithm is evaluated against other scheduling methods and shows improvements in reducing completion time and optimizing resource allocation for all tasks.
This document proposes a genetic algorithm called Workflow Scheduling for Public Cloud Using Genetic Algorithm (WSGA) to optimize the cost of executing workflows in the public cloud. It discusses how genetic algorithms can be applied to the workflow scheduling problem to generate optimal schedules. The WSGA represents potential scheduling solutions as chromosomes, uses a fitness function to evaluate scheduling costs, and applies genetic operators like selection, crossover and mutation to evolve new schedules over multiple iterations. The goal is to minimize total execution cost while meeting workflow dependencies and deadline constraints. An experimental setup is described and the WSGA approach is claimed to reduce costs more than other heuristic scheduling algorithms for communication-intensive workflows.
A novel scheduling algorithm for cloud computing environmentSouvik Pal
The document describes a proposed genetic algorithm-based scheduling approach for cloud computing environments. It aims to minimize waiting time and queue length. The algorithm first permutes task burst times and finds minimum waiting times using FCFS and genetic algorithms. It then applies a queuing model to the sequences with minimum waiting time from each approach. Experimental results on 4 sample tasks show the genetic algorithm reduces waiting time compared to FCFS. The genetic operators of selection, crossover and mutation are applied to evolve optimal task scheduling sequences.
Max Min Fair Scheduling Algorithm using In Grid Scheduling with Load Balancing IJORCS
This paper shows the importance of fair scheduling in grid environment such that all the tasks get equal amount of time for their execution such that it will not lead to starvation. The load balancing of the available resources in the computational grid is another important factor. This paper considers uniform load to be given to the resources. In order to achieve this, load balancing is applied after scheduling the jobs. It also considers the Execution Cost and Bandwidth Cost for the algorithms used here because in a grid environment, the resources are geographically distributed. The implementation of this approach the proposed algorithm reaches optimal solution and minimizes the make span as well as the execution cost and bandwidth cost.
This document presents a scheduling strategy that performs dynamic job grouping at runtime to optimize the execution of applications with many fine-grained tasks on global grids. The strategy groups individual jobs into larger "job groups" based on the processing requirements of each job, the capabilities of available grid resources, and a defined granularity size. It aims to minimize overall job execution time and cost while maximizing resource utilization. The strategy is evaluated through simulations using the GridSim toolkit, which models grid resources and application scheduling.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
An enhanced adaptive scoring job scheduling algorithm with replication strate...eSAT Publishing House
This document describes an enhanced adaptive scoring job scheduling algorithm with replication strategy for grid environments. The algorithm aims to improve upon an existing adaptive scoring job scheduling algorithm by identifying whether jobs are data-intensive or computation-intensive. It then divides large jobs into subtasks, replicates the subtasks, and allocates the replicas to clusters based on a computed cluster score in order to improve resource utilization and job completion times. The algorithm is evaluated through simulation using the GridSim toolkit.
GROUPING BASED JOB SCHEDULING ALGORITHM USING PRIORITY QUEUE AND HYBRID ALGOR...ijgca
This document describes a proposed grouping based job scheduling algorithm for grid computing that aims to maximize resource utilization and minimize job processing times. It discusses related work on job scheduling algorithms and then presents the steps of the proposed algorithm. The algorithm uses shortest job first, first-in first-out, and round robin scheduling to process jobs in groups. The algorithm is evaluated experimentally in MATLAB and shown to reduce total job processing time compared to using only first-in first-out scheduling. Graphs demonstrate the processing time improvements achieved by the combined scheduling approach.
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server Environmentrahulmonikasharma
This document summarizes a research paper that proposes a new algorithm called KD-Tree approach for efficient virtual machine (VM) allocation in cloud computing environments. The algorithm aims to minimize the response time for allocating VMs to user requests. It does this by adopting a KD-Tree data structure to index physical host machines, allowing the scheduler to quickly find the host that can accommodate a new VM request with the minimum latency in O(Log n) time. The proposed approach is evaluated through simulations using the CloudSim toolkit and is shown to outperform an existing linear scheduling strategy (LSTR) algorithm in terms of reducing VM allocation times.
The Cloud computing becomes an important topic
in the area of high performance distributed computing. On the
other hand, task scheduling is considered one the most significant
issues in the Cloud computing where the user has to pay for the
using resource based on the time. Therefore, distributing the
cloud resource among the users' applications should maximize
resource utilization and minimize task execution Time. The goal
of task scheduling is to assign tasks to appropriate resources that
optimize one or more performance parameters (i.e., completion
time, cost, resource utilization, etc.). In addition, the scheduling
belongs to a category of a problem known as an NP-complete
problem. Therefore, the heuristic algorithm could be applied to
solve this problem. In this paper, an enhanced dependent task
scheduling algorithm based on Genetic Algorithm (DTGA) has
been introduced for mapping and executing an application’s
tasks. The aim of this proposed algorithm is to minimize the
completion time. The performance of this proposed algorithm has
been evaluated using WorkflowSim toolkit and Standard Task
Graph Set (STG) benchmark.
The task scheduling is a key process in large-scale distributed systems like cloud computing infrastructures
which can have much impressed on system performance. This problem is referred to as a NP-hard problem
because of some reasons such as heterogeneous and dynamic features and dependencies among the
requests. Here, we proposed a bi-objective method called DWSGA to obtain a proper solution for
allocating the requests on resources. The purpose of this algorithm is to earn the response quickly, with
some goal-oriented operations. At first, it makes a good initial population by a special way that uses a bidirectional
tasks prioritization. Then the algorithm moves to get the most appropriate possible solution in a
conscious manner by focus on optimizing the makespan, and considering a good distribution of workload
on resources by using efficient parameters in the mentioned systems. Here, the experiments indicate that
the DWSGA amends the results when the numbers of tasks are increased in application graph, in order to
mentioned objectives. The results are compared with other studied algorithms.
Service Request Scheduling in Cloud Computing using Meta-Heuristic Technique:...IRJET Journal
This document discusses using the Teaching Learning Based Optimization (TLBO) meta-heuristic technique for service request scheduling between users and cloud service providers. TLBO is a nature-inspired algorithm that mimics the teacher-student learning process. It is compared to other meta-heuristic algorithms like Genetic Algorithm. The key steps of TLBO involve initializing a population, evaluating fitness, selecting the best solution as teacher, and updating the population through teacher and learner phases until termination criteria is met. The document proposes using number of users and virtual machines as parameters for TLBO scheduling in cloud computing. MATLAB simulation results show the initial and final iterations converging to an optimal scheduling solution.
Scheduling Divisible Jobs to Optimize the Computation and Energy Costsinventionjournals
ABSTRACT : The important challenge in cloud computing environment is to design a scheduling strategy to handle jobs, and to process them in a heterogeneous environment with shared data centers. In this paper, we attempt to investigate a new analytical framework model that enables an existing private cloud data-center for scheduling jobs and minimizing the overall computation and energy cost together. Our model is based on Divisible Load Theory (DLT) model to derive closed-form solution for the load fractions to be assigned to each machines considering computation and energy cost. Our analysis also attempts to schedule the jobs such a way that cloud provider can gain maximum benefit for his service and Quality of Service (QoS) requirement user’s job. Finally, we quantify the performance of the strategies via rigorous simulation studies.
Challenges in Dynamic Resource Allocation and Task Scheduling in Heterogeneou...rahulmonikasharma
This document discusses the challenges of dynamic resource allocation and task scheduling in heterogeneous cloud environments. It outlines that resource allocation involves deciding how to allocate resources to tasks to maximize utilization, while task scheduling assigns tasks to processors to minimize execution time. The major challenges are optimizing allocated resources to minimize costs while meeting customer demands and application requirements. Allocating resources dynamically in heterogeneous cloud environments is difficult due to issues like resource contention, scarcity, and fragmentation. The document also discusses approaches to resource modeling, allocation, offering, discovery and monitoring that algorithms must address to effectively allocate resources on demand.
The document discusses optimization of resource allocation in cloud environments using a modified particle swarm optimization (PSO) approach. It proposes a Modified Resource Allocation Mutation PSO (MRAMPSO) strategy that uses an Extended Multi Queue Scheduling algorithm to schedule tasks based on resource availability and reschedules failed tasks. The MRAMPSO strategy is compared to standard PSO and other algorithms to show it can reduce execution time, makespan, transmission cost, and round trip time.
A location based least-cost scheduling for data-intensive applicationsIAEME Publication
This document summarizes a research paper that proposes a location-based least-cost scheduling algorithm for transferring multiple data-intensive files simultaneously to multiple compute nodes in a grid environment. The proposed model includes an optimized meta-scheduler that receives multiple files, predicts the optimal number of parallel TCP streams to use for each file transfer based on sampling, and schedules the files to compute nodes using a greedy algorithm that considers location and cost. Experimental results showed the optimized model achieved better transfer times and throughput compared to non-optimized transfers.
The document discusses using a genetic algorithm to schedule tasks in a cloud computing environment. It aims to minimize task execution time and reduce computational costs compared to the traditional Round Robin scheduling algorithm. The proposed genetic algorithm mimics natural selection and genetics to evolve optimal task schedules. It was tested using the CloudSim simulation toolkit and results showed the genetic algorithm provided better performance than Round Robin scheduling.
A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...IDES Editor
The data flow model is gaining popularity as a
programming paradigm for multi-core processors. Efficient
scheduling of an application modeled by Directed Acyclic
Graph (DAG) is a key issue when performance is very
important. DAG represents computational solutions, in which
the nodes represent tasks to be executed and edges represent
precedence constraints among the tasks. The task scheduling
problem in general is a NP-complete problem[2]. Several static
scheduling heuristics have been proposed. But the major
problem in static list scheduling is the inherent difficulty in
exact estimation of task cost and edge cost in a DAG and also
its inability to consider and manage with runtime behavior of
tasks. This underlines the need for dynamic scheduling of a
DAG. This paper presents how in general, dynamic scheduling
of a DAG can be done. Also proposes 4 simple methods to
perform dynamic scheduling of a DAG. These methods have
been simulated and experimented using a representative set
of DAG structured computations from both synthetic and real
problems. The proposed dynamic scheduler performance is
found to be in comparable with that of static scheduling
methods. The performance comparison of the proposed
dynamic scheduling methods is also carried out.
Optimized Assignment of Independent Task for Improving Resources Performance ...ijgca
Grid computing has emerged from category of distributed and parallel computing where the
heterogeneous resources from different network are used simultaneously to solve a particular problem that
need huge amount of resources. Potential of Grid computing depends on my issues such as security of
resources, heterogeneity of resources, fault tolerance & resource discovery and job scheduling. Scheduling
is one of the core steps to efficiently exploit the capabilities of heterogeneous distributed computing
resources and is an NP-complete problem. To achieve the promising potential of grid computing, an
effective and efficient job scheduling algorithm is proposed, which will optimized two important criteria to
improve the performance of resources i.e. makespan time & resource utilization. With this, we have
classified various tasks scheduling heuristic in grid on the basis of their characteristics.
Effective and Efficient Job Scheduling in Grid ComputingAditya Kokadwar
The integration of remote and diverse resources and the increasing computational needs of Grand Challenges problems combined with the faster growth of the internet and communication technologies leads to the development of global computational grids. Grid computing is a prevailing technology, which unites underutilized resources in order to support sharing of resources and services distributed across numerous administrative region. An efficient and effective scheduling system is essentially required in order to achieve the promising capacity of grids. The main goal of scheduling is to maximize the resource utilization and minimize processing time and cost of the jobs. In this research, the objective is to prioritize the jobs based on execution cost and then allocate the resources with minimum cost by merging it with conventional job grouping strategy to provide the solution for better and more efficient job scheduling which is beneficial to both user and resource broker. The proposed scheduling approach in grid computing employs a dynamic cost-based job scheduling algorithm for making an efficient mapping of a job to available resources in the grid. It also improves communication to computation ratio (CCR) and utilization of available resources by grouping the user jobs before resource allocation.
This document evaluates the performance of a hybrid differential evolution-genetic algorithm (DE-GA) approach for load balancing in cloud computing. It first provides background on cloud computing and load balancing. It then describes the DE-GA approach, which uses differential evolution initially and switches to genetic algorithm if needed. The results show that the hybrid DE-GA approach improves performance over differential evolution and genetic algorithm alone, reducing makespan, average response time, and improving resource utilization. The study demonstrates the benefits of the hybrid evolutionary algorithm for an important problem in cloud computing.
Earlier stage for straggler detection and handling using combined CPU test an...IJECEIAES
This document summarizes a research paper that proposes a new framework called the combinatory late-machine (CLM) framework to facilitate early detection and handling of straggler tasks in MapReduce jobs. Straggler tasks significantly increase job execution time and energy consumption. The CLM framework combines CPU testing and the Longest Approximate Time to End (LATE) methodology to calculate a straggler tolerance threshold earlier. This allows for prompt mitigation actions. The paper reviews related work on straggler detection techniques and discusses the proposed methodology, which estimates task finish times based on progress scores. It aims to correlate straggler detection with system attributes like resource utilization that could cause delays.
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
RSDC (Reliable Scheduling Distributed in Cloud Computing)IJCSEA Journal
This document summarizes the PPDD algorithm for scheduling divisible loads originating from multiple sites in distributed computing environments. The PPDD algorithm is a two-phase approach that first derives a near-optimal load distribution and then considers actual communication delays when transferring load fractions. It guarantees a near-optimal solution and improved performance over previous algorithms like RSA by avoiding unnecessary load transfers between processors.
A Novel Framework and Policies for On-line Block of Cores Allotment for Multi...ijcsa
Computer industry has widely accepted that future performance increases must largely come from increasing the number of processing cores on a die. This has led to NoC processors. Task scheduling is one of the most challenging problems facing parallel programmers today which is known to be NP-complete. A good principle is space-sharing of cores and to schedule multiple DAGs simultaneously on NoC processor. Hence the need to find optimal number of cores for a DAG for a particular scheduling method and further which region of cores on NoC, to be allotted for a DAG . In this work, a method is proposed to find near-optimal minimal block of cores for a DAG on a NoC processor. Further, a time efficient framework and three on-line block allotment policies to the submitted DAGs are experimented. The objectives of the policies, is to improve the NoC throughput. The policies are experimented on a simulator and found to deliver better performance than the policies found in literature..
An Improved Parallel Activity scheduling algorithm for large datasetsIJERA Editor
Parallel processing is capable of executing a large number of tasks on a multiprocessor at the same time period, and it is also one of the emerging concepts. Complex and computational problems can be resolved in an efficient way with the help of parallel processing. The parallel processing system can be divided into two categories depending on the nature of tasks such are homogenous parallel system and the heterogeneous parallel processing system. In the homogeneous environment, the number of processors required for executing different tasks is similar in capacity. In case of heterogeneous environments, tasks are allocated to various processors with different capacity and speed. The main objective of parallel processing is to optimize the execution speed and to shorten the duration of task execution with independent of environment. In this proposed work, an optimized parallel project selection method was implemented to find the optimal resource utilization and project scheduling. The execution speeds of the task increases and the overall average execution time of the task decreases by allocating different tasks to various processors with the task scheduling algorithm.
A NOVEL METHODOLOGY FOR TASK DISTRIBUTION IN HETEROGENEOUS RECONFIGURABLE COM...ijesajournal
Modern embedded systems are being modeled as Heterogeneous Reconfigurable Computing Systems
(HRCS) where Reconfigurable Hardware i.e. Field Programmable Gate Array (FPGA) and soft core
processors acts as computing elements. So, an efficient task distribution methodology is essential for
obtaining high performance in modern embedded systems. In this paper, we present a novel methodology
for task distribution called Minimum Laxity First (MLF) algorithm that takes the advantage of runtime
reconfiguration of FPGA in order to effectively utilize the available resources. The MLF algorithm is a list
based dynamic scheduling algorithm that uses attributes of tasks as well computing resources as cost
function to distribute the tasks of an application to HRCS. In this paper, an on chip HRCS computing
platform is configured on Virtex 5 FPGA using Xilinx EDK. The real time applications JPEG, OFDM
transmitters are represented as task graph and then the task are distributed, statically as well dynamically,
to the platform HRCS in order to evaluate the performance of the designed task distribution model. Finally,
the performance of MLF algorithm is compared with existing static scheduling algorithms. The comparison
shows that the MLF algorithm outperforms in terms of efficient utilization of resources on chip and also
speedup an application execution.
A NOVEL METHODOLOGY FOR TASK DISTRIBUTION IN HETEROGENEOUS RECONFIGURABLE COM...ijesajournal
Modern embedded systems are being modeled as Heterogeneous Reconfigurable Computing Systems
(HRCS) where Reconfigurable Hardware i.e. Field Programmable Gate Array (FPGA) and soft core
processors acts as computing elements. So, an efficient task distribution methodology is essential for
obtaining high performance in modern embedded systems. In this paper, we present a novel methodology
for task distribution called Minimum Laxity First (MLF) algorithm that takes the advantage of runtime
reconfiguration of FPGA in order to effectively utilize the available resources. The MLF algorithm is a list
based dynamic scheduling algorithm that uses attributes of tasks as well computing resources as cost
function to distribute the tasks of an application to HRCS. In this paper, an on chip HRCS computing
platform is configured on Virtex 5 FPGA using Xilinx EDK. The real time applications JPEG, OFDM
transmitters are represented as task graph and then the task are distributed, statically as well dynamically,
to the platform HRCS in order to evaluate the performance of the designed task distribution model. Finally,
the performance of MLF algorithm is compared with existing static scheduling algorithms. The comparison
shows that the MLF algorithm outperforms in terms of efficient utilization of resources on chip and also
speedup an application execution.
A NOVEL METHODOLOGY FOR TASK DISTRIBUTION IN HETEROGENEOUS RECONFIGURABLE COM...ijesajournal
Modern embedded systems are being modeled as Heterogeneous Reconfigurable Computing Systems
(HRCS) where Reconfigurable Hardware i.e. Field Programmable Gate Array (FPGA) and soft core
processors acts as computing elements. So, an efficient task distribution methodology is essential for
obtaining high performance in modern embedded systems. In this paper, we present a novel methodology
for task distribution called Minimum Laxity First (MLF) algorithm that takes the advantage of runtime
reconfiguration of FPGA in order to effectively utilize the available resources. The MLF algorithm is a list
based dynamic scheduling algorithm that uses attributes of tasks as well computing resources as cost
function to distribute the tasks of an application to HRCS. In this paper, an on chip HRCS computing
platform is configured on Virtex 5 FPGA using Xilinx EDK. The real time applications JPEG, OFDM
transmitters are represented as task graph and then the task are distributed, statically as well dynamically,
to the platform HRCS in order to evaluate the performance of the designed task distribution model. Finally,
the performance of MLF algorithm is compared with existing static scheduling algorithms. The comparison
shows that the MLF algorithm outperforms in terms of efficient utilization of resources on chip and also
speedup an application execution.
GROUPING BASED JOB SCHEDULING ALGORITHM USING PRIORITY QUEUE AND HYBRID ALGOR...ijgca
This document describes a proposed grouping based job scheduling algorithm for grid computing that aims to maximize resource utilization and minimize job processing times. It discusses related work on job scheduling algorithms and then presents the steps of the proposed algorithm. The algorithm uses shortest job first, first-in first-out, and round robin scheduling to process jobs in groups. The algorithm is evaluated experimentally in MATLAB and shown to reduce total job processing time compared to using only first-in first-out scheduling. Graphs demonstrate the processing time improvements achieved by the combined scheduling approach.
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server Environmentrahulmonikasharma
This document summarizes a research paper that proposes a new algorithm called KD-Tree approach for efficient virtual machine (VM) allocation in cloud computing environments. The algorithm aims to minimize the response time for allocating VMs to user requests. It does this by adopting a KD-Tree data structure to index physical host machines, allowing the scheduler to quickly find the host that can accommodate a new VM request with the minimum latency in O(Log n) time. The proposed approach is evaluated through simulations using the CloudSim toolkit and is shown to outperform an existing linear scheduling strategy (LSTR) algorithm in terms of reducing VM allocation times.
The Cloud computing becomes an important topic
in the area of high performance distributed computing. On the
other hand, task scheduling is considered one the most significant
issues in the Cloud computing where the user has to pay for the
using resource based on the time. Therefore, distributing the
cloud resource among the users' applications should maximize
resource utilization and minimize task execution Time. The goal
of task scheduling is to assign tasks to appropriate resources that
optimize one or more performance parameters (i.e., completion
time, cost, resource utilization, etc.). In addition, the scheduling
belongs to a category of a problem known as an NP-complete
problem. Therefore, the heuristic algorithm could be applied to
solve this problem. In this paper, an enhanced dependent task
scheduling algorithm based on Genetic Algorithm (DTGA) has
been introduced for mapping and executing an application’s
tasks. The aim of this proposed algorithm is to minimize the
completion time. The performance of this proposed algorithm has
been evaluated using WorkflowSim toolkit and Standard Task
Graph Set (STG) benchmark.
The task scheduling is a key process in large-scale distributed systems like cloud computing infrastructures
which can have much impressed on system performance. This problem is referred to as a NP-hard problem
because of some reasons such as heterogeneous and dynamic features and dependencies among the
requests. Here, we proposed a bi-objective method called DWSGA to obtain a proper solution for
allocating the requests on resources. The purpose of this algorithm is to earn the response quickly, with
some goal-oriented operations. At first, it makes a good initial population by a special way that uses a bidirectional
tasks prioritization. Then the algorithm moves to get the most appropriate possible solution in a
conscious manner by focus on optimizing the makespan, and considering a good distribution of workload
on resources by using efficient parameters in the mentioned systems. Here, the experiments indicate that
the DWSGA amends the results when the numbers of tasks are increased in application graph, in order to
mentioned objectives. The results are compared with other studied algorithms.
Service Request Scheduling in Cloud Computing using Meta-Heuristic Technique:...IRJET Journal
This document discusses using the Teaching Learning Based Optimization (TLBO) meta-heuristic technique for service request scheduling between users and cloud service providers. TLBO is a nature-inspired algorithm that mimics the teacher-student learning process. It is compared to other meta-heuristic algorithms like Genetic Algorithm. The key steps of TLBO involve initializing a population, evaluating fitness, selecting the best solution as teacher, and updating the population through teacher and learner phases until termination criteria is met. The document proposes using number of users and virtual machines as parameters for TLBO scheduling in cloud computing. MATLAB simulation results show the initial and final iterations converging to an optimal scheduling solution.
Scheduling Divisible Jobs to Optimize the Computation and Energy Costsinventionjournals
ABSTRACT : The important challenge in cloud computing environment is to design a scheduling strategy to handle jobs, and to process them in a heterogeneous environment with shared data centers. In this paper, we attempt to investigate a new analytical framework model that enables an existing private cloud data-center for scheduling jobs and minimizing the overall computation and energy cost together. Our model is based on Divisible Load Theory (DLT) model to derive closed-form solution for the load fractions to be assigned to each machines considering computation and energy cost. Our analysis also attempts to schedule the jobs such a way that cloud provider can gain maximum benefit for his service and Quality of Service (QoS) requirement user’s job. Finally, we quantify the performance of the strategies via rigorous simulation studies.
Challenges in Dynamic Resource Allocation and Task Scheduling in Heterogeneou...rahulmonikasharma
This document discusses the challenges of dynamic resource allocation and task scheduling in heterogeneous cloud environments. It outlines that resource allocation involves deciding how to allocate resources to tasks to maximize utilization, while task scheduling assigns tasks to processors to minimize execution time. The major challenges are optimizing allocated resources to minimize costs while meeting customer demands and application requirements. Allocating resources dynamically in heterogeneous cloud environments is difficult due to issues like resource contention, scarcity, and fragmentation. The document also discusses approaches to resource modeling, allocation, offering, discovery and monitoring that algorithms must address to effectively allocate resources on demand.
The document discusses optimization of resource allocation in cloud environments using a modified particle swarm optimization (PSO) approach. It proposes a Modified Resource Allocation Mutation PSO (MRAMPSO) strategy that uses an Extended Multi Queue Scheduling algorithm to schedule tasks based on resource availability and reschedules failed tasks. The MRAMPSO strategy is compared to standard PSO and other algorithms to show it can reduce execution time, makespan, transmission cost, and round trip time.
A location based least-cost scheduling for data-intensive applicationsIAEME Publication
This document summarizes a research paper that proposes a location-based least-cost scheduling algorithm for transferring multiple data-intensive files simultaneously to multiple compute nodes in a grid environment. The proposed model includes an optimized meta-scheduler that receives multiple files, predicts the optimal number of parallel TCP streams to use for each file transfer based on sampling, and schedules the files to compute nodes using a greedy algorithm that considers location and cost. Experimental results showed the optimized model achieved better transfer times and throughput compared to non-optimized transfers.
The document discusses using a genetic algorithm to schedule tasks in a cloud computing environment. It aims to minimize task execution time and reduce computational costs compared to the traditional Round Robin scheduling algorithm. The proposed genetic algorithm mimics natural selection and genetics to evolve optimal task schedules. It was tested using the CloudSim simulation toolkit and results showed the genetic algorithm provided better performance than Round Robin scheduling.
A Framework and Methods for Dynamic Scheduling of a Directed Acyclic Graph on...IDES Editor
The data flow model is gaining popularity as a
programming paradigm for multi-core processors. Efficient
scheduling of an application modeled by Directed Acyclic
Graph (DAG) is a key issue when performance is very
important. DAG represents computational solutions, in which
the nodes represent tasks to be executed and edges represent
precedence constraints among the tasks. The task scheduling
problem in general is a NP-complete problem[2]. Several static
scheduling heuristics have been proposed. But the major
problem in static list scheduling is the inherent difficulty in
exact estimation of task cost and edge cost in a DAG and also
its inability to consider and manage with runtime behavior of
tasks. This underlines the need for dynamic scheduling of a
DAG. This paper presents how in general, dynamic scheduling
of a DAG can be done. Also proposes 4 simple methods to
perform dynamic scheduling of a DAG. These methods have
been simulated and experimented using a representative set
of DAG structured computations from both synthetic and real
problems. The proposed dynamic scheduler performance is
found to be in comparable with that of static scheduling
methods. The performance comparison of the proposed
dynamic scheduling methods is also carried out.
Optimized Assignment of Independent Task for Improving Resources Performance ...ijgca
Grid computing has emerged from category of distributed and parallel computing where the
heterogeneous resources from different network are used simultaneously to solve a particular problem that
need huge amount of resources. Potential of Grid computing depends on my issues such as security of
resources, heterogeneity of resources, fault tolerance & resource discovery and job scheduling. Scheduling
is one of the core steps to efficiently exploit the capabilities of heterogeneous distributed computing
resources and is an NP-complete problem. To achieve the promising potential of grid computing, an
effective and efficient job scheduling algorithm is proposed, which will optimized two important criteria to
improve the performance of resources i.e. makespan time & resource utilization. With this, we have
classified various tasks scheduling heuristic in grid on the basis of their characteristics.
Effective and Efficient Job Scheduling in Grid ComputingAditya Kokadwar
The integration of remote and diverse resources and the increasing computational needs of Grand Challenges problems combined with the faster growth of the internet and communication technologies leads to the development of global computational grids. Grid computing is a prevailing technology, which unites underutilized resources in order to support sharing of resources and services distributed across numerous administrative region. An efficient and effective scheduling system is essentially required in order to achieve the promising capacity of grids. The main goal of scheduling is to maximize the resource utilization and minimize processing time and cost of the jobs. In this research, the objective is to prioritize the jobs based on execution cost and then allocate the resources with minimum cost by merging it with conventional job grouping strategy to provide the solution for better and more efficient job scheduling which is beneficial to both user and resource broker. The proposed scheduling approach in grid computing employs a dynamic cost-based job scheduling algorithm for making an efficient mapping of a job to available resources in the grid. It also improves communication to computation ratio (CCR) and utilization of available resources by grouping the user jobs before resource allocation.
This document evaluates the performance of a hybrid differential evolution-genetic algorithm (DE-GA) approach for load balancing in cloud computing. It first provides background on cloud computing and load balancing. It then describes the DE-GA approach, which uses differential evolution initially and switches to genetic algorithm if needed. The results show that the hybrid DE-GA approach improves performance over differential evolution and genetic algorithm alone, reducing makespan, average response time, and improving resource utilization. The study demonstrates the benefits of the hybrid evolutionary algorithm for an important problem in cloud computing.
Earlier stage for straggler detection and handling using combined CPU test an...IJECEIAES
This document summarizes a research paper that proposes a new framework called the combinatory late-machine (CLM) framework to facilitate early detection and handling of straggler tasks in MapReduce jobs. Straggler tasks significantly increase job execution time and energy consumption. The CLM framework combines CPU testing and the Longest Approximate Time to End (LATE) methodology to calculate a straggler tolerance threshold earlier. This allows for prompt mitigation actions. The paper reviews related work on straggler detection techniques and discusses the proposed methodology, which estimates task finish times based on progress scores. It aims to correlate straggler detection with system attributes like resource utilization that could cause delays.
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
RSDC (Reliable Scheduling Distributed in Cloud Computing)IJCSEA Journal
This document summarizes the PPDD algorithm for scheduling divisible loads originating from multiple sites in distributed computing environments. The PPDD algorithm is a two-phase approach that first derives a near-optimal load distribution and then considers actual communication delays when transferring load fractions. It guarantees a near-optimal solution and improved performance over previous algorithms like RSA by avoiding unnecessary load transfers between processors.
A Novel Framework and Policies for On-line Block of Cores Allotment for Multi...ijcsa
Computer industry has widely accepted that future performance increases must largely come from increasing the number of processing cores on a die. This has led to NoC processors. Task scheduling is one of the most challenging problems facing parallel programmers today which is known to be NP-complete. A good principle is space-sharing of cores and to schedule multiple DAGs simultaneously on NoC processor. Hence the need to find optimal number of cores for a DAG for a particular scheduling method and further which region of cores on NoC, to be allotted for a DAG . In this work, a method is proposed to find near-optimal minimal block of cores for a DAG on a NoC processor. Further, a time efficient framework and three on-line block allotment policies to the submitted DAGs are experimented. The objectives of the policies, is to improve the NoC throughput. The policies are experimented on a simulator and found to deliver better performance than the policies found in literature..
An Improved Parallel Activity scheduling algorithm for large datasetsIJERA Editor
Parallel processing is capable of executing a large number of tasks on a multiprocessor at the same time period, and it is also one of the emerging concepts. Complex and computational problems can be resolved in an efficient way with the help of parallel processing. The parallel processing system can be divided into two categories depending on the nature of tasks such are homogenous parallel system and the heterogeneous parallel processing system. In the homogeneous environment, the number of processors required for executing different tasks is similar in capacity. In case of heterogeneous environments, tasks are allocated to various processors with different capacity and speed. The main objective of parallel processing is to optimize the execution speed and to shorten the duration of task execution with independent of environment. In this proposed work, an optimized parallel project selection method was implemented to find the optimal resource utilization and project scheduling. The execution speeds of the task increases and the overall average execution time of the task decreases by allocating different tasks to various processors with the task scheduling algorithm.
A NOVEL METHODOLOGY FOR TASK DISTRIBUTION IN HETEROGENEOUS RECONFIGURABLE COM...ijesajournal
Modern embedded systems are being modeled as Heterogeneous Reconfigurable Computing Systems
(HRCS) where Reconfigurable Hardware i.e. Field Programmable Gate Array (FPGA) and soft core
processors acts as computing elements. So, an efficient task distribution methodology is essential for
obtaining high performance in modern embedded systems. In this paper, we present a novel methodology
for task distribution called Minimum Laxity First (MLF) algorithm that takes the advantage of runtime
reconfiguration of FPGA in order to effectively utilize the available resources. The MLF algorithm is a list
based dynamic scheduling algorithm that uses attributes of tasks as well computing resources as cost
function to distribute the tasks of an application to HRCS. In this paper, an on chip HRCS computing
platform is configured on Virtex 5 FPGA using Xilinx EDK. The real time applications JPEG, OFDM
transmitters are represented as task graph and then the task are distributed, statically as well dynamically,
to the platform HRCS in order to evaluate the performance of the designed task distribution model. Finally,
the performance of MLF algorithm is compared with existing static scheduling algorithms. The comparison
shows that the MLF algorithm outperforms in terms of efficient utilization of resources on chip and also
speedup an application execution.
A NOVEL METHODOLOGY FOR TASK DISTRIBUTION IN HETEROGENEOUS RECONFIGURABLE COM...ijesajournal
Modern embedded systems are being modeled as Heterogeneous Reconfigurable Computing Systems
(HRCS) where Reconfigurable Hardware i.e. Field Programmable Gate Array (FPGA) and soft core
processors acts as computing elements. So, an efficient task distribution methodology is essential for
obtaining high performance in modern embedded systems. In this paper, we present a novel methodology
for task distribution called Minimum Laxity First (MLF) algorithm that takes the advantage of runtime
reconfiguration of FPGA in order to effectively utilize the available resources. The MLF algorithm is a list
based dynamic scheduling algorithm that uses attributes of tasks as well computing resources as cost
function to distribute the tasks of an application to HRCS. In this paper, an on chip HRCS computing
platform is configured on Virtex 5 FPGA using Xilinx EDK. The real time applications JPEG, OFDM
transmitters are represented as task graph and then the task are distributed, statically as well dynamically,
to the platform HRCS in order to evaluate the performance of the designed task distribution model. Finally,
the performance of MLF algorithm is compared with existing static scheduling algorithms. The comparison
shows that the MLF algorithm outperforms in terms of efficient utilization of resources on chip and also
speedup an application execution.
A NOVEL METHODOLOGY FOR TASK DISTRIBUTION IN HETEROGENEOUS RECONFIGURABLE COM...ijesajournal
Modern embedded systems are being modeled as Heterogeneous Reconfigurable Computing Systems
(HRCS) where Reconfigurable Hardware i.e. Field Programmable Gate Array (FPGA) and soft core
processors acts as computing elements. So, an efficient task distribution methodology is essential for
obtaining high performance in modern embedded systems. In this paper, we present a novel methodology
for task distribution called Minimum Laxity First (MLF) algorithm that takes the advantage of runtime
reconfiguration of FPGA in order to effectively utilize the available resources. The MLF algorithm is a list
based dynamic scheduling algorithm that uses attributes of tasks as well computing resources as cost
function to distribute the tasks of an application to HRCS. In this paper, an on chip HRCS computing
platform is configured on Virtex 5 FPGA using Xilinx EDK. The real time applications JPEG, OFDM
transmitters are represented as task graph and then the task are distributed, statically as well dynamically,
to the platform HRCS in order to evaluate the performance of the designed task distribution model. Finally,
the performance of MLF algorithm is compared with existing static scheduling algorithms. The comparison
shows that the MLF algorithm outperforms in terms of efficient utilization of resources on chip and also
speedup an application execution.
A novel methodology for task distributionijesajournal
Modern embedded systems are being modeled as Heterogeneous Reconfigurable Computing Systems
(HRCS) where Reconfigurable Hardware i.e. Field Programmable Gate Array (FPGA) and soft core
processors acts as computing elements. So, an efficient task distribution methodology is essential for
obtaining high performance in modern embedded systems. In this paper, we present a novel methodology
for task distribution called Minimum Laxity First (MLF) algorithm that takes the advantage of runtime
reconfiguration of FPGA in order to effectively utilize the available resources. The MLF algorithm is a list
based dynamic scheduling algorithm that uses attributes of tasks as well computing resources as cost
function to distribute the tasks of an application to HRCS. In this paper, an on chip HRCS computing
platform is configured on Virtex 5 FPGA using Xilinx EDK. The real time applications JPEG, OFDM
transmitters are represented as task graph and then the task are distributed, statically as well dynamically,
to the platform HRCS in order to evaluate the performance of the designed task distribution model. Finally,
the performance of MLF algorithm is compared with existing static scheduling algorithms. The comparison
shows that the MLF algorithm outperforms in terms of efficient utilization of resources on chip and also
speedup an application execution.
DGBSA : A BATCH JOB SCHEDULINGALGORITHM WITH GA WITH REGARD TO THE THRESHOLD ...IJCSEA Journal
In this paper , we will provide a scheduler on batch jobs with GA regard to the threshold detector. In The algorithm proposed in this paper, we will provide the batch independent jobs with a new technique ,so we can optimize the schedule them. To do this, we use a threshold detector then among the selected jobs, processing resources can process batch jobs with priority. Also hierarchy of tasks in each batch, will be determined with using DGBSA algorithm. Now , with the regard to the works done by previous ,we can provide an algorithm that by adding specific parameters to fitness function of the previous algorithms ,develop a optimum fitness function that in the proposed algorithm has been used. According to assessment done on DGBSA algorithm, in compare with the similar algorithms, it has more performance. The effective parameters that used in the proposed algorithm can reduce the total wasting time in compare with previous algorithms. Also this algorithm can improve the previous problems in batch processing with a new technique.
This document discusses scheduling in cloud computing. It proposes a priority-based scheduling protocol to improve resource utilization, server performance, and minimize makespan. The protocol assigns priorities to jobs, allocates jobs to processors based on completion time, and processes jobs in parallel queues to efficiently schedule jobs in cloud computing. Future work includes analyzing time complexity and completion times through simulation to validate the protocol's efficiency.
A popular programming model for running data intensive applications on the cloud is map reduce. In
the Hadoop usually, jobs are scheduled in FIFO order by default. There are many map reduce
applications which require strict deadline. In Hadoop framework, scheduler wi t h deadline
con s t ra in t s has not been implemented. Existing schedulers d o not guarantee that the job will be
completed by a specific deadline. Some schedulers address the issue of deadlines but focus more on
improving s y s t em utilization. We have proposed an algorithm which facilitates the user to
specify a jobs deadline and evaluates whether the job can be finished before the deadline.
Scheduler with deadlines for Hadoop, which ensures that only jobs, whose deadlines can be met are
scheduled for execution. If the job submitted does not satisfy the specified deadline, physical or
virtual nodes can be added dynamically to complete the job within deadline[8].
Qo s aware scientific application scheduling algorithm in cloud environmentAlexander Decker
The document describes a QoS-aware scientific application scheduling algorithm for cloud environments. It proposes an algorithm that ranks tasks in a workflow and uses a user preference fitness function to select resources based on the user's desired quality of service, such as time and cost. The algorithm is compared to other similar works through several scenarios, and results show the proposed algorithm has better efficiency. Key aspects considered include task dependencies, data sizes, compute times, data transfer times, workflow makespan, resource costs and attributes.
Qo s aware scientific application scheduling algorithm in cloud environmentAlexander Decker
This document summarizes a research paper that proposes a scheduling algorithm for scientific applications in cloud environments. The algorithm aims to schedule tasks in workflows based on user preferences for quality of service (QoS), like time and cost. It ranks tasks and uses an UPFF function to select resources that meet the user's desired QoS. The algorithm is compared to other similar algorithms through scenarios, and results show it has better efficiency. The full paper provides more details on scientific workflows, cloud computing, related work on workflow scheduling algorithms, and defines the problem of scheduling tasks to resources while considering costs and times.
GROUPING BASED JOB SCHEDULING ALGORITHM USING PRIORITY QUEUE AND HYBRID ALGOR...ijgca
Grid computing enlarge with computing platform which is collection of heterogeneous computing resources connected by a network across dynamic and geographically dispersed organization to form a distributed high performance computing infrastructure. Grid computing solves the complex computing
problems amongst multiple machines. Grid computing solves the large scale computational demands in a high performance computing environment. The main emphasis in the grid computing is given to the resource management and the job scheduler .The goal of the job scheduler is to maximize the resource utilization and minimize the processing time of the jobs. Existing approaches of Grid scheduling doesn’t give much emphasis on the performance of a Grid scheduler in processing time parameter. Schedulers allocate resources to the jobs to be executed using the First come First serve algorithm. In this paper, we have provided an optimize algorithm to queue of the scheduler using various scheduling methods like Shortest Job First, First in First out, Round robin. The job scheduling system is responsible to select best suitable machines in a grid for user jobs. The management and scheduling system generates job schedules for each machine in the grid by taking static restrictions and dynamic parameters of jobs and machines
into consideration. The main purpose of this paper is to develop an efficient job scheduling algorithm to maximize the resource utilization and minimize processing time of the jobs. Queues can be optimized by using various scheduling algorithms depending upon the performance criteria to be improved e.g. response
time, throughput. The work has been done in MATLAB using the parallel computing toolbox.
LOAD BALANCING LARGE DATA SETS IN A HADOOP CLUSTERijdpsjournal
With the interconnection of one to many computers online, the data shared by users is multiplying daily. As
a result, the amount of data to be processed by dedicated servers rises very quickly. However, the
instantaneous increase in the volume of data to be processed by the server comes up against latency during
processing. This requires a model to manage the distribution of tasks across several machines. This article
presents a study of load balancing for large data sets on a cluster of Hadoop nodes. In this paper, we use
Mapreduce to implement parallel programming and Yarn to monitor task execution and submission in a
node cluster.
Grid computing can involve lot of computational tasks which requires trustworthy computational nodes. Load balancing in grid computing is a technique which overall optimizes the whole process of assigning computational tasks to processing nodes. Grid computing is a form of distributed computing but different from conventional distributed computing in a manner that it tends to be heterogeneous, more loosely coupled and dispersed geographically. Optimization of this process must contains the overall maximization of resources utilization with balance load on each processing unit and also by decreasing the overall time or output. Evolutionary algorithms like genetic algorithms have studied so far for the implementation of load balancing across the grid networks. But problem with these genetic algorithm is that they are quite slow in cases where large number of tasks needs to be processed. In this paper we give a novel approach of parallel genetic algorithms for enhancing the overall performance and optimization of managing the whole process of load balancing across the grid nodes.
Reconfiguration Strategies for Online Hardware Multitasking in Embedded SystemsCSEIJJournal
An intensive use of reconfigurable hardware is expected in future embedded systems. This means that the
system has to decide which tasks are more suitable for hardware execution. In order to make an efficient
use of the FPGA it is convenient to choose one that allows hardware multitasking, which is implemented by
using partial dynamic reconfiguration. One of the challenges for hardware multitasking in embedded
systems is the online management of the only reconfiguration port of present FPGA devices. This paper
presents different online reconfiguration scheduling strategies which assign the reconfiguration interface
resource using different criteria: workload distribution or task’ deadline. The online scheduling strategies
presented take efficient and fast decisions based on the information available at each moment. Experiments
have been made in order to analyze the performance and convenience of these reconfiguration strategies.
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
Task Scheduling using Hybrid Algorithm in Cloud Computing Environmentsiosrjce
The document summarizes a proposed hybrid task scheduling algorithm called PSOCS that combines particle swarm optimization (PSO) and cuckoo search (CS) for scheduling tasks in cloud computing environments. The PSOCS algorithm aims to minimize task completion time (makespan) and improve resource utilization. It was tested in a simulation using CloudSim and showed reductions in makespan and increases in utilization compared to PSO and random scheduling algorithms.
This document summarizes a research paper that proposes a hybrid task scheduling algorithm for cloud computing environments called PSOCS. PSOCS combines the Particle Swarm Optimization (PSO) algorithm and Cuckoo Search (CS) algorithm to optimize task scheduling and minimize completion time while increasing resource utilization. The paper describes PSO and CS algorithms individually, then defines the proposed PSOCS algorithm. It evaluates PSOCS using a simulation and finds it reduces makespan and increases utilization compared to PSO and random allocation algorithms.
OPTIMIZED RESOURCE PROVISIONING METHOD FOR COMPUTATIONAL GRID ijgca
Grid computing is an accumulation of heterogeneous, dynamic resources from multiple administrative areas which are geographically distributed that can be utilized to reach a mutual end. Development of resource provisioning-based scheduling in large-scale distributed environments like grid computing brings in new requirement challenges that are not being believed in traditional distributed computing environments. Computational grid is applying the resources of many systems in a network to a single problem at the same time. Grid scheduling is the method by which work specified by some means is assigned to the resources that complete the work in the environment which cannot fulfill the user requirements considerably. The satisfaction of users while providing the resources might increase the beneficiary level of resource suppliers. Resource scheduling has to satisfy the multiple constraints specified by the user. The option of resource with the satisfaction of multiple constraints is the most tedious process. This trouble is solved by bringing out the particle swarm optimization based heuristic scheduling algorithm which attempts to select the most suitable resource from the set of available resources. The primary parameters that are taken in this work for selecting the most suitable resource are the makespan and cost. The experimental result shows that the proposed method yields optimal scheduling with the atonement of all user requirements.
Similar to MULTIPLE DAG APPLICATIONS SCHEDULING ON A CLUSTER OF PROCESSORS (20)
ANALYSIS OF LAND SURFACE DEFORMATION GRADIENT BY DINSAR cscpconf
The progressive development of Synthetic Aperture Radar (SAR) systems diversify the exploitation of the generated images by these systems in different applications of geoscience. Detection and monitoring surface deformations, procreated by various phenomena had benefited from this evolution and had been realized by interferometry (InSAR) and differential interferometry (DInSAR) techniques. Nevertheless, spatial and temporal decorrelations of the interferometric couples used, limit strongly the precision of analysis results by these techniques. In this context, we propose, in this work, a methodological approach of surface deformation detection and analysis by differential interferograms to show the limits of this technique according to noise quality and level. The detectability model is generated from the deformation signatures, by simulating a linear fault merged to the images couples of ERS1 / ERS2 sensors acquired in a region of the Algerian south.
4D AUTOMATIC LIP-READING FOR SPEAKER'S FACE IDENTIFCATIONcscpconf
A novel based a trajectory-guided, concatenating approach for synthesizing high-quality image real sample renders video is proposed . The lips reading automated is seeking for modeled the closest real image sample sequence preserve in the library under the data video to the HMM predicted trajectory. The object trajectory is modeled obtained by projecting the face patterns into an KDA feature space is estimated. The approach for speaker's face identification by using synthesise the identity surface of a subject face from a small sample of patterns which sparsely each the view sphere. An KDA algorithm use to the Lip-reading image is discrimination, after that work consisted of in the low dimensional for the fundamental lip features vector is reduced by using the 2D-DCT.The mouth of the set area dimensionality is ordered by a normally reduction base on the PCA to obtain the Eigen lips approach, their proposed approach by[33]. The subjective performance results of the cost function under the automatic lips reading modeled , which wasn’t illustrate the superior performance of the
method.
MOVING FROM WATERFALL TO AGILE PROCESS IN SOFTWARE ENGINEERING CAPSTONE PROJE...cscpconf
Universities offer software engineering capstone course to simulate a real world-working environment in which students can work in a team for a fixed period to deliver a quality product. The objective of the paper is to report on our experience in moving from Waterfall process to Agile process in conducting the software engineering capstone project. We present the capstone course designs for both Waterfall driven and Agile driven methodologies that highlight the structure, deliverables and assessment plans.To evaluate the improvement, we conducted a survey for two different sections taught by two different instructors to evaluate students’ experience in moving from traditional Waterfall model to Agile like process. Twentyeight students filled the survey. The survey consisted of eight multiple-choice questions and an open-ended question to collect feedback from students. The survey results show that students were able to attain hands one experience, which simulate a real world-working environment. The results also show that the Agile approach helped students to have overall better design and avoid mistakes they have made in the initial design completed in of the first phase of the capstone project. In addition, they were able to decide on their team capabilities, training needs and thus learn the required technologies earlier which is reflected on the final product quality
PROMOTING STUDENT ENGAGEMENT USING SOCIAL MEDIA TECHNOLOGIEScscpconf
This document discusses using social media technologies to promote student engagement in a software project management course. It describes the course and objectives of enhancing communication. It discusses using Facebook for 4 years, then switching to WhatsApp based on student feedback, and finally introducing Slack to enable personalized team communication. Surveys found students engaged and satisfied with all three tools, though less familiar with Slack. The conclusion is that social media promotes engagement but familiarity with the tool also impacts satisfaction.
A SURVEY ON QUESTION ANSWERING SYSTEMS: THE ADVANCES OF FUZZY LOGICcscpconf
In real world computing environment with using a computer to answer questions has been a human dream since the beginning of the digital era, Question-answering systems are referred to as intelligent systems, that can be used to provide responses for the questions being asked by the user based on certain facts or rules stored in the knowledge base it can generate answers of questions asked in natural , and the first main idea of fuzzy logic was to working on the problem of computer understanding of natural language, so this survey paper provides an overview on what Question-Answering is and its system architecture and the possible relationship and
different with fuzzy logic, as well as the previous related research with respect to approaches that were followed. At the end, the survey provides an analytical discussion of the proposed QA models, along or combined with fuzzy logic and their main contributions and limitations.
DYNAMIC PHONE WARPING – A METHOD TO MEASURE THE DISTANCE BETWEEN PRONUNCIATIONS cscpconf
Human beings generate different speech waveforms while speaking the same word at different times. Also, different human beings have different accents and generate significantly varying speech waveforms for the same word. There is a need to measure the distances between various words which facilitate preparation of pronunciation dictionaries. A new algorithm called Dynamic Phone Warping (DPW) is presented in this paper. It uses dynamic programming technique for global alignment and shortest distance measurements. The DPW algorithm can be used to enhance the pronunciation dictionaries of the well-known languages like English or to build pronunciation dictionaries to the less known sparse languages. The precision measurement experiments show 88.9% accuracy.
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS cscpconf
In education, the use of electronic (E) examination systems is not a novel idea, as Eexamination systems have been used to conduct objective assessments for the last few years. This research deals with randomly designed E-examinations and proposes an E-assessment system that can be used for subjective questions. This system assesses answers to subjective questions by finding a matching ratio for the keywords in instructor and student answers. The matching ratio is achieved based on semantic and document similarity. The assessment system is composed of four modules: preprocessing, keyword expansion, matching, and grading. A survey and case study were used in the research design to validate the proposed system. The examination assessment system will help instructors to save time, costs, and resources, while increasing efficiency and improving the productivity of exam setting and assessments.
TWO DISCRETE BINARY VERSIONS OF AFRICAN BUFFALO OPTIMIZATION METAHEURISTICcscpconf
African Buffalo Optimization (ABO) is one of the most recent swarms intelligence based metaheuristics. ABO algorithm is inspired by the buffalo’s behavior and lifestyle. Unfortunately, the standard ABO algorithm is proposed only for continuous optimization problems. In this paper, the authors propose two discrete binary ABO algorithms to deal with binary optimization problems. In the first version (called SBABO) they use the sigmoid function and probability model to generate binary solutions. In the second version (called LBABO) they use some logical operator to operate the binary solutions. Computational results on two knapsack problems (KP and MKP) instances show the effectiveness of the proposed algorithm and their ability to achieve good and promising solutions.
DETECTION OF ALGORITHMICALLY GENERATED MALICIOUS DOMAINcscpconf
In recent years, many malware writers have relied on Dynamic Domain Name Services (DDNS) to maintain their Command and Control (C&C) network infrastructure to ensure a persistence presence on a compromised host. Amongst the various DDNS techniques, Domain Generation Algorithm (DGA) is often perceived as the most difficult to detect using traditional methods. This paper presents an approach for detecting DGA using frequency analysis of the character distribution and the weighted scores of the domain names. The approach’s feasibility is demonstrated using a range of legitimate domains and a number of malicious algorithmicallygenerated domain names. Findings from this study show that domain names made up of English characters “a-z” achieving a weighted score of < 45 are often associated with DGA. When a weighted score of < 45 is applied to the Alexa one million list of domain names, only 15% of the domain names were treated as non-human generated.
GLOBAL MUSIC ASSET ASSURANCE DIGITAL CURRENCY: A DRM SOLUTION FOR STREAMING C...cscpconf
The document proposes a blockchain-based digital currency and streaming platform called GoMAA to address issues of piracy in the online music streaming industry. Key points:
- GoMAA would use a digital token on the iMediaStreams blockchain to enable secure dissemination and tracking of streamed content. Content owners could control access and track consumption of released content.
- Original media files would be converted to a Secure Portable Streaming (SPS) format, embedding watermarks and smart contract data to indicate ownership and enable validation on the blockchain.
- A browser plugin would provide wallets for fans to collect GoMAA tokens as rewards for consuming content, incentivizing participation and addressing royalty discrepancies by recording
IMPORTANCE OF VERB SUFFIX MAPPING IN DISCOURSE TRANSLATION SYSTEMcscpconf
This document discusses the importance of verb suffix mapping in discourse translation from English to Telugu. It explains that after anaphora resolution, the verbs must be changed to agree with the gender, number, and person features of the subject or anaphoric pronoun. Verbs in Telugu inflect based on these features, while verbs in English only inflect based on number and person. Several examples are provided that demonstrate how the Telugu verb changes based on whether the subject or pronoun is masculine, feminine, neuter, singular or plural. Proper verb suffix mapping is essential for generating natural and coherent translations while preserving the context and meaning of the original discourse.
EXACT SOLUTIONS OF A FAMILY OF HIGHER-DIMENSIONAL SPACE-TIME FRACTIONAL KDV-T...cscpconf
In this paper, based on the definition of conformable fractional derivative, the functional
variable method (FVM) is proposed to seek the exact traveling wave solutions of two higherdimensional
space-time fractional KdV-type equations in mathematical physics, namely the
(3+1)-dimensional space–time fractional Zakharov-Kuznetsov (ZK) equation and the (2+1)-
dimensional space–time fractional Generalized Zakharov-Kuznetsov-Benjamin-Bona-Mahony
(GZK-BBM) equation. Some new solutions are procured and depicted. These solutions, which
contain kink-shaped, singular kink, bell-shaped soliton, singular soliton and periodic wave
solutions, have many potential applications in mathematical physics and engineering. The
simplicity and reliability of the proposed method is verified.
AUTOMATED PENETRATION TESTING: AN OVERVIEWcscpconf
The document discusses automated penetration testing and provides an overview. It compares manual and automated penetration testing, noting that automated testing allows for faster, more standardized and repeatable tests but has limitations in developing new exploits. It also reviews some current automated penetration testing methodologies and tools, including those using HTTP/TCP/IP attacks, linking common scanning tools, a Python-based tool targeting databases, and one using POMDPs for multi-step penetration test planning under uncertainty. The document concludes that automated testing is more efficient than manual for known vulnerabilities but cannot replace manual testing for discovering new exploits.
CLASSIFICATION OF ALZHEIMER USING fMRI DATA AND BRAIN NETWORKcscpconf
Since the mid of 1990s, functional connectivity study using fMRI (fcMRI) has drawn increasing
attention of neuroscientists and computer scientists, since it opens a new window to explore
functional network of human brain with relatively high resolution. BOLD technique provides
almost accurate state of brain. Past researches prove that neuro diseases damage the brain
network interaction, protein- protein interaction and gene-gene interaction. A number of
neurological research paper also analyse the relationship among damaged part. By
computational method especially machine learning technique we can show such classifications.
In this paper we used OASIS fMRI dataset affected with Alzheimer’s disease and normal
patient’s dataset. After proper processing the fMRI data we use the processed data to form
classifier models using SVM (Support Vector Machine), KNN (K- nearest neighbour) & Naïve
Bayes. We also compare the accuracy of our proposed method with existing methods. In future,
we will other combinations of methods for better accuracy.
VALIDATION METHOD OF FUZZY ASSOCIATION RULES BASED ON FUZZY FORMAL CONCEPT AN...cscpconf
The document proposes a new validation method for fuzzy association rules based on three steps: (1) applying the EFAR-PN algorithm to extract a generic base of non-redundant fuzzy association rules using fuzzy formal concept analysis, (2) categorizing the extracted rules into groups, and (3) evaluating the relevance of the rules using structural equation modeling, specifically partial least squares. The method aims to address issues with existing fuzzy association rule extraction algorithms such as large numbers of extracted rules, redundancy, and difficulties with manual validation.
PROBABILITY BASED CLUSTER EXPANSION OVERSAMPLING TECHNIQUE FOR IMBALANCED DATAcscpconf
In many applications of data mining, class imbalance is noticed when examples in one class are
overrepresented. Traditional classifiers result in poor accuracy of the minority class due to the
class imbalance. Further, the presence of within class imbalance where classes are composed of
multiple sub-concepts with different number of examples also affect the performance of
classifier. In this paper, we propose an oversampling technique that handles between class and
within class imbalance simultaneously and also takes into consideration the generalization
ability in data space. The proposed method is based on two steps- performing Model Based
Clustering with respect to classes to identify the sub-concepts; and then computing the
separating hyperplane based on equal posterior probability between the classes. The proposed
method is tested on 10 publicly available data sets and the result shows that the proposed
method is statistically superior to other existing oversampling methods.
CHARACTER AND IMAGE RECOGNITION FOR DATA CATALOGING IN ECOLOGICAL RESEARCHcscpconf
Data collection is an essential, but manpower intensive procedure in ecological research. An
algorithm was developed by the author which incorporated two important computer vision
techniques to automate data cataloging for butterfly measurements. Optical Character
Recognition is used for character recognition and Contour Detection is used for imageprocessing.
Proper pre-processing is first done on the images to improve accuracy. Although
there are limitations to Tesseract’s detection of certain fonts, overall, it can successfully identify
words of basic fonts. Contour detection is an advanced technique that can be utilized to
measure an image. Shapes and mathematical calculations are crucial in determining the precise
location of the points on which to draw the body and forewing lines of the butterfly. Overall,
92% accuracy were achieved by the program for the set of butterflies measured.
SOCIAL MEDIA ANALYTICS FOR SENTIMENT ANALYSIS AND EVENT DETECTION IN SMART CI...cscpconf
Smart cities utilize Internet of Things (IoT) devices and sensors to enhance the quality of the city
services including energy, transportation, health, and much more. They generate massive
volumes of structured and unstructured data on a daily basis. Also, social networks, such as
Twitter, Facebook, and Google+, are becoming a new source of real-time information in smart
cities. Social network users are acting as social sensors. These datasets so large and complex
are difficult to manage with conventional data management tools and methods. To become
valuable, this massive amount of data, known as 'big data,' needs to be processed and
comprehended to hold the promise of supporting a broad range of urban and smart cities
functions, including among others transportation, water, and energy consumption, pollution
surveillance, and smart city governance. In this work, we investigate how social media analytics
help to analyze smart city data collected from various social media sources, such as Twitter and
Facebook, to detect various events taking place in a smart city and identify the importance of
events and concerns of citizens regarding some events. A case scenario analyses the opinions of
users concerning the traffic in three largest cities in the UAE
SOCIAL NETWORK HATE SPEECH DETECTION FOR AMHARIC LANGUAGEcscpconf
The anonymity of social networks makes it attractive for hate speech to mask their criminal
activities online posing a challenge to the world and in particular Ethiopia. With this everincreasing
volume of social media data, hate speech identification becomes a challenge in
aggravating conflict between citizens of nations. The high rate of production, has become
difficult to collect, store and analyze such big data using traditional detection methods. This
paper proposed the application of apache spark in hate speech detection to reduce the
challenges. Authors developed an apache spark based model to classify Amharic Facebook
posts and comments into hate and not hate. Authors employed Random forest and Naïve Bayes
for learning and Word2Vec and TF-IDF for feature selection. Tested by 10-fold crossvalidation,
the model based on word2vec embedding performed best with 79.83%accuracy. The
proposed method achieve a promising result with unique feature of spark for big data.
GENERAL REGRESSION NEURAL NETWORK BASED POS TAGGING FOR NEPALI TEXTcscpconf
This article presents Part of Speech tagging for Nepali text using General Regression Neural
Network (GRNN). The corpus is divided into two parts viz. training and testing. The network is
trained and validated on both training and testing data. It is observed that 96.13% words are
correctly being tagged on training set whereas 74.38% words are tagged correctly on testing
data set using GRNN. The result is compared with the traditional Viterbi algorithm based on
Hidden Markov Model. Viterbi algorithm yields 97.2% and 40% classification accuracies on
training and testing data sets respectively. GRNN based POS Tagger is more consistent than the
traditional Viterbi decoding technique.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
zkStudyClub - LatticeFold: A Lattice-based Folding Scheme and its Application...Alex Pruden
Folding is a recent technique for building efficient recursive SNARKs. Several elegant folding protocols have been proposed, such as Nova, Supernova, Hypernova, Protostar, and others. However, all of them rely on an additively homomorphic commitment scheme based on discrete log, and are therefore not post-quantum secure. In this work we present LatticeFold, the first lattice-based folding protocol based on the Module SIS problem. This folding protocol naturally leads to an efficient recursive lattice-based SNARK and an efficient PCD scheme. LatticeFold supports folding low-degree relations, such as R1CS, as well as high-degree relations, such as CCS. The key challenge is to construct a secure folding protocol that works with the Ajtai commitment scheme. The difficulty, is ensuring that extracted witnesses are low norm through many rounds of folding. We present a novel technique using the sumcheck protocol to ensure that extracted witnesses are always low norm no matter how many rounds of folding are used. Our evaluation of the final proof system suggests that it is as performant as Hypernova, while providing post-quantum security.
Paper Link: https://eprint.iacr.org/2024/257
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving
Manufacturing custom quality metal nameplates and badges involves several standard operations. Processes include sheet prep, lithography, screening, coating, punch press and inspection. All decoration is completed in the flat sheet with adhesive and tooling operations following. The possibilities for creating unique durable nameplates are endless. How will you create your brand identity? We can help!
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Essentials of Automations: Exploring Attributes & Automation ParametersSafe Software
Building automations in FME Flow can save time, money, and help businesses scale by eliminating data silos and providing data to stakeholders in real-time. One essential component to orchestrating complex automations is the use of attributes & automation parameters (both formerly known as “keys”). In fact, it’s unlikely you’ll ever build an Automation without using these components, but what exactly are they?
Attributes & automation parameters enable the automation author to pass data values from one automation component to the next. During this webinar, our FME Flow Specialists will cover leveraging the three types of these output attributes & parameters in FME Flow: Event, Custom, and Automation. As a bonus, they’ll also be making use of the Split-Merge Block functionality.
You’ll leave this webinar with a better understanding of how to maximize the potential of automations by making use of attributes & automation parameters, with the ultimate goal of setting your enterprise integration workflows up on autopilot.
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsDianaGray10
Join us to learn how UiPath Apps can directly and easily interact with prebuilt connectors via Integration Service--including Salesforce, ServiceNow, Open GenAI, and more.
The best part is you can achieve this without building a custom workflow! Say goodbye to the hassle of using separate automations to call APIs. By seamlessly integrating within App Studio, you can now easily streamline your workflow, while gaining direct access to our Connector Catalog of popular applications.
We’ll discuss and demo the benefits of UiPath Apps and connectors including:
Creating a compelling user experience for any software, without the limitations of APIs.
Accelerating the app creation process, saving time and effort
Enjoying high-performance CRUD (create, read, update, delete) operations, for
seamless data management.
Speakers:
Russell Alfeche, Technology Leader, RPA at qBotic and UiPath MVP
Charlie Greenberg, host
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
AppSec PNW: Android and iOS Application Security with MobSFAjin Abraham
Mobile Security Framework - MobSF is a free and open source automated mobile application security testing environment designed to help security engineers, researchers, developers, and penetration testers to identify security vulnerabilities, malicious behaviours and privacy concerns in mobile applications using static and dynamic analysis. It supports all the popular mobile application binaries and source code formats built for Android and iOS devices. In addition to automated security assessment, it also offers an interactive testing environment to build and execute scenario based test/fuzz cases against the application.
This talk covers:
Using MobSF for static analysis of mobile applications.
Interactive dynamic security assessment of Android and iOS applications.
Solving Mobile app CTF challenges.
Reverse engineering and runtime analysis of Mobile malware.
How to shift left and integrate MobSF/mobsfscan SAST and DAST in your build pipeline.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
2. 64 Computer Science & Information Technology (CS & IT)
distributed computing, powerful machines are now available at low prices. This is visible in large
spreading of cluster with hundreds of homogeneous/heterogeneous processors connected by high
speed interconnection network [2]. This democratization of cluster calls for new practical
administration tools.
The task scheduling problem is to allocate resources (processors) to the tasks and to establish an
order for the tasks to be executed by resources. There are two different types of task scheduling:
static and dynamic. Static strategies define a schedule at compile time based on estimated time
required to execute tasks and to communicate data. Static schedule can be generated only when
the application behaviour is fully deterministic and this has the advantage of being more efficient
and a small overhead during runtime. The full global knowledge of application in the form of
DAG will help to generate a better schedule. Dynamic strategies, on the other hand are applied
when tasks are generated during runtime. Tasks are assumed to be non-preemptive.
Workflows have recently emerged as a paradigm for representing complex scientific
computations [26]. Few widely used example workflows are Montage (Fig. 2), cybershake,
LIGO, SIPHT. Workflows represented by one of many workflow programming languages can be
translated into DAG, in general. Thus workflow scheduling is essentially a problem of scheduling
DAG. Although much work has been done in scheduling single workflow [3], multiple workflow
scheduling is not receiving deserved attention. Few initial studies are found in the literature [4, 5].
Because of huge computing power of a cluster and the inability of a single DAG to utilize all
processors on cluster, multiple DAG applications need to be executed concurrently. Thus a
scheduler to deal with multi-user jobs with the objectives of maximizing resource utilization and
minimizing overall DAG completion time is essential. The contributions of this paper are 1) a
new method to find minimum, optimal and maximum number of processors that can be allotted
for a DAG and this information is used to find one best way to share available processors among
multiple DAGs 2) a framework to receive submission of DAGs, find the allotment for each
submitted DAG and schedule tasks on allotted processors, with the objectives of maximizing
resource utilization and minimizing overall completion time.
1.1. Application Model
The data flow model is gaining popularity as a programming paradigm for parallel processors.
When the characteristics of an application is fully deterministic, including task's execution time,
size of data communication between tasks, and task dependencies, the application can be
represented by directed acyclic graph (DAG) as shown in Fig.1. Each node in DAG represents a
task to be performed and the edges indicate inter-task dependencies. Node weight stands for the
computation cost of the corresponding task and the edge cost represents the volume of data to be
communicated between the corresponding nodes. The node and edge weights are usually
obtained by estimation or profiling. Communication-to-Computation (CCR) is the ratio of
average communication cost to the average computation cost of a DAG. This characterizes the
nature of DAG. The objective of scheduling is to map tasks onto processors and order their
execution so that task dependencies are satisfied and minimum overall completion time is
achieved. Makespan is the total time required to complete a DAG.
1.2. Platform
A cluster with ‘P’ homogeneous processors, each of which is a schedulable resource is
considered. Processors are interconnected by a high speed and low latency network. A processor
can communicate with several other processors simultaneously with multi-port model.
3. Computer Science & Information Technology (CS & IT) 65
Figure 1. A Typical DAG Figure 2. Montage – a Workflow
2. RELATED WORK
Extensive work has been done on scheduling a single DAG [6, 7, 8]. Zhao et al.[4] have proposed
few methods to schedule multiple DAGs. One approach is to combine several DAGs into one by
making the entry nodes of all DAGs, immediate successors of new entry node and then use
standard methods to schedule the combined DAG. Another way is to consider tasks from each
DAG in round robin manner for scheduling. They have proposed other policies to optimize both
makespan and fairness. The key idea is to evaluate, after scheduling a task, the slowdown value of
each DAG against other DAGs and make a decision on which DAG must be considered next for
scheduling.
A list scheduling method to schedule multi-user jobs is developed by Barbosa et al. [9] with an
aim to maximize the resource usage by allowing a floating mapping of processors to a given job,
instead of the common mapping approach that assigns a fixed set of processors to a user job for a
period of time. A master DAG where each node is a user job and each edge representing a priority
of one job over another is constructed using all submitted DAGs. A list scheduling algorithm [6]
is used to schedule all tasks of Master DAG. The master DAG is created based on job priorities
and deadlines
Bittencourt et al. [10] have used Path Clustering Heuristic (PCH) to cluster tasks and the entire
cluster is assigned to a single machine. They have proposed four heuristics which differ in the
order tasks of multiple DAGs are considered for scheduling. The methods are sequential, Gap
search method, Interleave algorithm and Group DAGs method. A meta-scheduler for multiple
DAGs [11] merges multiple DAGs into one to improve the overall parallelism and optimize idle
time of resources. The efforts are limited to the static case and they do not deal with dynamic
workloads.
Duan et al. [12] have proposed a scheduling algorithm based on the adoption of game theory and
idea of sequential cooperative game. They provide two novel algorithms to schedule multiple
DAGs which work properly for applications that can be formulated as a typical solvable game.
Zhifeng et al. [13] addresses the problem of dynamic scheduling multiple DAGs from different
4. 66 Computer Science & Information Technology (CS & IT)
users. They expose a similar approach from Zhao et al. [4] without merging DAGs. Their
algorithm is similar to G-heft algorithm.
An application which can exploit both task and data parallelism can be structured as Parallel Data
Graph (PTG) in which task can be either sequential or data parallel. Data parallelism means
parallel execution of the same code segment but on different sections of data, distributed over
several processors in a network. A DAG is a special case of PTG where task can only be
sequential task. Thus PTG scheduling is quite similar to DAG scheduling. Not much work is
carried out in Multiple PTG scheduling. Tapke et al. [5] have proposed an approach where each
PTG is given a maximum constraint on number of processors it can use and tasks are scheduled
using a known PTG scheduling algorithm. The size of each processor subset is determined
statically according to various criteria pertaining to the characteristics of PTG like maximum
width, total absolute work to be done and proportional work to be carried out.
Sueter et al. [14] have focused on developing strategies that provide a fair distribution of
resources among Parallel Task Graphs (PTG), with the objectives of achieving fairness and
makespan minimization. Constraints are defined according to four general resource sharing
policies: unbounded Share(S), Equal Share (ES), Proportional Share (PS), and Weighted
Proportional Share (WPS). S policy uses all available resources. ES policy uses equal resources
for each PTG. PS and WPS use resources proportional to the work of each PTG, where the work
is considered as critical path cost by width of PTG.
A study of algorithms to schedule multiple PTGs on a single homogeneous cluster is carried out
by Casanova et al. [15]. Therein it is shown that best algorithms in terms of performance and
fairness all use the same principle of allocating a subset of processors to each PTG and that this
subset remains fixed throughout the execution of the whole batch of PTGs. The basic idea in job
schedulers [19] is to queue jobs and to schedule them one after the other using some simple rules
like FCFS (First Come First Served) with priorities. Jackson et al. [20] extended this model with
additional features like fairness and backfilling.
Online scheduling of multiple DAGs is addressed in [16]. Authors have proposed two strategies
based on aggregating DAGs into a single DAG. A modified FCFS and Service-On-Time (SOT)
scheduling are applied. FCFS appends arriving DAGs to an exit node of the single DAG, while
SOT appends arriving DAGs to a task whose predecessors have not completed execution. Once
the single DAG has been built, scheduling is carried out by HEFT.
An Online Workflow Management (OWM) strategy [18] for scheduling multiple mix-parallel
workflows is proposed. DAG tasks are labelled, sorted, and stored into independent buffers.
Labelling is based on the upward rank strategy. The sorting arranges tasks in descendent order of
the task rank. Task scheduling referred to as a rank hybrid phase determines the task execution
order. Tasks are sorted in descending order when all tasks in the queue belong to the same
workflow. Otherwise, they are sorted in ascending order. Allocation assigns idle processors to
tasks from the waiting queue.
Raphael et al. [23] have addressed online scheduling of several applications modelled as work-
flows. They have extended a well-known list scheduling heuristic (HEFT) and adapted it to the
multi-workflow context. Six different heuristics based on HEFT key ideas are proposed. These
heuristics have been designed to improve the slowdown of different applications sent from
multiple users.
Much work has not been done on scheduling multiple DAG applications. A common approach is
to schedule a single DAG on fixed number of processors [6] but methods to find the number of
processors to be used for a DAG, are not found in literature. Tapke et al. [5] have proposed
5. Computer Science & Information Technology (CS & IT) 67
methods to find maximum resource constraint for each PTG, while scheduling multiple PTGs.
But they have not restricted scheduling of a PTG to the fixed set of processors. A method
proposed by Barbosa et al. allows floating number of processors to a given job, instead of fixed
number of processors. Many existing workflow scheduling methods do not use fixed set of
processors for each DAG. Instead a task based on some heuristic is picked among tasks of all
DAGs and is scheduled on a processor where it can start earliest based on some heuristic. The
work proposed in this paper is similar to Barbosa [9] method, in the sense that a fixed set of
processor is allocated for each DAG which later can be varied during runtime with the objective
of maximizing resource usage. Their work does not address several issues like - how initial
processor allotment for each DAG is made, a method to decide number of DAGs to be scheduled
concurrently among several submitted DAGs, to deal with online submission of DAGs. This work
addresses all the above mentioned issues.
3. PROCESSOR ALLOTMENT FOR A DAG
A schedule for a DAG can be obtained with varied number of processors. By increasing the
number of processors allotted for a DAG, its makespan decreases. The gain in terms of reduction
in makespan, reduces as more number of processors are allotted to a DAG. This is due to
communication overhead and limited parallelism present in DAG. The optimal and maximum
number of processors for a DAG will help in finding processor allotment while scheduling
multiple DAGs concurrently on a cluster.
3.1. Maximum Number of Processor for a DAG
The maximum number of processors a DAG can utilize depends on its nature and degree of
parallelism present in it. The number of allotted processors, beyond which DAG’s makespan
does not decrease with any more additional processors, is the maximum number of processors
that can be utilized by a DAG. A brute force method can be used to find this, by making several
calls to scheduling method and recording the makespan for each case. But an efficient binary
search based method [24] is used in this work and its time complexity is O(log n) against O(n) of
the brute force method.
3.2. Optimal Number of Processor for a DAG
Optimal number of processors for a DAG is that number up to which every added processor is
utilized well and beyond it, they are underutilized. With increase in number of allotted
processors, DAG’s makespan decreases and average processor utilization decreases due to
communication overhead and limited parallelism. Average processor utilization can best be
measured using computing area, which is the product of makespan and the number of processors
used. In this work, computing area is used to find the optimal number of processors for a DAG.
As processor allotment to a DAG is increased, makespan decreases and computing area increases.
Initially decrease in makespan is more than increase in computing area, justifying the worthiness
of increase in processor allotment. After the processor allotment reaches a certain value, the
increase in computing area is more than the decrease in makespan for every added processor,
indicating that any further increase in processors allotment is not of significant use.
By successively increasing number of processors allotted for a DAG, makespan and computing
area are recorded. The number of processors for which decrease in makespan becomes less than
the increase in computing area, fixes the optimal number of processors for a DAG.
.
6. 68 Computer Science & Information Technology (CS & IT)
4. MULTIPLE DAGS SHARING CLUSTER
It is advantageous to schedule multiple DAGs simultaneously on a cluster instead of dedicating
the entire cluster to a single DAG, due to communication overhead. Furthermore, it is beneficial
to schedule more number of DAGs each with relatively less number of processors than
scheduling less number of DAGs each with large number of processors, because of
communication overhead. The returns, in terms of decrease in makespan, for each additional
processor differs for each DAG depending on its nature and number of processors already allotted
to it. Hence additional processors must be allotted to those DAGs which will be benefitted most
by means of reduction in makespan. To find the most benefitting DAGs, reduction in makespan
for the next added processor must be known for each DAG. Reduction in makespan for every
added processor can be best captured as an equation using regression analysis and are provided to
scheduler along with DAG. During regression analysis of large number of DAGs, it is observed
that for any DAG, makespan reduction follows either exponential or power curve. Thus for each
DAG, makespan reduction for each added processor is recorded and curve fitting is done. The
type of equation and its constants are stored along with each DAG, which then is used by the
scheduler while finding processor allotment for each DAG, while scheduling multiple DAGs. The
scheduler is invoked when a DAG arrives or a DAG completes execution. The minimum number
of processors to be allotted for each DAG is assumed to be four, by conducting the experiments
large number of times. The algorithm for the proposed scheduler is given below.
Algorithm multi_dag_scheduler()
// information submitted along each DAG – opt_proc, max_proc, eqn_type, eqn_const
//avail_proc – currently available number of free processors
// let min_core = 4
Input : submitted DAGs
Output : processor allotment and calling an instance of scheduler for each DAG
Step 1: if (arrival) then // DAG has arrived
Step 2: if (avail_proc < min_core ) then
Step 3: append to waiting queue
Step 4: else
Step 5 : allot (min_proc or max_proc or opt_proc) whichever best fits avail_proc
Step 6 : create an instance of scheduler for a DAG on allotted processors
Step 7 : endif
Step 8: else //DAG has completed
Step 9: if ( waiting queue is not empty) then
Step 10 : do_allot()
Step 11 : endif
Step 12 : end_algorithm
Algorithm do_allot()
Step 1: remove those many number of DAGs from queue beginning, whose sum of their
min_proc is less than avail_proc
Step 2 : if (sum of opt_proc of all removed DAGs is less than available processors) then
Step 3 : allot opt_proc to each removed DAG
Step 4 : else
Step 5 : for each removed DAG allot their min_proc number of processors
Step 6 : endif
7. Computer Science & Information Technology (CS & IT) 69
Step 7 : if (free processors are left) then
Step 8 : distribute those free processors among DAGs, in such a way that each processor is
added to that DAG for which it yields maximum reduction in makespan, using
equations types and their constants
Step 9 : endif
Step 10 : end_algorithm
5. EXPERIMENTAL SETUP AND RESULTS
5.1. Experiment Setup
A discrete-event based simulator is developed to simulate the arrival, scheduling, execution and
completion of DAGs. Simulation allows performing statistically significant number of
experiments for a wide range of application configurations. Poisson distribution is used to
simulate the arrival time of DAGs. Several kinds of benchmark DAGs from several sources are
used to experiment the proposed scheduler for different types of DAGs. Series-parallel DAGs
from Task Graphs For Free [22], random DAGs from Standard Task Graph Set [21], DAGs of
linear algebra applications like FFT, LU decomposition, Gauss-elimination, Laplace transform
and workflows like LIGO, cybershake, Montage, SIPHT[25]. DAGs with CCR values of 0.1,
0.4, 1 and 5 are used in experiments.
5.2. Results and Analysis
5.2.1. Optimal and Maximum Number of Processors for a DAG
An efficient binary search based method [24] with time complexity of O(log(n)) is used to find
the maximum number of processors a DAG can utilize. The decrease in makespan and increase in
computing area (decrease in average processor utilization) for every added processor is used to
fix the optimal number of processors for a DAG. The plot of decrease in makespan and increase
in computing area for different number of processors, for a DAG is given in Fig.3. The crossover
point gives the optimal number of processors for that DAG. The method can be used for any kind
of DAG.
Figure 3. To find Optimal Number of Processors for a DAG
8. 70 Computer Science & Information Technology (CS & IT)
5.2.1. Multiple DAGs Scheduling
Recent works on multiple DAG scheduling [4, 5, 9, 10, 11, 12, 14] have not considered allotment
of fixed set of processors to a DAG. Instead, tasks from all DAGs are scheduled on any processor
on which they can start earliest, using some heuristic. Hence initially, it is proved experimentally
that space partitioning of processors among multiple DAGs, delivers improvement in
performance compared to combined DAGs scheduling. To experiment this, a set of DAGs of all
kinds, were scheduled on a cluster with 100 numbers of processors. The metric used is the sum of
computing area of all scheduled DAGs. To study the effect on both computation intensive and
communication intensive applications, DAGs with both low and high CCR are considered. Two
sets of DAGs each with 8 and 16 number of DAGs, under each category are considered. Thus the
four categories of DAGs are labelled as ccrl_8, ccrl_16, ccrh_8 and ccrh_16. Since the behaviour
depends on the nature of DAG, 50 sets of DAGs are considered for each category. Care is taken
to consider all different types of DAGs in the sets of DAGs. The results obtained from 50 sets are
averaged and the same is shown in Fig. 4. The performance of the proposed method is better than
combined DAGs scheduling for all four categories of DAGs. For the category ccrh_8, proposed
method shows maximum improvement of 12%, since DAGs are communication intensive and
thus scheduling tasks on fixed set of processors reduces time to complete the DAG. Performance
improvement is only 9% for the category ccrh_18, as there is less scope for further improvement
due to large number of DAGs being scheduled together.
Figure 4. Combined DAGs scheduling vs Proposed Space-sharing Schedule
The benefits of space partitioning processors which cannot be measured for DAGs with dummy
tasks are 1) as tasks of a DAG are scheduled on the same set of processors, they will be benefitted
from cache-warm and secondary memory warm. 2) an online scheduler can be used for each
DAG, after allotting a set of processors to it. 3) processor allotment for a DAG can be varied
depending on availability of processor, with the objectives of maximizing resource utilization.
A highlight of this work is to find one best way to share available processors among multiple
DAGs, using regression analysis. The proposed work is compared against policies proposed by
Tapke et al. [14] - unbounded Share(S), Equal Share (ES), Proportional Share (PS), and Weighted
Proportional Share (WPS). The strategy S which is a selfish allocations and tasks of different
DAGs are not differentiated is used as a baseline performer for other strategies as it gives an
9. Computer Science & Information Technology (CS & IT) 71
indication of performance of heuristics originally designed for single DAG. Values obtained are
normalized with the value of S strategy, to help in comparison. Performance metric used is
average makespan and resource utilization which is measured as the sum of computing area of all
DAGs scheduled together. Five categories of DAGs each with 4, 8, 12, 16 and 20 number of
DAGs are considered. Random, series-parallel, linear algebra DAGs and various workflows like
montage, SIPHT, epigenemics, LIGO are considered. 100 sets of DAGs are considered for each
category and the results obtained are averaged. The result is shown in Fig. 5 and Fig. 6. The
proposed method is better than all policies found in literature.
For less number of DAGs, performance of all methods is almost the same, as there will not be
much conflict for resources. With more number of DAGs, resource conflicts increase and the
proposed method shows considerable good performance over previous methods.
Figure 5. Normalized Average Makespan of Set of DAGs
Figure 6. Normalized Sum of Computing Area of Set of DAGs
10. 72 Computer Science & Information Technology (CS & IT)
6. CONCLUSIONS
Multiple DAGs scheduling on a cluster is not receiving the deserved attention. Few methods
found in literature performing combined DAGs scheduling. But in this work, it is proposed to
allot a fixed number of processors to each DAG and an instance of local DAG scheduler to
schedule DAG’s tasks only on the allotted fixed set of processors. A method to find the maximum
and optimal number of processors that can be allotted to a DAG is given, which will be used to
find the processor allotment for each DAG while scheduling multiple DAGs. A new framework
to schedule multiple DAGs with the objectives of maximizing resource utilization and
minimizing DAGs completion time is proposed. Regression analysis is used to find the number of
processors to be allotted to each DAG while scheduling multiple DAGs. This method is proved to
outperform other methods found in literature by around 10-15%.
The other big advantage of the proposed approach is that instead of static schedule, an online
scheduler for each DAG can be used to schedule tasks, as they are generated, onto the allotted
processor. An Online scheduler overcomes drawbacks of static schedule and is more
advantageous. Also static DAG information can be used during online scheduling to further
improve performance. Because of space sharing of processors, the number of processors allotted
to each DAG can be varied during runtime, depending on the availability of free processors. This
will improve resource utilization, hence performance of the scheduler. In future work, the idea of
online scheduler and varied processor allotment for each DAG will be experimented.
REFERENCES
[1] Pinedo, M.L.: Scheduling: Theory, Algorithms, and Systems, 3rd edition. Springer (2008)
[2] D.E. Culler, J. P. Singh, and A. Gupta. Parallel Computer Architecture: A Hardware/Software
Approach. Morgan Kaufmann Publishers, inc., SanFrancisco, CA, 1999..
[3] Yu, J., Buyya, R.: A Taxonomy of Scientific Workflow Systems for Grid computing. SIGMOD
Records 34(3),44–49 (2005)
[4] Zhao, H., Sakellariou, R.: Scheduling Multiple DAGs onto Heterogeneous systems. In: Parallel and
Distributed Processing Symposium, 2006. (IPDPS 2006). 20th International, pp. 14–pp. IEEE (2006)
[5] N’takp´e, T., Suter, F., Casanova, H.: A Comparison of Scheduling Approaches for Mixed-Parallel
Applications on Heterogeneous Platforms. In: Parallel and Distributed Computing, 2007. ISPDC’07.
Sixth International Symposium on, pp. 35–35. IEEE (2007)
[6] Y.-K. Kwok and I. Ahmad. Static Scheduling Algorithms for Allocating Directed Task graphs to
Multiprocessors. ACM Computing Surveys, 31(4):406–471, 1999
[7] E. Ilavarasan and P. Thambidurai ,.Low Complexity Performance Effective Task Scheduling
Algorithm for Heterogeneous Computing Environments , Journal of Computer Sciences 3 (2): 94-103,
2007
[8] T. Hagras, J. Janeček . Static vs. Dynamic List-Scheduling Performance Comparison, Acta
Polytechnica Vol. 43 No. 6/2003
[9] J. Barbosa and A. P. Monteiro, A List Scheduling Algorithm for Scheduling Multi-user Jobs on
Clusters, High Performance Computing for Computational science- VECPA8 2008, Lecture Notes in
Computer Science volume 5336, 2008, pp123-136
[10] Bittencourt L.F., Madeira: Towards the Scheduling of Multiple Workflows on Computational Grids.
J. Grid Computing 8, 419–441 (2010)
[11] U. H. Nig and W. Schiffmann. A Meta-algorithm for Scheduling Multiple DAGs in Homogeneous
System Environments. In Proceedings of the 18th International Conference on Parallel and
Distributed Computing and Systems (PDCS’06), 2006
[12] R. Duan, R. Prodan, and T. Fahringer. Performance and Cost optimization for Multiple large-scale
Grid Workflow Applications. In Proceedings of the 2007 ACM/IEEE conference on Supercomputing,
pages 1–12, New York, NY, USA, 2007.
[13] Zhifeng Yu and Weisong Shi. A Planner-guided Scheduling Strategy for Multiple work-flow
Applications. Parallel Processing Workshops, International Conference on, 0:1–8,2008.
11. Computer Science & Information Technology (CS & IT) 73
[14] N’takpé, T., Suter, F.: Concurrent scheduling of parallel task graphs on multi-clusters using
constrained resource allocations. In: International Parallel and Distributed Processing
Symposium/International Parallel Processing Symposium, pp. 1–8 (2009)
[15]Casanova, H., Desprez, F., Suter, F.: On cluster Resource Allocation for Multiple Parallel Task
Graphs. Journal of Parallel and Distributed Computing 70(12), 1193–1203 (2010)
[16] Zhu, L., Sun, Z., Guo, W., Jin, Y., Sun, W., Hu, W.: Dynamic Multi DAG Scheduling Algorithm for
Optical Grid Environment. New Architecture Management Applications. V 6784(1), 1122 (2007)
[17] H Topcuoglu, S. Hariri and M. Y. Yu, Performance Effective and Low-complexity Task Scheduling
for Heterogene Computing, IEEE TPDS, 13(3):260-274, 2002
[18] Hsu, C.-C., Huang, K.-C., Wang, F.-J.: Online scheduling of workflow applications. In Grid
environments. Future Gen. Comput. Syst. 27, 860–870 (2011)
[19] R. L. Henderson, Job Scheduling under the Portable Batch System, Job Scheduling Strategies for
Parallel Processing, volume 949 of LNCS, pages 279–294, 1995.
[20] D. Jackson, Q. Snell, and M. J. Clement. Core algorithms of Job Schedule, Job Scheduling Strategies
for Parallel Processing, volume 2221 of LNCS, pages 87–102, 2001.
[21] Standard Task Graph Set http://www.kasahara.elec.waseda.ac.jp/schedule/.
[22] Task Graphs for Free http://ziyang.eecs.umich.edu/~dickrp/tgff
[23] Raphael Bolze, Frederic Desprez and Benjamin Insard, Evaluation of Online Multi-workflow
Heuristics based on List Scheduling Methods, Gwendia ANR-06-MDCA-009
[24] Uma B , C R Venugopal, A Novel Binary Search based method to find minimal number of cores
required to obtain minimal makespan, given makespan and given utilization of cores for a DAG
application, 4th International Conference on Computer and Automation Engineering, (ICCAE 2012)
January 14–15, 2012, Mumbai, India.
[25] https://confluence.pegasus.isi.edu/display/pegasus/WorkflowGenerator
[26] Taylor, I.J., Deelman, E., Gannon, D.B., Shields, Workflows for e-Science - Scientific Workflows for
Grids. Springer, New York (2007)
AUTHORS
C R Venugopal received his Ph. D from IIT, Bombay. Presently serving as a Professor
in Sri Jayachamarajendra College of Engineering, Mysore, India. His main research
interests are Cloud computing, High Performance computing, VLSI Technology and
File System development. Has authored more than 50 international conference and
journal papers.
Uma B completed M. Tech. in Computer Science and Engineering from IIT, Delhi.
Currently working as a Associate Professor in Malnad College of engineering, Hassan,
India. Her research interests are Parallel Programming, High Performance Computing
and Task scheduling.