Abstract Cloud Computing as the name suggests, it is a style of computing where different users uses the resources on the go i.e. over the Internet. In the recent era, this technology has emerged as a strong option for not only large scale organizations but also for small scale organizations that only access/use the resources what they want. In recent research study, many organizations lose significant part of their revenues in handling the requests given by the clients over the web servers i.e. unable to balance the load for web servers which results in loss of data, delay in time and increased costs. This Paper gives a new enhanced load balancing algorithm by which the performance of their web application can be increased. This Algorithm works on the major drawbacks such as delay in time, response to request ratio etc.
Modified Active Monitoring Load Balancing with Cloud Computingijsrd.com
Cloud computing is internet-based computing in which large groups of remote servers are networked to allow the centralized data storage, and online access to computer services or resources. Load Balancing is essential for efficient operations in distributed environments. As Cloud Computing is growing rapidly and clients are demanding more services and better results, load balancing for the Cloud has become a very interesting and important research area. In the absence of proper load balancing strategy/technique the growth of CC will never go as per predictions. The main focus of this paper is to verify the approach that has been proposed in the model paper [3]. An efficient load balancing algorithm has the ability to reduce the data center processing time, overall response time and to cope with the dynamic changes of cloud computing environments. The traditional load balancing Active Monitoring algorithm has been modified to achieve better data center processing time and overall response time. The algorithm presented in this paper efficiently distributes the requests to all the VMs for their execution, considering the CPU utilization of all VMs.
Analysis of a Pool Management Scheme for Cloud Computing Centres by Using Par...IJERA Editor
A monolithic model may suffer from and poor scalability due to large number of parameters. A cloud user may
submit a super task at once. The user request is sent to the global queue and then to the Resource Assigning
Module (RAM). A number of heterogeneous server pools placed in the RAM. First is Hot, in which the servers
will be handling the jobs currently, second is Warm, in which the servers are kept in ideal state, then Finally
Cold, in which the servers are Turned Off state. Initially the request is send to Hot, if those servers are busy the
request is forwarded to warm, then finally if required to Cold if both the hot and warm server pools are busy.
The user submitted supertask may split so that the individual task run on different physical machines, this is
called as partial acceptance policy. So the supertask rejection ratio will be reduced.
A Comparative Study of Load Balancing Algorithms for Cloud ComputingIJERA Editor
Cloud Computing is fast growing technology in both industry research and academy. User can access the cloud
service and pay based on the usage of resource. Balancing the load is major task of cloud service provider with
minimum response time, maximum throughput and better resource utilization. There are many load balancing
algorithms proposed to assign a user request to cloud resource in efficient manner. In this paper three load balancing
algorithms are simulated in Cloud Analyst and results are compared.
Cloud computing is that ensuing generation of computation. In all probability folks can have everything they need on the cloud. Cloud computing provides resources to shopper on demand. The resources also are code package resources or hardware resources. Cloud computing architectures unit distributed, parallel and serves the requirements of multiple purchasers in various things. This distributed style deploys resources distributive to deliver services with efficiency to users in various geographical channels. Purchasers in a very distributed setting generate request haphazardly in any processor. So the most important disadvantage of this randomness is expounded to task assignment. The unequal task assignment to the processor creates imbalance i.e., variety of the processors sq. measure over laden and many of them unit of measurement to a lower place loaded. The target of load equalisation is to transfer the load from over laden technique to a lower place loaded technique transparently. Load equalisation is one altogether the central issues in cloud computing. To comprehend high performance, minimum interval and high resource utilization relation we want to transfer the tasks between nodes in cloud network. Load equalisation technique is utilized to distribute tasks from over loaded nodes to a lower place loaded or idle nodes. In following sections we have a tendency to tend to stand live discuss concerning cloud computing, load equalisation techniques and additionally the planned work of our load equalisation system. Proposed load equalisation rule is simulated on Cloud Analyst toolkit. Performance is analyzed on the parameters of overall interval, knowledge transfer, average knowledge center mating time and total value of usage. Results area unit compared with 3 existing load equalisation algorithms specifically spherical Robin, Equally unfold Current Execution Load, and Throttled. Results on the premise of case studies performed shows additional knowledge transfer with minimum interval.
Cloud computing is an on demand service in which shared resources, information, software and other devices are provided to the end user as per their requirement at a specific time. A cloud consists of several elements such as clients, datacenters and distributed servers. There are n number of clients and end users involved in cloud environment. These clients may make requests to the cloud system simultaneously, making it difficult for the cloud to manage the entire load at a time. The load can be CPU load, memory load, delay or network load. This might cause inconvenience to the clients as there may be delay in the response time or it might affect the performance and efficiency of the cloud environment. So, the concept of load balancing is very important in cloud computing to improve the efficiency of the cloud. Good load balancing makes cloud computing more efficient and improves user satisfaction. This paper gives an approach to balance the incoming load in cloud environment by making partitions of the public cloud.
Modified Active Monitoring Load Balancing with Cloud Computingijsrd.com
Cloud computing is internet-based computing in which large groups of remote servers are networked to allow the centralized data storage, and online access to computer services or resources. Load Balancing is essential for efficient operations in distributed environments. As Cloud Computing is growing rapidly and clients are demanding more services and better results, load balancing for the Cloud has become a very interesting and important research area. In the absence of proper load balancing strategy/technique the growth of CC will never go as per predictions. The main focus of this paper is to verify the approach that has been proposed in the model paper [3]. An efficient load balancing algorithm has the ability to reduce the data center processing time, overall response time and to cope with the dynamic changes of cloud computing environments. The traditional load balancing Active Monitoring algorithm has been modified to achieve better data center processing time and overall response time. The algorithm presented in this paper efficiently distributes the requests to all the VMs for their execution, considering the CPU utilization of all VMs.
Analysis of a Pool Management Scheme for Cloud Computing Centres by Using Par...IJERA Editor
A monolithic model may suffer from and poor scalability due to large number of parameters. A cloud user may
submit a super task at once. The user request is sent to the global queue and then to the Resource Assigning
Module (RAM). A number of heterogeneous server pools placed in the RAM. First is Hot, in which the servers
will be handling the jobs currently, second is Warm, in which the servers are kept in ideal state, then Finally
Cold, in which the servers are Turned Off state. Initially the request is send to Hot, if those servers are busy the
request is forwarded to warm, then finally if required to Cold if both the hot and warm server pools are busy.
The user submitted supertask may split so that the individual task run on different physical machines, this is
called as partial acceptance policy. So the supertask rejection ratio will be reduced.
A Comparative Study of Load Balancing Algorithms for Cloud ComputingIJERA Editor
Cloud Computing is fast growing technology in both industry research and academy. User can access the cloud
service and pay based on the usage of resource. Balancing the load is major task of cloud service provider with
minimum response time, maximum throughput and better resource utilization. There are many load balancing
algorithms proposed to assign a user request to cloud resource in efficient manner. In this paper three load balancing
algorithms are simulated in Cloud Analyst and results are compared.
Cloud computing is that ensuing generation of computation. In all probability folks can have everything they need on the cloud. Cloud computing provides resources to shopper on demand. The resources also are code package resources or hardware resources. Cloud computing architectures unit distributed, parallel and serves the requirements of multiple purchasers in various things. This distributed style deploys resources distributive to deliver services with efficiency to users in various geographical channels. Purchasers in a very distributed setting generate request haphazardly in any processor. So the most important disadvantage of this randomness is expounded to task assignment. The unequal task assignment to the processor creates imbalance i.e., variety of the processors sq. measure over laden and many of them unit of measurement to a lower place loaded. The target of load equalisation is to transfer the load from over laden technique to a lower place loaded technique transparently. Load equalisation is one altogether the central issues in cloud computing. To comprehend high performance, minimum interval and high resource utilization relation we want to transfer the tasks between nodes in cloud network. Load equalisation technique is utilized to distribute tasks from over loaded nodes to a lower place loaded or idle nodes. In following sections we have a tendency to tend to stand live discuss concerning cloud computing, load equalisation techniques and additionally the planned work of our load equalisation system. Proposed load equalisation rule is simulated on Cloud Analyst toolkit. Performance is analyzed on the parameters of overall interval, knowledge transfer, average knowledge center mating time and total value of usage. Results area unit compared with 3 existing load equalisation algorithms specifically spherical Robin, Equally unfold Current Execution Load, and Throttled. Results on the premise of case studies performed shows additional knowledge transfer with minimum interval.
Cloud computing is an on demand service in which shared resources, information, software and other devices are provided to the end user as per their requirement at a specific time. A cloud consists of several elements such as clients, datacenters and distributed servers. There are n number of clients and end users involved in cloud environment. These clients may make requests to the cloud system simultaneously, making it difficult for the cloud to manage the entire load at a time. The load can be CPU load, memory load, delay or network load. This might cause inconvenience to the clients as there may be delay in the response time or it might affect the performance and efficiency of the cloud environment. So, the concept of load balancing is very important in cloud computing to improve the efficiency of the cloud. Good load balancing makes cloud computing more efficient and improves user satisfaction. This paper gives an approach to balance the incoming load in cloud environment by making partitions of the public cloud.
ANALYSIS ON LOAD BALANCING ALGORITHMS IMPLEMENTATION ON CLOUD COMPUTING ENVIR...AM Publications
Cloud computing means storing and accessing data and programs over the Internet instead of your computer's hard drive. The cloud is just a metaphor for the Internet. The elements involved in cloud computing are clients, data center and distributed server. One of the main problems in cloud computing is load balancing. Balancing the load means to distribute the workload among several nodes evenly so that no single node will be overloaded. Load can be of any type that is it can be CPU load, memory capacity or network load. In this paper we presented an architecture of load balancing and algorithm which will further improve the load balancing problem by minimizing the response time. In this paper, we have proposed the enhanced version of existing regulated load balancing approach for cloud computing by comping the Randomization and greedy load balancing algorithm. To check the performance of proposed approach, we have used the cloud analyst simulator (Cloud Analyst). Through simulation analysis, it has been found that proposed improved version of regulated load balancing approach has shown better performance in terms of cost, response time and data processing time.
Virtualization Technology using Virtual Machines for Cloud ComputingIJMER
Cloud computing is the delivery of computing and storage capacity as a service to a community of end users. The name “cloud computing” comes from the use of a cloud-shaped symbol as an abstraction for the complex infrastructure it contains in system diagrams. Cloud computing entrusts services with a user's software, data and computation over a network. End users access cloud-based applications through a web browser or mobile application or a light-weight desktop while the business software and user's data are stored on servers at a remote location. Proponents claim that cloud computing environment allows enterprises to get their applications up and running faster, with improved manageability and less maintenance, and enables IT industry to more rapidly adjust resources to meet fluctuating and unpredictable business demand. In this paper, we present a system that uses virtualization technology to allocate the data center resources dynamically based on the application demands and support green computing by optimizing the number of servers in use. This method multiplexes virtual to physical resources adaptively based on the changing demand. We use the concept of skewness metric to combine virtual machines with different resource characteristics appropriately so that the capacities of servers are well utilized.
A Novel Switch Mechanism for Load Balancing in Public CloudIJMER
In cloud computing environment, one of the core design principles is dynamic scalability,
which guarantees cloud storage service to handle the growing amounts of application data in a flexible
manner or to be readily enlarged. By integrating several private and public cloud services, the hybrid
clouds can effectively provide dynamic scalability of service and data migration. A load balancing is a
method of dividing computing loads among numerous hardware resources. Due to unpredictable job
arrival pattern and the capacities of the nodes in cloud differ for the load balancing problem. In this load
control is very crucial to improve system performance and maintenance. This paper presents a switch
mechanism for load balancing in cloud computing. The load balancing model given in this work is aimed
at the public cloud which has numerous nodes with distributed computing resources in many different
geographical areas. Thus, this model divides the public cloud environment into several cloud partitions.
When the cloud environment is very large and complex, these divisions simplify the load balancing. The
cloud environment has a main controller that chooses the suitable partitions for arriving jobs while the
balancer for each cloud partition chooses the best load balancing strategy
Public Cloud Partition Using Load Status Evaluation and Cloud Division RulesIJSRD
with growth of cloud computing load balancing is important impact on performance. Cloud computing efficiency depends on good load balancer. Many type of situation occur that time cloud partitioning is done by load balancer. Different type of situation needed different type of strategies for public cloud portioning using load balancer.in this paper we work on, partition of public cloud using two type of situation first is load status evaluation and second is cloud division rules. Load status evaluation is measure in number of cloudlets arrives at datacenter and cloud divisions rules are based on cloudlet come from which geographical location. On the basis of geographical location we partition public cloud and improve performance of load balancing in cloud computing. We implement proposed system with help of cloudsim3.0 simulator.
Hybrid Scheduling Algorithm for Efficient Load Balancing In Cloud ComputingEswar Publications
In cloud computing environment, various users send requests for the transmission of data for different demands. The access to different number of users increase load on the cloud servers. Due to this, the cloud server does not provide best efficiency. To provide best efficiency, load has to be balanced. The highlight of this work is the division of different jobs into tasks. The job dependency checking is done on the basis of directed acyclic graph. The dependency checking the make span has to be created on the basis of first come first serve and priority based scheduling algorithms. In this paper, each scheduling algorithm has been implemented sequentially and the hybrid algorithm (round robin and priority based) has also been compared with other scheduling algorithms.
In this research paper, we have conducted work on modeling of local broker policy based on workload profile in Network cloud. For this we are using workload based applications. To handle workload based applications, two Scheduling Policies Random Non-overlap and Workload profile based be used. We compare these two scheduling policies based on three parameters Execution (mean) time, Response (mean) time and Waiting (mean) time. Workload based profile policy gave better results than Random-Non overlap policy in terms of time performance parameter.
Load Balancing in Cloud Computing Environment: A Comparative Study of Service...Eswar Publications
Load balancing is a computer networking method to distribute workload across multiple computers or a computer cluster, network links, central processing units, disk drives, or other resources, to achieve optimal resource utilization, maximize throughput, minimize response time, and avoid overload. Using multiple components with load balancing, instead of a single component, may increase reliability through redundancy. The
load balancing service is usually provided by dedicated software or hardware, such as a multilayer switch or a Domain Name System server. In this paper, the existing static algorithms used for simple cloud load balancing have been identified and also a hybrid algorithm for developments in the future is suggested.
Cloud Partitioning of Load Balancing Using Round Robin ModelIJCERT
Abstract: The purpose of load balancing is to look up the performance of a cloud environment through an appropriate
circulation strategy. Good load balancing will construct cloud computing for more stability and efficiency. This paper
introduces a better round robin model for the public cloud based on the cloud partitioning concept with a switch mechanism
to choose different strategies for different situations. Load balancing is the process of giving out of workload among
different nodes or processor. It will introduces a enhanced approach for public cloud load distribution using screening and
game theory concept to increase the presentation of the system.
Area And Power Efficient LMS Adaptive Filter With Low Adaptation DelayEditor IJMTER
We present an efficient architecture for the implementation of delayed least mean square
adaptive filter. We use a novel partial product generator and propose a strategy for optimized
balanced pipelining across the time consuming combinational blocks of the structure .An efficient
systolic architecture of the delayed least mean square adaptive filter based on the processing element
.We propose an efficient fixed point implementation of the proposed architecture ,and derive the
expression for steady-state error. The architecture is synthesized by using a number of function
preserving transformations on the signal flow graph representation of the delayed LMS
algorithm.
Enhancing minimal virtual machine migration in cloud environmenteSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Cloud Computing Load Balancing Algorithms Comparison Based SurveyINFOGAIN PUBLICATION
Cloud computing is an online primarily based computing. This computing paradigm has increased the employment of network wherever the potential of 1 node may be used by alternative node. Cloud provides services on demand to distributive resources like info, servers, software, infrastructure etc. in pay as you go basis. Load reconciliation is one amongst the vexing problems in distributed atmosphere. Resources of service supplier have to be compelled to balance the load of shopper request. Totally different load reconciliation algorithms are planned so as to manage the resources of service supplier with efficiency and effectively. This paper presents a comparison of assorted policies used for load reconciliation.
Load Balancing Algorithm to Improve Response Time on Cloud Computingneirew J
Load balancing techniques in cloud computing can be applied at different levels. There are two main
levels: load balancing on physical server and load balancing on virtual servers. Load balancing on a
physical server is policy of allocating physical servers to virtual machines. And load balancing on virtual
machines is a policy of allocating resources from physical server to virtual machines for tasks or
applications running on them. Depending on the requests of the user on cloud computing is SaaS (Software
as a Service), PaaS (Platform as a Service) or IaaS (Infrastructure as a Service) that has a proper load
balancing policy. When receiving the task, the cloud data center will have to allocate these tasks efficiently
so that the response time is minimized to avoid congestion. Load balancing should also be performed
between different datacenters in the cloud to ensure minimum transfer time. In this paper, we propose a
virtual machine-level load balancing algorithm that aims to improve the average response time and
average processing time of the system in the cloud environment. The proposed algorithm is compared to the
algorithms of Avoid Deadlocks [5], Maxmin [6], Throttled [8] and the results show that our algorithms
have optimized response times.
ANALYSIS ON LOAD BALANCING ALGORITHMS IMPLEMENTATION ON CLOUD COMPUTING ENVIR...AM Publications
Cloud computing means storing and accessing data and programs over the Internet instead of your computer's hard drive. The cloud is just a metaphor for the Internet. The elements involved in cloud computing are clients, data center and distributed server. One of the main problems in cloud computing is load balancing. Balancing the load means to distribute the workload among several nodes evenly so that no single node will be overloaded. Load can be of any type that is it can be CPU load, memory capacity or network load. In this paper we presented an architecture of load balancing and algorithm which will further improve the load balancing problem by minimizing the response time. In this paper, we have proposed the enhanced version of existing regulated load balancing approach for cloud computing by comping the Randomization and greedy load balancing algorithm. To check the performance of proposed approach, we have used the cloud analyst simulator (Cloud Analyst). Through simulation analysis, it has been found that proposed improved version of regulated load balancing approach has shown better performance in terms of cost, response time and data processing time.
Virtualization Technology using Virtual Machines for Cloud ComputingIJMER
Cloud computing is the delivery of computing and storage capacity as a service to a community of end users. The name “cloud computing” comes from the use of a cloud-shaped symbol as an abstraction for the complex infrastructure it contains in system diagrams. Cloud computing entrusts services with a user's software, data and computation over a network. End users access cloud-based applications through a web browser or mobile application or a light-weight desktop while the business software and user's data are stored on servers at a remote location. Proponents claim that cloud computing environment allows enterprises to get their applications up and running faster, with improved manageability and less maintenance, and enables IT industry to more rapidly adjust resources to meet fluctuating and unpredictable business demand. In this paper, we present a system that uses virtualization technology to allocate the data center resources dynamically based on the application demands and support green computing by optimizing the number of servers in use. This method multiplexes virtual to physical resources adaptively based on the changing demand. We use the concept of skewness metric to combine virtual machines with different resource characteristics appropriately so that the capacities of servers are well utilized.
A Novel Switch Mechanism for Load Balancing in Public CloudIJMER
In cloud computing environment, one of the core design principles is dynamic scalability,
which guarantees cloud storage service to handle the growing amounts of application data in a flexible
manner or to be readily enlarged. By integrating several private and public cloud services, the hybrid
clouds can effectively provide dynamic scalability of service and data migration. A load balancing is a
method of dividing computing loads among numerous hardware resources. Due to unpredictable job
arrival pattern and the capacities of the nodes in cloud differ for the load balancing problem. In this load
control is very crucial to improve system performance and maintenance. This paper presents a switch
mechanism for load balancing in cloud computing. The load balancing model given in this work is aimed
at the public cloud which has numerous nodes with distributed computing resources in many different
geographical areas. Thus, this model divides the public cloud environment into several cloud partitions.
When the cloud environment is very large and complex, these divisions simplify the load balancing. The
cloud environment has a main controller that chooses the suitable partitions for arriving jobs while the
balancer for each cloud partition chooses the best load balancing strategy
Public Cloud Partition Using Load Status Evaluation and Cloud Division RulesIJSRD
with growth of cloud computing load balancing is important impact on performance. Cloud computing efficiency depends on good load balancer. Many type of situation occur that time cloud partitioning is done by load balancer. Different type of situation needed different type of strategies for public cloud portioning using load balancer.in this paper we work on, partition of public cloud using two type of situation first is load status evaluation and second is cloud division rules. Load status evaluation is measure in number of cloudlets arrives at datacenter and cloud divisions rules are based on cloudlet come from which geographical location. On the basis of geographical location we partition public cloud and improve performance of load balancing in cloud computing. We implement proposed system with help of cloudsim3.0 simulator.
Hybrid Scheduling Algorithm for Efficient Load Balancing In Cloud ComputingEswar Publications
In cloud computing environment, various users send requests for the transmission of data for different demands. The access to different number of users increase load on the cloud servers. Due to this, the cloud server does not provide best efficiency. To provide best efficiency, load has to be balanced. The highlight of this work is the division of different jobs into tasks. The job dependency checking is done on the basis of directed acyclic graph. The dependency checking the make span has to be created on the basis of first come first serve and priority based scheduling algorithms. In this paper, each scheduling algorithm has been implemented sequentially and the hybrid algorithm (round robin and priority based) has also been compared with other scheduling algorithms.
In this research paper, we have conducted work on modeling of local broker policy based on workload profile in Network cloud. For this we are using workload based applications. To handle workload based applications, two Scheduling Policies Random Non-overlap and Workload profile based be used. We compare these two scheduling policies based on three parameters Execution (mean) time, Response (mean) time and Waiting (mean) time. Workload based profile policy gave better results than Random-Non overlap policy in terms of time performance parameter.
Load Balancing in Cloud Computing Environment: A Comparative Study of Service...Eswar Publications
Load balancing is a computer networking method to distribute workload across multiple computers or a computer cluster, network links, central processing units, disk drives, or other resources, to achieve optimal resource utilization, maximize throughput, minimize response time, and avoid overload. Using multiple components with load balancing, instead of a single component, may increase reliability through redundancy. The
load balancing service is usually provided by dedicated software or hardware, such as a multilayer switch or a Domain Name System server. In this paper, the existing static algorithms used for simple cloud load balancing have been identified and also a hybrid algorithm for developments in the future is suggested.
Cloud Partitioning of Load Balancing Using Round Robin ModelIJCERT
Abstract: The purpose of load balancing is to look up the performance of a cloud environment through an appropriate
circulation strategy. Good load balancing will construct cloud computing for more stability and efficiency. This paper
introduces a better round robin model for the public cloud based on the cloud partitioning concept with a switch mechanism
to choose different strategies for different situations. Load balancing is the process of giving out of workload among
different nodes or processor. It will introduces a enhanced approach for public cloud load distribution using screening and
game theory concept to increase the presentation of the system.
Area And Power Efficient LMS Adaptive Filter With Low Adaptation DelayEditor IJMTER
We present an efficient architecture for the implementation of delayed least mean square
adaptive filter. We use a novel partial product generator and propose a strategy for optimized
balanced pipelining across the time consuming combinational blocks of the structure .An efficient
systolic architecture of the delayed least mean square adaptive filter based on the processing element
.We propose an efficient fixed point implementation of the proposed architecture ,and derive the
expression for steady-state error. The architecture is synthesized by using a number of function
preserving transformations on the signal flow graph representation of the delayed LMS
algorithm.
Enhancing minimal virtual machine migration in cloud environmenteSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Cloud Computing Load Balancing Algorithms Comparison Based SurveyINFOGAIN PUBLICATION
Cloud computing is an online primarily based computing. This computing paradigm has increased the employment of network wherever the potential of 1 node may be used by alternative node. Cloud provides services on demand to distributive resources like info, servers, software, infrastructure etc. in pay as you go basis. Load reconciliation is one amongst the vexing problems in distributed atmosphere. Resources of service supplier have to be compelled to balance the load of shopper request. Totally different load reconciliation algorithms are planned so as to manage the resources of service supplier with efficiency and effectively. This paper presents a comparison of assorted policies used for load reconciliation.
Load Balancing Algorithm to Improve Response Time on Cloud Computingneirew J
Load balancing techniques in cloud computing can be applied at different levels. There are two main
levels: load balancing on physical server and load balancing on virtual servers. Load balancing on a
physical server is policy of allocating physical servers to virtual machines. And load balancing on virtual
machines is a policy of allocating resources from physical server to virtual machines for tasks or
applications running on them. Depending on the requests of the user on cloud computing is SaaS (Software
as a Service), PaaS (Platform as a Service) or IaaS (Infrastructure as a Service) that has a proper load
balancing policy. When receiving the task, the cloud data center will have to allocate these tasks efficiently
so that the response time is minimized to avoid congestion. Load balancing should also be performed
between different datacenters in the cloud to ensure minimum transfer time. In this paper, we propose a
virtual machine-level load balancing algorithm that aims to improve the average response time and
average processing time of the system in the cloud environment. The proposed algorithm is compared to the
algorithms of Avoid Deadlocks [5], Maxmin [6], Throttled [8] and the results show that our algorithms
have optimized response times.
LOAD BALANCING ALGORITHM TO IMPROVE RESPONSE TIME ON CLOUD COMPUTINGijccsa
Load balancing techniques in cloud computing can be applied at different levels. There are two main
levels: load balancing on physical server and load balancing on virtual servers. Load balancing on a
physical server is policy of allocating physical servers to virtual machines. And load balancing on virtual
machines is a policy of allocating resources from physical server to virtual machines for tasks or
applications running on them. Depending on the requests of the user on cloud computing is SaaS (Software
as a Service), PaaS (Platform as a Service) or IaaS (Infrastructure as a Service) that has a proper load
balancing policy. When receiving the task, the cloud data center will have to allocate these tasks efficiently
so that the response time is minimized to avoid congestion. Load balancing should also be performed
between different datacenters in the cloud to ensure minimum transfer time. In this paper, we propose a
virtual machine-level load balancing algorithm that aims to improve the average response time and
average processing time of the system in the cloud environment. The proposed algorithm is compared to the
algorithms of Avoid Deadlocks [5], Maxmin [6], Throttled [8] and the results show that our algorithms
have optimized response times.
Performance Comparision of Dynamic Load Balancing Algorithm in Cloud ComputingEswar Publications
Cloud computing as a distributed paradigm, it has the latent to make over a large part of the Cooperative industry. In cloud computing it’s automatically describe more technologies like distributed computing, virtualization, software, web services and networking. We review the new cloud computing technologies, and indicate the main challenges for their development in future, among which load balancing problem stands out and attracts our attention Concept of load balancing in networking and in cloud environment both are widely different. Load balancing in networking its complete concern to avoid the problem of overloading and under loading in any sever networking cloud computing its complete different its involves different elements metrics such as security, reliability, throughput, tolerance, on demand services, cost etc. Through these elements we avoiding various node problem of distributing system where many services waiting for request and others are heavily loaded and through these its increase response time and degraded performance optimization. In this paper first we classify algorithms in static and dynamic. Then we analyzed the dynamic algorithms applied in dynamics environments in cloud. Through this paper we have been show compression of various dynamics algorithm in which we include honey bee algorithm, throttled algorithm, Biased random algorithm with different elements and describe how and which is best in cloud environment with different metrics mainly used elements are performance, resource utilization and minimum cost. Our main focus of paper is in the analyze various load
balancing algorithms and their applicability in cloud environment.
LOAD BALANCING ALGORITHM ON CLOUD COMPUTING FOR OPTIMIZE RESPONE TIMEijccsa
To improve the performance of cloud computing, there are many parameters and issues that we should consider, including resource allocation, resource responsiveness, connectivity to resources, unused resources exploration, corresponding resource mapping and planning for resource. The planning for the use of resources can be based on many kinds of parameters, and the service response time is one of them.
The users can easily figure out the response time of their requests, and it becomes one of the important QoSs. When we discover and explore more on this, response time can provide solutions for the distribution, the load balancing of resources with better efficiency. This is one of the most promising
research directions for improving the cloud technology. Therefore, this paper proposes a load balancing algorithm based on response time of requests on cloud with the name APRA (ARIMA Prediction of Response Time Algorithm), the main idea is to use ARIMA algorithms to predict the coming response time, thus giving a better way of effectively resolving resource allocation with threshold value. The experiment
result outcomes are potential and valuable for load balancing with predicted response time, it shows that prediction is a great direction for load balancing.
Cloud computing is that ensuing generation of computation. In all probability folks can have everything they need on the cloud. Cloud computing provides resources to shopper on demand. The resources also are code package resources or hardware resources. Cloud computing architectures unit distributed, parallel and serves the requirements of multiple purchasers in various things. This distributed style deploys resources distributive to deliver services with efficiency to users in various geographical channels. Purchasers in a very distributed setting generate request haphazardly in any processor. So the most important disadvantage of this randomness is expounded to task assignment. The unequal task assignment to the processor creates imbalance i.e., variety of the processors sq. measure over laden and many of them unit of measurement to a lower place loaded. The target of load equalisation is to transfer the load from over laden technique to a lower place loaded technique transparently. Load equalisation is one altogether the central issues in cloud computing. To comprehend high performance, minimum interval and high resource utilization relation we want to transfer the tasks between nodes in cloud network. Load equalisation technique is utilized to distribute tasks from over loaded nodes to a lower place loaded or idle nodes. In following sections we have a tendency to tend to stand live discuss concerning cloud computing, load equalisation techniques and additionally the planned work of our load equalisation system. Proposed load equalisation rule is simulated on Cloud Analyst toolkit. Performance is analyzed on the parameters of overall interval, knowledge transfer, average knowledge center mating time and total value of usage. Results area unit compared with 3 existing load equalisation algorithms specifically spherical Robin, Equally unfold Current Execution Load, and Throttled. Results on the premise of case studies performed shows additional knowledge transfer with minimum interval.
PROPOSED LOAD BALANCING ALGORITHM TO REDUCE RESPONSE TIME AND PROCESSING TIME...IJCNCJournal
Cloud computing is a new technology that brings new challenges to all organizations around the world.
Improving response time for user requests on cloud computing is a critical issue to combat bottlenecks. As
for cloud computing, bandwidth to from cloud service providers is a bottleneck. With the rapid development
of the scale and number of applications, this access is often threatened by overload. Therefore, this paper
our proposed Throttled Modified Algorithm(TMA) for improving the response time of VMs on cloud
computing to improve performance for end-user. We have simulated the proposed algorithm with the
CloudAnalyts simulation tool and this algorithm has improved response times and processing time of the
cloud data center.
Dynamic Cloud Partitioning and Load Balancing in Cloud Shyam Hajare
Cloud computing is the emerging and transformational paradigm in the field of information technology. It mostly focuses in providing various services on demand and resource allocation and secure data storage are some of them. To store huge amount of data and accessing data from such metadata is new challenge. Distributing and balancing of the load over a cloud using cloud partitioning can ease the situation. Implementing load balancing by considering static as well as dynamic parameters can improve the performance cloud service provider and can improve the user satisfaction. Implementation the model can provide dynamic way of resource selection de-pending upon different situation of cloud environment at the time of accessing cloud provisions based on cloud partitioning. This model can provide effective load balancing algorithm over the cloud environment, better refresh time methods and better load status evaluation methods.
dynamic resource allocation using virtual machines for cloud computing enviro...Kumar Goud
Abstract—Cloud computing allows business customers to scale up and down their resource usage based on needs., we present a system that uses virtualization technology to allocate data center resources dynamically based on application demands and support green computing by optimizing the number of servers in use. We introduce the concept of “skewness” to measure the unevenness in the multidimensional resource utilization of a server. By minimizing imbalance, we will mix completely different of workloads nicely and improve the overall utilization of server resources. We develop a set of heuristics that prevent overload in the system effectively while saving energy used. Many of the touted gains in the cloud model come from resource multiplexing through virtualization technology. In this paper Trace driven simulation and experiment results demonstrate that our algorithm achieves good performance.
Index Terms—Cloud computing, resource management, virtualization, green computing.
Similar to Enhanced equally distributed load balancing algorithm for cloud computing (20)
Mechanical properties of hybrid fiber reinforced concrete for pavementseSAT Journals
Abstract
The effect of addition of mono fibers and hybrid fibers on the mechanical properties of concrete mixture is studied in the present
investigation. Steel fibers of 1% and polypropylene fibers 0.036% were added individually to the concrete mixture as mono fibers and
then they were added together to form a hybrid fiber reinforced concrete. Mechanical properties such as compressive, split tensile and
flexural strength were determined. The results show that hybrid fibers improve the compressive strength marginally as compared to
mono fibers. Whereas, hybridization improves split tensile strength and flexural strength noticeably.
Keywords:-Hybridization, mono fibers, steel fiber, polypropylene fiber, Improvement in mechanical properties.
Material management in construction – a case studyeSAT Journals
Abstract
The objective of the present study is to understand about all the problems occurring in the company because of improper application
of material management. In construction project operation, often there is a project cost variance in terms of the material, equipments,
manpower, subcontractor, overhead cost, and general condition. Material is the main component in construction projects. Therefore,
if the material management is not properly managed it will create a project cost variance. Project cost can be controlled by taking
corrective actions towards the cost variance. Therefore a methodology is used to diagnose and evaluate the procurement process
involved in material management and launch a continuous improvement was developed and applied. A thorough study was carried
out along with study of cases, surveys and interviews to professionals involved in this area. As a result, a methodology for diagnosis
and improvement was proposed and tested in selected projects. The results obtained show that the main problem of procurement is
related to schedule delays and lack of specified quality for the project. To prevent this situation it is often necessary to dedicate
important resources like money, personnel, time, etc. To monitor and control the process. A great potential for improvement was
detected if state of the art technologies such as, electronic mail, electronic data interchange (EDI), and analysis were applied to the
procurement process. These helped to eliminate the root causes for many types of problems that were detected.
Managing drought short term strategies in semi arid regions a case studyeSAT Journals
Abstract
Drought management needs multidisciplinary action. Interdisciplinary efforts among the experts in various fields of the droughts
prone areas are helpful to achieve tangible and permanent solution for this recurring problem. The Gulbarga district having the total
area around 16, 240 sq.km, and accounts 8.45 per cent of the Karnataka state area. The district has been situated with latitude 17º 19'
60" North and longitude of 76 º 49' 60" east. The district is situated entirely on the Deccan plateau positioned at a height of 300 to
750 m above MSL. Sub-tropical, semi-arid type is one among the drought prone districts of Karnataka State. The drought
management is very important for a district like Gulbarga. In this paper various short term strategies are discussed to mitigate the
drought condition in the district.
Keywords: Drought, South-West monsoon, Semi-Arid, Rainfall, Strategies etc.
Life cycle cost analysis of overlay for an urban road in bangaloreeSAT Journals
Abstract
Pavements are subjected to severe condition of stresses and weathering effects from the day they are constructed and opened to traffic
mainly due to its fatigue behavior and environmental effects. Therefore, pavement rehabilitation is one of the most important
components of entire road systems. This paper highlights the design of concrete pavement with added mono fibers like polypropylene,
steel and hybrid fibres for a widened portion of existing concrete pavement and various overlay alternatives for an existing
bituminous pavement in an urban road in Bangalore. Along with this, Life cycle cost analyses at these sections are done by Net
Present Value (NPV) method to identify the most feasible option. The results show that though the initial cost of construction of
concrete overlay is high, over a period of time it prove to be better than the bituminous overlay considering the whole life cycle cost.
The economic analysis also indicates that, out of the three fibre options, hybrid reinforced concrete would be economical without
compromising the performance of the pavement.
Keywords: - Fatigue, Life cycle cost analysis, Net Present Value method, Overlay, Rehabilitation
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materialseSAT Journals
Abstract
The issue of growing demand on our nation’s roadways over that past couple of decades, decreasing budgetary funds, and the need to
provide a safe, efficient, and cost effective roadway system has led to a dramatic increase in the need to rehabilitate our existing
pavements and the issue of building sustainable road infrastructure in India. With these emergency of the mentioned needs and this
are today’s burning issue and has become the purpose of the study.
In the present study, the samples of existing bituminous layer materials were collected from NH-48(Devahalli to Hassan) site.The
mixtures were designed by Marshall Method as per Asphalt institute (MS-II) at 20% and 30% Reclaimed Asphalt Pavement (RAP).
RAP material was blended with virgin aggregate such that all specimens tested for the, Dense Bituminous Macadam-II (DBM-II)
gradation as per Ministry of Roads, Transport, and Highways (MoRT&H) and cost analysis were carried out to know the economics.
Laboratory results and analysis showed the use of recycled materials showed significant variability in Marshall Stability, and the
variability increased with the increase in RAP content. The saving can be realized from utilization of recycled materials as per the
methodology, the reduction in the total cost is 19%, 30%, comparing with the virgin mixes.
Keywords: Reclaimed Asphalt Pavement, Marshall Stability, MS-II, Dense Bituminous Macadam-II
Laboratory investigation of expansive soil stabilized with natural inorganic ...eSAT Journals
Abstract
Soil stabilization has proven to be one of the oldest techniques to improve the soil properties. Literature review conducted revealed
that uses of natural inorganic stabilizers are found to be one of the best options for soil stabilization. In this regard an attempt has
been made to evaluate the influence of RBI-81 stabilizer on properties of black cotton soil through laboratory investigations. Black
cotton soil with varying percentages of RBI-81 viz., 0, 0.5, 1, 1.5, 2, and 2.5 percent were studied for moisture density relationships
and strength behaviour of soils. Also the effect of curing period was evaluated as literature review clearly emphasized the strength
gain of soils stabilized with RBI-81 over a period of time. The results obtained shows that the unconfined compressive strength of
specimens treated with RBI-81 increased approximately by 250% for a curing period of 28 days as compared to virgin soil. Further
the CBR value improved approximately by 400%. The studies indicated an increasing trend for soil strength behaviour with
increasing percentage of RBI-81 suggesting its potential applications in soil stabilization.
Influence of reinforcement on the behavior of hollow concrete block masonry p...eSAT Journals
Abstract
Reinforced masonry was developed to exploit the strength potential of masonry and to solve its lack of tensile strength. Experimental
and analytical studies have been carried out to investigate the effect of reinforcement on the behavior of hollow concrete block
masonry prisms under compression and to predict ultimate failure compressive strength. In the numerical program, three dimensional
non-linear finite elements (FE) model based on the micro-modeling approach is developed for both unreinforced and reinforced
masonry prisms using ANSYS (14.5). The proposed FE model uses multi-linear stress-strain relationships to model the non-linear
behavior of hollow concrete block, mortar, and grout. Willam-Warnke’s five parameter failure theory has been adopted to model the
failure of masonry materials. The comparison of the numerical and experimental results indicates that the FE models can successfully
capture the highly nonlinear behavior of the physical specimens and accurately predict their strength and failure mechanisms.
Keywords: Structural masonry, Hollow concrete block prism, grout, Compression failure, Finite element method,
Numerical modeling.
Influence of compaction energy on soil stabilized with chemical stabilizereSAT Journals
Abstract
Increase in traffic along with heavier magnitude of wheel loads cause rapid deterioration in pavements. There is a need to improve
density, strength of soil subgrade and other pavement layers. In this study an attempt is made to improve the properties of locally
available loamy soil using twin approaches viz., i) increasing the compaction of soil and ii) treating the soil with chemical stabilizer.
Laboratory studies are carried out on both untreated and treated soil samples compacted by different compaction efforts. Studies
show that increase in compaction effort results in increase in density of soil. However in soil treated with chemical stabilizer, rate of
increase in density is not significant. The soil treated with chemical stabilizer exhibits improvement in both strength and performance
properties.
Keywords: compaction, density, subgradestabilization, resilient modulus
Geographical information system (gis) for water resources managementeSAT Journals
Abstract
Water resources projects are inherited with overlapping and at times conflicting objectives. These projects are often of varied sizes
ranging from major projects with command areas of millions of hectares to very small projects implemented at the local level. Thus,
in all these projects there is seldom proper coordination which is essential for ensuring collective sustainability.
Integrated watershed development and management is the accepted answer but in turn requires a comprehensive framework that can
enable planning process involving all the stakeholders at different levels and scales is compulsory. Such a unified hydrological
framework is essential to evaluate the cause and effect of all the proposed actions within the drainage basins.
The present paper describes a hydrological framework developed in the form of a Hydrologic Information System (HIS) which is
intended to meet the specific information needs of the various line departments of a typical State connected with water related aspects.
The HIS consist of a hydrologic information database coupled with tools for collating primary and secondary data and tools for
analyzing and visualizing the data and information. The HIS also incorporates hydrological model base for indirect assessment of
various entities of water balance in space and time. The framework would be maintained and updated to reflect fully the most
accurate ground truth data and the infrastructure requirements for planning and management.
Keywords: Hydrological Information System (HIS); WebGIS; Data Model; Web Mapping Services
Forest type mapping of bidar forest division, karnataka using geoinformatics ...eSAT Journals
Abstract
The study demonstrate the potentiality of satellite remote sensing technique for the generation of baseline information on forest types
including tree plantation details in Bidar forest division, Karnataka covering an area of 5814.60Sq.Kms. The Total Area of Bidar
forest division is 5814Sq.Kms analysis of the satellite data in the study area reveals that about 84% of the total area is Covered by
crop land, 1.778% of the area is covered by dry deciduous forest, 1.38 % of mixed plantation, which is very threatening to the
environmental stability of the forest, future plantation site has been mapped. With the use of latest Geo-informatics technology proper
and exact condition of the trees can be observed and necessary precautions can be taken for future plantation works in an appropriate
manner
Keywords:-RS, GIS, GPS, Forest Type, Tree Plantation
Factors influencing compressive strength of geopolymer concreteeSAT Journals
Abstract
To study effects of several factors on the properties of fly ash based geopolymer concrete on the compressive strength and also the
cost comparison with the normal concrete. The test variables were molarities of sodium hydroxide(NaOH) 8M,14M and 16M, ratio of
NaOH to sodium silicate (Na2SiO3) 1, 1.5, 2 and 2.5, alkaline liquid to fly ash ratio 0.35 and 0.40 and replacement of water in
Na2SiO3 solution by 10%, 20% and 30% were used in the present study. The test results indicated that the highest compressive
strength 54 MPa was observed for 16M of NaOH, ratio of NaOH to Na2SiO3 2.5 and alkaline liquid to fly ash ratio of 0.35. Lowest
compressive strength of 27 MPa was observed for 8M of NaOH, ratio of NaOH to Na2SiO3 is 1 and alkaline liquid to fly ash ratio of
0.40. Alkaline liquid to fly ash ratio of 0.35, water replacement of 10% and 30% for 8 and 16 molarity of NaOH and has resulted in
compressive strength of 36 MPa and 20 MPa respectively. Superplasticiser dosage of 2 % by weight of fly ash has given higher
strength in all cases.
Keywords: compressive strength, alkaline liquid, fly ash
Experimental investigation on circular hollow steel columns in filled with li...eSAT Journals
Abstract
Composite Circular hollow Steel tubes with and without GFRP infill for three different grades of Light weight concrete are tested for
ultimate load capacity and axial shortening , under Cyclic loading. Steel tubes are compared for different lengths, cross sections and
thickness. Specimens were tested separately after adopting Taguchi’s L9 (Latin Squares) Orthogonal array in order to save the initial
experimental cost on number of specimens and experimental duration. Analysis was carried out using ANN (Artificial Neural
Network) technique with the assistance of Mini Tab- a statistical soft tool. Comparison for predicted, experimental & ANN output is
obtained from linear regression plots. From this research study, it can be concluded that *Cross sectional area of steel tube has most
significant effect on ultimate load carrying capacity, *as length of steel tube increased- load carrying capacity decreased & *ANN
modeling predicted acceptable results. Thus ANN tool can be utilized for predicting ultimate load carrying capacity for composite
columns.
Keywords: Light weight concrete, GFRP, Artificial Neural Network, Linear Regression, Back propagation, orthogonal
Array, Latin Squares
Experimental behavior of circular hsscfrc filled steel tubular columns under ...eSAT Journals
Abstract
This paper presents an outlook on experimental behavior and a comparison with predicted formula on the behaviour of circular
concentrically loaded self-consolidating fibre reinforced concrete filled steel tube columns (HSSCFRC). Forty-five specimens were
tested. The main parameters varied in the tests are: (1) percentage of fiber (2) tube diameter or width to wall thickness ratio (D/t
from 15 to 25) (3) L/d ratio from 2.97 to 7.04 the results from these predictions were compared with the experimental data. The
experimental results) were also validated in this study.
Keywords: Self-compacting concrete; Concrete-filled steel tube; axial load behavior; Ultimate capacity.
Evaluation of punching shear in flat slabseSAT Journals
Abstract
Flat-slab construction has been widely used in construction today because of many advantages that it offers. The basic philosophy in
the design of flat slab is to consider only gravity forces; this method ignores the effect of punching shear due to unbalanced moments
at the slab column junction which is critical. An attempt has been made to generate generalized design sheets which accounts both
punching shear due to gravity loads and unbalanced moments for cases (a) interior column; (b) edge column (bending perpendicular
to shorter edge); (c) edge column (bending parallel to shorter edge); (d) corner column. These design sheets are prepared as per
codal provisions of IS 456-2000. These design sheets will be helpful in calculating the shear reinforcement to be provided at the
critical section which is ignored in many design offices. Apart from its usefulness in evaluating punching shear and the necessary
shear reinforcement, the design sheets developed will enable the designer to fix the depth of flat slab during the initial phase of the
design.
Keywords: Flat slabs, punching shear, unbalanced moment.
Evaluation of performance of intake tower dam for recent earthquake in indiaeSAT Journals
Abstract
Intake towers are typically tall, hollow, reinforced concrete structures and form entrance to reservoir outlet works. A parametric
study on dynamic behavior of circular cylindrical towers can be carried out to study the effect of depth of submergence, wall thickness
and slenderness ratio, and also effect on tower considering dynamic analysis for time history function of different soil condition and
by Goyal and Chopra accounting interaction effects of added hydrodynamic mass of surrounding and inside water in intake tower of
dam
Key words: Hydrodynamic mass, Depth of submergence, Reservoir, Time history analysis,
Evaluation of operational efficiency of urban road network using travel time ...eSAT Journals
Abstract
Efficiency of the road network system is analyzed by travel time reliability measures. The study overlooks on an important measure of
travel time reliability and prioritizing Tiruchirappalli road network. Traffic volume and travel time were collected using license plate
matching method. Travel time measures were estimated from average travel time and 95th travel time. Effect of non-motorized vehicle
on efficiency of road system was evaluated. Relation between buffer time index and traffic volume was created. Travel time model has
been developed and travel time measure was validated. Then service quality of road sections in network were graded based on
travel time reliability measures.
Keywords: Buffer Time Index (BTI); Average Travel Time (ATT); Travel Time Reliability (TTR); Buffer Time (BT).
Estimation of surface runoff in nallur amanikere watershed using scs cn methodeSAT Journals
Abstract
The development of watershed aims at productive utilization of all the available natural resources in the entire area extending from
ridge line to stream outlet. The per capita availability of land for cultivation has been decreasing over the years. Therefore, water and
the related land resources must be developed, utilized and managed in an integrated and comprehensive manner. Remote sensing and
GIS techniques are being increasingly used for planning, management and development of natural resources. The study area, Nallur
Amanikere watershed geographically lies between 110 38’ and 110 52’ N latitude and 760 30’ and 760 50’ E longitude with an area of
415.68 Sq. km. The thematic layers such as land use/land cover and soil maps were derived from remotely sensed data and overlayed
through ArcGIS software to assign the curve number on polygon wise. The daily rainfall data of six rain gauge stations in and around
the watershed (2001-2011) was used to estimate the daily runoff from the watershed using Soil Conservation Service - Curve Number
(SCS-CN) method. The runoff estimated from the SCS-CN model was then used to know the variation of runoff potential with different
land use/land cover and with different soil conditions.
Keywords: Watershed, Nallur watershed, Surface runoff, Rainfall-Runoff, SCS-CN, Remote Sensing, GIS.
Estimation of morphometric parameters and runoff using rs & gis techniqueseSAT Journals
Abstract
Land and water are the two vital natural resources, the optimal management of these resources with minimum adverse environmental
impact are essential not only for sustainable development but also for human survival. Satellite remote sensing with geographic
information system has a pragmatic approach to map and generate spatial input layers of predicting response behavior and yield of
watershed. Hence, in the present study an attempt has been made to understand the hydrological process of the catchment at the
watershed level by drawing the inferences from moprhometric analysis and runoff. The study area chosen for the present study is
Yagachi catchment situated in Chickamaglur and Hassan district lies geographically at a longitude 75⁰52’08.77”E and
13⁰10’50.77”N latitude. It covers an area of 559.493 Sq.km. Morphometric analysis is carried out to estimate morphometric
parameters at Micro-watershed to understand the hydrological response of the catchment at the Micro-watershed level. Daily runoff
is estimated using USDA SCS curve number model for a period of 10 years from 2001 to 2010. The rainfall runoff relationship of the
study shows there is a positive correlation.
Keywords: morphometric analysis, runoff, remote sensing and GIS, SCS - method
-
Effect of variation of plastic hinge length on the results of non linear anal...eSAT Journals
Abstract The nonlinear Static procedure also well known as pushover analysis is method where in monotonically increasing loads are applied to the structure till the structure is unable to resist any further load. It is a popular tool for seismic performance evaluation of existing and new structures. In literature lot of research has been carried out on conventional pushover analysis and after knowing deficiency efforts have been made to improve it. But actual test results to verify the analytically obtained pushover results are rarely available. It has been found that some amount of variation is always expected to exist in seismic demand prediction of pushover analysis. Initial study is carried out by considering user defined hinge properties and default hinge length. Attempt is being made to assess the variation of pushover analysis results by considering user defined hinge properties and various hinge length formulations available in literature and results compared with experimentally obtained results based on test carried out on a G+2 storied RCC framed structure. For the present study two geometric models viz bare frame and rigid frame model is considered and it is found that the results of pushover analysis are very sensitive to geometric model and hinge length adopted. Keywords: Pushover analysis, Base shear, Displacement, hinge length, moment curvature analysis
Effect of use of recycled materials on indirect tensile strength of asphalt c...eSAT Journals
Abstract
Depletion of natural resources and aggregate quarries for the road construction is a serious problem to procure materials. Hence
recycling or reuse of material is beneficial. On emphasizing development in sustainable construction in the present era, recycling of
asphalt pavements is one of the effective and proven rehabilitation processes. For the laboratory investigations reclaimed asphalt
pavement (RAP) from NH-4 and crumb rubber modified binder (CRMB-55) was used. Foundry waste was used as a replacement to
conventional filler. Laboratory tests were conducted on asphalt concrete mixes with 30, 40, 50, and 60 percent replacement with RAP.
These test results were compared with conventional mixes and asphalt concrete mixes with complete binder extracted RAP
aggregates. Mix design was carried out by Marshall Method. The Marshall Tests indicated highest stability values for asphalt
concrete (AC) mixes with 60% RAP. The optimum binder content (OBC) decreased with increased in RAP in AC mixes. The Indirect
Tensile Strength (ITS) for AC mixes with RAP also was found to be higher when compared to conventional AC mixes at 300C.
Keywords: Reclaimed asphalt pavement, Foundry waste, Recycling, Marshall Stability, Indirect tensile strength.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
Overview of the fundamental roles in Hydropower generation and the components involved in wider Electrical Engineering.
This paper presents the design and construction of hydroelectric dams from the hydrologist’s survey of the valley before construction, all aspects and involved disciplines, fluid dynamics, structural engineering, generation and mains frequency regulation to the very transmission of power through the network in the United Kingdom.
Author: Robbie Edward Sayers
Collaborators and co editors: Charlie Sims and Connor Healey.
(C) 2024 Robbie E. Sayers
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptxR&R Consult
CFD analysis is incredibly effective at solving mysteries and improving the performance of complex systems!
Here's a great example: At a large natural gas-fired power plant, where they use waste heat to generate steam and energy, they were puzzled that their boiler wasn't producing as much steam as expected.
R&R and Tetra Engineering Group Inc. were asked to solve the issue with reduced steam production.
An inspection had shown that a significant amount of hot flue gas was bypassing the boiler tubes, where the heat was supposed to be transferred.
R&R Consult conducted a CFD analysis, which revealed that 6.3% of the flue gas was bypassing the boiler tubes without transferring heat. The analysis also showed that the flue gas was instead being directed along the sides of the boiler and between the modules that were supposed to capture the heat. This was the cause of the reduced performance.
Based on our results, Tetra Engineering installed covering plates to reduce the bypass flow. This improved the boiler's performance and increased electricity production.
It is always satisfying when we can help solve complex challenges like this. Do your systems also need a check-up or optimization? Give us a call!
Work done in cooperation with James Malloy and David Moelling from Tetra Engineering.
More examples of our work https://www.r-r-consult.dk/en/cases-en/
Quality defects in TMT Bars, Possible causes and Potential Solutions.PrashantGoswami42
Maintaining high-quality standards in the production of TMT bars is crucial for ensuring structural integrity in construction. Addressing common defects through careful monitoring, standardized processes, and advanced technology can significantly improve the quality of TMT bars. Continuous training and adherence to quality control measures will also play a pivotal role in minimizing these defects.
Automobile Management System Project Report.pdfKamal Acharya
The proposed project is developed to manage the automobile in the automobile dealer company. The main module in this project is login, automobile management, customer management, sales, complaints and reports. The first module is the login. The automobile showroom owner should login to the project for usage. The username and password are verified and if it is correct, next form opens. If the username and password are not correct, it shows the error message.
When a customer search for a automobile, if the automobile is available, they will be taken to a page that shows the details of the automobile including automobile name, automobile ID, quantity, price etc. “Automobile Management System” is useful for maintaining automobiles, customers effectively and hence helps for establishing good relation between customer and automobile organization. It contains various customized modules for effectively maintaining automobiles and stock information accurately and safely.
When the automobile is sold to the customer, stock will be reduced automatically. When a new purchase is made, stock will be increased automatically. While selecting automobiles for sale, the proposed software will automatically check for total number of available stock of that particular item, if the total stock of that particular item is less than 5, software will notify the user to purchase the particular item.
Also when the user tries to sale items which are not in stock, the system will prompt the user that the stock is not enough. Customers of this system can search for a automobile; can purchase a automobile easily by selecting fast. On the other hand the stock of automobiles can be maintained perfectly by the automobile shop manager overcoming the drawbacks of existing system.
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Enhanced equally distributed load balancing algorithm for cloud computing
1. IJRET: International Journal of Research in Engineering and Technology ISSN: 2319-1163
__________________________________________________________________________________________
Volume: 02 Issue: 06 | Jun-2013, Available @ http://www.ijret.org 976
ENHANCED EQUALLY DISTRIBUTED LOAD BALANCING
ALGORITHM FOR CLOUD COMPUTING
Shreyas Mulay1
, Sanjay Jain2
1
Student, 2
Assistant Professor Computer Science and Engineering Department,
Amity School of Engineering and Technology (ASET), Amity University Rajasthan, AUR, Rajasthan, India,
shreyasmulay23@gmail.com, jainsanjay17@yahoo.co.in
Abstract
Cloud Computing as the name suggests, it is a style of computing where different users uses the resources on the go i.e. over the
Internet. In the recent era, this technology has emerged as a strong option for not only large scale organizations but also for small
scale organizations that only access/use the resources what they want. In recent research study, many organizations lose significant
part of their revenues in handling the requests given by the clients over the web servers i.e. unable to balance the load for web servers
which results in loss of data, delay in time and increased costs. This Paper gives a new enhanced load balancing algorithm by which
the performance of their web application can be increased. This Algorithm works on the major drawbacks such as delay in time,
response to request ratio etc.
Index Terms: Cloud Computing, Public Cloud, Load Balancing, Load Balancing Algorithms, Enhanced Equally
Distributed Load Balancing Algorithm.
-----------------------------------------------------------------------***-----------------------------------------------------------------------
1. INTRODUCTION
Cloud Computing uses distributed technologies to satisfy the
user needs. Cloud Services includes the sharing of resources,
delivery of software, infrastructure and storage over the
internet based on their demands. The main functions of cloud
computing are reduced cost, better performance and satisfy the
needs of user at a great extent. Now taking all these functions
into consideration web servers are designed which can give
the best performance but many times the performance drops
drastically why? The answer for this question is the balancing
of the load on the servers appropriately by some mechanism
which will improve the performance of total system. The
Simple logic behind the balancing the load over the servers
(nodes) is that distribute total load in a systematic manner i.e.
balancing the load on the overloaded node to under loaded
node so that the response time from the server will decrease
and performance of the servers increased. These algorithms
does not take the previous state or behavior into consideration,
it depends on the current state of the system because of its
dynamic behavior. Some Algorithms works on circular order
by handling the process without any priority but enhanced
equally distributed load balancing algorithm handles the
requests on priority.
2. NEED OF LOAD BALANCING IN CLOUD
COMPUTING
Load balancing in clouds is a mechanism that spreads the
excess dynamic workload evenly across all the Servers
(Virtual Machines). It is used to achieve a high user
satisfaction and resource utilization ratio, making sure that no
single node is overloaded or under loaded, hence improving
the overall performance of the system. Proper load balancing
can help in utilizing the availability of given resources,
thereby minimizing the resource consumption. It also helps in
implementing fail-over (fault tolerance), enabling scalability,
reducing response time, time delay, reduces cost etc.
3. EXISTING LOAD BALANCING ALGORITHMS
/ TECHNIQUES IN CLOUD COMPUTING
There are many load balancing techniques given by the
researchers over time to time some have advantages over other
and vice versa. Load Balancing is required to achieve the
maximum throughput, performance and decrease the response
time. Here in this paper we will discuss the 5 main load
balancing algorithms/techniques given by the researchers.
Following are the Load Balancing Techniques which are in
use:-
3.1 Round Robin Load Balancing Algorithm:
Round Robin algorithm is random sampling based algorithm.
It means that it selects the request one by one and randomly
place them to servers (nodes) irrespective of whether it is
heavily loaded or lightly.
2. IJRET: International Journal of Research in Engineering and Technology ISSN: 2319-1163
__________________________________________________________________________________________
Volume: 02 Issue: 06 | Jun-2013, Available @ http://www.ijret.org 977
3.2 Server-based Load Balancing for Internet
Distributed Services:
M. Nakai et al. [1] proposed a new server based load
balancing policy for web servers which spread all over the
world. It helps in reducing the service response times by using
some set of rules that limits the redirection of requests to the
closest remote servers without overloading them. A
middleware is described to implement this protocol. It also
uses a heuristic to help web servers to endure overloads.
3.3 Scheduling Strategy on Load Balancing of Virtual
Machine Resources:
J. Hu et al. [2] proposed a scheduling strategy on load
balancing of VM resources that uses historical data and
current state of the system. This strategy achieves load
balancing and reduced dynamic migration by using a genetic
algorithm. It helps in resolving the issue of load-imbalance
and high cost of migration thus achieving better resource
utilization.
3.4 Central Load Balancing Policy for Virtual
Machines:
A.Bhadani et al. [3] proposed a Central Load Balancing Policy
for Virtual Machines (CLBVM) that balances the load evenly
in a distributed virtual machine/cloud computing environment.
This Load Balancing technique improves the overall
performance of the system but does not consider the systems
that are fault-tolerant.
3.5 A Task Scheduling Algorithm Based on Load
Balancing:
Y. Fang et al. [4] discussed a two-level task scheduling
mechanism based on load balancing to meet dynamic
requirements of users and obtain high resource utilization. It
achieves load balancing by first mapping tasks to virtual
machines and then virtual machines to host resources thereby
improving the resource utilization and overall performance of
the cloud computing environment but it does not improves the
response to request ratio.
4. PROPOSED WORK
4.1 Enhanced Equally Distributed Load Balancing
Algorithm
Enhanced equally distributed load balancing algorithm
handles the requests with priorities. It is a distributed
algorithm by which the load can be distributed not only in a
balanced manner but also it allocates the load systematically
by checking the counter variable of each data center. After
checking, it transfers the load accordingly i.e. the minimum
value of the counter variable will be chosen and the request is
handled easily and takes less time, and gives maximum
throughput. It is a distributed technique in which the load
balancer allocates the load of the job in hand into multiple web
servers. The randomly transfer of load can cause some server
to heavily loaded while other server is lightly loaded. If the
load is equally distributed it not only improves performance
also reduces the time delay. So the analysis on the load
distribution algorithms the efficient scheduling and resource
allocation is a critical task in case of cloud computing and also
to improve the response time and processing time. While
considering the impact of cost optimization one has to think
about the solution to this problem. This algorithm not only
balances the load also it increases the response time for the
cloud.
Fig -1: Enhanced Equally Distributed Load Balancing
Algorithm
3. IJRET: International Journal of Research in Engineering and Technology ISSN: 2319-1163
__________________________________________________________________________________________
Volume: 02 Issue: 06 | Jun-2013, Available @ http://www.ijret.org 978
4.2 The Steps for efficient Distributed Load
balancing through this Algorithm
This algorithm simply allocates request which is coming from
the client nodes to the lightly loaded server cluster (Data
Center) and gives the response in less amount of time by doing
this, it makes the algorithm efficient for response to request
ratio. We can see that the clients at a same time make requests
to access the cloud application over the internet. Now in this
algorithm all the request goes through the load balancer
system by which the Node A checks the counter variable
which is set to the maximum requests handled by a server
cluster e.g. 300 requests per server cluster i.e. each server in
the cluster can handle up to 100 request simultaneously. Let us
assume that the cluster 1 is having a counter value to 250,
cluster 2 is having the counter variable to 220 and cluster 3 to
270 i.e. cluster 2 is handling the smaller number of requests
compared to cluster 1 and cluster 3, so here the load balancer
will balances the load (requests) to cluster 2 as it is less hence
the balancing is done at this level. Till now we have handled
the request but how the counter variable will gets updated?
The answer is the servers which the counter variable is
associated with, will simultaneously changes (updates) the
counter variable i.e. when a response is given back to the
client the server automatically decreases its counter variable
by the number 1,so that every time the algorithm will have the
updated value of counter variable. Hence with the help of this
algorithm the requests are handled easily by Server Clusters.
The Strength of server can be increased or decreased by the
service provider on request and for data centers too. So no
need of Round Robin Balancing or any other technique where
time is consumed and response to request ratio is low for vast
number of requests. Clearly we can see over here if we assume
our example that this load balancer can balances 900 requests
at a time without any delay in time and responses can be given
back to the clients.
5. PERFORMANCE ANALYSIS
When it comes to performance analysis of the cloud system,
there are some metrics on which the analysis can be done.
Some of the important metrics like Performance, Resource
Utilization, Scalability, Fault Tolerance and Response Time.
For a better system these are some important attributes which
has to be maintained for a perfect system.
5.1 Performance:
It is used to check how efficient the system and how the
performance remains steady under lot of pressure. In this
algorithm the performance of whole system has increased. As
we have said earlier that this system can perform many
requests simultaneously i.e. by taking the help of above
example 900 requests in a time so the performance of this
algorithm not only stays steady but also efficient to under
pressure because of the dynamic behavior of the Algorithm.
5.2 Resource Utilization:
It is used to check the utilization of the available resources
given to the cloud. In this metric the algorithm works smooth
i.e. on utilization of resources this algorithm uses the resources
wisely and performs the task efficiently.
5.3 Scalability:
It is used to check whether the system is functional after it is
scaled to any amount. Every Algorithm must comply with this
metric i.e. if there is a need to expand the system to an extent
the algorithm has to adjust itself to continue giving its
services. This algorithm is made for such scalable
requirements, because here the counter variable plays the key
role when there is an increase in the number of servers or data
centers this algorithm dynamically adjusts itself to that
scenario and performance is maintained.
5.4 Fault Tolerance:
It is type of metric in which it is measured that how the system
will perform under the node (server) failure. In this algorithm
the counter variable is feeding back the requests handled by
the data centers. Now, if any node fails or any node is having a
fault this algorithm automatically updates the counter variable
and stops itself from updating further, by doing this the
algorithm itself judges that some fault occurred in that node or
datacenter so the algorithm maintains its flexibility.
5.5 Response Time:
It is an amount of time taken by the system to give the
response of the request given by the client. The main prospect
why this algorithm is proposed is this metric. In Cloud
Computing the client need the data to be accessed as fast as
like he is accessing it in his home computer so response time
in this algorithm is dramatically reduced from the rest of the
algorithms. This can be seen by the above given examples that
at a time the algorithm is giving the 900 responses back to the
client which is a great figure. So the Response time for this
algorithm is very good.
6. RESULTS
The results for this algorithm have been observed on the basis
of the above explained scenario. We have used 3 different data
centers in our cloud, and the operation is performed. For this
we have applied above explained 5 algorithms and response
time has been noted down. Now here on Y-axis we have
Requests on servers and on X axis we have 6 different
algorithms and their respective performances are shown in
Bars. We can see here that each load balancing algorithm
though produces good response to request ratio but Enhanced
Equally Load Balancing Algorithm (EEDLBA) has the highest
number responses to the requests compared to others.
4. IJRET: International Journal of Research in Engineering and Technology ISSN: 2319-1163
__________________________________________________________________________________________
Volume: 02 Issue: 06 | Jun-2013, Available @ http://www.ijret.org 979
Chart -1: Response to Request Ratio Chart
Similarly, after observing the results of response time, rest
four matrices are also observed and we discovered that the
equally enhanced distributed load balancing algorithm is more
superior to the other load balancing algorithms though there
are some issues also e.g. Energy Management, Carbon
Emission etc. but still the said algorithm works fine. The
Performance, Resource Utilization, Scalability, Fault
Tolerance and the response time is much better than the
others.
CONCLUSIONS
Cloud Computing System has widely been adopted by the
industry, though there are many existing issues like Load
Balancing, Migration of Virtual Machines, Server unification,
etc. which have not been yet fully addressed. On the Contrary
the Load Balancing is the most central issue in the System i.e.
to distribute the load in an efficient manner. It also ensures
that every computing resource is distributed efficiently and
fairly. Existing Load Balancing techniques/Algorithms that
have been studied mainly focus on reducing overhead,
reducing the migration time and improving performance etc.,
but none of them have considered the response to request
ratio. The response time is a challenge of every engineer to
develop the products that can increase the throughput in the
cloud based sector. The several strategies lack efficient
scheduling and load balancing resource allocation techniques
leading to increased operational cost. This proposed algorithm
not only rectifies the said issues but also reduces the request to
response ratio.
REFERENCES:
[1]. A. M. Nakai, E. Madeira, and L. E. Buzato, “Load
Balancing for Internet Distributed Services Using Limited
Redirection Rates”, 5th IEEE Latin-American Symposium on
Dependable Computing (LADC), 2011, pages 156-165.
[2]. J. Hu, J. Gu, G. Sun, and T. Zhao, “A Scheduling Strategy
on Load Balancing of Virtual Machine Resources in Cloud
Computing Environment”, Third International Symposium on
Parallel Architectures, Algorithms and Programming (PAAP),
2010, pages 89-96.
[3]. A. Bhadani, and S. Chaudhary, “Performance evaluation
of web servers using central load balancing policy over virtual
machines on cloud”, Proceedings of the Third Annual ACM
Bangalore Conference (COMPUTE), January 2010.
[4]. Y. Fang, F. Wang, and J. Ge, “A Task Scheduling
Algorithm Based on Load Balancing in Cloud Computing”,
Web Information Systems and Mining, Lecture Notes in
Computer Science, Vol. 6318, 2010, pages 271-277.
[5]. Nidhi Jain Kansal and Inderveer Chana,” Cloud Load
Balancing Techniques: A Step towards Green”, IJCSI
International Journal of Computer Science Issues, Vol. 9,
Issue 1, No 1, January 2012,pages 238-246.
[6]. Nikita, Shaveta and Guarav Raj, “Comparative Analysis
of Load Balancing Algorithms in Cloud Computing”,
International Journal of Advanced Research in Computer
Engineering & Technology Volume 1, Issue 3, May 2012,
pages 120-124.
[7]. Amazon Elastic Compute Cloud (EC2),
http://www.amazon.com/gp/browse.html?node=201590011
[8]. Load Balancing (Computing) Wikipedia,
http://en.wikipedia.org/wiki/Load_balancing_ (computing)
[9]. Load Balancing and Cloud Computing (SCVMM 2012),
http://social.technet.microsoft.com/wiki/contents/articles/3937
.load-balancing-and-cloud-computing-scvmm-2012.aspx
[10].Cloud Computing Patterns, Elastic Load Balancer,
http://cloudcomputingpatterns.org/?page_id=269
[11]. Y. Zhao, and W. Huang, “Adaptive Distributed Load
Balancing Algorithm based on Live Migration of Virtual
Machines in Cloud”, Proceedings of 5th IEEE International
Joint Conference on INC, IMS and IDC, Seoul, Republic of
Korea, August 2009, pages 170-175.
[12]. M. Randles, D. Lamb, and A. Taleb-Bendiab, “A
Comparative Study into Distributed Load Balancing
Algorithms for Cloud Computing”, Proceedings of 24th IEEE
International Conference on Advanced Information
Networking and Applications Workshops, Perth, Australia,
April 2010, pages 551-556.
[13]. Cloud Load Balancing, The KEMP GEO LoadMaster
and the hybrid cloud a perfect cloud load balancing
combination.
http://www.kemptechnologies.com/products/solutions/kem
p-geo-loadmaster-and-hybrid-cloud-perfect-cloud-load-
balancing-combination
[14]. K. M. Nagothu, B. Kelley, J. Prevost, and M. Jamshidi,
“Ultra low energy cloud computing using adaptive load
prediction”, Proceedings of IEEE World Automation
Congress(WAC) , Kobe, September 2010, pages 1-7.
[15]. B. P. Rima, E. Choi, and I. Lumb, “A Taxonomy and
Survey of Cloud Computing Systems”, Proceedings of 5th
IEEE International Joint Conference on INC, IMS and IDC,
Seoul, Korea, August 2009, pages 44-51.
5. IJRET: International Journal of Research in Engineering and Technology ISSN: 2319-1163
__________________________________________________________________________________________
Volume: 02 Issue: 06 | Jun-2013, Available @ http://www.ijret.org 980
BIOGRAPHIES:
Shreyas Mulay received his Bachelor’s
Degree in Computer Science and
Engineering From Rajiv Gandhi Proudyogiki
Vishwavidyalaya, Bhopal, Madhya Pradesh,
India in 2011.At present, he is an M.Tech.
Candidate in Computer Science &
Engineering Department at Amity University
Rajasthan (AUR), Jaipur, Rajasthan, India His Research
Interest lies in Cloud Computing.
Sanjay Jain working as an Assistant
Professor in Amity University Rajasthan
(AUR) .He has published many National and
International Papers. His Research interests
include Cloud Computing, its Security etc
and are having 14 years of teaching and
Research experience