The new relatively concept of cloud computing & its associated methodologies has many advantages in the
world of today. Such advantages range between providing solutions for integration of the miscellaneous
systems & presenting as well guarantees for distribution of searching means & integration of software
tools which are used by consumers & different providers. In this paper, we have constructed an ontologybased
cloud framework with a view to identifying its external agent’s interoperability. The proposed
framework has been designed using the blackboard design style. This framework is composed of mainly
two components: controller and cloud ontology blackboard environment. The function of the controller is
to interact with consumers after receipt of the subject request where it spontaneously uses the ontology
base to distribute it & constitute the required related responses whereas the function of the second
framework component is to interact with different cloud providers and systems, using the meta-ontology
framework to restructure data via using AI reasoning tools and map them to its corresponding
redistributed request. Finally, E-tourism case study can be applicable will be explored.
A Novel Switch Mechanism for Load Balancing in Public CloudIJMER
In cloud computing environment, one of the core design principles is dynamic scalability,
which guarantees cloud storage service to handle the growing amounts of application data in a flexible
manner or to be readily enlarged. By integrating several private and public cloud services, the hybrid
clouds can effectively provide dynamic scalability of service and data migration. A load balancing is a
method of dividing computing loads among numerous hardware resources. Due to unpredictable job
arrival pattern and the capacities of the nodes in cloud differ for the load balancing problem. In this load
control is very crucial to improve system performance and maintenance. This paper presents a switch
mechanism for load balancing in cloud computing. The load balancing model given in this work is aimed
at the public cloud which has numerous nodes with distributed computing resources in many different
geographical areas. Thus, this model divides the public cloud environment into several cloud partitions.
When the cloud environment is very large and complex, these divisions simplify the load balancing. The
cloud environment has a main controller that chooses the suitable partitions for arriving jobs while the
balancer for each cloud partition chooses the best load balancing strategy
In this research paper, we have conducted work on modeling of local broker policy based on workload profile in Network cloud. For this we are using workload based applications. To handle workload based applications, two Scheduling Policies Random Non-overlap and Workload profile based be used. We compare these two scheduling policies based on three parameters Execution (mean) time, Response (mean) time and Waiting (mean) time. Workload based profile policy gave better results than Random-Non overlap policy in terms of time performance parameter.
Tremendous usage of internet has made huge data on the network, without compromising on the
performance of network the end-users must obtain best service. As cloud provides different services on
leasing basis many companies are migrating from their own Infrastructure to cloud,This migration should
not compromise on performance of the cloud, The performance of the cloud can be improved by having
excellent load balancing strategy such that the end user is satisfied. The paper reveals the method by which
a cloud can be partitioned and a study of different algorithm with comparative study to balance the
dynamic load. The comparative study between Ant Colony and Honey Bee algorithm gives the result which
algorithm is optimal in normal load condition also the simplest round robin algorithm is applied when the
partition are in Idle state
Cloud computing is a technological paradigm that enables the consumer to enjoy the benefits of computing
services and applications without necessarily worrying about the investment and maintenance costs. This paper focuses on
the applicability of a new fully homomorphic encryption scheme (FHE) in solving data security in cloud computing. Different types
of existing homomorphic encryption schemes, including both partial and fully homomorphic encryption schemes are reviewed. The
study was aimed at constructing a fully homomorphic encryption scheme that lessens the computational strain on the computing
assets as compared to Gentry’s contribution on partial homomorphic encryption schemes where he constructed homomorphic
encryption based on ideal lattices using both additive and multiplicative Homomorphisms. In this study both addition and
composition operations implementing a fully homomorphic encryption scheme that secures data within cloud computing is used. The
work is founded on mathematical theory that is translated into an algorithm implementable in JAVA. The work was tested by a single
computing hardware to ascertain its suitability. The newly developed FHE scheme posted better results that confirmed its suitability
for data security in cloud computing.
A Novel Switch Mechanism for Load Balancing in Public CloudIJMER
In cloud computing environment, one of the core design principles is dynamic scalability,
which guarantees cloud storage service to handle the growing amounts of application data in a flexible
manner or to be readily enlarged. By integrating several private and public cloud services, the hybrid
clouds can effectively provide dynamic scalability of service and data migration. A load balancing is a
method of dividing computing loads among numerous hardware resources. Due to unpredictable job
arrival pattern and the capacities of the nodes in cloud differ for the load balancing problem. In this load
control is very crucial to improve system performance and maintenance. This paper presents a switch
mechanism for load balancing in cloud computing. The load balancing model given in this work is aimed
at the public cloud which has numerous nodes with distributed computing resources in many different
geographical areas. Thus, this model divides the public cloud environment into several cloud partitions.
When the cloud environment is very large and complex, these divisions simplify the load balancing. The
cloud environment has a main controller that chooses the suitable partitions for arriving jobs while the
balancer for each cloud partition chooses the best load balancing strategy
In this research paper, we have conducted work on modeling of local broker policy based on workload profile in Network cloud. For this we are using workload based applications. To handle workload based applications, two Scheduling Policies Random Non-overlap and Workload profile based be used. We compare these two scheduling policies based on three parameters Execution (mean) time, Response (mean) time and Waiting (mean) time. Workload based profile policy gave better results than Random-Non overlap policy in terms of time performance parameter.
Tremendous usage of internet has made huge data on the network, without compromising on the
performance of network the end-users must obtain best service. As cloud provides different services on
leasing basis many companies are migrating from their own Infrastructure to cloud,This migration should
not compromise on performance of the cloud, The performance of the cloud can be improved by having
excellent load balancing strategy such that the end user is satisfied. The paper reveals the method by which
a cloud can be partitioned and a study of different algorithm with comparative study to balance the
dynamic load. The comparative study between Ant Colony and Honey Bee algorithm gives the result which
algorithm is optimal in normal load condition also the simplest round robin algorithm is applied when the
partition are in Idle state
Cloud computing is a technological paradigm that enables the consumer to enjoy the benefits of computing
services and applications without necessarily worrying about the investment and maintenance costs. This paper focuses on
the applicability of a new fully homomorphic encryption scheme (FHE) in solving data security in cloud computing. Different types
of existing homomorphic encryption schemes, including both partial and fully homomorphic encryption schemes are reviewed. The
study was aimed at constructing a fully homomorphic encryption scheme that lessens the computational strain on the computing
assets as compared to Gentry’s contribution on partial homomorphic encryption schemes where he constructed homomorphic
encryption based on ideal lattices using both additive and multiplicative Homomorphisms. In this study both addition and
composition operations implementing a fully homomorphic encryption scheme that secures data within cloud computing is used. The
work is founded on mathematical theory that is translated into an algorithm implementable in JAVA. The work was tested by a single
computing hardware to ascertain its suitability. The newly developed FHE scheme posted better results that confirmed its suitability
for data security in cloud computing.
MCCVA: A NEW APPROACH USING SVM AND KMEANS FOR LOAD BALANCING ON CLOUDijccsa
Nowadays, the demand of using resources, using services via the intranet system or on the Internet is rapidly growing. The respective problem coming is how to use these resources effectively in terms of time and quality. Therefore, the network QoS and its economy are people concerns, cloud computing was born in an inevitable trend. However, managing resources and scheduling tasks in virtualized data centres on the cloud are challenging tasks. Currently, there are a lot of Load Balancing algorithms applied in clouds and proposed by many authors, scholars, and experts. These existing methods are more about natural and heuristic, but the application of AI, or modern datamining technologies, in load balancing is not too popular due to the different characteristics of cloud. In this paper, we propose an algorithm to reduce the processing time (makespan) on cloud computing, helping the load balancing work more efficiency. Here, we use the SVM algorithm to classify the coming Requests, K - Mean to cluster the VMs in cloud, then the LB will allocate the requests into the VMs in the most reasonable way. In this way, request with the least processing time will be allocated to the VMs with the lowest usage. We name this new proposal as MCCVA - Makespan Classification & Clustering VM Algorithm. We have experimented and evaluated this algorithm in CloudSim, a cloud simulation environment, we obtained better results than some other wellknown algorithms. With this MCCVA, we can see the big potential of AI and datamining in Load Balancing, we can further develop LB with AI to achieve better and better results of QoS.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
AN OPEN JACKSON NETWORK MODEL FOR HETEROGENEOUS INFRASTRUCTURE AS A SERVICE O...IJCNCJournal
Cloud computing is an environment which provides services for user demand such as software, platform, infrastructure. Applications which are deployed on cloud computing have become more varied and complex to adapt to increase end-user quantity and fluctuating workload. One popular characteristic of
cloud computing is the heterogeneity of network, hosts and virtual machines (VM). There were many studies on cloud computing modeling based on queuing theory, but most studies have focused on homogeneity characteristic. In this study, we propose a cloud computing model based on open Jackson
network for multi-tier application systems which are deployed on heterogeneous VMs of IaaS cloud computing. The important metrics are analyzed in our experiments such as mean waiting time; mean request quantity, the throughput of the system. Besides that, metrics in model is used to modify number VMs
allocated for applications. Result of experiments shows that open queue network provides high efficiency.
A BAYE'S THEOREM BASED NODE SELECTION FOR LOAD BALANCING IN CLOUD ENVIRONMENThiij
Cloud computing is a popular computing model as it renders service to large number of users request on
the fly and has lead to the proliferation of large number of cloud users. This has lead to the overloaded
nodes in the cloud environment along with the problem of load imbalance among the cloud servers and
thereby impacts the performance. Hence, in this paper a heuristic Baye's theorem approach is considered
along with clustering to identify the optimal node for load balancing. Experiments using the proposed
approach are carried out on cloudsim simulator and are compared with the existing approach. Results
demonstrates that task deployment performed using this approach has improved performance in terms of
utilization and throughput when compared to the existing approaches
A Baye's Theorem Based Node Selection for Load Balancing in Cloud Environmentneirew J
Cloud computing is a popular computing model as it renders service to large number of users request on
the fly and has lead to the proliferation of large number of cloud users. This has lead to the overloaded
nodes in the cloud environment along with the problem of load imbalance among the cloud servers and
thereby impacts the performance. Hence, in this paper a heuristic Baye's theorem approach is considered
along with clustering to identify the optimal node for load balancing. Experiments using the proposed
approach are carried out on cloudsim simulator and are compared with the existing approach. Results
demonstrates that task deployment performed using this approach has improved performance in terms of
utilization and throughput when compared to the existing approaches.
The advent of Big Data has seen the emergence of new processing and storage challenges. These challenges are often solved by distributed processing. Distributed systems are inherently dynamic and unstable, so it is realistic to expect that some resources will fail during use. Load balancing and task scheduling is an important step in determining the performance of parallel applications. Hence the need to design load balancing algorithms adapted to grid computing. In this paper, we propose a dynamic and hierarchical load balancing strategy at two levels: Intrascheduler load balancing, in order to avoid the use of the large-scale communication network, and interscheduler load balancing, for a load regulation of our whole system. The strategy allows improving the average response time of CLOAK-Reduce application tasks with minimal communication. We first focus on the three performance indicators, namely response time, process latency and running time of MapReduce tasks.
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server Environmentrahulmonikasharma
Cloud computing is an incipient and quickly evolving model, with new expenses and capabilities being proclaimed frequently. The increases of user on cloud with the expansion of variety of services, with that the complete allocation of resource with the minimum latent time for Virtual machine is necessary. To allocate this virtual cloud computing resources to the cloud user is a key technical issue because user demand is dynamic in nature that required dynamic allocation of resource too. To improve the allocation there must be a correct balanced algorithmic scheduling for Resource Allocation Technique. The aim of this work is to allocate resource to scientific experiment request coming from multiple users, wherever customized Virtual machines (VM) are aloft in applicable host out there in cloud. Therefore, properly programmed scheduling cloud is extremely vital and it’s significant to develop efficient scheduling methods for appropriately allocation of VMs into physical resource. The planned formulas minimize the time interval quality so as of O (Log n) by adopting KD-Tree.
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENTIJCNCJournal
Cloud computing has an indispensable role in the modern digital scenario. The fundamental challenge of cloud systems is to accommodate user requirements which keep on varying. This dynamic cloud environment demands the necessity of complex algorithms to resolve the trouble of task allotment. The overall performance of cloud systems is rooted in the efficiency of task scheduling algorithms. The dynamic property of cloud systems makes it challenging to find an optimal solution satisfying all the evaluation metrics. The new approach is formulated on the Round Robin and the Shortest Job First algorithms. The Round Robin method reduces starvation, and the Shortest Job First decreases the average waiting time. In this work, the advantages of both algorithms are incorporated to improve the makespan of user tasks.
Public Cloud Partition Using Load Status Evaluation and Cloud Division RulesIJSRD
with growth of cloud computing load balancing is important impact on performance. Cloud computing efficiency depends on good load balancer. Many type of situation occur that time cloud partitioning is done by load balancer. Different type of situation needed different type of strategies for public cloud portioning using load balancer.in this paper we work on, partition of public cloud using two type of situation first is load status evaluation and second is cloud division rules. Load status evaluation is measure in number of cloudlets arrives at datacenter and cloud divisions rules are based on cloudlet come from which geographical location. On the basis of geographical location we partition public cloud and improve performance of load balancing in cloud computing. We implement proposed system with help of cloudsim3.0 simulator.
NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...ijccsa
Cloud collaboration is an emerging technology which enables sharing of computer files using cloud
computing. Here the cloud resources are assembled and cloud services are provided using these resources.
Cloud collaboration technologies are allowing users to share documents. Resource allocation in the cloud
is challenging because resources offer different Quality of Service (QoS) and services running on these
resources are risky for user demands. We propose a solution for resource allocation based on multi
attribute QoS Scoring considering parameters such as distance to the resource from user site, reputation of
the resource, task completion time, task completion ratio, and load at the resource. The proposed algorithm
referred to as Multi Attribute QoS scoring (MAQS) uses Neuro Fuzzy system. We have also included a
speculative manager to handle fault tolerance. In this paper it is shown that the proposed algorithm
perform better than others including power trust reputation based algorithms and harmony method which
use single attribute to compute the reputation score of each resource allocated.
Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...neirew J
Cloud collaboration is an emerging technology which enables sharing of computer files using cloud
computing. Here the cloud resources are assembled and cloud services are provided using these resources.
Cloud collaboration technologies are allowing users to share documents. Resource allocation in the cloud
is challenging because resources offer different Quality of Service (QoS) and services running on these
resources are risky for user demands. We propose a solution for resource allocation based on multi
attribute QoS Scoring considering parameters such as distance to the resource from user site, reputation of
the resource, task completion time, task completion ratio, and load at the resource. The proposed algorithm
referred to as Multi Attribute QoS scoring (MAQS) uses Neuro Fuzzy system. We have also included a
speculative manager to handle fault tolerance. In this paper it is shown that the proposed algorithm
perform better than others including power trust reputation based algorithms and harmony method which
use single attribute to compute the reputation score of each resource allocated.
CONTENT BASED VIDEO CATEGORIZATION USING RELATIONAL CLUSTERING WITH LOCAL SCA...ijcsit
This paper introduces a novel approach for efficient video categorization. It relies on two main
components. The first one is a new relational clustering technique that identifies video key frames by
learning cluster dependent Gaussian kernels. The proposed algorithm, called clustering and Local Scale
Learning algorithm (LSL) learns the underlying cluster dependent dissimilarity measure while finding
compact clusters in the given dataset. The learned measure is a Gaussian dissimilarity function defined
with respect to each cluster. We minimize one objective function to optimize the optimal partition and the
cluster dependent parameter. This optimization is done iteratively by dynamically updating the partition
and the local measure. The kernel learning task exploits the unlabeled data and reciprocally, the
categorization task takes advantages of the local learned kernel. The second component of the proposed
video categorization system consists in discovering the video categories in an unsupervised manner using
the proposed LSL. We illustrate the clustering performance of LSL on synthetic 2D datasets and on high
dimensional real data. Also, we assess the proposed video categorization system using a real video
collection and LSL algorithm.
Public Integrity Auditing for Shared Dynamic Cloud Data with Group User Revoc...1crore projects
IEEE PROJECTS 2015
1 crore projects is a leading Guide for ieee Projects and real time projects Works Provider.
It has been provided Lot of Guidance for Thousands of Students & made them more beneficial in all Technology Training.
Dot Net
DOTNET Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
Java Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
ECE IEEE Projects 2015
1. Matlab project
2. Ns2 project
3. Embedded project
4. Robotics project
Eligibility
Final Year students of
1. BSc (C.S)
2. BCA/B.E(C.S)
3. B.Tech IT
4. BE (C.S)
5. MSc (C.S)
6. MSc (IT)
7. MCA
8. MS (IT)
9. ME(ALL)
10. BE(ECE)(EEE)(E&I)
TECHNOLOGY USED AND FOR TRAINING IN
1. DOT NET
2. C sharp
3. ASP
4. VB
5. SQL SERVER
6. JAVA
7. J2EE
8. STRINGS
9. ORACLE
10. VB dotNET
11. EMBEDDED
12. MAT LAB
13. LAB VIEW
14. Multi Sim
CONTACT US
1 CRORE PROJECTS
Door No: 214/215,2nd Floor,
No. 172, Raahat Plaza, (Shopping Mall) ,Arcot Road, Vadapalani, Chennai,
Tamin Nadu, INDIA - 600 026
Email id: 1croreprojects@gmail.com
website:1croreprojects.com
Phone : +91 97518 00789 / +91 72999 51536
A Prolific Scheme for Load Balancing Relying on Task Completion Time IJECEIAES
In networks with lot of computation, load balancing gains increasing significance. To offer various resources, services and applications, the ultimate aim is to facilitate the sharing of services and resources on the network over the Internet. A key issue to be focused and addressed in networks with large amount of computation is load balancing. Load is the number of tasks„t‟ performed by a computation system. The load can be categorized as network load and CPU load. For an efficient load balancing strategy, the process of assigning the load between the nodes should enhance the resource utilization and minimize the computation time. This can be accomplished by a uniform distribution of load of to all the nodes. A Load balancing method should guarantee that, each node in a network performs almost equal amount of work pertinent to their capacity and availability of resources. Relying on task subtraction, this work has presented a pioneering algorithm termed as E-TS (Efficient-Task Subtraction). This algorithm has selected appropriate nodes for each task. The proposed algorithm has improved the utilization of computing resources and has preserved the neutrality in assigning the load to the nodes in the network.
THRESHOLD BASED VM PLACEMENT TECHNIQUE FOR LOAD BALANCED RESOURCE PROVISIONIN...IJCNCJournal
The unbalancing load issue is a multi-variation, multi-imperative issue that corrupts the execution and productivity of processing assets. Workload adjusting methods give solutions of load unbalancing circumstances for two bothersome aspects over-burdening and under-stacking. Cloud computing utilizes planning and workload balancing for a virtualized environment, resource partaking in cloud foundation. These two factors must be handled in an improved way in cloud computing to accomplish ideal resource sharing. Henceforth, there requires productive resource, asset reservation for guaranteeing load advancement in the cloud. This work aims to present an incorporated resource, asset reservation, and workload adjusting calculation for effective cloud provisioning. The strategy develops a Priority-based Resource Scheduling Model to acquire the resource, asset reservation with threshold-based load balancing for improving the proficiency in cloud framework. Extending utilization of Virtual Machines through the suitable and sensible outstanding task at hand modifying is then practiced by intensely picking a job from submitting jobs using Priority-based Resource Scheduling Model to acquire resource asset reservation. Experimental evaluations represent, the proposed scheme gives better results by reducing execution time, with minimum resource cost and improved resource utilization in dynamic resource provisioning conditions.
Challenges in Dynamic Resource Allocation and Task Scheduling in Heterogeneou...rahulmonikasharma
Resource Allocation and Task scheduling are the most important key words in today’s dynamic cloud based applications. Task scheduling involves assigning tasks to available processors with the aim of producing minimum execution time, whereas resource allocation involves deciding on an allocation policy to allocate resources to various tasks so as to have maximum resource utilization. Algorithms used for scheduling resources for virtual machines are designed for both homogeneous and heterogeneous environments. Majority of the algorithms focus on processing ability often neglecting other features such as network bandwidth and actual resource requirements. One of the major pitfalls in cloud computing is related to optimizing the resources being allocated. Because of the uniqueness of the model, resource allocation is performed with the objective of minimizing the costs associated with it. The other challenges of resource allocation are meeting customer demands and application requirements. In this paper we will focus on the challenges faced in task scheduling and resource allocation in dynamic heterogeneous clouds.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
Providing a multi-objective scheduling tasks by Using PSO algorithm for cost ...Editor IJCATR
This article is intended to use the multi-PSO algorithm for scheduling tasks for cost management in cloud computing. This means that
any migration costs due to supply failure consider as a one objective and each task is a little particle and recognize by use of the
appropriate fitness schedule function (how the particles arrangement) that cost at least amount of total expense. In addition to, the weight
is granted to the each expenditure that reflects the importance of cost. The data which is used to simulate proposed method are series of
academic and research data that are prepared from the Internet and MATLAB software is used for simulation. We simulate two issues,
in the first issue, consider four task by four vehicles and divide tasks. In the second issue, make the issue more complicated and consider
six tasks by four vehicles. We write PSO's output for each two issues of various iterations. Finally, the particles dispersion and as well
as the output of the cost function were computed for each pa
MCCVA: A NEW APPROACH USING SVM AND KMEANS FOR LOAD BALANCING ON CLOUDijccsa
Nowadays, the demand of using resources, using services via the intranet system or on the Internet is rapidly growing. The respective problem coming is how to use these resources effectively in terms of time and quality. Therefore, the network QoS and its economy are people concerns, cloud computing was born in an inevitable trend. However, managing resources and scheduling tasks in virtualized data centres on the cloud are challenging tasks. Currently, there are a lot of Load Balancing algorithms applied in clouds and proposed by many authors, scholars, and experts. These existing methods are more about natural and heuristic, but the application of AI, or modern datamining technologies, in load balancing is not too popular due to the different characteristics of cloud. In this paper, we propose an algorithm to reduce the processing time (makespan) on cloud computing, helping the load balancing work more efficiency. Here, we use the SVM algorithm to classify the coming Requests, K - Mean to cluster the VMs in cloud, then the LB will allocate the requests into the VMs in the most reasonable way. In this way, request with the least processing time will be allocated to the VMs with the lowest usage. We name this new proposal as MCCVA - Makespan Classification & Clustering VM Algorithm. We have experimented and evaluated this algorithm in CloudSim, a cloud simulation environment, we obtained better results than some other wellknown algorithms. With this MCCVA, we can see the big potential of AI and datamining in Load Balancing, we can further develop LB with AI to achieve better and better results of QoS.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
AN OPEN JACKSON NETWORK MODEL FOR HETEROGENEOUS INFRASTRUCTURE AS A SERVICE O...IJCNCJournal
Cloud computing is an environment which provides services for user demand such as software, platform, infrastructure. Applications which are deployed on cloud computing have become more varied and complex to adapt to increase end-user quantity and fluctuating workload. One popular characteristic of
cloud computing is the heterogeneity of network, hosts and virtual machines (VM). There were many studies on cloud computing modeling based on queuing theory, but most studies have focused on homogeneity characteristic. In this study, we propose a cloud computing model based on open Jackson
network for multi-tier application systems which are deployed on heterogeneous VMs of IaaS cloud computing. The important metrics are analyzed in our experiments such as mean waiting time; mean request quantity, the throughput of the system. Besides that, metrics in model is used to modify number VMs
allocated for applications. Result of experiments shows that open queue network provides high efficiency.
A BAYE'S THEOREM BASED NODE SELECTION FOR LOAD BALANCING IN CLOUD ENVIRONMENThiij
Cloud computing is a popular computing model as it renders service to large number of users request on
the fly and has lead to the proliferation of large number of cloud users. This has lead to the overloaded
nodes in the cloud environment along with the problem of load imbalance among the cloud servers and
thereby impacts the performance. Hence, in this paper a heuristic Baye's theorem approach is considered
along with clustering to identify the optimal node for load balancing. Experiments using the proposed
approach are carried out on cloudsim simulator and are compared with the existing approach. Results
demonstrates that task deployment performed using this approach has improved performance in terms of
utilization and throughput when compared to the existing approaches
A Baye's Theorem Based Node Selection for Load Balancing in Cloud Environmentneirew J
Cloud computing is a popular computing model as it renders service to large number of users request on
the fly and has lead to the proliferation of large number of cloud users. This has lead to the overloaded
nodes in the cloud environment along with the problem of load imbalance among the cloud servers and
thereby impacts the performance. Hence, in this paper a heuristic Baye's theorem approach is considered
along with clustering to identify the optimal node for load balancing. Experiments using the proposed
approach are carried out on cloudsim simulator and are compared with the existing approach. Results
demonstrates that task deployment performed using this approach has improved performance in terms of
utilization and throughput when compared to the existing approaches.
The advent of Big Data has seen the emergence of new processing and storage challenges. These challenges are often solved by distributed processing. Distributed systems are inherently dynamic and unstable, so it is realistic to expect that some resources will fail during use. Load balancing and task scheduling is an important step in determining the performance of parallel applications. Hence the need to design load balancing algorithms adapted to grid computing. In this paper, we propose a dynamic and hierarchical load balancing strategy at two levels: Intrascheduler load balancing, in order to avoid the use of the large-scale communication network, and interscheduler load balancing, for a load regulation of our whole system. The strategy allows improving the average response time of CLOAK-Reduce application tasks with minimal communication. We first focus on the three performance indicators, namely response time, process latency and running time of MapReduce tasks.
Time Efficient VM Allocation using KD-Tree Approach in Cloud Server Environmentrahulmonikasharma
Cloud computing is an incipient and quickly evolving model, with new expenses and capabilities being proclaimed frequently. The increases of user on cloud with the expansion of variety of services, with that the complete allocation of resource with the minimum latent time for Virtual machine is necessary. To allocate this virtual cloud computing resources to the cloud user is a key technical issue because user demand is dynamic in nature that required dynamic allocation of resource too. To improve the allocation there must be a correct balanced algorithmic scheduling for Resource Allocation Technique. The aim of this work is to allocate resource to scientific experiment request coming from multiple users, wherever customized Virtual machines (VM) are aloft in applicable host out there in cloud. Therefore, properly programmed scheduling cloud is extremely vital and it’s significant to develop efficient scheduling methods for appropriately allocation of VMs into physical resource. The planned formulas minimize the time interval quality so as of O (Log n) by adopting KD-Tree.
DYNAMIC TASK SCHEDULING BASED ON BURST TIME REQUIREMENT FOR CLOUD ENVIRONMENTIJCNCJournal
Cloud computing has an indispensable role in the modern digital scenario. The fundamental challenge of cloud systems is to accommodate user requirements which keep on varying. This dynamic cloud environment demands the necessity of complex algorithms to resolve the trouble of task allotment. The overall performance of cloud systems is rooted in the efficiency of task scheduling algorithms. The dynamic property of cloud systems makes it challenging to find an optimal solution satisfying all the evaluation metrics. The new approach is formulated on the Round Robin and the Shortest Job First algorithms. The Round Robin method reduces starvation, and the Shortest Job First decreases the average waiting time. In this work, the advantages of both algorithms are incorporated to improve the makespan of user tasks.
Public Cloud Partition Using Load Status Evaluation and Cloud Division RulesIJSRD
with growth of cloud computing load balancing is important impact on performance. Cloud computing efficiency depends on good load balancer. Many type of situation occur that time cloud partitioning is done by load balancer. Different type of situation needed different type of strategies for public cloud portioning using load balancer.in this paper we work on, partition of public cloud using two type of situation first is load status evaluation and second is cloud division rules. Load status evaluation is measure in number of cloudlets arrives at datacenter and cloud divisions rules are based on cloudlet come from which geographical location. On the basis of geographical location we partition public cloud and improve performance of load balancing in cloud computing. We implement proposed system with help of cloudsim3.0 simulator.
NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...ijccsa
Cloud collaboration is an emerging technology which enables sharing of computer files using cloud
computing. Here the cloud resources are assembled and cloud services are provided using these resources.
Cloud collaboration technologies are allowing users to share documents. Resource allocation in the cloud
is challenging because resources offer different Quality of Service (QoS) and services running on these
resources are risky for user demands. We propose a solution for resource allocation based on multi
attribute QoS Scoring considering parameters such as distance to the resource from user site, reputation of
the resource, task completion time, task completion ratio, and load at the resource. The proposed algorithm
referred to as Multi Attribute QoS scoring (MAQS) uses Neuro Fuzzy system. We have also included a
speculative manager to handle fault tolerance. In this paper it is shown that the proposed algorithm
perform better than others including power trust reputation based algorithms and harmony method which
use single attribute to compute the reputation score of each resource allocated.
Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...neirew J
Cloud collaboration is an emerging technology which enables sharing of computer files using cloud
computing. Here the cloud resources are assembled and cloud services are provided using these resources.
Cloud collaboration technologies are allowing users to share documents. Resource allocation in the cloud
is challenging because resources offer different Quality of Service (QoS) and services running on these
resources are risky for user demands. We propose a solution for resource allocation based on multi
attribute QoS Scoring considering parameters such as distance to the resource from user site, reputation of
the resource, task completion time, task completion ratio, and load at the resource. The proposed algorithm
referred to as Multi Attribute QoS scoring (MAQS) uses Neuro Fuzzy system. We have also included a
speculative manager to handle fault tolerance. In this paper it is shown that the proposed algorithm
perform better than others including power trust reputation based algorithms and harmony method which
use single attribute to compute the reputation score of each resource allocated.
CONTENT BASED VIDEO CATEGORIZATION USING RELATIONAL CLUSTERING WITH LOCAL SCA...ijcsit
This paper introduces a novel approach for efficient video categorization. It relies on two main
components. The first one is a new relational clustering technique that identifies video key frames by
learning cluster dependent Gaussian kernels. The proposed algorithm, called clustering and Local Scale
Learning algorithm (LSL) learns the underlying cluster dependent dissimilarity measure while finding
compact clusters in the given dataset. The learned measure is a Gaussian dissimilarity function defined
with respect to each cluster. We minimize one objective function to optimize the optimal partition and the
cluster dependent parameter. This optimization is done iteratively by dynamically updating the partition
and the local measure. The kernel learning task exploits the unlabeled data and reciprocally, the
categorization task takes advantages of the local learned kernel. The second component of the proposed
video categorization system consists in discovering the video categories in an unsupervised manner using
the proposed LSL. We illustrate the clustering performance of LSL on synthetic 2D datasets and on high
dimensional real data. Also, we assess the proposed video categorization system using a real video
collection and LSL algorithm.
Public Integrity Auditing for Shared Dynamic Cloud Data with Group User Revoc...1crore projects
IEEE PROJECTS 2015
1 crore projects is a leading Guide for ieee Projects and real time projects Works Provider.
It has been provided Lot of Guidance for Thousands of Students & made them more beneficial in all Technology Training.
Dot Net
DOTNET Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
Java Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
ECE IEEE Projects 2015
1. Matlab project
2. Ns2 project
3. Embedded project
4. Robotics project
Eligibility
Final Year students of
1. BSc (C.S)
2. BCA/B.E(C.S)
3. B.Tech IT
4. BE (C.S)
5. MSc (C.S)
6. MSc (IT)
7. MCA
8. MS (IT)
9. ME(ALL)
10. BE(ECE)(EEE)(E&I)
TECHNOLOGY USED AND FOR TRAINING IN
1. DOT NET
2. C sharp
3. ASP
4. VB
5. SQL SERVER
6. JAVA
7. J2EE
8. STRINGS
9. ORACLE
10. VB dotNET
11. EMBEDDED
12. MAT LAB
13. LAB VIEW
14. Multi Sim
CONTACT US
1 CRORE PROJECTS
Door No: 214/215,2nd Floor,
No. 172, Raahat Plaza, (Shopping Mall) ,Arcot Road, Vadapalani, Chennai,
Tamin Nadu, INDIA - 600 026
Email id: 1croreprojects@gmail.com
website:1croreprojects.com
Phone : +91 97518 00789 / +91 72999 51536
A Prolific Scheme for Load Balancing Relying on Task Completion Time IJECEIAES
In networks with lot of computation, load balancing gains increasing significance. To offer various resources, services and applications, the ultimate aim is to facilitate the sharing of services and resources on the network over the Internet. A key issue to be focused and addressed in networks with large amount of computation is load balancing. Load is the number of tasks„t‟ performed by a computation system. The load can be categorized as network load and CPU load. For an efficient load balancing strategy, the process of assigning the load between the nodes should enhance the resource utilization and minimize the computation time. This can be accomplished by a uniform distribution of load of to all the nodes. A Load balancing method should guarantee that, each node in a network performs almost equal amount of work pertinent to their capacity and availability of resources. Relying on task subtraction, this work has presented a pioneering algorithm termed as E-TS (Efficient-Task Subtraction). This algorithm has selected appropriate nodes for each task. The proposed algorithm has improved the utilization of computing resources and has preserved the neutrality in assigning the load to the nodes in the network.
THRESHOLD BASED VM PLACEMENT TECHNIQUE FOR LOAD BALANCED RESOURCE PROVISIONIN...IJCNCJournal
The unbalancing load issue is a multi-variation, multi-imperative issue that corrupts the execution and productivity of processing assets. Workload adjusting methods give solutions of load unbalancing circumstances for two bothersome aspects over-burdening and under-stacking. Cloud computing utilizes planning and workload balancing for a virtualized environment, resource partaking in cloud foundation. These two factors must be handled in an improved way in cloud computing to accomplish ideal resource sharing. Henceforth, there requires productive resource, asset reservation for guaranteeing load advancement in the cloud. This work aims to present an incorporated resource, asset reservation, and workload adjusting calculation for effective cloud provisioning. The strategy develops a Priority-based Resource Scheduling Model to acquire the resource, asset reservation with threshold-based load balancing for improving the proficiency in cloud framework. Extending utilization of Virtual Machines through the suitable and sensible outstanding task at hand modifying is then practiced by intensely picking a job from submitting jobs using Priority-based Resource Scheduling Model to acquire resource asset reservation. Experimental evaluations represent, the proposed scheme gives better results by reducing execution time, with minimum resource cost and improved resource utilization in dynamic resource provisioning conditions.
Challenges in Dynamic Resource Allocation and Task Scheduling in Heterogeneou...rahulmonikasharma
Resource Allocation and Task scheduling are the most important key words in today’s dynamic cloud based applications. Task scheduling involves assigning tasks to available processors with the aim of producing minimum execution time, whereas resource allocation involves deciding on an allocation policy to allocate resources to various tasks so as to have maximum resource utilization. Algorithms used for scheduling resources for virtual machines are designed for both homogeneous and heterogeneous environments. Majority of the algorithms focus on processing ability often neglecting other features such as network bandwidth and actual resource requirements. One of the major pitfalls in cloud computing is related to optimizing the resources being allocated. Because of the uniqueness of the model, resource allocation is performed with the objective of minimizing the costs associated with it. The other challenges of resource allocation are meeting customer demands and application requirements. In this paper we will focus on the challenges faced in task scheduling and resource allocation in dynamic heterogeneous clouds.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
Providing a multi-objective scheduling tasks by Using PSO algorithm for cost ...Editor IJCATR
This article is intended to use the multi-PSO algorithm for scheduling tasks for cost management in cloud computing. This means that
any migration costs due to supply failure consider as a one objective and each task is a little particle and recognize by use of the
appropriate fitness schedule function (how the particles arrangement) that cost at least amount of total expense. In addition to, the weight
is granted to the each expenditure that reflects the importance of cost. The data which is used to simulate proposed method are series of
academic and research data that are prepared from the Internet and MATLAB software is used for simulation. We simulate two issues,
in the first issue, consider four task by four vehicles and divide tasks. In the second issue, make the issue more complicated and consider
six tasks by four vehicles. We write PSO's output for each two issues of various iterations. Finally, the particles dispersion and as well
as the output of the cost function were computed for each pa
CLOUD COMPUTING: A NEW VISION OF THE DISTRIBUTED SYSTEM cscpconf
Cloud computing is a new emerging system which offers information technologies via Internet. Clients use services they need when they need and at the place they want and pay only for what they have consumed. So, cloud computing offers many advantages especially for business. A deep study and understanding of this emerging system and the inherent components help a lot in identifying what should we do in order to improve its performance. In this work, we present first cloud computing and its components then we describe an idea which attempts to optimize the management of cloud computing system that are composed of many data centers.
DYNAMIC TENANT PROVISIONING AND SERVICE ORCHESTRATION IN HYBRID CLOUDijccsa
The advent of container orchestration and cloud computing, as well as associated security and compliance complexities, make it challenging for the enterprises to develop robust, secure, manageable and extendable architectures which would be applicable to the public and private cloud. The main challenges stem from the fact that on-premises, private cloud and third-party, public cloud services often have seemingly different and sometimes conflicting requirements to tenant provisioning, service deployment, security and compliance and that can lead to rather different architectures which still have a lot of commonalities but evolve independently. Understanding and bridging the functionality gaps between such architectures is highly desirable in terms of common approaches, API/SPI as well as maintainability and extendibility. The authors discuss and propose common architectural approaches to the dynamic tenant provisioning and
service orchestration in public, private and hybrid clouds focusing on deployment, security, compliance, scalability and extendibility of stateful Kubernetes runtimes.
DYNAMIC TENANT PROVISIONING AND SERVICE ORCHESTRATION IN HYBRID CLOUDijccsa
The advent of container orchestration and cloud computing, as well as associated security and compliance complexities, make it challenging for the enterprises to develop robust, secure, manageable and extendable architectures which would be applicable to the public and private cloud. The main challenges stem from the fact that on-premises, private cloud and third-party, public cloud services often have seemingly different and sometimes conflicting requirements to tenant provisioning, service deployment, security and compliance and that can lead to rather different architectures which still have a lot of commonalities but evolve independently. Understanding and bridging the functionality gaps between such architectures is highly desirable in terms of common approaches, API/SPI as well as maintainability and extendibility. The authors discuss and propose common architectural approaches to the dynamic tenant provisioning and service orchestration in public, private and hybrid clouds focusing on deployment, security, compliance, scalability and extendibility of stateful Kubernetes runtimes.
The concept of Genetic algorithm is specifically useful in load balancing for best virtual
machines distribution across servers. In this paper, we focus on load balancing and also on
efficient use of resources to reduce the energy consumption without degrading cloud
performance. Cloud computing is an on demand service in which shared resources, information,
software and other devices are provided according to the clients requirement at specific time. It‟s
a term which is generally used in case of Internet. The whole Internet can be viewed as a cloud.
Capital and operational costs can be cut using cloud computing. Cloud computing is defined as a
large scale distributed computing paradigm that is driven by economics of scale in which a pool
of abstracted virtualized dynamically scalable , managed computing power ,storage , platforms
and services are delivered on demand to external customer over the internet. cloud computing is
a recent field in the computational intelligence techniques which aims at surmounting the
computational complexity and provides dynamically services using very large scalable and
virtualized resources over the Internet. It is defined as a distributed system containing a
collection of computing and communication resources located in distributed data enters which
are shared by several end users. It has widely been adopted by the industry, though there are
many existing issues like Load Balancing, Virtual Machine Migration, Server Consolidation,
Energy Management, etc.
Support for Goal Oriented Requirements Engineering in Elastic Cloud Applicationszillesubhan
Businesses have already started to exploit potential uses of cloud computing as a new paradigm for promoting their services. Although the general concepts they practically focus on are: viability, survivability, adaptability, etc., however, on the ground, there is still a lack for forming mechanisms to sustain viability with adaptation of new requirements in cloud-based applications. This has inspired a pressing need to adopt new methodologies and abstract models which support system acquisition for self-adaptation, thus guaranteeing autonomic cloud application behavior. This paper relies over state-of-the-art Neptune framework as runtime adaptive software development environment supported with intention-oriented modeling language in the representation and adaptation of goal based model artifacts and their intrinsic properties requirements. Such an approach will in turn support distributed service based applications virtually over the cloud to sustain a self-adaptive behavior with respect to its functional and non-functional characteristics.
WiSANCloud: a set of UML-based specifications for the integration of Wireless...Priscill Orue Esquivel
Giving the current trend to combine the advantages of Wireless Sensor and Actor Networks (WSANs) with the Cloud Computing technology, this work proposes a set of specifications, based on the Unified Modeling Language - UML, in order to provide the general framework for the design of the integration of said components. One of the keys of the integration is the architecture of the WSAN, due to its structural relationship with the Cloud in the definition of the combination. Regarding the standard applied in the integration, UML and its subset, Systems Modeling Language - SysML , are proposed by the Object Management Group - OMG to deal with cloud applications; so, this indicates the starting point of the process of the design of specifications for WSAN-Cloud Integration. Based on the current state of UML tools for analysis and design, there are several aspects to take into account in order to define the integration process.
A CLOUD BROKER APPROACH WITH QOS ATTENDANCE AND SOA FOR HYBRID CLOUD COMPUTIN...cscpconf
Cloud Computing is the industry whose demand has been growing continuously since its appearance as a solution that offers different types of computing resources as a service over the Internet. The number of cloud computing providers grows into a run, while the end user is currently in the position of having many pricing options, distinct features and performance for the same required service. This work is inserted in the cloud computing task scheduling research field to hybrid cloud environments with service-oriented architecture (SOA), dynamic allocation and control of services and QoS requirements attendance. Therefore, it is proposed the QBroker Architecture, representing a cloud broker with trading features that implement the intermediation services, defined by the NIST Cloud Computing Reference Model. An experimental design was created in order to demonstrate compliance to the QoS requirement of maximum task execution time, the differentiation of services and dynamic allocation of services. The experimental results obtained by simulation with CloudSim prove that QBroker has the necessary requirements to provide QoS improvement in hybrid cloud computing environments based on SOA.
A cloud broker approach with qos attendance and soa for hybrid cloud computin...csandit
Cloud Computing is the industry whose demand has been growing continuously since its
appearance as a solution that offers different types of computing resources as a service over the
Internet. The number of cloud computing providers grows into a run, while the end user is
currently in the position of having many pricing options, distinct features and performance for
the same required service. This work is inserted in the cloud computing task scheduling
research field to hybrid cloud environments with service-oriented architecture (SOA), dynamic
allocation and control of services and QoS requirements attendance. Therefore, it is proposed
the QBroker Architecture, representing a cloud broker with trading features that implement the
intermediation services, defined by the NIST Cloud Computing Reference Model. An
experimental design was created in order to demonstrate compliance to the QoS requirement of
maximum task execution time, the differentiation of services and dynamic allocation of services.
The experimental results obtained by simulation with CloudSim prove that QBroker has the
necessary requirements to provide QoS improvement in hybrid cloud computing environments
based on SOA.
An Efficient Queuing Model for Resource Sharing in Cloud Computingtheijes
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
Student information management system project report ii.pdfKamal Acharya
Our project explains about the student management. This project mainly explains the various actions related to student details. This project shows some ease in adding, editing and deleting the student details. It also provides a less time consuming process for viewing, adding, editing and deleting the marks of the students.
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptxR&R Consult
CFD analysis is incredibly effective at solving mysteries and improving the performance of complex systems!
Here's a great example: At a large natural gas-fired power plant, where they use waste heat to generate steam and energy, they were puzzled that their boiler wasn't producing as much steam as expected.
R&R and Tetra Engineering Group Inc. were asked to solve the issue with reduced steam production.
An inspection had shown that a significant amount of hot flue gas was bypassing the boiler tubes, where the heat was supposed to be transferred.
R&R Consult conducted a CFD analysis, which revealed that 6.3% of the flue gas was bypassing the boiler tubes without transferring heat. The analysis also showed that the flue gas was instead being directed along the sides of the boiler and between the modules that were supposed to capture the heat. This was the cause of the reduced performance.
Based on our results, Tetra Engineering installed covering plates to reduce the bypass flow. This improved the boiler's performance and increased electricity production.
It is always satisfying when we can help solve complex challenges like this. Do your systems also need a check-up or optimization? Give us a call!
Work done in cooperation with James Malloy and David Moelling from Tetra Engineering.
More examples of our work https://www.r-r-consult.dk/en/cases-en/
Quality defects in TMT Bars, Possible causes and Potential Solutions.PrashantGoswami42
Maintaining high-quality standards in the production of TMT bars is crucial for ensuring structural integrity in construction. Addressing common defects through careful monitoring, standardized processes, and advanced technology can significantly improve the quality of TMT bars. Continuous training and adherence to quality control measures will also play a pivotal role in minimizing these defects.
Event Management System Vb Net Project Report.pdfKamal Acharya
In present era, the scopes of information technology growing with a very fast .We do not see any are untouched from this industry. The scope of information technology has become wider includes: Business and industry. Household Business, Communication, Education, Entertainment, Science, Medicine, Engineering, Distance Learning, Weather Forecasting. Carrier Searching and so on.
My project named “Event Management System” is software that store and maintained all events coordinated in college. It also helpful to print related reports. My project will help to record the events coordinated by faculties with their Name, Event subject, date & details in an efficient & effective ways.
In my system we have to make a system by which a user can record all events coordinated by a particular faculty. In our proposed system some more featured are added which differs it from the existing system such as security.
Vaccine management system project report documentation..pdfKamal Acharya
The Division of Vaccine and Immunization is facing increasing difficulty monitoring vaccines and other commodities distribution once they have been distributed from the national stores. With the introduction of new vaccines, more challenges have been anticipated with this additions posing serious threat to the already over strained vaccine supply chain system in Kenya.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
Automobile Management System Project Report.pdfKamal Acharya
The proposed project is developed to manage the automobile in the automobile dealer company. The main module in this project is login, automobile management, customer management, sales, complaints and reports. The first module is the login. The automobile showroom owner should login to the project for usage. The username and password are verified and if it is correct, next form opens. If the username and password are not correct, it shows the error message.
When a customer search for a automobile, if the automobile is available, they will be taken to a page that shows the details of the automobile including automobile name, automobile ID, quantity, price etc. “Automobile Management System” is useful for maintaining automobiles, customers effectively and hence helps for establishing good relation between customer and automobile organization. It contains various customized modules for effectively maintaining automobiles and stock information accurately and safely.
When the automobile is sold to the customer, stock will be reduced automatically. When a new purchase is made, stock will be increased automatically. While selecting automobiles for sale, the proposed software will automatically check for total number of available stock of that particular item, if the total stock of that particular item is less than 5, software will notify the user to purchase the particular item.
Also when the user tries to sale items which are not in stock, the system will prompt the user that the stock is not enough. Customers of this system can search for a automobile; can purchase a automobile easily by selecting fast. On the other hand the stock of automobiles can be maintained perfectly by the automobile shop manager overcoming the drawbacks of existing system.
About
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Technical Specifications
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
Key Features
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface
• Compatible with MAFI CCR system
• Copatiable with IDM8000 CCR
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
Application
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdfKamal Acharya
The College Bus Management system is completely developed by Visual Basic .NET Version. The application is connect with most secured database language MS SQL Server. The application is develop by using best combination of front-end and back-end languages. The application is totally design like flat user interface. This flat user interface is more attractive user interface in 2017. The application is gives more important to the system functionality. The application is to manage the student’s details, driver’s details, bus details, bus route details, bus fees details and more. The application has only one unit for admin. The admin can manage the entire application. The admin can login into the application by using username and password of the admin. The application is develop for big and small colleges. It is more user friendly for non-computer person. Even they can easily learn how to manage the application within hours. The application is more secure by the admin. The system will give an effective output for the VB.Net and SQL Server given as input to the system. The compiled java program given as input to the system, after scanning the program will generate different reports. The application generates the report for users. The admin can view and download the report of the data. The application deliver the excel format reports. Because, excel formatted reports is very easy to understand the income and expense of the college bus. This application is mainly develop for windows operating system users. In 2017, 73% of people enterprises are using windows operating system. So the application will easily install for all the windows operating system users. The application-developed size is very low. The application consumes very low space in disk. Therefore, the user can allocate very minimum local disk space for this application.
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
The Benefits and Techniques of Trenchless Pipe Repair.pdf
Cobe framework cloud ontology blackboard environment for enhancing discovery behavior
1. International Journal on Cloud Computing: Services and Architecture (IJCCSA) ,Vol. 4, No. 5, October 2014
COBE Framework: Cloud Ontology Blackboard
Environment for Enhancing Discovery Behavior
Ahmed Ghoneim1, 3 Amr Tolba2, 3
1College of Computers and Information Sciences, King Saud University, Saudi Arabia,
2Riyadh Community College, King Saud University, Saudi Arabia
3Faculty of Science, Menoufia University, Egypt
ABSTRACT
The new relatively concept of cloud computing its associated methodologies has many advantages in the
world of today. Such advantages range between providing solutions for integration of the miscellaneous
systems presenting as well guarantees for distribution of searching means integration of software
tools which are used by consumers different providers. In this paper, we have constructed an ontology-based
cloud framework with a view to identifying its external agent’s interoperability. The proposed
framework has been designed using the blackboard design style. This framework is composed of mainly
two components: controller and cloud ontology blackboard environment. The function of the controller is
to interact with consumers after receipt of the subject request where it spontaneously uses the ontology
base to distribute it constitute the required related responses whereas the function of the second
framework component is to interact with different cloud providers and systems, using the meta-ontology
framework to restructure data via using AI reasoning tools and map them to its corresponding
redistributed request. Finally, E-tourism case study can be applicable will be explored.
KEYWORDS
Ontology, cloud computing, agent systems, AI reasoning, design style.
1. INTRODUCTION
Cloud computing has been formulated as a general term to describe a class of complex computing
services on demand, which was introduced by several commercial providers, such as Amazon,
Google, and Microsoft. It denotes a model that looks at the infrastructure of computing as a
cloud, through which companies and individuals can access applications from anywhere in the
world based on theirdemand. The main principle behind this computing model is to provide
computing, storage, and software as a service [8].
The most common service models introduced in the cloud computing environment are IaaS
(Infrastructure as a Service), PaaS (Platform as a Service), and SaaS (Software as a Service). In
IaaS model, the service providers outsource the equipment used to support operations, including
hardware, storage, networking components and servers over the internet. In PaaS model, the
service providers allow the cloud consumers to use operating systems, and virtualized servers
DOI : 10.5121/ijccsa.2014.4504 29
2. International Journal on Cloud Computing: Services and Architecture (IJCCSA) ,Vol. 4, No. 5, October 2014
together with its associated services for running applications over the internet. Finally, In SaaS
model, the service providers allow cloud users to remote access of applications or software as a
web-based service to avoid the cost of buying software [10, 11].
Ontology aspect is used to constitute basic- and processed- information which describe the data
semantics based on the consumers request. It has a set of concepts that deal with information
retrievals to process the consumers and providers requests [4, 7]. Most of the existing cloud
service discovery frameworks are using the ontology reasoning for matchmaking the suitable
requests and responses for both consumers and providers [4, 9].
In this paper, we introduce an ontology-based cloud framework for discovery their external
agent’s interoperability. The proposed framework is designed based on blackboard design style.
This framework is composed of mainly two components: controller and cloud ontology
blackboard environment. Both components are using ontology for discovering the best suit set of
responses regarding the consumer requests.
The rest of this paper is organized as follows: In Section 2, the proposed framework is described;
its structural and behavioral is illustrated in details. Section 3 illustrates the related works. Case
study is described in section 4. Finally, In Section 5, conclusions and future work are discussed.
30
2. PROPOSED FRAMEWORK
In this section, we have been introducing both the structure and behavior of our ontology-based
cloud framework for discovery their external agent’s interoperability. The proposed framework is
designed based on blackboard design style. This framework is composed of mainly two
components: controller and cloud ontology blackboard environment as shown in Figure 1.
The role of controller is to directly interact with consumers. The controller functionality starts,
after receiving the consumer request, and then it uses its built-in controller-ontology for
decomposing the request into set of mini-requests (i.e. distributing the request to its primary basic
components). After that, send this set of requests to the cloud ontology blackboard environment
component and waiting for responses. Finally, after receiving the responses from the (COBE)
component, it starts his final process by constitute these set of responses into one consistence
response and send it back to its consumer.
The (COBE) component is responsible for interacting to all possible clouds, different agent
systems, and controller. It uses meta-ontology for restructuring data, using AI reasoning methods,
and mapping data to its equivalent redistributed request. The behavior of (COBE) consists of the
following sequence: first, receive the distributed set of request from the controller, then using the
meta-ontology for restructuring cloud and agent data and runtime updating this data for any
changes. After that, it uses AI reasoning methods such as forward or backward reasoning to
extract the acceptable responses from the data streaming flow between COBE and its external
agents. Next, COBE starts its processes in this stage by mapping the responses data to its
equivalent request. Finally, COBE sends the response sets and its mapping parameters to the
controller. The implemented behavior of COBE is illustrated as pseudo code in table 1.
The COBE starts its functionalities by checking the request type; if the request type is simple this
means the COBE neglecting the distribution option. In case the request type is composite this
3. International Journal on Cloud Computing: Services and Architecture (IJCCSA) ,Vol. 4, No. 5, October 2014
implies the distribute function of COBE will be fire, and based on the result of distribution, the
COBE starts the second function by mapping each sub request to its partner.
The main component of the proposed framework and its interaction are shown in Figure 2, which
represents the structure part of the COBE. The dynamic behavior of the COBE is illustrated using
Figure 3, which focuses on business processes and dataflow of the system Moreover each
component of the core uses the business rules to handle the flow or apply the reasoning methods.
31
Figure 1 – COBE Framework
Figure 2: Class diagram of the proposed framework
4. International Journal on Cloud Computing: Services and Architecture (IJCCSA) ,Vol. 4, No. 5, October 2014
32
Figure 3: Business processes that represent the internal functionality of COBE framework
Table 1: Pseudo code implemented behavior of COBE
5. International Journal on Cloud Computing: Services and Architecture (IJCCSA) ,Vol. 4, No. 5, October 2014
33
3. RELATED WORK
In the following, we will illustrate some related work for applying ontology concept in a cloud
computing environment. Lamia Youseff et al. [1] proposed the use of ontology inside the cloud.
They try to establish a knowledge domain in the field of cloud computing. They present cloud
ontology as a stack of layers discussed each layer in the strength point of view and neglect the
integration and communication issue between layers.
N.M.Karia et al. [2] have focused on introducing an ontological framework that provides a
structure of different cloud environments and, describing its components. This work doesn’t
address the heterogeneity of terms used by vendors, number of standards, and a structure of the
different cloud environments. Teodor-Forin Forties et al. [3] attempted to achieve ontology
concept in the cloud services. They introduce mOSAIC as a middleware to facilitate the
communication with various cloud providers. They suggest the use of ontology to facilitate the
selection, execution and discovery of resources and services. They introduce a comprehensive
model of ontology cloud the can work only for public clouds.
Taekgyeong Han and KwangMongSim. [4] havepresented a Cloud Service Discovery System
(CSDS) based on ontology. They use ontology to enhance performance of the system also, the
use of software agent to consult ontology when selecting information retrieved from a public
cloud. Miranda Zhang et al. [5] concentrated on using ontology for Cloud Based IaaS services.
They implement ontology in the CloudRecommender system with focusing on IaaS services and
neglect PaaS and SaaS configurations in a public cloud. On the other hand, lot of work done to
utilize a software agent in cloud computing environment to enhance the performance of cloud
services. Toshihiro Uchibayashi et al. [6] introduced a framework of an agent-based support
system for discovering services in a public cloud. They use agents implemented by JADE only to
measure network usage against user requirements information.
Based on the above related works, we can realize that most of the above models and frameworks
are using the AI reasoning methods and ontology without providing an approach on how to
distribute the request into request list, restructure the cloud data at runtime and mapping the parts
of the request list into its equivalent partner. Our proposed COBE framework addresses the above
challenges.
4. Case Study: E-Tourism
In this section, we will approve the applicability of our COBE framework by illustrating how the
different systems and agents are communicating together through the blackboard style. Moreover,
how to distribute the visitor request to asset of requests and mapping these set into its equivalent
agents. In figure 4, the E-tourism data inputs and its processed output is designed using
component based style. The required services and the provided services of the system are
identified.
6. International Journal on Cloud Computing: Services and Architecture (IJCCSA) ,Vol. 4, No. 5, October 2014
34
Figure 4: E-tourism system with required and provided data
The data center interoperability between the E-tourism systems is shown in figure 5, which
represent the dataflow and the control flow from COBE main component into all interacted
agents using blackboard style. COBE of E-tourism has the ability to detect and identify the set of
agents based on the number of distributed requests. Moreover, COBE also identify the type of
communication with these agents and play the role of controller to those agents. The number of
agents depends on the number of sub-requests. The COBE sometimes find that one of the
identified agents which is not connected to its functions not responding. In this case the COBE
has a wide cloud search function to search for alternative agents with the same feature and
compare between them and to select the best one.
Figure 5: COBE Role on E-tourism
7. International Journal on Cloud Computing: Services and Architecture (IJCCSA) ,Vol. 4, No. 5, October 2014
In Figure 6, we provide a sample of visitor request as XML message and the role of COBE to
distribute it into set of sub-requests and its equivalent agents (Hotels-agent, Airlines-Agent, Car
Rental -Company Agent). The COBE uses AI forward reasoning techniques to distribute the
request to a set of consistent sub-requests as shown in samples at figure 6.After that, it waits the
responses from all connected agents. Next, it uses the validation and verification techniques to
check that each response is valid to its sub-request. Then, the COBE uses AI backward reasoning
techniques for constituting the set of responses into one response including some features or plan
options. Finally, The COBE provides this response to the identified client.
35
Figure 6: Interactivity between COBE of E-tourism Agent using XML messages
5. CONCLUSION AND FUTURE WORK
In this paper, we presented the COBE framework which provides the ability to manage the client
request by distributing it to mini-requests and use the ontology with reasoning methods for
identifyingthe responses for all the request sets by mapping it to its equivalent cloud agents. The
propose framework is composed of two core components: Controller and COBE, both
components are responsible for driving the framework functionalities. The structure and the
business processes simulation of the proposed framework are designed. Moreover, the
implemented pseudo code as algorithmic environment of the COBE framework is illustrated. In
future work we are planning to use the quality of service (QoS) criteria to measure the response-time,
availability and throughput. Moreover, we will integrate with CloudSim tool as a simulation
environment for the COBE framework.
8. International Journal on Cloud Computing: Services and Architecture (IJCCSA) ,Vol. 4, No. 5, October 2014
6. REFERENCES
[1] Youseff, Lamia, Maria Butrico
Butrico, and Dilma Da Silva. Toward a unified ontology of cloud
computing. Grid Computing Environments Workshop
Workshop, 2008.GCE'08. IEEE, 2008.
An Ontological Framework for a Cloud Forensic
[2] Karie, Nickson M., and Hein S. Venter.
Environment. (2013).
[3] Fortis, T-F., Victor Ion Munteanu, and ViorelNegru. Towards an ontology for cloud services. 2012
Systems (CISIS),
Sixth International Conference on Complex, Intelligent and Software Intensive Syst
IEEE, 2012.
[4] Han, Taekgyeong, and KwangMongSim
KwangMongSim. An ontology-enhanced cloud service discovery system
Proceedings of the International MultiConference of Engineers and Computer Scientists
system.
Scientists.Vol. 1. 2010.
Ontology based System for Cloud Infrastructure Services Discovery
[5] Zhang, Miranda, et al. An Onto
arXiv preprint arXiv:1212.0156
logy Discovery.
(2012).
[6] Uchibayashi, Toshihiro, Bernady O. Apduhan, and Norio Shiratori. A framework of an agent
agent-based
International Conference on Computational Science
support system for IaaS service discovery. 13th Int
and Its Applications (ICCSA),
, IEEE, 2013.
[7] Fensel, Dieter, McGuiness, D.L. , Schulten, E. , Wee Keong Ng , Ge Peng Lim , and Guanghao Yan.
Ontologies and electronic commerce. Intelligent Systems, IEEE 1
16.1, 2001.
Agent-Based Cloud Framework”,
[8] Amr Tolba, and Ahmed Ghoneim “IABCF Smarter: An Intelligent Agent
IJCST Vol. 3, Issue 4, Oct - Dec 2012.
[9] Liu, L., Yao, X., Qin, L., Zhang, M. “Ontology
Ontology-based service matching in cloud computing”. IEEE
International Conference on In Fuzzy Systems (FUZZ
nal FUZZ-IEEE), (pp. 2544-2550). IEEE. 2014.
[10] Rafique, Ansar, et al. Towards portability and interoperability support in middleware for hybrid
clouds. IEEE Conference on Computer Communications Workshops (INFOCOM WKSH
2014.
[11] Mosbah, Mohamed M., Hany Sol
Sol, and Mohamad Abou El-Nasr. Cloud Computing Framework for
Solving Egyptian Higher Education
Computing and Communications
Authors
Education. Fourth International IEEE Conference on Advances in
(ICACC), IEEE, 2014.
Dr. Ahmed Ghoneim received his M.Sc. degree in software modeling from
University of Menoufia, Egypt, and the Ph.D. degree from the University of
Magdeburg (Germany) in the area of software engineering, in 1999 and 2007
respectively. He is currently an assistant professor at the department of software
engineering, king Saud University. His research activities address software
evolution; service oriented engineering, software development methodologies,
scripting languages.
Dr. Amr Tolba received his M.Sc. c. and Ph.D. degrees in computer science, Menoufia
University, Egypt, in the area of databases and Multi
Multi-Agent systems, in 2002 and
2006 respectively. He joined King Saud University (KSU) in August 2006 as an
assistant professor at the department of comput
computer science. His research interests
include internet of things, intelligent systems, mobile social networks, recommender
systems, e-learning, and cloud computing.
36
ernational WKSHPS), IEEE.