Data center have the largest consumption amounts of energy in sharing the power. The public cloud workloads of different
priorities and performance requirements of various applications [4]. Cloud data center have capable of sensing an opportunity to present
different programs. In my proposed construction and the name of the security level of imperturbable privacy leakage rarely distributed
cloud system to deal with the persistent characteristics there is a substantial increases and information that can be used to augment the
profit, retrenchment overhead or both. Data Mining Analysis of data from different perspectives and summarizing it into useful
information is a process. Three empirical algorithms have been proposed assignments estimate the ratios are dissected theoretically and
compared using real Internet latency data recital of testing methods
35 content distribution with dynamic migration of services for minimum cost u...INFOGAIN PUBLICATION
Content Delivery Networks are the key for today’s internet content delivery. Users are knowingly or unknowingly accessing the CDN via internet. No matter how much the data retrieved by the user it may contain the CDN hand behind every character of text and every pixel of image. CDN came into existence to solve the delay problem. The moment when a user requests for a web page and the response delivered to the corresponding users web browser facing a huge delay. The main goal of this paper is content distribution of web services to multiple data centers placed in different geographical locations and providing security. A content distribution service is a major part of popular Internet applications. In proposed system hybrid clouds are used i.e., both private cloud as well as public cloud. One data center is allocated to each region. Providing security to the data is always an important issue because of the critical nature of the cloud and very large amount of complicated data it carries. To provide security cipher text policy algorithm is used. Authentication technique is used to verify the user authentication. If the user is authorized to access services then and only he receives configuration key to use.
DISTRIBUTED AND BIG DATA STORAGE MANAGEMENT IN GRID COMPUTINGijgca
Big data storage management is one of the most challenging issues for Grid computing environments, since large amount of data intensive applications frequently involve a high degree of data access locality. Grid applications typically deal with large amounts of data. In traditional approaches high-performance computing consists dedicated servers that are used to data storage and data replication. In this paper we present a new mechanism for distributed and big data storage and resource discovery services. Here we proposed an architecture named Dynamic and Scalable Storage Management (DSSM) architecture in grid environments. This allows in grid computing not only sharing the computational cycles, but also share the storage space. The storage can be transparently accessed from any grid machine, allowing easy data sharing among grid users and applications. The concept of virtual ids that, allows the creation of virtual spaces has been introduced and used. The DSSM divides all Grid Oriented Storage devices (nodes) into multiple geographically distributed domains and to facilitate the locality and simplify the intra-domain storage management. Grid service based storage resources are adopted to stack simple modular service piece by piece as demand grows. To this end, we propose four axes that define: DSSM architecture and algorithms description, Storage resources and resource discovery into Grid service, Evaluate purpose prototype system, dynamically, scalability, and bandwidth, and Discuss results. Algorithms at bottom and upper level for standardization dynamic and scalable storage management, along with higher bandwidths have been designed.
CONTENT BASED DATA TRANSFER MECHANISM FOR EFFICIENT BULK DATA TRANSFER IN GRI...ijgca
A new class of Data Grid infrastructure is needed to support management, transport, distributed access, and analysis of terabyte and peta byte of data collections by thousands of users. Even though some of the existing data management systems (DMS) of Grid computing infrastructures provides methodologies to handle bulk data transfer. These technologies are not usable in addressing some kind of simultaneous data
access requirements. Often, in most of the scientific computing environments, a common data will be needed to access from different locations. Further, most of such computing entities will wait for a common scientific data (such as a data belonging to an astronomical phenomenon) which will be published only
when it is available. These kinds of data access needs were not yet addressed in the design of data component Grid Access to Secondary Storage (GASS) or GridFTP. In this paper, we address an application layer content based data transfer scheme for grid computing environments. By using the
proposed scheme in a grid computing environment, we can simultaneously move bulk data in an efficient way using simple subscribe and publish mechanism.
‘Grids’areanapproachforbuildingdynamicallyconstructedproblem-solvingenvironmentsusing
geographically and organizationally dispersed,
high-performance computing and
data handling resources.
Gridsalsoprovideimportantinfrastructuresupportingmulti-institutionalcollaboration.
Survey on cloud backup services of personal storageeSAT Journals
Abstract In widespread cloud environment cloud services is tremendously growing due to large amount of personal computation data. Deduplication process is used for avoiding the redundant data. A cloud storage environment for data backup in personal computing devices facing various challenge, of source deduplication for the cloud backup services with low deduplication efficiency. Challenges facing in the process of deduplication for cloud backup service are-1)Low deduplication efficiency due to exclusive access to large amount of data and limited system resources of PC based client site.2)Low data transfer efficiency due to transferring deduplicate data from source to backup server are typically small but that can be often across the WAN. Keywords- Cloud computing, Deduplication, cloud backup, application awareness
35 content distribution with dynamic migration of services for minimum cost u...INFOGAIN PUBLICATION
Content Delivery Networks are the key for today’s internet content delivery. Users are knowingly or unknowingly accessing the CDN via internet. No matter how much the data retrieved by the user it may contain the CDN hand behind every character of text and every pixel of image. CDN came into existence to solve the delay problem. The moment when a user requests for a web page and the response delivered to the corresponding users web browser facing a huge delay. The main goal of this paper is content distribution of web services to multiple data centers placed in different geographical locations and providing security. A content distribution service is a major part of popular Internet applications. In proposed system hybrid clouds are used i.e., both private cloud as well as public cloud. One data center is allocated to each region. Providing security to the data is always an important issue because of the critical nature of the cloud and very large amount of complicated data it carries. To provide security cipher text policy algorithm is used. Authentication technique is used to verify the user authentication. If the user is authorized to access services then and only he receives configuration key to use.
DISTRIBUTED AND BIG DATA STORAGE MANAGEMENT IN GRID COMPUTINGijgca
Big data storage management is one of the most challenging issues for Grid computing environments, since large amount of data intensive applications frequently involve a high degree of data access locality. Grid applications typically deal with large amounts of data. In traditional approaches high-performance computing consists dedicated servers that are used to data storage and data replication. In this paper we present a new mechanism for distributed and big data storage and resource discovery services. Here we proposed an architecture named Dynamic and Scalable Storage Management (DSSM) architecture in grid environments. This allows in grid computing not only sharing the computational cycles, but also share the storage space. The storage can be transparently accessed from any grid machine, allowing easy data sharing among grid users and applications. The concept of virtual ids that, allows the creation of virtual spaces has been introduced and used. The DSSM divides all Grid Oriented Storage devices (nodes) into multiple geographically distributed domains and to facilitate the locality and simplify the intra-domain storage management. Grid service based storage resources are adopted to stack simple modular service piece by piece as demand grows. To this end, we propose four axes that define: DSSM architecture and algorithms description, Storage resources and resource discovery into Grid service, Evaluate purpose prototype system, dynamically, scalability, and bandwidth, and Discuss results. Algorithms at bottom and upper level for standardization dynamic and scalable storage management, along with higher bandwidths have been designed.
CONTENT BASED DATA TRANSFER MECHANISM FOR EFFICIENT BULK DATA TRANSFER IN GRI...ijgca
A new class of Data Grid infrastructure is needed to support management, transport, distributed access, and analysis of terabyte and peta byte of data collections by thousands of users. Even though some of the existing data management systems (DMS) of Grid computing infrastructures provides methodologies to handle bulk data transfer. These technologies are not usable in addressing some kind of simultaneous data
access requirements. Often, in most of the scientific computing environments, a common data will be needed to access from different locations. Further, most of such computing entities will wait for a common scientific data (such as a data belonging to an astronomical phenomenon) which will be published only
when it is available. These kinds of data access needs were not yet addressed in the design of data component Grid Access to Secondary Storage (GASS) or GridFTP. In this paper, we address an application layer content based data transfer scheme for grid computing environments. By using the
proposed scheme in a grid computing environment, we can simultaneously move bulk data in an efficient way using simple subscribe and publish mechanism.
‘Grids’areanapproachforbuildingdynamicallyconstructedproblem-solvingenvironmentsusing
geographically and organizationally dispersed,
high-performance computing and
data handling resources.
Gridsalsoprovideimportantinfrastructuresupportingmulti-institutionalcollaboration.
Survey on cloud backup services of personal storageeSAT Journals
Abstract In widespread cloud environment cloud services is tremendously growing due to large amount of personal computation data. Deduplication process is used for avoiding the redundant data. A cloud storage environment for data backup in personal computing devices facing various challenge, of source deduplication for the cloud backup services with low deduplication efficiency. Challenges facing in the process of deduplication for cloud backup service are-1)Low deduplication efficiency due to exclusive access to large amount of data and limited system resources of PC based client site.2)Low data transfer efficiency due to transferring deduplicate data from source to backup server are typically small but that can be often across the WAN. Keywords- Cloud computing, Deduplication, cloud backup, application awareness
Ensuring secure transfer, access and storage over the cloud storageeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Ensuring secure transfer, access and storage over the cloud storageeSAT Journals
Abstract The main concern in today’s growing IT sectors is the storage and maintenance of the data. As the data keep on updating according to the needs of users, there is a huge overhead in the maintenance of the hardware by the company. One of the solutions to this problem is the use of cloud storage for this enormous data. The cloud storage uses the huge data centers, which are remotely located, to store the data. In addition to the easy storage of the data, this huge data center also reduces the cost of maintenance of the data .However this distinct feature of cloud storage leads to many security issues which should be lucidly understood by the IT sectors. One of the emerging security issue would be the integrity of the data stored in the data center i.e. to check whether the cloud provider misuses the data or not .The cloud provider can misuses the data in many ways like they can copy or modify the file .Due to the storage of data on the data center, user is not able to access the data physically thus there should be the way by which user can check the reliability of the data on the cloud. In this paper we provide the scheme to check the reliability of the data and this scheme can be agreed upon by both the user and the cloud provider. Keywords: Cloud Security, Masking, Cloud Storage Security, Data Center
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
Survey on Division and Replication of Data in Cloud for Optimal Performance a...IJSRD
Outsourcing information to an outsider authoritative control, as is done in distributed computing, offers ascend to security concerns. The information trade off may happen because of assaults by different clients and hubs inside of the cloud. Hence, high efforts to establish safety are required to secure information inside of the cloud. On the other hand, the utilized security procedure should likewise consider the advancement of the information recovery time. In this paper, we propose Division and Replication of Data in the Cloud for Optimal Performance and Security (DROPS) that all in all methodologies the security and execution issues. In the DROPS procedure, we partition a record into sections, and reproduce the divided information over the cloud hubs. Each of the hubs stores just a itary part of a specific information record that guarantees that even in the event of a fruitful assault, no important data is uncovered to the assailant. Additionally, the hubs putting away the sections are isolated with certain separation by method for diagram T-shading to restrict an assailant of speculating the areas of the sections. Moreover, the DROPS procedure does not depend on the customary cryptographic procedures for the information security; in this way alleviating the arrangement of computationally costly approaches. We demonstrate that the likelihood to find and bargain the greater part of the hubs putting away the sections of a solitary record is to a great degree low. We likewise analyze the execution of the DROPS system with ten different plans. The more elevated amount of security with slight execution overhead was watched.
The advent of Big Data has seen the emergence of new processing and storage challenges. These challenges are often solved by distributed processing. Distributed systems are inherently dynamic and unstable, so it is realistic to expect that some resources will fail during use. Load balancing and task scheduling is an important step in determining the performance of parallel applications. Hence the need to design load balancing algorithms adapted to grid computing. In this paper, we propose a dynamic and hierarchical load balancing strategy at two levels: Intrascheduler load balancing, in order to avoid the use of the large-scale communication network, and interscheduler load balancing, for a load regulation of our whole system. The strategy allows improving the average response time of CLOAK-Reduce application tasks with minimal communication. We first focus on the three performance indicators, namely response time, process latency and running time of MapReduce tasks.
An Exploration of Grid Computing to be Utilized in Teaching and Research at TUEswar Publications
Taiz University (TU) has a hundreds of computing resources on different campuses for use in areas from offices work to general access student labs. However, these resources are not used to their full potential. Grid computing is a technology that is capable to unify these resources and utilize them in very significant way. The difficulties of funding a complete grid computing environment and also, the difficulties of grid tools makes teachers and researchers in TU unable to involve in teaching and research in grid computing or in distributed computing. These problems raised up our awareness to mitigate this problem by build a simple environment for Grid
computing from resources are available in TU and the built environment we can use it for teaching and research.
The objective of this paper is to build, implement and testing a grid computing environment (Globus Toolkit). To achieving this objective we built the hardware and software parts, and configured several basic grid services commands line and web portal. The test result for basic grid services have been indicated that our proposed grid computing model is promising and can use in teaching and research in TU. The paper takes a look at how grid computing is realizing this aim and have created unbelievable opportunities for students, teachers and
researchers at TU in addition the result of this paper will make TU a pilot to the other universities in whole Yemen in field of Grid and distributing computing.
Load balancing in public cloud by division of cloud based on the geographical...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
The past decade has seen increasingly ambitious and successful methods for outsourcing computing. Approaches such as utility computing, on-demand computing, grid computing, software as a service, and cloud computing all seek to free computer applications from the limiting confines of a single computer. Software that thus runs "outside the box" can be more powerful (think Google, TeraGrid), dynamic (think Animoto, caBIG), and collaborative (think FaceBook, myExperiment). It can also be cheaper, due to economies of scale in hardware and software. The combination of new functionality and new economics inspires new applications, reduces barriers to entry for application providers, and in general disrupts the computing ecosystem. I discuss the new applications that outside-the-box computing enables, in both business and science, and the hardware and software architectures that make these new applications possible.
Data Partitioning Technique In Cloud: A Survey On Limitation And BenefitsIJERA Editor
In recent years,increment in the growth and popularity of cloud services has lead the enterprises to an increase in the capability to handle, store and retrieve critical data. This technology access a shared group of configurable computing resources, which are- servers,storage and applications. Cloud computing is a succeeding generation architecture of IT enterprise, which convert the application software and databaseto large data hubs.Data security and storage of data is an essential functionality of cloud services.It allows data storage in the cloud server efficiently without any worry. Cloud services includes request service, wide web access, measured services, just single click away ,easy usage, just pay for the services you use and location independent.All these features poses many security challenges.The data partitioning techniques are used in literature, for privacy conserving and security of data, using third party auditor (TPA). Objective of the current workis to review all available partitioning technique in literature and analyze them. Through this work authors will compare and identify the limitations and benefits of the available and widely used partitioning techniques.
Excellent Manner of Using Secure way of data storage in cloud computingEditor IJMTER
The major challenging issue in Cloud computing is Security. Providing Security is big issue
towards protecting data from third person as well as in Internet. This mainly deals the Security how it is
provided. Various type of services are there to protect our data and Various Services are available in Cloud
Computing to Utilize effective manner as Software as a Service (SaaS), Platform as a Service (PaaS),
Hardware as a Service (HaaS). Cloud computing is the use of computing resources (hardware and
software) that are delivered as a service over Internet network. Cloud Computing moves the Application
software and databases to the large data centres, where the administration of the data and services may not
be fully trustworthy that is in third party here the party has to get certified and authorized. Since Cloud
Computing share distributed resources via network in the open environment thus it makes new security
risks towards the correctness of the data in cloud. I propose in this paper flexibility of data storage
mechanism in the distributed environment by using the homomorphism token generation. In the proposed
system, users need to allow auditing the cloud storage with lightweight communication. While using
Encryption and Decryption methods it is very burden for a single processor. Than the processing
Capabilities can we utilize from Cloud Computing.
Ensuring secure transfer, access and storage over the cloud storageeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Ensuring secure transfer, access and storage over the cloud storageeSAT Journals
Abstract The main concern in today’s growing IT sectors is the storage and maintenance of the data. As the data keep on updating according to the needs of users, there is a huge overhead in the maintenance of the hardware by the company. One of the solutions to this problem is the use of cloud storage for this enormous data. The cloud storage uses the huge data centers, which are remotely located, to store the data. In addition to the easy storage of the data, this huge data center also reduces the cost of maintenance of the data .However this distinct feature of cloud storage leads to many security issues which should be lucidly understood by the IT sectors. One of the emerging security issue would be the integrity of the data stored in the data center i.e. to check whether the cloud provider misuses the data or not .The cloud provider can misuses the data in many ways like they can copy or modify the file .Due to the storage of data on the data center, user is not able to access the data physically thus there should be the way by which user can check the reliability of the data on the cloud. In this paper we provide the scheme to check the reliability of the data and this scheme can be agreed upon by both the user and the cloud provider. Keywords: Cloud Security, Masking, Cloud Storage Security, Data Center
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
Survey on Division and Replication of Data in Cloud for Optimal Performance a...IJSRD
Outsourcing information to an outsider authoritative control, as is done in distributed computing, offers ascend to security concerns. The information trade off may happen because of assaults by different clients and hubs inside of the cloud. Hence, high efforts to establish safety are required to secure information inside of the cloud. On the other hand, the utilized security procedure should likewise consider the advancement of the information recovery time. In this paper, we propose Division and Replication of Data in the Cloud for Optimal Performance and Security (DROPS) that all in all methodologies the security and execution issues. In the DROPS procedure, we partition a record into sections, and reproduce the divided information over the cloud hubs. Each of the hubs stores just a itary part of a specific information record that guarantees that even in the event of a fruitful assault, no important data is uncovered to the assailant. Additionally, the hubs putting away the sections are isolated with certain separation by method for diagram T-shading to restrict an assailant of speculating the areas of the sections. Moreover, the DROPS procedure does not depend on the customary cryptographic procedures for the information security; in this way alleviating the arrangement of computationally costly approaches. We demonstrate that the likelihood to find and bargain the greater part of the hubs putting away the sections of a solitary record is to a great degree low. We likewise analyze the execution of the DROPS system with ten different plans. The more elevated amount of security with slight execution overhead was watched.
The advent of Big Data has seen the emergence of new processing and storage challenges. These challenges are often solved by distributed processing. Distributed systems are inherently dynamic and unstable, so it is realistic to expect that some resources will fail during use. Load balancing and task scheduling is an important step in determining the performance of parallel applications. Hence the need to design load balancing algorithms adapted to grid computing. In this paper, we propose a dynamic and hierarchical load balancing strategy at two levels: Intrascheduler load balancing, in order to avoid the use of the large-scale communication network, and interscheduler load balancing, for a load regulation of our whole system. The strategy allows improving the average response time of CLOAK-Reduce application tasks with minimal communication. We first focus on the three performance indicators, namely response time, process latency and running time of MapReduce tasks.
An Exploration of Grid Computing to be Utilized in Teaching and Research at TUEswar Publications
Taiz University (TU) has a hundreds of computing resources on different campuses for use in areas from offices work to general access student labs. However, these resources are not used to their full potential. Grid computing is a technology that is capable to unify these resources and utilize them in very significant way. The difficulties of funding a complete grid computing environment and also, the difficulties of grid tools makes teachers and researchers in TU unable to involve in teaching and research in grid computing or in distributed computing. These problems raised up our awareness to mitigate this problem by build a simple environment for Grid
computing from resources are available in TU and the built environment we can use it for teaching and research.
The objective of this paper is to build, implement and testing a grid computing environment (Globus Toolkit). To achieving this objective we built the hardware and software parts, and configured several basic grid services commands line and web portal. The test result for basic grid services have been indicated that our proposed grid computing model is promising and can use in teaching and research in TU. The paper takes a look at how grid computing is realizing this aim and have created unbelievable opportunities for students, teachers and
researchers at TU in addition the result of this paper will make TU a pilot to the other universities in whole Yemen in field of Grid and distributing computing.
Load balancing in public cloud by division of cloud based on the geographical...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
The past decade has seen increasingly ambitious and successful methods for outsourcing computing. Approaches such as utility computing, on-demand computing, grid computing, software as a service, and cloud computing all seek to free computer applications from the limiting confines of a single computer. Software that thus runs "outside the box" can be more powerful (think Google, TeraGrid), dynamic (think Animoto, caBIG), and collaborative (think FaceBook, myExperiment). It can also be cheaper, due to economies of scale in hardware and software. The combination of new functionality and new economics inspires new applications, reduces barriers to entry for application providers, and in general disrupts the computing ecosystem. I discuss the new applications that outside-the-box computing enables, in both business and science, and the hardware and software architectures that make these new applications possible.
Data Partitioning Technique In Cloud: A Survey On Limitation And BenefitsIJERA Editor
In recent years,increment in the growth and popularity of cloud services has lead the enterprises to an increase in the capability to handle, store and retrieve critical data. This technology access a shared group of configurable computing resources, which are- servers,storage and applications. Cloud computing is a succeeding generation architecture of IT enterprise, which convert the application software and databaseto large data hubs.Data security and storage of data is an essential functionality of cloud services.It allows data storage in the cloud server efficiently without any worry. Cloud services includes request service, wide web access, measured services, just single click away ,easy usage, just pay for the services you use and location independent.All these features poses many security challenges.The data partitioning techniques are used in literature, for privacy conserving and security of data, using third party auditor (TPA). Objective of the current workis to review all available partitioning technique in literature and analyze them. Through this work authors will compare and identify the limitations and benefits of the available and widely used partitioning techniques.
Excellent Manner of Using Secure way of data storage in cloud computingEditor IJMTER
The major challenging issue in Cloud computing is Security. Providing Security is big issue
towards protecting data from third person as well as in Internet. This mainly deals the Security how it is
provided. Various type of services are there to protect our data and Various Services are available in Cloud
Computing to Utilize effective manner as Software as a Service (SaaS), Platform as a Service (PaaS),
Hardware as a Service (HaaS). Cloud computing is the use of computing resources (hardware and
software) that are delivered as a service over Internet network. Cloud Computing moves the Application
software and databases to the large data centres, where the administration of the data and services may not
be fully trustworthy that is in third party here the party has to get certified and authorized. Since Cloud
Computing share distributed resources via network in the open environment thus it makes new security
risks towards the correctness of the data in cloud. I propose in this paper flexibility of data storage
mechanism in the distributed environment by using the homomorphism token generation. In the proposed
system, users need to allow auditing the cloud storage with lightweight communication. While using
Encryption and Decryption methods it is very burden for a single processor. Than the processing
Capabilities can we utilize from Cloud Computing.
ABSTRACT
Cloud computing utilizes large scale computing infrastructure that has been radically changing the IT landscape enabling remote access to computing resources with low service cost, high scalability , availability and accessibility. Serving tasks from multiple users where the tasks are of different characteristics with variation in the requirement of computing power may cause under or over utilization of resources.Therefore maintaining such mega-scale datacenter requires efficient resource management procedure to increase resource utilization. However, while maintaining efficiency in service provisioning it is necessary to ensure the maximization of profit for the cloud providers. Most of the current research works aims at how providers can offer efficient service provisioning to the user and improving system performance. There are comparatively fewer specific works regarding resource management which also deals with the economic section that considers profit maximization for the provider. In this paper we represent a model that deals with both efficient resource utilization and pricing of the resources. The joint resource management model combines the work of user assignment, task scheduling and load balancing on the fact of CPU power endorsement. We propose four algorithms respectively for user assignment, task scheduling, load balancing and pricing that works on group based resources offering reduction in task execution time(56.3%),activated physical machines(41.44%),provisioning cost(23%) . The cost is calculated over a time interval involving the number of served customer at this time and the amount of resources used within this time
GROUP BASED RESOURCE MANAGEMENT AND PRICING MODEL IN CLOUD COMPUTINGijcsit
Cloud computing utilizes large scale computing infrastructure that has been radically changing the IT landscape enabling remote access to computing resources with low service cost, high scalability , availability and accessibility. Serving tasks from multiple users where the tasks are of different characteristics with variation in the requirement of computing power may cause under or over utilization of resources.Therefore maintaining such mega-scale datacenter requires efficient resource management procedure to increase resource utilization. However, while maintaining efficiency in service provisioning it is necessary to ensure the maximization of profit for the cloud providers. Most of the current research works aims at how providers can offer efficient service provisioning to the user and improving system performance. There are comparatively fewer specific works regarding resource management which also deals with the economic section that considers profit maximization for the provider. In this paper we represent a model that deals with both efficient resource utilization and pricing of the resources. The joint resource management model combines the work of user assignment, task scheduling and load balancing on the fact of CPU power endorsement. We propose four algorithms respectively for user assignment, task scheduling, load balancing and pricing that works on group based resources offering reduction in task execution time(56.3%),activated physical machines(41.44%),provisioning cost(23%) . The cost is calculated over a time interval involving the number of served customer at this time and the amount of resources used within this time.
A Secure Cloud Storage System with Data Forwarding using Proxy Re-encryption ...IJTET Journal
Cloud computing provides the facility to access shared resources and common support which contributes services on
demand over the network to perform operations that meet changing business needs. A cloud storage system, consisting of a collection
of storage servers, affords long-term storage services over the internet. Storing the data in a third party cloud system cause serious
concern over data confidentiality, without considering the local infrastructure limitations, the cloud services allow the user to enjoy the
cloud applications. As the different users may be working in the collaborative relationship, the data sharing becomes significant to
achieve productive benefit during the data accessing. The existing security system only focuses on the authentication; it shows that
user’s private data cannot be accessed by the fake users. To address the above cloud storage privacy issue shared authority based
privacy-preserving authentication protocol is used. In the SAPA, the shared access authority is achieved by anonymous access request
and privacy consideration, attribute based access control allows the user to access their own data fields. To provide the data sharing
among the multiple users proxy re-encryption scheme is applied by the cloud server. The privacy-preserving data access authority
sharing is attractive for multi-user collaborative cloud applications.
NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...ijccsa
Cloud collaboration is an emerging technology which enables sharing of computer files using cloud
computing. Here the cloud resources are assembled and cloud services are provided using these resources.
Cloud collaboration technologies are allowing users to share documents. Resource allocation in the cloud
is challenging because resources offer different Quality of Service (QoS) and services running on these
resources are risky for user demands. We propose a solution for resource allocation based on multi
attribute QoS Scoring considering parameters such as distance to the resource from user site, reputation of
the resource, task completion time, task completion ratio, and load at the resource. The proposed algorithm
referred to as Multi Attribute QoS scoring (MAQS) uses Neuro Fuzzy system. We have also included a
speculative manager to handle fault tolerance. In this paper it is shown that the proposed algorithm
perform better than others including power trust reputation based algorithms and harmony method which
use single attribute to compute the reputation score of each resource allocated.
Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...neirew J
Cloud collaboration is an emerging technology which enables sharing of computer files using cloud
computing. Here the cloud resources are assembled and cloud services are provided using these resources.
Cloud collaboration technologies are allowing users to share documents. Resource allocation in the cloud
is challenging because resources offer different Quality of Service (QoS) and services running on these
resources are risky for user demands. We propose a solution for resource allocation based on multi
attribute QoS Scoring considering parameters such as distance to the resource from user site, reputation of
the resource, task completion time, task completion ratio, and load at the resource. The proposed algorithm
referred to as Multi Attribute QoS scoring (MAQS) uses Neuro Fuzzy system. We have also included a
speculative manager to handle fault tolerance. In this paper it is shown that the proposed algorithm
perform better than others including power trust reputation based algorithms and harmony method which
use single attribute to compute the reputation score of each resource allocated.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
A Secure Cloud Storage System with Data Forwarding using Proxy Re-encryption ...IJTET Journal
Cloud computing provides the facility to access shared resources and common support which contributes services on demand over the network to perform operations that meet changing business needs. A cloud storage system, consisting of a collection of storage servers, affords long-term storage services over the internet. Storing the data in a third party cloud system cause serious concern over data confidentiality, without considering the local infrastructure limitations, the cloud services allow the user to enjoy the cloud applications. As the different users may be working in the collaborative relationship, the data sharing becomes significant to achieve productive benefit during the data accessing. The existing security system only focuses on the authentication; it shows that user’s private data cannot be accessed by the fake users. To address the above cloud storage privacy issue shared authority based privacy-preserving authentication protocol is used. In the SAPA, the shared access authority is achieved by anonymous access request and privacy consideration, attribute based access control allows the user to access their own data fields. To provide the data sharing among the multiple users proxy re-encryption scheme is applied by the cloud server. The privacy-preserving data access authority sharing is attractive for multi-user collaborative cloud applications.
An Efficient Cloud Scheduling Algorithm for the Conservation of Energy throug...IJECEIAES
Method of broadcasting is the well known operation that is used for providing support to different computing protocols in cloud computing. Attaining energy efficiency is one of the prominent challenges, that is quite significant in the scheduling process that is used in cloud computing as, there are fixed limits that have to be met by the system. In this research paper, we are particularly focusing on the cloud server maintenance and scheduling process and to do so, we are using the interactive broadcasting energy efficient computing technique along with the cloud computing server. Additionally, the remote host machines used for cloud services are dissipating more power and with that they are consuming more and more energy. The effect of the power consumption is one of the main factors for determining the cost of the computing resources. With the idea of using the avoidance technology for assigning the data center resources that dynamically depend on the application demands and supports the cloud computing with the optimization of the servers in use.
Hybrid Based Resource Provisioning in CloudEditor IJCATR
The data centres and energy consumption characteristics of the various machines are often noted with different capacities.
The public cloud workloads of different priorities and performance requirements of various applications when analysed we had noted
some invariant reports about cloud. The Cloud data centres become capable of sensing an opportunity to present a different program.
In out proposed work, we are using a hybrid method for resource provisioning in data centres. This method is used to allocate the
resources at the working conditions and also for the energy stored in the power consumptions. Proposed method is used to allocate the
process behind the cloud storage.
Cloud Computing is the revolution in current generation IT enterprise. Cloud computing displaces database and application software to the large data centres, where the management of services and data may not be predictable, where as the conventional solutions, for IT services are under proper logical, physical and personal controls. This aspect attribute, however comprises different security challenges which have not been well understood. It concentrates on cloud data storage security which has always been an important aspect of quality of service (QOS). In this paper, we designed and simulated an adaptable and efficient scheme to guarantee the correctness of user data stored in the cloud and also with some prominent features. Homomorphic token is used for distributed verification of erasure – coded data. By using this scheme, we can identify misbehaving servers. In spite of past works, our scheme supports effective and secure dynamic operations on data blocks such as data insertion, deletion and modification. In contrast to traditional solutions, where the IT services are under proper physical, logical and personnel controls, cloud computing moves the application software and databases to the large data centres, where the data management and services may not be absolutely truthful. This effective security and performance analysis describes that the proposed scheme is extremely flexible against malicious data modification, convoluted failures and server clouding attacks.
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTINGijcsit
Cloud Computing is the revolution in current generation IT enterprise. Cloud computing displaces
database and application software to the large data centres, where the management of services and data
may not be predictable, where as the conventional solutions, for IT services are under proper logical,
physical and personal controls. This aspect attribute, however comprises different security challenges
which have not been well understood. It concentrates on cloud data storage security which has always been
an important aspect of quality of service (QOS). In this paper, we designed and simulated an adaptable and
efficient scheme to guarantee the correctness of user data stored in the cloud and also with some prominent
features. Homomorphic token is used for distributed verification of erasure – coded data. By using this
scheme, we can identify misbehaving servers. In spite of past works, our scheme supports effective and
secure dynamic operations on data blocks such as data insertion, deletion and modification. In contrast to
traditional solutions, where the IT services are under proper physical, logical and personnel controls,
cloud computing moves the application software and databases to the large data centres, where the data
management and services may not be absolutely truthful. This effective security and performance analysis
describes that the proposed scheme is extremely flexible against malicious data modification, convoluted
failures and server clouding attacks.
Cloud Computing is the revolution in current generation IT enterprise. Cloud computing displaces database and application software to the large data centres, where the management of services and data may not be predictable, where as the conventional solutions, for IT services are under proper logical, physical and personal controls. This aspect attribute, however comprises different security challenges which have not been well understood. It concentrates on cloud data storage security which has always been an important aspect of quality of service (QOS). In this paper, we designed and simulated an adaptable and efficient scheme to guarantee the correctness of user data stored in the cloud and also with some prominent features. Homomorphic token is used for distributed verification of erasure – coded data. By using this scheme, we can identify misbehaving servers. In spite of past works, our scheme supports effective and secure dynamic operations on data blocks such as data insertion, deletion and modification. In contrast to traditional solutions, where the IT services are under proper physical, logical and personnel controls, cloud computing moves the application software and databases to the large data centres, where the data management and services may not be absolutely truthful. This effective security and performance analysis describes that the proposed scheme is extremely flexible against malicious data modification, convoluted failures and server clouding attacks.
Virtual Machine Migration and Allocation in Cloud Computing: A Reviewijtsrd
Cloud computing is an emerging computing technology that maintains computational resources on large data centers and accessed through internet, rather than on local computers. VM migration provides the capability to balance the load, system maintenance, etc. Virtualization technology gives power to cloud computing. The virtual machine migration techniques can be divided into two categories that is pre copy and post copy approach. The process to move running applications or VMs from one physical machine to another is known as VM migration. In migration process the processor state, storage, memory and network connection are moved from one host to another.. Two important performance metrics are downtime and total migration time that the users care about most, because these metrics deals with service degradation and the time during which the service is unavailable. This paper focus on the analysis of live VM migration Techniques in cloud computing. Khushbu Singh Chandel | Dr. Avinash Sharma "Virtual Machine Migration and Allocation in Cloud Computing: A Review" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-1 , December 2019, URL: https://www.ijtsrd.com/papers/ijtsrd29556.pdfPaper URL: https://www.ijtsrd.com/computer-science/computer-network/29556/virtual-machine-migration-and-allocation-in-cloud-computing-a-review/khushbu-singh-chandel
Effective & Flexible Cryptography Based Scheme for Ensuring User`s Data Secur...ijsrd.com
Cloud computing has been envisioned as the next-generation architecture of IT enterprise. In contrast to traditional solutions, where the IT services are under proper physical, logical and personnel controls, cloud computing moves the application software and databases to the large data centers, where the management of the data and services may not be fully trustworthy. This unique attribute, however, poses many new security challenges which have not been well understood. In this article, we focus on cloud data storage security, which has always been an important aspect of quality of service. To ensure the correctness of users' data in the cloud, we propose an effective and flexible cryptography based scheme. Extensive security and performance analysis shows that the proposed scheme is highly efficient and resilient against malicious data modification attack.
Text Mining in Digital Libraries using OKAPI BM25 ModelEditor IJCATR
The emergence of the internet has made vast amounts of information available and easily accessible online. As a result, most libraries have digitized their content in order to remain relevant to their users and to keep pace with the advancement of the internet. However, these digital libraries have been criticized for using inefficient information retrieval models that do not perform relevance ranking to the retrieved results. This paper proposed the use of OKAPI BM25 model in text mining so as means of improving relevance ranking of digital libraries. Okapi BM25 model was selected because it is a probability-based relevance ranking algorithm. A case study research was conducted and the model design was based on information retrieval processes. The performance of Boolean, vector space, and Okapi BM25 models was compared for data retrieval. Relevant ranked documents were retrieved and displayed at the OPAC framework search page. The results revealed that Okapi BM 25 outperformed Boolean model and Vector Space model. Therefore, this paper proposes the use of Okapi BM25 model to reward terms according to their relative frequencies in a document so as to improve the performance of text mining in digital libraries.
Green Computing, eco trends, climate change, e-waste and eco-friendlyEditor IJCATR
This study focused on the practice of using computing resources more efficiently while maintaining or increasing overall performance. Sustainable IT services require the integration of green computing practices such as power management, virtualization, improving cooling technology, recycling, electronic waste disposal, and optimization of the IT infrastructure to meet sustainability requirements. Studies have shown that costs of power utilized by IT departments can approach 50% of the overall energy costs for an organization. While there is an expectation that green IT should lower costs and the firm’s impact on the environment, there has been far less attention directed at understanding the strategic benefits of sustainable IT services in terms of the creation of customer value, business value and societal value. This paper provides a review of the literature on sustainable IT, key areas of focus, and identifies a core set of principles to guide sustainable IT service design.
Policies for Green Computing and E-Waste in NigeriaEditor IJCATR
Computers today are an integral part of individuals’ lives all around the world, but unfortunately these devices are toxic to the environment given the materials used, their limited battery life and technological obsolescence. Individuals are concerned about the hazardous materials ever present in computers, even if the importance of various attributes differs, and that a more environment -friendly attitude can be obtained through exposure to educational materials. In this paper, we aim to delineate the problem of e-waste in Nigeria and highlight a series of measures and the advantage they herald for our country and propose a series of action steps to develop in these areas further. It is possible for Nigeria to have an immediate economic stimulus and job creation while moving quickly to abide by the requirements of climate change legislation and energy efficiency directives. The costs of implementing energy efficiency and renewable energy measures are minimal as they are not cash expenditures but rather investments paid back by future, continuous energy savings.
Performance Evaluation of VANETs for Evaluating Node Stability in Dynamic Sce...Editor IJCATR
Vehicular ad hoc networks (VANETs) are a favorable area of exploration which empowers the interconnection amid the movable vehicles and between transportable units (vehicles) and road side units (RSU). In Vehicular Ad Hoc Networks (VANETs), mobile vehicles can be organized into assemblage to promote interconnection links. The assemblage arrangement according to dimensions and geographical extend has serious influence on attribute of interaction .Vehicular ad hoc networks (VANETs) are subclass of mobile Ad-hoc network involving more complex mobility patterns. Because of mobility the topology changes very frequently. This raises a number of technical challenges including the stability of the network .There is a need for assemblage configuration leading to more stable realistic network. The paper provides investigation of various simulation scenarios in which cluster using k-means algorithm are generated and their numbers are varied to find the more stable configuration in real scenario of road.
Optimum Location of DG Units Considering Operation ConditionsEditor IJCATR
The optimal sizing and placement of Distributed Generation units (DG) are becoming very attractive to researchers these days. In this paper a two stage approach has been used for allocation and sizing of DGs in distribution system with time varying load model. The strategic placement of DGs can help in reducing energy losses and improving voltage profile. The proposed work discusses time varying loads that can be useful for selecting the location and optimizing DG operation. The method has the potential to be used for integrating the available DGs by identifying the best locations in a power system. The proposed method has been demonstrated on 9-bus test system.
Analysis of Comparison of Fuzzy Knn, C4.5 Algorithm, and Naïve Bayes Classifi...Editor IJCATR
Early detection of diabetes mellitus (DM) can prevent or inhibit complication. There are several laboratory test that must be done to detect DM. The result of this laboratory test then converted into data training. Data training used in this study generated from UCI Pima Database with 6 attributes that were used to classify positive or negative diabetes. There are various classification methods that are commonly used, and in this study three of them were compared, which were fuzzy KNN, C4.5 algorithm and Naïve Bayes Classifier (NBC) with one identical case. The objective of this study was to create software to classify DM using tested methods and compared the three methods based on accuracy, precision, and recall. The results showed that the best method was Fuzzy KNN with average and maximum accuracy reached 96% and 98%, respectively. In second place, NBC method had respective average and maximum accuracy of 87.5% and 90%. Lastly, C4.5 algorithm had average and maximum accuracy of 79.5% and 86%, respectively.
Web Scraping for Estimating new Record from Source SiteEditor IJCATR
Study in the Competitive field of Intelligent, and studies in the field of Web Scraping, have a symbiotic relationship mutualism. In the information age today, the website serves as a main source. The research focus is on how to get data from websites and how to slow down the intensity of the download. The problem that arises is the website sources are autonomous so that vulnerable changes the structure of the content at any time. The next problem is the system intrusion detection snort installed on the server to detect bot crawler. So the researchers propose the use of the methods of Mining Data Records and the method of Exponential Smoothing so that adaptive to changes in the structure of the content and do a browse or fetch automatically follow the pattern of the occurrences of the news. The results of the tests, with the threshold 0.3 for MDR and similarity threshold score 0.65 for STM, using recall and precision values produce f-measure average 92.6%. While the results of the tests of the exponential estimation smoothing using ? = 0.5 produces MAE 18.2 datarecord duplicate. It slowed down to 3.6 datarecord from 21.8 datarecord results schedule download/fetch fix in an average time of occurrence news.
Evaluating Semantic Similarity between Biomedical Concepts/Classes through S...Editor IJCATR
Most of the existing semantic similarity measures that use ontology structure as their primary source can measure semantic similarity between concepts/classes using single ontology. The ontology-based semantic similarity techniques such as structure-based semantic similarity techniques (Path Length Measure, Wu and Palmer’s Measure, and Leacock and Chodorow’s measure), information content-based similarity techniques (Resnik’s measure, Lin’s measure), and biomedical domain ontology techniques (Al-Mubaid and Nguyen’s measure (SimDist)) were evaluated relative to human experts’ ratings, and compared on sets of concepts using the ICD-10 “V1.0” terminology within the UMLS. The experimental results validate the efficiency of the SemDist technique in single ontology, and demonstrate that SemDist semantic similarity techniques, compared with the existing techniques, gives the best overall results of correlation with experts’ ratings.
Semantic Similarity Measures between Terms in the Biomedical Domain within f...Editor IJCATR
The techniques and tests are tools used to define how measure the goodness of ontology or its resources. The similarity between biomedical classes/concepts is an important task for the biomedical information extraction and knowledge discovery. However, most of the semantic similarity techniques can be adopted to be used in the biomedical domain (UMLS). Many experiments have been conducted to check the applicability of these measures. In this paper, we investigate to measure semantic similarity between two terms within single ontology or multiple ontologies in ICD-10 “V1.0” as primary source, and compare my results to human experts score by correlation coefficient.
A Strategy for Improving the Performance of Small Files in Openstack Swift Editor IJCATR
This is an effective way to improve the storage access performance of small files in Openstack Swift by adding an aggregate storage module. Because Swift will lead to too much disk operation when querying metadata, the transfer performance of plenty of small files is low. In this paper, we propose an aggregated storage strategy (ASS), and implement it in Swift. ASS comprises two parts which include merge storage and index storage. At the first stage, ASS arranges the write request queue in chronological order, and then stores objects in volumes. These volumes are large files that are stored in Swift actually. During the short encounter time, the object-to-volume mapping information is stored in Key-Value store at the second stage. The experimental results show that the ASS can effectively improve Swift's small file transfer performance.
Integrated System for Vehicle Clearance and RegistrationEditor IJCATR
Efficient management and control of government's cash resources rely on government banking arrangements. Nigeria, like many low income countries, employed fragmented systems in handling government receipts and payments. Later in 2016, Nigeria implemented a unified structure as recommended by the IMF, where all government funds are collected in one account would reduce borrowing costs, extend credit and improve government's fiscal policy among other benefits to government. This situation motivated us to embark on this research to design and implement an integrated system for vehicle clearance and registration. This system complies with the new Treasury Single Account policy to enable proper interaction and collaboration among five different level agencies (NCS, FRSC, SBIR, VIO and NPF) saddled with vehicular administration and activities in Nigeria. Since the system is web based, Object Oriented Hypermedia Design Methodology (OOHDM) is used. Tools such as Php, JavaScript, css, html, AJAX and other web development technologies were used. The result is a web based system that gives proper information about a vehicle starting from the exact date of importation to registration and renewal of licensing. Vehicle owner information, custom duty information, plate number registration details, etc. will also be efficiently retrieved from the system by any of the agencies without contacting the other agency at any point in time. Also number plate will no longer be the only means of vehicle identification as it is presently the case in Nigeria, because the unified system will automatically generate and assigned a Unique Vehicle Identification Pin Number (UVIPN) on payment of duty in the system to the vehicle and the UVIPN will be linked to the various agencies in the management information system.
Assessment of the Efficiency of Customer Order Management System: A Case Stu...Editor IJCATR
The Supermarket Management System deals with the automation of buying and selling of good and services. It includes both sales and purchase of items. The project Supermarket Management System is to be developed with the objective of making the system reliable, easier, fast, and more informative.
Energy-Aware Routing in Wireless Sensor Network Using Modified Bi-Directional A*Editor IJCATR
Energy is a key component in the Wireless Sensor Network (WSN)[1]. The system will not be able to run according to its function without the availability of adequate power units. One of the characteristics of wireless sensor network is Limitation energy[2]. A lot of research has been done to develop strategies to overcome this problem. One of them is clustering technique. The popular clustering technique is Low Energy Adaptive Clustering Hierarchy (LEACH)[3]. In LEACH, clustering techniques are used to determine Cluster Head (CH), which will then be assigned to forward packets to Base Station (BS). In this research, we propose other clustering techniques, which utilize the Social Network Analysis approach theory of Betweeness Centrality (BC) which will then be implemented in the Setup phase. While in the Steady-State phase, one of the heuristic searching algorithms, Modified Bi-Directional A* (MBDA *) is implemented. The experiment was performed deploy 100 nodes statically in the 100x100 area, with one Base Station at coordinates (50,50). To find out the reliability of the system, the experiment to do in 5000 rounds. The performance of the designed routing protocol strategy will be tested based on network lifetime, throughput, and residual energy. The results show that BC-MBDA * is better than LEACH. This is influenced by the ways of working LEACH in determining the CH that is dynamic, which is always changing in every data transmission process. This will result in the use of energy, because they always doing any computation to determine CH in every transmission process. In contrast to BC-MBDA *, CH is statically determined, so it can decrease energy usage.
Security in Software Defined Networks (SDN): Challenges and Research Opportun...Editor IJCATR
In networks, the rapidly changing traffic patterns of search engines, Internet of Things (IoT) devices, Big Data and data centers has thrown up new challenges for legacy; existing networks; and prompted the need for a more intelligent and innovative way to dynamically manage traffic and allocate limited network resources. Software Defined Network (SDN) which decouples the control plane from the data plane through network vitalizations aims to address these challenges. This paper has explored the SDN architecture and its implementation with the OpenFlow protocol. It has also assessed some of its benefits over traditional network architectures, security concerns and how it can be addressed in future research and related works in emerging economies such as Nigeria.
Measure the Similarity of Complaint Document Using Cosine Similarity Based on...Editor IJCATR
Report handling on "LAPOR!" (Laporan, Aspirasi dan Pengaduan Online Rakyat) system depending on the system administrator who manually reads every incoming report [3]. Read manually can lead to errors in handling complaints [4] if the data flow is huge and grows rapidly, it needs at least three days to prepare a confirmation and it sensitive to inconsistencies [3]. In this study, the authors propose a model that can measure the identities of the Query (Incoming) with Document (Archive). The authors employed Class-Based Indexing term weighting scheme, and Cosine Similarities to analyse document similarities. CoSimTFIDF, CoSimTFICF and CoSimTFIDFICF values used in classification as feature for K-Nearest Neighbour (K-NN) classifier. The optimum result evaluation is pre-processing employ 75% of training data ratio and 25% of test data with CoSimTFIDF feature. It deliver a high accuracy 84%. The k = 5 value obtain high accuracy 84.12%
Hangul Recognition Using Support Vector MachineEditor IJCATR
The recognition of Hangul Image is more difficult compared with that of Latin. It could be recognized from the structural arrangement. Hangul is arranged from two dimensions while Latin is only from the left to the right. The current research creates a system to convert Hangul image into Latin text in order to use it as a learning material on reading Hangul. In general, image recognition system is divided into three steps. The first step is preprocessing, which includes binarization, segmentation through connected component-labeling method, and thinning with Zhang Suen to decrease some pattern information. The second is receiving the feature from every single image, whose identification process is done through chain code method. The third is recognizing the process using Support Vector Machine (SVM) with some kernels. It works through letter image and Hangul word recognition. It consists of 34 letters, each of which has 15 different patterns. The whole patterns are 510, divided into 3 data scenarios. The highest result achieved is 94,7% using SVM kernel polynomial and radial basis function. The level of recognition result is influenced by many trained data. Whilst the recognition process of Hangul word applies to the type 2 Hangul word with 6 different patterns. The difference of these patterns appears from the change of the font type. The chosen fonts for data training are such as Batang, Dotum, Gaeul, Gulim, Malgun Gothic. Arial Unicode MS is used to test the data. The lowest accuracy is achieved through the use of SVM kernel radial basis function, which is 69%. The same result, 72 %, is given by the SVM kernel linear and polynomial.
Application of 3D Printing in EducationEditor IJCATR
This paper provides a review of literature concerning the application of 3D printing in the education system. The review identifies that 3D Printing is being applied across the Educational levels [1] as well as in Libraries, Laboratories, and Distance education systems. The review also finds that 3D Printing is being used to teach both students and trainers about 3D Printing and to develop 3D Printing skills.
Survey on Energy-Efficient Routing Algorithms for Underwater Wireless Sensor ...Editor IJCATR
In underwater environment, for retrieval of information the routing mechanism is used. In routing mechanism there are three to four types of nodes are used, one is sink node which is deployed on the water surface and can collect the information, courier/super/AUV or dolphin powerful nodes are deployed in the middle of the water for forwarding the packets, ordinary nodes are also forwarder nodes which can be deployed from bottom to surface of the water and source nodes are deployed at the seabed which can extract the valuable information from the bottom of the sea. In underwater environment the battery power of the nodes is limited and that power can be enhanced through better selection of the routing algorithm. This paper focuses the energy-efficient routing algorithms for their routing mechanisms to prolong the battery power of the nodes. This paper also focuses the performance analysis of the energy-efficient algorithms under which we can examine the better performance of the route selection mechanism which can prolong the battery power of the node
Comparative analysis on Void Node Removal Routing algorithms for Underwater W...Editor IJCATR
The designing of routing algorithms faces many challenges in underwater environment like: propagation delay, acoustic channel behaviour, limited bandwidth, high bit error rate, limited battery power, underwater pressure, node mobility, localization 3D deployment, and underwater obstacles (voids). This paper focuses the underwater voids which affects the overall performance of the entire network. The majority of the researchers have used the better approaches for removal of voids through alternate path selection mechanism but still research needs improvement. This paper also focuses the architecture and its operation through merits and demerits of the existing algorithms. This research article further focuses the analytical method of the performance analysis of existing algorithms through which we found the better approach for removal of voids
Decay Property for Solutions to Plate Type Equations with Variable CoefficientsEditor IJCATR
In this paper we consider the initial value problem for a plate type equation with variable coefficients and memory in
1 n R n ), which is of regularity-loss property. By using spectrally resolution, we study the pointwise estimates in the spectral
space of the fundamental solution to the corresponding linear problem. Appealing to this pointwise estimates, we obtain the global
existence and the decay estimates of solutions to the semilinear problem by employing the fixed point theorem
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Let's dive deeper into the world of ODC! Ricardo Alves (OutSystems) will join us to tell all about the new Data Fabric. After that, Sezen de Bruijn (OutSystems) will get into the details on how to best design a sturdy architecture within ODC.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...
Dynamic Resource Provisioning with Authentication in Distributed Database
1. International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 791 - 794, 2014, ISSN:- 2319–8656
www.ijcat.com 791
Dynamic Resource Provisioning with Authentication in
Distributed Database
Anju Aravind K
Shree Venkateshwara Hi-Tech
Engineering College
Gobi, India
Dr. T. Senthil Prakash
Shree Venkateshwara Hi-Tech
Engineering College
Gobi, India
M. Rajesh
Shree Venkateshwara Hi-Tech
Engineering College
Gobi, India
Abstract: Data center have the largest consumption amounts of energy in sharing the power. The public cloud workloads of different
priorities and performance requirements of various applications [4]
. Cloud data center have capable of sensing an opportunity to present
different programs. In my proposed construction and the name of the security level of imperturbable privacy leakage rarely distributed
cloud system to deal with the persistent characteristics there is a substantial increases and information that can be used to augment the
profit, retrenchment overhead or both. Data Mining Analysis of data from different perspectives and summarizing it into useful
information is a process. Three empirical algorithms have been proposed assignments estimate the ratios are dissected theoretically and
compared using real Internet latency data recital of testing methods.
Keywords: Mining, MD5, green computing, workload imitation, power consumption
1. INTRODUCTION
Most data centers today have a three or four tier hierarchical
networking structure. Three tier network architectures were
designed around client-server applications and single-purpose
of application server. Client and server applications are
caused traffic to flow primarily in patterns: from a server up
to the data center core, to the environment core where it
moves out to the internet. These large core switches usually
contain the vast majority of the intelligence in the network.
The cost of builds and operates the large computer platform
and a focus on service quality and cost-efficient driving will
require cost estimation and on the capacity of processing and
storage.
The cloud consists of
Private Cloud: The infrastructure is provisioned for exclusive
use by a single organization.
Public Cloud: The infrastructure is provisioned for open use
by the general public.
Hybrid Cloud: There is a system infrastructure of two or more
distinct cloud Infrastructures (private, community, or public)
that remain unique entities, but are bound together by
standardized or proprietary technology that enables data and
application portability.
Community Cloud: Consumers who share the concerns of the
infrastructure for personal use by the provision of a particular
community.
While Cloud computing is not equivalent to virtualization,
virtualization technology is heavily used to operate a cloud
environment system. In the virtualization have one of the host
that ran a single operating system now has the ability to run
multiple guest operating systems as virtual machines [VMs].
The VMs are created for fast and easily data storage in a cloud
environment. The infrastructure environment is invisible for
abstracted from the consumer. The computing of firmware
system to provide virtual machine and it allows to operating
directly on underlying hardware but within specified
constraint. It is the software that manages communications
between the physical server memory and CPU or processing
capability and the virtual machines that are running. The
software allows VMs to be quickly provisioned or
decommissioned. So the data center contains delay process,
security process, mining process and also cost efficiency.
First, in data centre have different perspectives and useful
information of data mining [DM]. In DM information that can
be used to increase the revenue and it is the process of
summarizing the costs or both are reduced. It allows the user
to analyze the data with different dimensions or angles. To
categorize and summarize the relationships are identified the
data. The proposed method of Support Vector Machine for
data mining system [DMS] is designed to take the advantage
of powerful processors and shared pools to the calculation is
performed using the message passing paradigm for data that is
distributed across processors. The calculation results are then
collected, and the process is repeated with new data processor.
Second, data center security concerns to secure all aspects of
cloud data [2]
. Many of these features are not unique in cloud
system: Irrespective of the data stored on it is vulnerable
attack. Therefore the cloud computing have security, access
control, malware protection structures, reducing attack
surfaces, safety design and implementation of all data
including computing security. In this proposed method having
MD5 is used to secure the data in data center.
Third, resource provisioning in cloud computing over the last
few years, has emerged as new computing model to allowing
utility based delivery of services to end users[1]
. Cloud
computing relies on virtualization technologies to provide on-
demand resources according to the end user needs but
problems with resource allocation pose an issue for the
management of several data centers. In my proposed gossip
protocol of green computing based virtualization can be used
to increase energy efficiency in substrate networks, allowing
consolidation through the virtual hosting of different
resources on the same substrate resource. Migrating resources
virtually allows the network to balance its overall energy load
and reduce the total power consumption of the data center.
The dynamic provisioning concept is used to allocate the
resources dynamically.
2. International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 791 - 794, 2014, ISSN:- 2319–8656
www.ijcat.com 792
2. RELATED WORK
2.1 Datacenter
The datacenter is the collection of servers, the cloud
computing and information technology between physical
servers to migrate the cloud computing services, virtualized
data center, emerging paradigm changes and executives of the
largest independent provider [7]
. VM virtualizations and
migration capabilities to integrate their computer services and
the minimum number of physical servers used to process the
data center such as mining processing, security purpose, load
balancing, server establishment, online maintenance,
proactive fault tolerance and VM migration. Information’s are
use of the cloud computing to provide services for worldwide
users. The consumer scientific controls the commercial
domains hosting gives pervasive applications have costs and
environmental contribution to the carbon footprint of data
centers and cloud hosting applications that consume huge
amounts of electrical energy. Therefore, reduces the
environmental impact of cloud computing with the help of
green computing and gossip protocol.
2.2 Database
Database systems serving cloud platforms must serve large
numbers of applications. In addition to managing tolerant with
small data footprints, different textures and shapes with
variable load patterns such data platforms must minimize their
operating costs by efficient resource sharing. The persistent
database have the files are stored in network attached storage.
VM migrate the database cache and the state of active
transactions to ensure minimal impact on transaction
execution while allowing transactions active during migration
to continue executions and also guarantee the concurrency
while ensuring correctness during failures [8]
.
2.3 Resource Allocation
Dynamic Resource Management for Cloud Computing
paradigm is an active area of research. The cost varies are
considerably depending upon the configuration of resources
by using them [6]
. Efficient management of resources, cloud
providers and users have the prime interests depending upon
the size, capacity and flexibility for the cloud have been
managing software which is able to use the hardware
resources to succeed, and argued that the critical alone to
provide the desired performance [5]
. The successful resources
management in the context of the resource constraints for the
best private cloud solution, initial placement and load
balancing when the resources to offers rich set. For example,
during the peak of banking applications based on customer
needs, they have number of servers that can be accessed from
the cloud. In a moment have shut down the server and it can
be used to save the energy.
3. METHODOLOGY
3.1 Support Vector Machine
The support vector machine (SVM) is a training
algorithm for learning classification and regression rules
of data [3]
. The SVM is used to find the best classification
functions to distinguish between members of the two classes
in the trained data. The metrics for the best classification
function can be realized geometrically and a linear
classification function corresponds to a separating hyperplane
that passes through the middle and separating of the two
classes. A function is determined the new data instance of
can be classified by simply testing the sign of the function
belongs to the positive class if .
3.2 MD5 Method
Step 1 – Append padded bits:
– The length of the messages 448 modulo of 512 that is
similar to the modular padded.
• 64 bits to 512 bits long means they are just shy of being
extended.
– The value of 1 bit is appended to the message and then the
value 0 bit also appended so that the length in bits are equals
448 modulo 512.
Step 2 – Append length:
– 64 bit representation of b is appended to the result of the
previous step.
– Right that has a length of 512 bits of message.
Step 3 –Initialize MD Buffer
• A four-word buffer (A, B, C, D) is used to compute the
message digest.
– Here every one has A, B, C, D is a 32-bit register
- Word A: 01 23 45 67 etc., the following hexadecimal values
of these registers are initialized.
Step 4 –The 16 word blocks for process message.
– The input of three 32-bit words and getting the output of one
32-bit word four sub-functions.
F(X,Y,Z) = XY v not(X) Z
The bits X,Y and Z are independent and also unbiased. The
each bit of F(X,Y,Z), G(X,Y,Z), H(X,Y,Z), and I(X,Y,Z) will
be independent and unbiased.
Step 5 –output
– Output A,B,C,D of the message digest.
– The ouput is start with the low order byte A and end with
the high-order byte D.
3.3 GREEN AND GOSSIP
To initiate more environmental - friendly computing practices.
There are some steps take toward a green computing strategy.
Green Resource Allocator: It is act as the interface between
3. International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 791 - 794, 2014, ISSN:- 2319–8656
www.ijcat.com 793
the Cloud infrastructure and consumers. The interactions of
the following components to support energy efficient resource
management. The components are Arbitrator, Overhaul-
Prediction, Client Focused, Budget Processing, Power
Consuming, Overhaul-Router, Storage Leadership and Gauge.
Figure: 3.1 Green Cloud Computing Architecture
Gossip huge clouds of green computing resource allocation
based they aim to reduce server power consumption by
integration with the specific objectives are sensible the
resource allocation protocol and propose a common rumors.
Under load the ethical and fair allocation of CPU resources to
clients. The simulation result and the key performance of
metrics for resource allocation process and suggest that it is
the body has do not change with increasing size.
4. SYSTEM IMPLEMENTATION
4.1 General
In general, user interface design for consumer requests and
getting response from the server. The consumer is valid
provider means they are getting the further processes. It is
used to interact with client and server for request and
responding from the cloud storage.
4.2 Implementation of Physical Machine
In this module, to implement n number of physical server and
which is interconnected with data storage. Each and every
physical server have been separate identification like server
IP, server port and server instance name. in this all the data
stored in the storage device which the data can be applicable
to source through physical server.
4.3 Cloud Environment mining setup
system
It is provide the dynamic allocation of consumer request to the
particular physical services and the physical server retrieve
the consumers request based information from the storage
device. The response of physical server is applicable for
available data only. Cloud environment use the concept of
SVM to populate the data from data centers.
4.4 Highly Secure Client Processing
In this system using MD5 based concept of achieving the
securable data transmission between the consumer and the
server. This MD5 is converted into the data normal format to
undefined format. In this application apply the highly security
for data transmission client request processing physical
machine data population server mining user identification by
server using the request name.
4.5 Efficient Server Provisioning
Dynamic allocation has been user request to the physical
server in done by cloud environment by using the concept of
protocol named as gossip [5]
. This protocol is sufficient
protocol for dynamic resource allocation and it gives response
to the client at exact query matching server provisioning
approach for minimized the cost, reduce the time and quick
response. In order to assemble the cloud environmental setup
and physical server storage device is very expensive but they
are applying the mining setup. Show that it is must like
expensive.
5. EXPERIMENTAL RESULT
The implementation of my concept is to create the n number
of virtual machines and physical machines. In this machines
have n number of information’s are stored. This physical
machine contains java connection based classes and service
based concepts. In cloud environment system distributed set
up of mined servers. The cloud server’s shows all the
information of the physical server in the data center.
In data center having mining setup for retrieve the data from
the data storage. In data storage wants to store the number of
files with the help of query processing from server. In this
method using SVM for classify the data for the user query
searched from the server.
The data center is the large storage network. The network
wants to secure the stored information for the storage devices
by using the cryptography technique. In this concept am using
MD5 method for creating number of keys to secure the data
from the storage devices. The valid user’s are only views the
information about the user queries. The key based concept is
achieving the more secure for storage devices.
Finally, the server client communication is the very large
process. In this having number of resources and it also
intermediate process for file transfers to the user’s and data
storage. It is the nonstop processes so here using gossip
protocol for green computing process. It is an automatically
allocate the resources for file sharing in the data center.
6. CONCLUSION
To conclude that the concept have so many process in the
cloud environment. The cloud environment has number of
virtual machines and physical machines. These machines are
used to store the number of data in the storage devices. The
storage devices have been number of processes to mining the
data.
In this the data can be find out from the storage device. The
data retrieval by the user’s from the cloud storage devices. In
this the user given the request and get the response from the
server. The SVM method is used for mined the data from the
storage device. This method is used to classify the data and it
is very efficient to gather some information without any
unknown data.
The second process is to secure the data by using the method
of MD5. The cryptographic technique is used to secure the
information storage by using the key values. So the
information storage is getting more secure without any
leakage in the cloud environment.
The third process to find the delay provisioning for data
center. The data center having so many request and response
4. International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 791 - 794, 2014, ISSN:- 2319–8656
www.ijcat.com 794
process for user’s from the storage devices. In this device
having continue processes so the machines want to overcome
the delay processing by using the method of gossip protocol
for green computing.
7. REFERENCES
[1] Qi Zhang, Mohamed Faten Zhani, Raouf Boutaba,
Joseph L. Hellerstein “Dynamic Heterogeneity Aware
Resource Provisioning in the Cloud” Proc.IEEE Trans.
Cloud computing,2014.
[2] Jun Zhou, Xiaodong Lin, Xiaolei Dong, Zhenfu Cao
“PSMPA: Patient Self-controllable and Multi-level
Privacy-preserving Cooperative Authentication in
Distributed m-Healthcare Cloud Computing System”
Proc. IEEE Tans. Parallel and Distributed System, 2014.
[3] Lu Zhang and Xueyan Tang “The Client Assignment
Problem for Continuous Distributed Interactive
Applications: Analysis, Algorithms, and Evaluation”
Proc. IEEE Trans. Parallel and Distributed System, 2014.
[4] Q. Zhang, M.F. Zhani, R. Boutaba, and J.L. Hellerstein,
“HARMONY: Dynamic Heterogeneity-Aware Resource
Provisioning in the Cloud,”Proc. IEEE Int’l Conf.
Distributed Computing Systems (ICDCS), 2013.
[5] Q. Zhang, M.F. Zhani, Q. Zhu, S. Zhang, R. Boutaba,
and J.L. Hellerstein, “Dynamic Energy-Aware Capacity
Provisioning for Cloud Computing Environments,” Proc.
ACM Int’l Conf. Autonomic Computing (ICAC), 2012.
[6] Lu Zhang, Xueyan Tang, “Optimizing Client Assignment
for Enhancing Interactivity in Distributed Interactive
Applications,” Proc. IEEE/ACM Transaction on
Networking, 2012.
[7] P. Morillo, J. Orduna, M. Fernandez, and J. Duato,
“Improving the performance of distributed virtual
environment systems,” IEEE Trans. Parallel Distrib.
Syst., vol. 16, no. 7, pp. 637–649, Jul. 2005.
[8] J. Sun and Y. Fang, Cross-domain Data Sharing in
Distributed Electronic Health Record System, IEEE
Transactions on Parallel and Distributed Systems, vol.
21, No. 6, 2010.
Authors
MS. Anju Aravind K, received the Bachelor of Engineering in Anna University,
TamilNadu, India in 2011. PG Scholar Currently persuing her M.E CSE degree in
shree Venkateshwara Hi-Tech Engg College, Gobi, TamilNadu, India.
Dr.T.Senthil Prakash received the Ph.D. degree from the PRIST University,
Thanjavur, India in 2013 and M.E(CSE) degree from Vinayaka Mission’s University,
Salem, India in 2007 and M.Phil.,MCA.,B.Sc(CS) degrees from Bharathiyar
University, Coimbatore India, in 2000,2003 and 2006 respectively, all in Computer
Science and Engineering. He is a Member in ISTE New Delhi, India, IAENG, Hong
Kong..IACSIT, Singapore SDIWC, USA. He has the experience in Teaching of
10+Years and in Industry 2 Years. Now He is currently working as a Professor and
Head of the Department of Computer Science and Engineering in Shree Venkateshwara
Hi-Tech Engineering College, Gobi, Tamil Nadu, and India. His research interests
include Data Mining, Data Bases, Artificial Intelligence, Software Engineering etc.,He
has published several papers in 17 International Journals, 43 International and National
Conferences.
Mr.M.Rajesh, Received the Bachelor of Engineering in Anna University, TamilNadu,
India in 2007 and Master of Engineering from Kongu Engineering College of India in
2012. Currently he is doing Ph.D at Bharath University, Chennai. His research
interests include cloud computing in resource provisioning.