Cloud computing
Upcoming SlideShare
Loading in...5
×

Like this? Share it with your network

Share

Cloud computing

  • 4,329 views
Uploaded on

IEEE Final Year Projects 2011-2012 :: Elysium Technologies Pvt Ltd...

IEEE Final Year Projects 2011-2012 :: Elysium Technologies Pvt Ltd
IEEE projects, final year projects, students project, be project, engineering projects, academic project, project center in madurai, trichy, chennai, kollam, coimbatore

  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Be the first to comment
No Downloads

Views

Total Views
4,329
On Slideshare
4,329
From Embeds
0
Number of Embeds
0

Actions

Shares
Downloads
121
Comments
0
Likes
1

Embeds 0

No embeds

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
    No notes for slide

Transcript

  • 1. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 Abstract CLOUD COMPUTING 2011 - 201201 A Hybrid Shared-nothing/Shared-data Storage Architecture for Large Scale Databases Shared-nothing and shared-disk are two widely-used storage architectures in current parallel database systems, and each of them has its own merits for different query patterns. However, there is no much effort in investigating the integration of these two architectures and exploiting their merits together. In this study, we propose a novel hybrid shared-nothing/shared- data storage scheme for large-scale databases, to leverage the benefits of both shared-nothing and shared-disk architectures. We adopt a shared-nothing architecture as the hardware layer and leverage a parallel file system as the storage layer. The proposed hybrid storage scheme can provide a high degree of parallelism in both I/O and computing, like that in a shared-nothing system. In the meantime, it can achieve convenient and high-speed data sharing across multiple database nodes, like that in a shared-disk system. The hybrid scheme is more appropriate for large-scale and dataintensive applications than each of the two individual types of systems.02 A performance goal oriented processor allocation technique for centralized heterogeneous multi-cluster environments A performance goal oriented processor allocation technique for centralized heterogeneous multi-cluster environments. This paper proposes a processor allocation technique named temporal look-ahead processor allocation (TLPA) that makes allocation decision by evaluating the allocation effects on subsequent jobs in the waiting queue. TLPA has two strengths. First, it takes multiple performance factors into account when making allocation decision. Second, it can be used to optimize different performance metrics. To evaluate the performance of TLPA, we compare TLPA with best-fit and fastest- first algorithms. Simulation results show that TLPA has up to 32.75% performance improvement over conventional processor allocation algorithms in terms of average turnaround time in various system configurations03 A Petri Net Approach to Analyzing Behavioral Compatibility and Similarity of Web Services Web services have become the technology of choice for service-oriented computing implementation, where Web services can be composed in response to some users’ needs. It is critical to verify the compatibility of component Web services to ensure the correctness of the whole composition in which these components participate. Traditionally, two conditions need to be satisfied during the verification of compatibility: reachable termination and proper termination. Unfortunately, it is complex and time consuming to verify those two conditions. To reduce the complexity of this verification, we model Web services using colored Petri nets (PNs) so that a specific property of their structures is looked into, namely, well structuredness. We prove that only reachable termination needs to be satisfied when verifying behavioral compatibility among well-structured Web services. When a composition is declared as valid and in the case where one of its component Web services fails at run time, an alternative one with similar behavior needs to come into play as a substitute. Thus, it is important to develop effective approaches that permit one to analyze the similarity of Web services. Although many existing approaches utilize PNs to analyze behavioral compatibility, few of them explore further appropriate definitions of behavioral similarity and provide a user-friendly tool with automatic verification. In this paper, we introduce a formal definition of context-independent similarity and show that a Web service can be substituted by an alternative peer of similar behavior without intervening otherWeb services in the composition. Therefore, the cost of verifying service substitutability is largely reduced. We also provide an algorithm for the verification and implement it in a tool. Using the tool, the verification of behavioral similarity of Web services can be performed in an automatic way.Madurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 1
  • 2. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-201204 A Privacy Preserving Repository for Securing Data across the Cloud Popularity of cloud computing is increasing day by day in distributed computing environment. There is a growing trend of using cloud environments for storage and data processing needs. Cloud computing is an Internet-based computing, whereby shared resources, software, and information are provided to computers and other devices on demand. However, adopting a cloud computing paradigm may have positive as well as negative effects on the data security of service consumers. This paper primarily highlights some major security issues existing in current cloud computing environments. The primary issue that has to be dealt with when talking about data security in a cloud is protection of the data. The idea is to construct a privacy preserving repository where data sharing services can update and control the access and limit the usage of their shared data, instead of submitting data to central authorities, and, hence, the repository will promote data sharing and privacy of data. This paper aims at simultaneously achieving data confidentiality while still keeping the harmonizing relations intact in the cloud. Our proposed scheme enables the data owner to delegate most of computation intensive tasks to cloud servers without disclosing data contents or user access privilege information.05 A Scalable Method for Signalling Dynamic Reconfiguration Events with OpenSM Rerouting around faulty components, on-the-fly policy changes, and migration of jobs all require reconfiguration of data structures in the Queue Pairs residing in the hosts on an InfiniBand cluster. In addition to a proper implementation at the host, the subnet manager needs to implement a scalable method for signaling reconfiguration events to the hosts. In this paper we propose and evaluate three different implementations for signalling dynamic reconfiguration events with OpenSM. Through our evaluation we demonstrate a scalable solution for signalling host-side reconfiguration events in an InfiniBand network based on an example where dynamic network reconfiguration combined with a topology-agnostic routing function is used to avoid malfunctioning components. Through measurements on our test-cluster and an analytical study we show that our best proposal reduces reconfiguration latency by more than 90%and in certain situations eliminates it completely. Furthermore, the processing overhead in the subnet manager is shown to be minimal.06 A Segment-Level Adaptive Data Layout Scheme for Improved Load Balance in Parallel File Systems Parallel file systems are designed to mask the everincreasing gap between CPU and disk speeds via parallel I/O processing. While they have become an indispensable component of modern high-end computing systems, their inadequate performance is a critical issue facing the HPC community today. Conventionally, a parallel file system stripes a file across multiple file servers with a fixed stripe size. The stripe size is a vital performance parameter, but the optimal value for it is often application dependent. How to determine the optimal stripe size is a difficult research problem. Based on the observation that many applications have different data-access clusters in one file, with each cluster having a distinguished data access pattern, we propose in this paper a segmented data layout scheme for parallel file systems. The basic idea behind the segmented approach is to divide a file logically into segments such that an optimal stripe size can be identified for each segment. A five-step method is introduced to conduct the segmentation, to identify the appropriate stripe size for each segment, and to carry out the segmented data layout scheme automatically. Experimental results show that the proposed layout scheme is feasible and effective, and it improves performance up to 163% for writing and 132% for reading on the widely used IOR and IOzone benchmarks07 A Sketch-based Architecture for Mining Frequent Items and Itemsets from Distributed Data StreamsMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 2
  • 3. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 Parallel file systems are designed to mask the everincreasing gap between CPU and disk speeds via parallel I/O processing. While they have become an indispensable component of modern high-end computing systems, their inadequate performance is a critical issue facing the HPC community today. Conventionally, a parallel file system stripes a file across multiple file servers with a fixed stripe size. The stripe size is a vital performance parameter, but the optimal value for it is often application dependent. How to determine the optimal stripe size is a difficult research problem. Based on the observation that many applications have different data-access clusters in one file, with each cluster having a distinguished data access pattern, we propose in this paper a segmented data layout scheme for parallel file systems. The basic idea behind the segmented approach is to divide a file logically into segments such that an optimal stripe size can be identified for each segment. A five-step method is introduced to conduct the segmentation, to identify the appropriate stripe size for each segment, and to carry out the segmented data layout scheme automatically. Experimental results show that the proposed layout scheme is feasible and effective, and it improves performance up to 163% for writing and 132% for reading on the widely used IOR and IOzone benchmarks.08 A Trustworthiness Fusion Model for Service Cloud Platform Based on D-S Evidence Theory Gold Trustworthiness plays an important role in service selection and usage. However, it is not easy to define and compute the service trustworthiness because of its subject meaning and also the different views on it. In this paper, we describe the meaning of trustworthiness and the computation method for trustworthiness fusion. Through extracting trustworthiness from service provider, service requestor and service broker, we creatively adopted D-S (Dempster-Shafer) evident theory to fuse the tripartite trustworthiness. Finally, we completed some comparison experiments on our web service supermarket platform and certified the efficiency of our method.09 Addressing Resource Fragmentation in Grids Through Network–Aware Meta–Scheduling in Advance Grids are made of heterogeneous computing resources geographically dispersed where providing Quality of Service (QoS) is a challenging task. One way of enhancing the QoS perceived by users is by performing scheduling of jobs in advance, since reservations of resources are not always possible. This way, it becomes more likely that the appropriate resources are available to run the job when needed. One drawback of this scenario is that fragmentation appears as a well known effect in job allocations into resources and becomes the cause for poor resource utilization. So, a new technique has been developed to tackle fragmentation problems, which consists of rescheduling already scheduled tasks. To this end, some heuristics are implemented to calculate the intervals to be replanned and to select the jobs involved in the process. Moreover, another heuristic is implemented to put rescheduled jobs as close together as possible to minimize the fragmentation. This technique has been tested using a real testbed.10 APP: Minimizing Interference Using Aggressive Pipelined Prefetching In Multi-Level Buffer Caches As services become more complex with multiple interactions, and storage servers are shared by multiple services, the different I/O streams arising from these multiple services compete for disk attention. Aggressive Pipelined Prefetching (APP) enabled storage clients are designed to manage the buffer cache and I/O streams to minimize the disk I/O-interference arising from competing streams. Due to the large number of streams serviced by a storage server, most of the disk time is spent seeking, leading to degradation in response times. The goal of APP is to decrease application execution time by increasing the throughput of individual I/O streams and utilizing idle capacity on remote nodes along with idle network times thus effectively avoiding alternating bursts of activity followed by periods of inactivity. APP significantly increases overall I/O throughput and decreases overall messaging overhead between servers. In APP, the intelligence is embedded in the clients and they automatically infer parameters in order to achieve the maximum throughput. APP clients make use of aggressive prefetching and data offloading to remote buffer caches in multi-level buffer cache hierarchies in an effort to minimize disk interference and tranquilize the effects of aggressive prefetching. We used an extremely I/O-intensive Radix-k application employed in studies on the scalability of parallel image composition andMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 3
  • 4. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 particle tracing developed at the Argonne National Laboratory with data sets of up to 128 GB and implemented our scheme on a 16- node Linux cluster. We observed that the execution time of the application decreased by 68%on average when using our scheme11 ASDF: An Autonomous and Scalable Distributed File System The demand for huge storage space on data-intensive applications and high-performance scientific computing continues to grow. To integrate massive distributed storage resources for providing huge storage space is an important and challenging issue in Cloud and Grid computing. In this paper, we propose a distributed file system, called ASDF, to meet the demands of not only data-intensive applications but also end users, developers and administrators. While sharing many of the same goals as previous distributed file systems such as scalability, reliability, and performance, it is also designed with the emphasis on compatibility, extensibility and autonomy. With the design goals in minds, we address several issues and present our design by adopting peer-to-peer technology, replication, multi-source data transfer, metadata caching and service-oriented architecture. The experimental results show the proposed distributed file system meet our design goals and will be useful in Cloud and Grid computing.12 A Assertion Based Parallel Debugging Programming languages have advanced tremendously over the years, but program debuggers have hardly changed. Sequential debuggers do little more than allow a user to control the flow of a program and examine its state. Parallel ones support the same operations on multiple processes, which are adequate with a small number of processors, but become unwieldy and ineffective on very large machines. Typical scientific codes have enormous multidimensional data structures and it is impractical to expect a user to view the data using traditional display techniques. In this paper we discuss the use of debug-time assertions, and show that these can be used to debug parallel programs. The techniques reduce the debugging complexity because they reason about the state of large arrays without requiring the user to know the expected value of every element. Assertions can be expensive to evaluate, but their performance can be improved by running them in parallel. We demonstrate the system with a case study finding errors in a parallel version of the Shallow Water Equations, and evaluate the performance of the tool on a 4,096 cores Cray XE6.13 Autonomic SLA-driven Provisioning for Cloud Applications Significant achievements have been made for automated allocation of cloud resources. However, the performance of applications may be poor in peak load periods, unless their cloud resources are dynamically adjusted. Moreover, although cloud resources dedicated to different applications are virtually isolated, performance fluctuations do occur because of resource sharing, and software or hardware failures (e.g. unstable virtual machines, power outages, etc.). In this paper, we propose a decentralized economic approach for dynamically adapting the cloud resources of various applications, so as to statistically meet their SLA performance and availability goals in the presence of varying loads or failures. According to our approach, the dynamic economic fitness of a Web service determines whether it is replicated or migrated to another server, or deleted. The economic fitness of a Web service depends on its individual performance constraints, its load, and the utilization of the resources where it resides. Cascading performance objectives are dynamically calculated for individual tasks in the application workflow according to the user requirements. By fully implementing our framework, we experimentally proved that our adaptive approach statistically meets the performance objectives under peak load periods or failures, as opposed to static resource settings.14 BAR: An Efficient Data Locality Driven Task Scheduling Algorithm for Cloud Computing Large scale data processing is increasingly common in cloud computing systems like MapReduce, Hadoop, and Dryad inMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 4
  • 5. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 recent years. In these systems, files are split into many small blocks and all blocks are replicated over several servers. To process files efficiently, each job is divided into many tasks and each task is allocated to a server to deals with a file block. Because network bandwidth is a scarce resource in these systems, enhancing task data locality(placing tasks on servers that contain their input blocks) is crucial for the job completion time. Although there have been many approaches on improving data locality, most of them either are greedy and ignore global optimization, or suffer from high computation complexity. To address these problems, we propose a heuristic task scheduling algorithm called BAlance-Reduce(BAR) , in which an initial task allocation will be produced at first, then the job completion time can be reduced gradually by tuning the initial task allocation. By taking a global view, BAR can adjust data locality dynamically according to network state and cluster workload. The simulation results show that BAR is able to deal with large problem instances in a few seconds and outperforms previous related algorithms in term of the job completion time.15 Building an online domain-specific computing service over non-dedicated grid and cloud resources: the Superlink-online experience Linkage analysis is a statistical method used by geneticists in everyday practice for mapping disease-susceptibility genes in the study of complex diseases. An essential first step in the study of genetic diseases, linkage computations may require years of CPU time. The recent DNA sampling revolution enabled unprecedented sampling density, but made the analysis even more computationally demanding. In this paper we describe a high performance online service for genetic linkage analysis, called Superlink-online. The system enables anyone with Internet access to submit genetic data and analyze it as easily and quickly as if using a supercomputer. The analyses are automatically parallelized and executed on tens of thousands distributed CPUs in multiple clouds and grids. The first version of the system, which employed up to 3,000 CPUs in UW Madison and Technion Condor pools, has been successfully used since 2006 by hundreds of geneticists worldwide, with over 40 citations in the genetics literature. Here we describe the second version, which substantially improves the scalability and performance of first: it uses over 45,000 non-dedicated hosts, in 10 different grids and clouds, including EC2 and the Superlink@Technion community grid. Improved system performance is obtained through a virtual grid hierarchy with dynamic load balancing and multi-grid overlay via the GridBot system, parallel pruning of short tasks for overhead minimization, and cost-efficient use of cloud resources in reliability-critical execution periods. These enhancements enabled execution of many previously infeasible analyses, which can now be completed within a few hours. The new version of the system, in production since 2009, has completed over 6500 different runs of over 10 million tasks, with total consumption of 420 CPU years.16 Cheetah: A Framework for Scalable Hierarchical Collective Operations Collective communication operations, used by many scientific applications, tend to limit overall parallel application performance and scalability. Computer systems are becoming more heterogeneous with increasing node and core-per-node counts. Also, a growing number of data-access mechanisms, of varying characteristics, are supported within a single computer system. We describe a new hierarchical collective communication framework that takes advantage of hardware- specific data-access mechanisms. It is flexible, with run-time hierarchy specification, and sharing of collective communication primitives between collective algorithms. Data buffers are shared between levels in the hierarchy reducing collective communication management overhead. We have implemented several versions of the Message Passing Interface (MPI) collective operations, MPI Barrier() and MPI Bcast(), and run experiments using up to 49,152 processes on a Cray XT5, and a small InfiniBand based cluster. At 49,152 processes our barrier implementation outperforms the optimized native implementation by 75%. 32 Byte and one Mega-Byte broadcasts outperform it by 62% and 11%, respectively, with better scalability characteristics. Improvements relative to the default Open MPI implementation are much larger.17 Classification and Composition of QoS Attributes in Distributed, Heterogeneous Systems In large-scale distributed systems the selection of services and data sources to respond to a given request is a crucial task. Non-functional or Quality of Service (QoS) attributes need to be considered when there are several candidate services with identical functionality. Before applying any service selection optimization strategy, the system has to be analyzed in terms ofMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 5
  • 6. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 QoS metrics, comparable to the statistics needed by a database query optimizer. This paper presents a classification approach for QoS attributes of system components, from which aggregation functions for composite services are derived. The applicability and usefulness of the approach is shown in a distributed system from a High-Energy Physics experiment posing a complex service selection challenge.. 18 Cloud MapReduce: a MapReduce Implementation on top of a Cloud Operating System This study presents a fully automatedmembrane segmentation technique for immunohistochemical tissue images with membrane staining, which is a critical task in computerized immunohistochemistry (IHC). Membrane segmentation is particularly tricky in immunohistochemical tissue images because the cellular membranes are visible only in the stained tracts of the cell, while the unstained tracts are not visible. Our automated method provides accurate segmentation of the cellularmembranes in the stained tracts and reconstructs the approximate location of the unstained tracts using nuclear membranes as a spatial reference. Accurate cell-by-cell membrane segmentation allows per cell morphological analysis and quantification of the target membrane proteins that is fundamental in several medical applications such as cancer characterization and classification, personalized therapy design, and for any other applications requiring cell morphology characterization. Experimental results on real datasets from different anatomical locations demonstrate the wide applicability and high accuracy of our approach in the context of IHC analysis.19 CloudSpider: Combining Replication with Scheduling for Optimizing Live Migration of Virtual Machines Across Wide Area Networks Exact information about the shape of a lumbar pedicle can increase operation accuracy and safety during computeraided spinal fusion surgery, which requires extreme caution on the part of the surgeon, due to the complexity and delicacy of the procedure. In this paper, a robust framework for segmenting the lumbar pedicle in computed tomography (CT) images is presented. The framework that has been designed takes a CT image, which includes the lumbar pedicle as input, and provides the segmented lumbar pedicle in the form of 3-D voxel sets. This multistep approach begins with 2-D dynamic thresholding using local optimal thresholds, followed by procedures to recover the spine geometry in a high curvature environment. A subsequent canal reference determination using proposed thinning-based integrated cost is then performed. Based on the obtained segmented vertebra and canal reference, the edge of the spinal pedicle is segmented. This framework has been tested on 84 lumbar vertebrae of 19 patients requiring spinal fusion. It was successfully applied, resulting in an average success rate of 93.22% and a final mean error of 0.14±0.05 mm. Precision errors were smaller than 1% for spine pedicle volumes. Intra- and interoperator precision errors were not significantly different.20 Automatic and Unsupervised Snore Sound Extraction From Respiratory Sound Signals In this paper, an automatic and unsupervised snore detection algorithm is proposed. The respiratory sound signals of 30 patients with different levels of airway obstruction were recorded by twomicrophones: one placed over the trachea (the tracheal microphone), and the other was a freestanding microphone (the ambient microphone). All the recordings were done simultaneously with full-night polysomnography during sleep. The sound activity episodes were identified using the vertical box (V-Box) algorithm. The 500-Hz subband energy distribution and principal component analysis were used to extract discriminative features from sound episodes. An unsupervised fuzzy C-means clustering algorithm was then deployed to label the sound episodes as either snore or no-snore class, which could be breath sound, swallowing sound, or any other noise. The algorithm was evaluated using manual annotation of the sound signals. The overall accuracy of the proposed algorithm was found to be 98.6% for tracheal sounds recordings, and 93.1% for the sounds recorded by the ambient microphone.Madurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 6
  • 7. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-201221 Dealing with Grid-Computing Authorization using Identity-Based Certificateless Proxy Signature In this paper, we propose a new Identity-Based Certificateless Proxy Signature scheme, for the grid environment, in order to enable attribute-based authorization, finegrained delegation and enhanced delegation chain establishment and validation, all without relying on any kind of PKI Certificates or proxy certificates. We show that our scheme is correct and secure. We also give an evaluation of the computational and communication overhead of the proposed scheme. Simulations shows satisfying results.22 Debunking Real-Time Pricing in Cloud Computing Under Elasticity of cloud computing eases the burden of capacity planning. Cloud computing users dynamically provision IT resources tracking their uctuating demand, and only pay for their usage. Therefore, cloud comput- ing essentially shifts the burden of capacity planning from users side to providers side. On the other hand, providers take this burden with the optimistic assumption that di- verse workloads from various users will atten the overall demand curve. However, this optimistic hypothesis has not been proved yet in the real world cases. In fact, counter evidences have been raised. December 2009, Amazon Web Services (AWS), a leading infrastructure cloud service provider, started to o er a real-time pricing for computing resources {Amazon EC2 Spot Instances (SIs). Real-time pricing, in princi- ple, encourages users to shift their flexible workloads from providers peak hours to o -peak hours with monetary incentives. Interestingly, from our observation on AWSs one-year SI price history datasets, we conclude that the observed monetary incentive is not large enough to motivate users to shift their workloads. It is reasonable for users to choose SIs over on-demand instances because SIs are 52.3% cheaper on average. After that, shifting the workload to cheaper period provides only 3.7 % additional cost savings at best. Moreover, both average cost savings and price fluctuation have not been meaningfully changed over time.. 23 DELMA: Dynamically ELastic MApReduce Framework for CPU-Intensive Applications Since its introduction, MapReduce implementations have been primarily focused towards static compute cluster sizes. In this paper, we introduce the concept of dynamic elasticity to MapReduce. We present the design decisions and implementation tradeoffs for DELMA, (Dynamically ELastic MApReduce), a framework that follows the MapReduce paradigm, just like Hadoop MapReduce, but that is capable of growing and shrinking its cluster size, as jobs are underway. In our study, we test DELMA in diverse performance scenarios, ranging from diverse node additions to node additions at various points in the application run-time with various dataset sizes. The applicability of the MapReduce paradigm extends far beyond its use with large-scale data intensive applications, and can also be brought to bear in processing long running distributed applications executing on small-sized clusters. In this work, we focus both on the performance of processing hierarchical data in distributed scientific applications, as well as the processing of smaller but demanding input sizes primarily used in small clusters. We run experiments for datasets that require CPU intensive processing, ranging in size from Millions of input data elements to process, up to over half a billion elements, and observe the positive scalability patterns exhibited by the system. We show that for such sizes, performance increases accordingly with data and cluster size increases. We conclude on the benefits of providing MapReduce with the capability of dynamically growing and shrinking its cluster configuration by adding and removing nodes during jobs, and explain the possibilities presented by this model.24 Detection and Protection against Distributed Denial of Service Attacks in Accountable Grid Computing SystemsMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 7
  • 8. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 By exploiting existing vulnerabilities, malicious parties can take advantage of resources made available by grid systems to attack mission-critical websites or the grid itself. In this paper, we present two approaches for protecting against attacks targeting sites outside or inside the grid. Our approach is based on special-purpose software agents that collect provenance and resource usage data in order to perform detection and protection. We show the effectiveness and the efficiency of our approach by conducting various experiments on an emulated grid test-bed.25 DHTbd: A Reliable Block-based Storage System for High Performance Clusters Large, reliable and efficient storage systems are becoming increasingly important in enterprise environments. Our research in storage system design is oriented towards the exploita- tion of commodity hardware for building a high performance, resilient and scalable storage system. We present the design and implementation of DHTbd, a general purpose decentralized storage system where storage nodes support a distributed hash table based interface and clients are implemented as in- kernel device drivers. DHTbd, unlike most storage systems proposed to date, is implemented at the block device level of the I/O stack, a simple yet efficient design. The experimental evaluation of the proposed system demonstrates its very good I/O performance, its ability to scale to large clusters, as well as its robustness, even when massive failures occur.26 Diagnosing Anomalous Network Performance with Confidence Variability in network performance is a major obstacle in effectively analyzing the throughput of modern high performance computer systems. High performance interconnection networks offer excellent best-case network latencies; however, highly parallel applications running on parallel machines typically require consistently high levels of performance to adequately leverage the massive amounts of available computing power. Performance analysts have usually quantified network performance using traditional summary statistics that assume the observational data is sampled from a normal distribution. In our examinations of network performance, we have found this method of analysis often provides too little data to understand anomalous network performance. In particular, we examine a multi-modal performance scenario encountered with an Infiniband interconnection network and we explore the performance repeatability on the custom Cray SeaStar2 interconnection network after a set of software and driver updates.27 Enabling Multi-Physics Coupled Simulations within the PGAS Programming Framework Complex coupled multi-physics simulations are playing increasingly important roles in scientific and engineering applications such as fusion plasma and climate modeling. At the same time, extreme scales, high levels of concurrency and the advent of multicore and manycore technologies are making the high-end parallel computing systems on which these simulations run, hard to program. While the Partitioned Global Address Space (PGAS) languages is attempting to address the problem, the PGAS model does not easily support the coupling of multiple application codes, which is necessary for the coupled multi-physics simulations. Furthermore, existing frameworks that support coupled simulations have been developed for fragmented programming models such as message passing, and are conceptually mismatched with the shared memory address space abstraction in the PGAS programming model. This paper explores how multi-physics coupled simulations can be supported within the PGAS programming framework. Specifically, in this paper, we present the design and implementation of the XpressSpace programming system, which enables efficient and productive development of coupled simulations across multiple independent PGAS Unified Parallel C (UPC) executables. XpressSpace provides the global-view style programming interface that is consistent with the memory model in UPC, and provides an efficient runtime system that can dynamically capture the data decomposition of global-view arrays and enable fast exchange of parallel data structures between coupled codes. In addition, XpressSpace provides the flexibility to define the coupling process in specification file that is independent of the program source codes. We evaluate the performance and scalability of XpressSpace prototype implementation using different coupling patterns extracted from real world multi-physics simulation scenarios, on the Jaguar Cray XT5 system of Oak Ridge National LaboratoryMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 8
  • 9. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-201228 EZTrace: a generic framework for performance analysis Modern supercomputers with multi-core nodes enhanced by accelerators, as well as hybrid programming models introduce more complexity in modern applications. Exploiting efficiently all the resources requires a complex analysis of the performance of applications in order to detect time-consuming sections. We present EZTRACE, a generic trace generation framework that aims at providing a simple way to analyze applications. EZTRACE is based on plugins that allow it to trace different programming models such as MPI, pthread or OpenMP as well as user-defined libraries or applications. EZTRACE uses two steps: one to collect the basic information during execution and one post-mortem analysis. This permits tracing the execution of applications with low overhead while allowing to refine the analysis after the execution. We also present a script language for EZTRACE that gives the user the opportunity to easily define the functions to instrument without modifying the source code of the application.29 Failure Avoidance through Fault Prediction Based on Synthetic Transactions System logs are an important tool in studying the conditions (e.g., environment misconfigurations, resource status, erroneous user input) that cause failures. However, production system logs are complex, verbose, and lack structural stability over time. These traits make them hard to use, and make solutions that rely on them susceptible to high maintenance costs. Additionally, logs record failures after they occur: by the time logs are investigated, users have already experienced the failures’ consequences. To detect the environment conditions that are correlated with failures without dealing with the complexities associated with processing production logs, and to prevent failure-causing conditions from occurring before the system goes live, this research suggests a three step methodology: i) using synthetic transactions, i.e., simplified workloads, in pre-production environments that emulate user behavior, ii) recording the result of executing these transactions in logs that are compact, simple to analyze, stable over time, and specifically tailored to the fault metrics of interest, and iii) mining these specialized logs to understand the conditions that correlate to failures. This allows system administrators to configure the system to prevent these conditions from happening. We evaluate the effectiveness of this approach by replicating the behavior of a service used in production at Microsoft, and testing the ability to predict failures using a synthetic workload on a 650 million events production trace. The synthetic prediction system is able to predict 91% of real production failures using 50-fold fewer transactions and logs that are 10,000-fold more compact than their production counterparts.30 GeoServ: A Distributed Urban Sensing Platform Urban sensing where mobile users continuously gather, process, and share location-sensitive sensor data (e.g., street images, road condition, traffic flow) is emerging as a new network paradigm of sensor information sharing in urban environments. The key enablers are the smartphones (e.g., iPhones and Android phones) equipped with onboard sensors (e.g., cameras, accelerometer, compass, GPS) and various wireless devices (e.g., WiFi and 2/3G). The goal of this paper is to design a scalable sensor networking platform where millions of users on the move can participate in urban sensing and share locationaware information using always-on cellular data connections. We propose a two-tier sensor networking platform called GeoServ where mobile users publish/access sensor data via an Internetbased distributed P2P overlay network. The main contribution of this paper is two-fold: a location-aware sensor data retrieval scheme that supports geographic range queries, and a locationaware publish-subscribe scheme that enables efficient multicast routing over a group of subscribed users. We prove that GeoServ protocols preserve locality and validate their performance via extensive simulations. 31 GPGPU-Accelerated Parallel and Fast Simulation of Thousand-core PlatformsMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 9
  • 10. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 The multicore revolution and the ever-increasing complexity of computing systems is dramatically changing system design, analysis and programming of computing platforms. Future architectures will feature hundreds to thousands of simple processors and on-chip memories connected through a network-on-chip. Architectural simulators will remain primary tools for design space exploration, software development and performance evaluation of these massively parallel architectures. However, architectural simulation performance is a serious concern, as virtual platforms and simulation technology are not able to tackle the complexity of thousands of core future scenarios. The main contribution of this paper is the development of a new simulation approach and technology for many core processors which exploit the enormous parallel processing capability of low-cost and widely available General Purpose Graphic Processing Units (GPGPU). The simulation of many- core architectures exhibits indeed a high level of parallelism and is inherently parallelizable, but GPGPU acceleration of architectural simulation requires an in-depth revision of the data structures and functional partitioning traditionally used in parallel simulation. We demonstrate our GPGPU simulator on a target architecture composed by several cores (i.e. ARM ISA based), with instruction and data caches, connected through a Network-on-Chip (NoC). Our experiments confirm the feasibility of our approach.32 Grid Global Behavior Prediction Complexity has always been one of the most important issues in distributed computing. From the first clusters to grid and now cloud computing, dealing correctly and efficiently with system complexity is the key to taking technology a step further. In this sense, global behavior modeling is an innovative methodology aimed at understanding the grid behavior. The main objective of this methodology is to synthesize the grid’s vast, heterogeneous nature into a simple but powerful behavior model, represented in the form of a single, abstract entity, with a global state. Global behavior modeling has proved to be very useful in effectively managing grid complexity but, in many cases, deeper knowledge is needed. It generates a descriptive model that could be greatly improved if extended not only to explain behavior, but also to predict it. In this paper we present a prediction methodology whose objective is to define the techniques needed to create global behavior prediction models for grid systems. This global behavior prediction can benefit grid management, specially in areas such as fault tolerance or job scheduling. The paper presents experimental results obtained in real scenarios in order to validate this approach.33 High Performance Pipelined Process Migration with RDMA Coordinated Checkpoint/Restart (C/R) is a widely deployed strategy to achieve fault-tolerance. However, C/R by itself is not capable enough to meet the demands of upcoming exascale systems, due to its heavy I/O overhead. Process migration has already been proposed in literature as a pro-active fault-tolerance mechanism to complement C/R. Several popular MPI implementations have provided support for process migration, including MVAPICH2 and OpenMPI. But these existing solutions cannot yield a satisfactory performance. In this paper we conduct extensive profiling on several process migration mechanisms, and reveal that inefficient I/O and network transfer are the principal factors responsible for the high overhead. We then propose a new approach, Pipelined Process Migration with RDMA (PPMR), to overcome these overheads. Our new protocol fully pipelines data writing, data transfer, and data read operations during different phases of a migration cycle. PPMR aggregates data writes on the migration source node and transfers data to the target node via high throughput RDMA transport. It implements an efficient process restart mechanism at the target node to restart processes from the RDMA data streams. We have implemented this Pipelined Process Migration protocol in MVAPICH2 and studied the performance benefits. Experimental results show that PPMR achieves a 10.7X speedup to complete a process migration over the conventional approach at a moderate (8MB) memory usage. Process migration overhead on the application is significantly minimized from 38% to 5% by PPMR when three migrations are performed in succession.34 Improving Utilization of Infrastructure Clouds A key advantage of infrastructure-as-a-service (IaaS) clouds is providing users on-demand access to resources. To provideMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 10
  • 11. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 on-demand access, however, cloud providers must either significantly overprovision their infrastructure (and pay a high price for operating resources with low utilization) or reject a large proportion of user requests (in which case the access is no longer on-demand). At the same time, not all users require truly on-demand access to resources. Many applications and workflows are designed for recoverable systems where interruptions in service are expected. For instance, many scientists utilize high-throughput computing (HTC)-enabled resources, such as Condor, where jobs are dispatched to available resources and terminated when the resource is no longer available. We propose a cloud infrastructure that combines on- demand allocation of resources with opportunistic provisioning of cycles from idle cloud nodes to other processes by deploying backfill virtual machines (VMs). For demonstration and experimental evaluation, we extend the Nimbus cloud computing toolkit to deploy backfill VMs on idle cloud nodes for processing an HTC workload. Initial tests show an increase in IaaS cloud utilization from 37.5% to 100% during a portion of the evaluation trace but only 6.39% overhead cost for processing the HTC workload. We demonstrate that a shared infrastructure between IaaS cloud providers and an HTC job management system can be highly beneficial to both the IaaS cloud provider and HTC users by increasing the utilization of the cloud infrastructure (thereby decreasing the overall cost) and contributing cycles that would otherwise be idle to processing HTC jobs.35 Development Inferring Network Topologies in Infrastructure as a Service Cloud Infrastructure as a Service (IaaS) clouds are gaining increasing popularity as a platform for distributed computations. The virtualization layers of those clouds offer new possibilities for rapid resource provisioning, but also hide aspects of the underlying IT infrastructure which have often been exploited in classic cluster environments. One of those hidden aspects is the network topology, i.e. the way the rented virtual machines are physically interconnected inside the cloud. We propose an approach to infer the network topology connecting a set of virtual machines in IaaS clouds and exploit it for data-intensive distributed applications. Our inference approach relies on delay-based end-to-end measurements and can be combined with traditional IP-level topology information, if available. We evaluate the inference accuracy using the popular hypervisors KVM as well as XEN and highlight possible performance gains for distributed applications.36 Directed Differential Connectivity Graph of Interictal Epileptiform Discharges In this paper, we study temporal couplings between interictal events of spatially remote regions in order to localize the leading epileptic regions from intracerebral EEG (iEEG). We aim to assess whether quantitative epileptic graph analysis during interictal period may be helpful to predict the seizure onset zone of ictal iEEG. Using wavelet transform, cross- correlation coefficient, and multiple hypothesis test, we propose a differential connectivity graph (DCG) to represent the connections that change significantly between epileptic and nonepileptic states as defined by the interictal events. Postprocessings based on mutual information and multiobjective optimization are proposed to localize the leading epileptic regions through DCG. The suggested approach is applied on iEEG recordings of five patients suffering from focal epilepsy. Quantitative comparisons of the proposed epileptic regions within ictal onset zones detected by visual inspection and using electrically stimulated seizures, reveal good performance of the present method.37 Driver Drowsiness Managing distributed files with RNS in heterogeneous Data Grids This paper describes the management of files distributed in heterogeneous Data Grids by using RNS (Resource Namespace Service). RNS provides hierarchical namespace management for name-to-resource mapping as a key technology to use Grid resources for different kinds of middleware. RNS directory entries and junction entries can contain their own XML messages as metadata. We define attribute expressions in XML for the RNS entries and give an algorithm to access distributed files stored within different kinds of Data Grids. The example in this paper shows how our Grid application can retrieve the actual locations of files from the RNS server. An application can also access the distributed files as though they were files in theMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 11
  • 12. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 local file system without worrying about the underlying Data Grids. This approach can be used in a Grid computing system to handle distributed Grid resources.. 38 Multi-Cloud Deployment of Computing Clusters for Loosely-Coupled MTC Applications Cloud computing is gaining acceptance in many IT organizations, as an elastic, flexible and variable-cost way to deploy their service platforms using outsourced resources. Unlike traditional utilities where a single provider scheme is a common practice, the ubiquitous access to cloud resources easily enables the simultaneous use of different clouds. In this paper we explore this scenario to deploy a computing cluster on top of a multi-cloud infrastructure, for solving loosely-coupled Many- Task Computing (MTC) applications. In this way, the cluster nodes can be provisioned with resources from different clouds to improve the cost-effectiveness of the deployment, or to implement high-availability strategies. We prove the viability of this kind of solutions by evaluating the scalability, performance, and cost of different configurations of a Sun Grid Engine cluster, deployed on a multi-cloud infrastructure spanning a local data-center and three different cloud sites: Amazon EC2 Europe, Amazon EC2 USA, and ElasticHosts. Although the testbed deployed in this work is limited to a reduced number of computing resources (due to hardware and budget limitations), we have complemented our analysis with a simulated infrastructure model, which includes a larger number of resources, and runs larger problem sizes. Data obtained by simulation show that performance and cost results can be extrapolated to large scale problems and cluster infrastructures.39 Dynamic Brain Phantom for Intracranial Volume Measurements Knowledge of intracranial ventricular volume is important for the treatment of hydrocephalus, a disease in which cerebrospinal fluid (CSF) accumulates in the brain. Current monitoring options involve MRI or pressure monitors (InSite, Medtronic). However, there are no existing methods for continuous cerebral ventricle volume measurements. In order to test a novel impedance sensor for direct ventricular volume measurements, we present a model that emulates the expansion of the lateral ventricles seen in hydrocephalus. To quantify the ventricular volume, sensor prototypes were fabricated and tested with this experimental model. Fluidwas injected andwithdrawn cyclically in a controlledmanner and volume measurements were tracked over 8 h. Pressure measurements were also comparable to conditions seen clinically. The results from the bench-top model served to calibrate the sensor for preliminary animal experiments. A hydrocephalic rat model was used to validate a scaled-down, microfabricated prototype sensor. CSF was removed from the enlarged ventricles and a dynamic volume decrease was properly recorded. This method of testing new designs on brain phantoms prior to animal experimentation accelerates medical device design by determining sensor specifications and optimization in a rational process.40 Multiple Services Throughput Optimization in a Hierarchical Middleware Accessing the power of distributed resources can nowadays easily be done using a middleware based on a client/server approach. Several architectures exist for those middlewares. The most scalable ones rely on a hierarchical design. Determining the best shape for the hierarchy, the one giving the best throughput of services, is not an easy task. We first propose a computation and communication model for such hierarchical middleware. Our model takes into account the deployment of several services in the hierarchy. Then, based on this model, we propose algorithms for automatically constructing a hierarchy on two kinds of heterogeneous platforms: communication homogeneous/computation heterogeneous platforms, and fully heterogeneous platforms. The proposed algorithms aim at offering the users the best obtained to requested throughput ratio, while providing fairness on this ratio for the different kinds of services, and using as few resources as possible for the hierarchy. For each kind of platforms, we compare our model with experimental results onMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 12
  • 13. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 a real middleware called DIET (Distributed Interactive Engineering Toolbox).41 Network-Friendly One-Sided Communication Through Multinode Cooperation on Petascale Cray XT5 Systems One-sided communication is important to enable asynchronous communication and data movement for Global Address Space (GAS) programming models. Such communication is typically realized through direct messages between initiator and target processes. For petascale systems with 10,000s of nodes and 100,000s of cores, these direct messages require dedicated communication buffers and/or channels, which can lead to significant scalability challenges for GAS programming models. In this paper, we describe a network-friendly communication model, multinode cooperation, to enable indirect one- sided communication. Compute nodes work together to handle one-side requests through (1) request forwarding in which one node can intercept a request and forward it to a target node, and (2) request aggregation in which one node can aggregate many requests to a target node. We have implemented multinode cooperation for a popular GAS runtime library, Aggregate Remote Memory Copy Interface (ARMCI). Our experimental results on a largescale Cray XT5 system demonstrate that multinode cooperation is able to greatly increase memory scalability by reducing communication buffers required on each node. In addition, multinode cooperation improves the resiliency of GAS runtime system to network contention. Furthermore, multinode cooperation can benefit the performance of scientific applications. In one case, it reduces the total execution time of an NWChem application by 52%.42 Non-Cooperative Scheduling Considered Harmful in Collaborative Volunteer Computing Environments Advances in inter-networking technology and computing components have enabled Volunteer Computing (VC) systems that allows volunteers to donate their computers’ idle CPU cycles to a given project. BOINC is the most popular VC infrastructure today with over 580,000 hosts that deliver over 2,300 TeraFLOP per day. BOINC projects usually have hundreds of thousands of independent tasks and are interested in overall throughput. Each project has its own server which is responsible for distributing work units to clients, recovering results and validating them. The BOINC scheduling algorithms are complex and have been used for many years now. Their efficiency and fairness have been assessed in the context of throughput oriented projects. Yet, recently, burst projects, with fewer tasks and interested in response time, have emerged. Many works have proposed new scheduling algorithms to optimize individual response time but their use may be problematic in presence of other projects. In this article we show that the commonly used BOINC scheduling algorithms are unable to enforce fairness and project isolation. Burst projects may dramatically impact the performance of all other projects (burst or non-burst). To study such interactions, we perform a detailed, multi-player and multi-objective game theoretic study. Our analysis and experiments provide a good understanding on the impact of the different scheduling parameters and show that the non- cooperative optimization may result in inefficient and unfair share of the resources.43 Finite-Element-Based Discretization and Regularization Strategies for 3-D Inverse Electrocardiography We consider the inverse electrocardiographic problem of computing epicardial potentials from a body-surface potential map. We study how to improve numerical approximation of the inverse problem when the finite-element method is used. Being ill- posed, the inverse problem requires different discretization strategies from its corresponding forward problem. We propose refinement guidelines that specifically address the ill-posedness of the problem. The resulting guidelines necessitate the use of hybrid finite elements composed of tetrahedra and prism elements. Also, in order to maintain consistent numerical quality when the inverse problem is discretized into different scales, we propose a new family of regularizers using the variational principle underlying finiteelement methods. These variational-formed regularizers serve as an alternative to the traditional Tikhonov regularizers, but preserves the L2 norm and thereby achieves consistent regularization in multiscale simulations. The variational formulation also enables a simple construction of the discrete gradient operator over irregularmeshes, which is difficult to define in traditional discretization schemes.We validated our hybrid element technique and the variational regularizers by simulations on a realistic 3-D torso/heart model with empirical heart data. Results show that discretization based on our proposed strategies mitigates the ill-conditioning and improves the inverse solution, and that the variationalMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 13
  • 14. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 formulation may benefit a broader range of potential-based bioelectric problems.44 On the Performance Variability of Production Cloud Services Cloud computing is an emerging infrastructure paradigm that promises to eliminate the need for companies to maintain expensive computing hardware. Through the use of virtualization and resource time-sharing, clouds address with a single set of physical resources a large user base with diverse needs. Thus, clouds have the potential to provide their owners the benefits of an economy of scale and, at the same time, become an alternative for both the industry and the scientific community to self-owned clusters, grids, and parallel production environments. For this potential to become reality, the first generation of commercial clouds need to be proven to be dependable. In this work we analyze the dependability of cloud services. Towards this end, we analyze long-term performance traces from Amazon Web Services and Google App Engine, currently two of the largest commercial clouds in production. We find that the performance of about half of the cloud services we investigate exhibits yearly and daily patterns, but also that most services have periods of especially stable performance. Last, through tracebased simulation we assess the impact of the variability observed for the studied cloud services on three large-scale applications, job execution in scientific computing, virtual goods trading in social networks, and state management in social gaming. We show that the impact of performance variability depends on the application, and give evidence that performance variability can be an important factor in cloud provider selection.45 On the Relation Between Congestion Control, Switch Arbitration and Fairness In lossless interconnection networks such as Infini- Band, congestion control (CC) can be an effective mechanism to achieve high performance and good utilization of network resources. The InfiniBand standard describes CC functionality for detecting and resolving congestion, but the design decisions on how to implement this functionallity is left to the hardware designer. One must be cautious when making these design decisions not to introduce fairness problems, as our study shows. In this paper we study the relationship between congestion control, switch arbitration, and fairness. Specifically, we look at fairness among different traffic flows arriving at a hot spot switch on different input ports, as CC is turned on. In addition we study the fairness among traffic flows at a switch where some flows are exclusive users of their input ports while other flows are sharing an input port (the parking lot problem). Our results show that the implementation of congestion control in a switch is vulnerable to unfairness if care is not taken. In detail, we found that a threshold hysteresis of more than one MTU is needed to resolve arbitration unfairness. Furthermore, to fully solve the parking lot problem, proper configuration of the CC parameters are required.46 On the Scheduling of Checkpoints in Desktop Grids Frequent resources failures are a major challenge for the rapid completion of batch jobs. Checkpointing and migration is one approach to accelerate job completion avoiding deadlock. We study the problem of scheduling checkpoints of sequential jobs in the context of Desktop Grids, consisting of volunteered distributed resources. We craft a checkpoint scheduling algorithm that is provably optimal for discrete time when failures obey any general probability distribution. We show using simulations with parameters based on real-world systems that this optimal strategy scales and outperforms other strategies significantly in terms of checkpointing costs and batch completion times.47 Parameter Exploration in Science and Engineering Using Many-Task Computing Robust scientific methods require the exploration of the parameter space of a system (some of which can be run in parallel on distributed resources), and may involve complete state space exploration, experimental design, or numerical optimizationMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 14
  • 15. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 techniques. Many-Task Computing (MTC) provides a framework for performing robust design, because it supports the execution of a large number of otherwise independent processes. Further, scientific workflow engines facilitate the specification and execution of complex software pipelines, such as those found in real science and engineering design problems. However, most existing workflow engines do not support a wide range of experimentation techniques, nor do they support a large number of independent tasks. In this paper, we discuss Nimrod/K—a set of add in components and a new run time machine for a general workflow engine, Kepler. Nimrod/K provides an execution architecture based on the tagged dataflow concepts, developed in 1980s for highly parallel machines. This is embodied in a new Kepler “Director” that supports many-task computing by orchestrating execution of tasks on on clusters, Grids, and Clouds. Further, Nimrod/K provides a set of “Actors” that facilitate the various modes of parameter exploration discussed above. We demonstrate the power of Nimrod/K to solve real problems in cardiac science.48 Performance Analysis of Cloud Computing Services for Many-Tasks Scientific Computing Cloud computing is an emerging commercial infrastructure paradigm that promises to eliminate the need for maintaining expensive computing facilities by companies and institutes alike. Through the use of virtualization and resource time sharing, clouds serve with a single set of physical resources a large user base with different needs. Thus, clouds have the potential to provide to their owners the benefits of an economy of scale and, at the same time, become an alternative for scientists to clusters, grids, and parallel production environments. However, the current commercial clouds have been built to support web and small database workloads, which are very different from typical scientific computing workloads. Moreover, the use of virtualization and resource time sharing may introduce significant performance penalties for the demanding scientific computing workloads. In this work, we analyze the performance of cloud computing services for scientific computing workloads. We quantify the presence in real scientific computing workloads of Many-Task Computing (MTC) users, that is, of users who employ loosely coupled applications comprising many tasks to achieve their scientific goals. Then, we perform an empirical evaluation of the performance of four commercial cloud computing services including Amazon EC2, which is currently the largest commercial cloud. Last, we compare through trace-based simulation the performance characteristics and cost models of clouds and other scientific computing platforms, for general and MTC-based scientific computing workloads. Our results indicate that the current clouds need an order of magnitude in performance improvement to be useful to the scientific community, and show which improvements should be considered first to address this discrepancy between offer and demand.49 Predictive Data Grouping and Placement for Cloud-based Elastic Server Infrastructures Workload variations on Internet platforms such as YouTube, Flickr, LastFM require novel approaches to dynamic resource provisioning in order to meet QoS requirements, while reducing the Total Cost of Ownership (TCO) of the infrastructures. The economy of scale promise of cloud computing is a great opportunity to approach this problem, by developing elastic large scale server infrastructures. However, a proactive approach to dynamic resource provisioning requires prediction models forecasting future load patterns. On the other hand, unexpected volume and data spikes require reactive provisioning for serving unexpected surges in workloads. When workload can not be predicted, adequate data grouping and placement algorithms may facilitate agile scaling up and down of an infrastructure. In this paper, we analyze a dynamic workload of an on-line music portal and present an elastic Web infrastructure that adapts to workload variations by dynamically scaling up and down servers. The workload is predicted by an autoregressive model capturing trends and seasonal patterns. Further, for enhancing data locality, we propose a predictive data grouping based on the history of content access of a user community. Finally, in order to facilitate agile elasticity, we present a data placement based on workload and access pattern prediction. The experimental results demonstrate that our forecasting model predicts workload with a high precision. Further, the predictive data grouping and placement methods provide high locality, load balance and high utilization of resources, allowing a server infrastructure to scale up and down depending on workload.Madurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 15
  • 16. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-201250 Resource Allocation for Security Services in Mobile Cloud Computing Mobile cloud is a machine-to-machine service model, where a mobile device can use the cloud for searching, data mining, and multimedia processing. To protect the processed data, security services, i.e., encryption, decryption, authentications, etc., are performed in the cloud. In general, we can classify cloud security services in two categories: Critical Security (CS) service and Normal Security (NS) service. CS service provides strong security protection such as using longer key size, strict security access policies, isolations for protecting data, and so on. The CS service usually occupies more cloud computing resources, however it generates more rewards to the cloud provider since the CS service users need to pay more for using the CS service. With the increase of the number of CS and NS service users, it is important to allocate the cloud resource to maximize the system rewards with the considerations of the cloud resource consumption and incomes generated from cloud users. To address this issue, we propose a Security Service Admission Model (SSAM) based on Semi- Markov Decision Process to model the system reward for the cloud provider. We, first, define system states by a tuple represented by the numbers of cloud users and their associated security service categories, and current event type (i.e., arrival or departure).We then derive the system steady-state probability and service request blocking probability by using the proposed SSAM. Numerical results show that the obtained theoretic probabilities are consistent with our simulation results.51 Resource and Revenue Sharing with Coalition Formation of Cloud Providers: Game Theoretic Approach In cloud computing, multiple cloud providers can cooperate to establish a resource pool to support internal users and to offer services to public cloud users. In this paper, we study the cooperative behavior of multiple cloud providers. The hierarchical cooperative game model is presented. First, given a group (i.e., coalition) of cloud providers, the resource and revenue sharing of a resource pool is presented. To obtain the solution, we develop the stochastic linear programming game model which takes the uncertainty of internal users from each provider into account. We show that the solution of the stochastic linear programming game is the core of cooperation. Second, we analyze the stability of the coalition formation among cloud providers based on coalitional game. The dynamic model of coalition formation is used to obtain stable coalitional structures. The resource and revenue sharing and coalition formation of cloud providers are intertwined in which the proposed hierarchical cooperative game model can be used to obtain the solution. An extensive performance evaluation is performed to investigate the decision making of cloud providers when cooperation can lead to the higher profit.52 Robust Execution of Service Workflows Using Redundancy and Advance Reservations In this paper, we develop a novel algorithm that allows service consumers to execute business processes (or workflows) of interdependent services in a dependable manner within tight time-constraints. In particular, we consider large interorganizational service-oriented systems, where services are offered by external organizations that demand financial remuneration and where their use has to be negotiated in advance using explicit service-level agreements (as is common in Grids and cloud computing). Here, different providers often offer the same type of service at varying levels of quality and price. Furthermore, some providers may be less trustworthy than others, possibly failing to meet their agreements. To control this unreliability and ensure end-to-end dependability while maximizing the profit obtained from completing a business process, our algorithm automatically selects the most suitable providers. Moreover, unlike existing work, it reasons about the dependability properties of a workflow, and it controls these by using service redundancy for critical tasks and by planning for contingencies. Finally, our algorithm reserves services for only parts of its workflow at any time, in order to retain flexibility when failures occur. We show empirically that our algorithm consistently outperforms existing approaches, achieving up to a 35-fold increase in profit and successfully completing most workflows, even when the majority of providers fail.Madurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 16
  • 17. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-201253 Modified Kinematic Technique for Measuring Pathological Hyperextension and Hypermobility of the Interphalangeal Joints Dynamic finger joint motion is difficult to measure using optical motion analysis techniques due to the limited surface area allowed for adequate marker placement. This paper describes an extension of a previously validated kinematic measurement technique using a reduced surface marker set and outlines the required calculations based on a specific surface marker placement to calculate flexion/extension and hyperextension of the metacarpophalangeal, proximal interphalangeal, and distal interphalangeal joints. The modified technique has been assessed for accuracy using a series of static reference frames (absolute residual error = ±3.7◦, cross correlation between new method and reference frames; r = 0.99). The method was then applied to a small group of participantswith rheumatoid arthritis (seven females, one male; mean age = 62.8 years ± 12.04) and illustrated congruent strategies of movement for a participant and a large range of finger joint movement over the sample (5.8–71.1◦, smallest to largest active range of motion). This method used alongside the previous paper [1] provides a comprehensive, validated method for calculating 3-D wrist, hand, fingers, and thumb kinematics to date and provides a valuable measurement tool for clinical research.54 Role-Based Access-Control Using Reference Ontology in Clouds In cloud computing, security is an important issue due to the increasing scale of users. Current approaches to access control on clouds do not scale well to multi-tenancy requirements because they are mostly based on individual user IDs at different granularity levels. However, the number of users can be enormous and causes significant overhead in managing security. RBAC (Role-Based Access Control) is attractive because the number of roles is significantly less, and users can be classified according to their roles. This paper proposes a RBAC model using a role ontology for Multi- Tenancy Architecture (MTA) in clouds. The ontology is used to build up the role hierarchy for a specific domain. Ontology transformation operations algorithms are provided to compare the similarity of different ontology. The proposed framework can ease the design of security system in cloud and reduce the complexity of system design and implementation.55 SLA-based Resource Allocation for Software as a Service Provider (SaaS) in Cloud Computing Environments Cloud computing has been considered as a solution for solving enterprise application distribution and configuration challenges in the traditional software sales model. Migrating from traditional software to Cloud enables on-going revenue for software providers. However, in order to deliver hosted services to customers, SaaS companies have to either maintain their own hardware or rent it from infrastructure providers. This requirement means that SaaS providers will incur extra costs. In order to minimize the cost of resources, it is also important to satisfy a minimum service level to customers. Therefore, this paper proposes resource allocation algorithms for SaaS providers who want to minimize infrastructure cost and SLA violations. Our proposed algorithms are designed in a way to ensure that Saas providers are able to manage the dynamic change of customers, mapping customer requests to infrastructure level parameters and handling heterogeneity of Virtual Machines. We take into account the customers’ Quality of Service parameters such as response time, and infrastructure level parameters such as service initiation time. This paper also presents an extensive evaluation study to analyze and demonstrate that our proposed algorithms minimize the SaaS provider’s cost and the number of SLA violations in a dynamic resource sharing Cloud environmentMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 17
  • 18. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-201256 Small Discrete Fourier Transforms on GPUs Efficient implementations of the Discrete Fourier Transform (DFT) for GPUs provide good performance with large data sizes, but are not competitive with CPU code for small data sizes. On the other hand, several applications perform multiple DFTs on small data sizes. In fact, even algorithms for large data sizes use a divide-andconquer approach, where eventually small DFTs need to be performed. We discuss our DFT implementation, which is efficient for multiple small DFTs. One feature of our implementation is the use of the asymptotically slow matrix multiplication approach for small data sizes, which improves performance on the GPU due to its regular memory access and computational patterns. We combine this algorithm with the mixed radix algorithm for 1-D, 2-D, and 3-D complex DFTs. We also demonstrate the effect of different optimization techniques. When GPUs are used to accelerate a component of an application running on the host, it is important that decisions taken to optimize the GPU performance not affect the performance of the rest of the application on the host. One feature of our implementation is that we use a data layout that is not optimal for the GPU so that the overall effect on the application is better. Our implementation performs up to two orders of magnitude faster than cuFFT on an NVIDIA GeForce 9800 GTX GPU and up to one to two orders of magnitude faster than FFTW on a CPU for multiple small DFTs. Furthermore, we show that our implementation can accelerate the performance of a Quantum Monte Carlo application for which cuFFT is not effective. The primary contributions of this work lie in demonstrating the utility of the matrix multiplication approach and also in providing an implementation that is efficient for small DFTs when a GPU is used to accelerate an application running on the host.57 Neural Control of Posture During Small Magnitude Perturbations: Effects of Aging and Localized Muscle Fatigue This study investigated the effects of aging and localized muscle fatigue on the neural control of upright stance during small postural perturbations. Sixteen young (aged 18–24 years) and 16 older (aged 55–74 years) participants were exposed to small magnitude, anteriorly-directed postural perturbations before and after fatiguing exercises (lumbar extensors and ankle plantar flexors). A single degree of freedom model of the human body was used to simulate recovery kinematics following the perturbations. Central to the model was a simulated neural controller that multiplied time-delayed kinematics by invariant feedback gains. Feedback gains and time delay were optimized for each participant based on measured kinematics, and a novel delay margin analysis was performed to assess system robustness. A 10.9% longer effective time delay (p = 0.010) was found among the older group, who also showed a greater reliance upon velocity feedback information (31.1% higher differential gain, p = 0.001) to control upright stance. Based on delay margins, older participants adopted a more robust control scheme to accommodate the small perturbations, potentially compensating for longer time delays or degraded sensory feedback. No fatigue-induced changes in neural controller gains, time delay, or delay margin were found in either age group, indicating that integration of this feedback information was not altered by muscle fatigue. The sensitivity of this approach to changes with fatigue may have been limited by model simplifications.58 Techniques for fine-grained, multi-site computation offloading Increasingly, mobile devices are becoming the preferred platform for computation for many users. Unfortunately, the resource limitations, in battery life, computation power and storage, restricts the richness of applications that can be run on such devices. To alleviate these concerns, a popular approach that has gained currency in recent years is computation offloading, where a portion of an application is run off-site, leveraging the far greater resources of the cloud. Prior work in this area has focused on a constrained form of the problem: a single mobile device offloading computation to a single server. However, with the increased popularity of cloud computing and storage, it is moreMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 18
  • 19. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 common for the data that an application accesses to be distributed among several servers. This paper describes algorithmic approaches for performing fine-grained, multi-site offloading. This allows portions of an application to be offloaded in a data-centric manner, even if that data exists at multiple sites. Our approach is based on a novel partitioning algorithm, and a novel data representation. We demonstrate that our partitioning algorithm outperforms existing multi-site offloading algorithms, and that our data representation provides for more efficient, fine-grained offloading than prior approaches.59 The Benefits of Estimated Global Information in DHT Load Balancing Distributed hash tables (DHT) often rely on uniform hashing for balancing the load among their nodes. However, the most overloaded node may still have a load up to O(logN) times higher than the average load [1]. DHTs with support for range queries cannot rely on hashing to fairly balance the system’s load since hashing destroys the order of the stored items. Ensuring a fair load distribution is vital to avoid individual nodes becoming overloaded, potentially leading to node crashes or an incentive not to participate in the system. In both scenarios explicit load balancing schemes can help to spread the load more evenly. In this paper, we improve on existing algorithms for itembased active load balancing by relying on approximations of global properties. We show that the algorithms can be made more efficient by incorporating estimates of properties such as the average load and the standard deviation. Our algorithms reduce the network traffic induced by load balancing while achieving a better load balance than standard algorithms. We also show that these improvements can be applied to passive load balancing algorithms. Compared to DHTs without explicit load balancing, both variants are able to reduce the total maintenance traffic, i.e. item movements due to churn and load balancing, by up to 18%. Simultaneously the system is being balanced achieving a better load distribution.60 Towards Real-Time, Volunteer Distributed Computing Many large-scale distributed computing applications demand real-time responses by soft deadlines. To enable such real-time task distribution and execution on the volunteer resources, we previously proposed the design of the realtime volunteer computing platform called RT-BOINC. The system gives low O(1) worst-case execution time for task management operations, such as task scheduling, state transitioning, and validation. In this work, we present a full implementation RT-BOINC, adding new features including deadline timer and parameter-based admission control. We evaluate RT-BOINC at large scale using two real-time applications, namely, the games Go and Chess. The results of our case study show that RT-BOINC provides much better performance than the original BOINC in terms of average and worst-case response time, scalability and efficiency.. 61 Towards Reliable, PerformantWorkflows for Streaming-Applications on Cloud Platforms Scientific workflows are commonplace in eScience applications. Yet, the lack of integrated support for data models, including streaming data, structured collections and files, is limiting the ability of workflows to support emerging applications in energy informatics that are stream oriented. This is compounded by the absence of Cloud data services that support reliable and performant streams. In this paper, we propose and present a scientific workflow framework that supports streams as first-class data, and is optimized for performant and reliable execution across desktop and Cloud platforms. The workflow framework features and its empirical evaluation on a private Eucalyptus Cloud are presented.62 Towards Secure and Dependable Storage Services in Cloud Computing Cloud storage enables users to remotely store their data and enjoy the on-demand high quality cloud applicationsMadurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 19
  • 20. Elysium Technologies Private Limited ISO 9001:2008 A leading Research and Development Division Madurai | Chennai | Trichy | Coimbatore | Kollam| Singapore Website: elysiumtechnologies.com, elysiumtechnologies.info Email: info@elysiumtechnologies.com IEEE Final Year Project List 2011-2012 without the burden of local hardware and software management. Though the benefits are clear, such a service is also relinquishing users’ physical possession of their outsourced data, which inevitably poses new security risks towards the correctness of the data in cloud. In order to address this new problem and further achieve a secure and dependable cloud storage service, we propose in this paper a flexible distributed storage integrity auditing mechanism, utilizing the homomorphic token and distributed erasure-coded data. The proposed design allows users to audit the cloud storage with very lightweight communication and computation cost. The auditing result not only ensures strong cloud storage correctness guarantee, but also simultaneously achieves fast data error localization, i.e., the identification of misbehaving server. Considering the cloud data are dynamic in nature, the proposed design further supports secure and efficient dynamic operations on outsourced data, including block modification, deletion, and append. Analysis shows the proposed scheme is highly efficient and resilient against Byzantine failure, malicious data modification attack, and even server colluding attacks.. 63 Utilizing “Opaque” Resources for Revenue Enhancement on Clouds and Grids Clouds and grids are distributed resource infrastructures that are subjected to both On Demand (OD) as well as Advance reservation (AR) requests. This paper focuses on the revenue that can be earned on cloud and grid systems using different matchmaking strategies that map requests to resources. Existing research on matchmaking uses a priori knowledge of the local scheduling policy used at the resources. However, in these dynamic and heterogeneous systems having detailed a priori knowledge about all the resources is not always feasible. Thus, traditional matchmaking strategies cannot use the resources for which the local scheduling policies are unknown. This paper discusses how a novel matchmaking strategy that does not use knowledge of the local scheduling policies used at the resources may be deployed in order to utilize these un-used resources and improve the revenue earned by the service providing enterprise. Based on analytic bounds and simulation, insights gained into system behaviour and revenue earned are presented.Madurai Trichy KollamElysium Technologies Private Limited Elysium Technologies Private Limited Elysium Technologies Private Limited230, Church Road, Annanagar, 3rd Floor,SI Towers, Surya Complex,Vendor junction,Madurai , Tamilnadu – 625 020. 15 ,Melapudur , Trichy, kollam,Kerala – 691 010.Contact : 91452 4390702, 4392702, 4394702. Tamilnadu – 620 001. Contact : 91474 2723622.eMail: info@elysiumtechnologies.com Contact : 91431 - 4002234. eMail: elysium.kollam@gmail.com eMail: elysium.trichy@gmail.com 20