We propose the Bio-UnaGrid infrastructure to facilitate the automatic execution of intensive-computing workflows that require the use of existing application suites and distributed computing infrastructures. With Bio-UnaGrid, bioinformatics workflows are easily created and executed, with a simple click and in a transparent manner, on different cluster and grid computing infrastructures (line command is not used). To provide more processing capabilities, at low cost, Bio-UnaGrid use the idle processing capabilities of computer labs with Windows, Linux and Mac desktop computers, using a key virtualization strategy. We implement Bio-UnaGrid in a dedicated cluster and a computer lab. Results of performance tests evidence the gain obtained by our researchers.
This document provides an abstract and biography for a talk on cloud operational visibility and analytics. The abstract discusses how cloud platforms and services are growing rapidly in complexity, which creates challenges for operational visibility and security. Traditional monitoring solutions are becoming ineffective. The talk will present an approach for deep, seamless visibility into cloud instances using agentless system crawlers. It will also discuss how this visibility can be used to develop operational and security analytics for the cloud. Specific projects for IBM containers called Agentless System Crawler and Vulnerability Advisor will be overviewed. The biography introduces the speaker, Dr. Canturk Isci, as a research manager working on cloud monitoring and security analytics at IBM Research.
Cooperative hierarchical based edge-computing approach for resources allocati...IJECEIAES
Using mobile and Internet of Things (IoT) applications is becoming very popular and obtained researchers’ interest and commercial investment, in order to fulfill future vision and the requirements for smart cities. These applications have common demands such as fast response, distributed nature, and awareness of service location. However, these requirements’ nature cannot be satisfied by central systems services that reside in the clouds. Therefore, edge computing paradigm has emerged to satisfy such demands, by providing an extension for cloud resources at the network edge, and consequently, they become closer to end-user devices. In this paper, exploiting edge resources is studied; therefore, a cooperative-hierarchical approach for executing the pre-partitioned applications’ modules between edges resources is proposed, in order to reduce traffic between the network core and the cloud, where this proposed approach has a polynomial-time complexity. Furthermore, edge computing increases the efficiency of providing services, and improves end-user experience. To validate our proposed cooperative-hierarchical approach for modules placement between edge nodes’ resources, iFogSim toolkit is used. The obtained simulation results show that the proposed approach reduces network’s load and the total delay compared to a baseline approach for modules’ placement, moreover, it increases the network’s overall throughput.
This document discusses the development of a Virtual Computing Lab (VCL) using a private cloud at Terna Engineering College in India. It begins with an abstract that outlines how a VCL and private cloud can help meet increasing demands on limited resources. It then reviews related work on VCL implementations at other universities. The proposed system would use the open-source Eucalyptus framework to create the private cloud infrastructure and provide on-demand access to virtual machines and applications through a web portal. Setting up the private cloud involves configuring controller and node machines running Eucalyptus services. Once complete, students and faculty could remotely launch and manage virtual environments. The conclusion discusses potential future work like expanding capabilities and mobile access.
Virtual Machine Allocation Policy in Cloud Computing Environment using CloudSim IJECEIAES
This document discusses virtual machine allocation policies in cloud computing environments using the CloudSim simulation tool. It begins with an introduction to cloud computing and discusses challenges related to resource management and energy consumption. It then reviews previous research on modeling approaches, energy optimization techniques, and network topologies. A UML class model is presented for analyzing energy consumption when accessing cloud servers arranged in a step network topology. The methodology section outlines how energy consumption by system components like processors, RAM, hard disks, and motherboards will be calculated. Simulation results will depict response times and cost details for different data center configurations and allocation policies.
Developing Web-based Interactive Teaching System for Core Network Technology ...drboon
This paper presents an in-house created software application that emulates and virtualizes computer networking laboratory hardware and proprietary equipment on PCs and servers. A user friendly remote laboratory manage system is developed using .NET technology. Through the presented lab management system, the students are able to conduct more than 200 lab assignments covering 10 courses from both lower level routing and switching classes to upper level network management and security classes that demand a large amount of physical equipment and a great number of student lab contact hours prior to the emulated lab. This emulation solution eliminates our dependencies on expensive proprietary networking equipment. It also reduces the burden of system administration. This remote lab has been used and tested in three networking courses. The feedbacks show that the lab delivers similar or better lab experiences for the DE students while the lab construction and maintenance cost is reduced to the minimum.
Abstract:
The explosive growth of online services powered by data centres (web search, cloud computing, etc.) has motivated intense research into data centre network (DCN) design over the past decade. Computational demands are testing the limits of current engineering capabilities, and a new field in theoretical research has emerged with its roots in the area of traditional interconnection networks; adaptations of well-understood topologies such as generalized hypercubes, fat-trees, random regular graphs, WK-recursive networks, butterfly networks, as well as topologies geared explicitly towards DCNs, have all been proposed as DCN topologies in the last 7 years. Along with these custom-built (graph) topologies comes the need for theoretical analysis in scenarios that can be radically different from those ordinarily expected in traditional networks. I will give an overview of this emerging field.
I will then present a generic method of adapting a suitably chosen graph G to build a server-centric DCN topology G*, in such a way that any good networking properties of G can be preserved. In particular, routing algorithms on G can be used in G*, and when G is regular, the bisection width of G* (a well known throughput metric) can be obtained from the solution to an edge-isoperimetric problem on G.
This document provides an abstract and biography for a talk on cloud operational visibility and analytics. The abstract discusses how cloud platforms and services are growing rapidly in complexity, which creates challenges for operational visibility and security. Traditional monitoring solutions are becoming ineffective. The talk will present an approach for deep, seamless visibility into cloud instances using agentless system crawlers. It will also discuss how this visibility can be used to develop operational and security analytics for the cloud. Specific projects for IBM containers called Agentless System Crawler and Vulnerability Advisor will be overviewed. The biography introduces the speaker, Dr. Canturk Isci, as a research manager working on cloud monitoring and security analytics at IBM Research.
Cooperative hierarchical based edge-computing approach for resources allocati...IJECEIAES
Using mobile and Internet of Things (IoT) applications is becoming very popular and obtained researchers’ interest and commercial investment, in order to fulfill future vision and the requirements for smart cities. These applications have common demands such as fast response, distributed nature, and awareness of service location. However, these requirements’ nature cannot be satisfied by central systems services that reside in the clouds. Therefore, edge computing paradigm has emerged to satisfy such demands, by providing an extension for cloud resources at the network edge, and consequently, they become closer to end-user devices. In this paper, exploiting edge resources is studied; therefore, a cooperative-hierarchical approach for executing the pre-partitioned applications’ modules between edges resources is proposed, in order to reduce traffic between the network core and the cloud, where this proposed approach has a polynomial-time complexity. Furthermore, edge computing increases the efficiency of providing services, and improves end-user experience. To validate our proposed cooperative-hierarchical approach for modules placement between edge nodes’ resources, iFogSim toolkit is used. The obtained simulation results show that the proposed approach reduces network’s load and the total delay compared to a baseline approach for modules’ placement, moreover, it increases the network’s overall throughput.
This document discusses the development of a Virtual Computing Lab (VCL) using a private cloud at Terna Engineering College in India. It begins with an abstract that outlines how a VCL and private cloud can help meet increasing demands on limited resources. It then reviews related work on VCL implementations at other universities. The proposed system would use the open-source Eucalyptus framework to create the private cloud infrastructure and provide on-demand access to virtual machines and applications through a web portal. Setting up the private cloud involves configuring controller and node machines running Eucalyptus services. Once complete, students and faculty could remotely launch and manage virtual environments. The conclusion discusses potential future work like expanding capabilities and mobile access.
Virtual Machine Allocation Policy in Cloud Computing Environment using CloudSim IJECEIAES
This document discusses virtual machine allocation policies in cloud computing environments using the CloudSim simulation tool. It begins with an introduction to cloud computing and discusses challenges related to resource management and energy consumption. It then reviews previous research on modeling approaches, energy optimization techniques, and network topologies. A UML class model is presented for analyzing energy consumption when accessing cloud servers arranged in a step network topology. The methodology section outlines how energy consumption by system components like processors, RAM, hard disks, and motherboards will be calculated. Simulation results will depict response times and cost details for different data center configurations and allocation policies.
Developing Web-based Interactive Teaching System for Core Network Technology ...drboon
This paper presents an in-house created software application that emulates and virtualizes computer networking laboratory hardware and proprietary equipment on PCs and servers. A user friendly remote laboratory manage system is developed using .NET technology. Through the presented lab management system, the students are able to conduct more than 200 lab assignments covering 10 courses from both lower level routing and switching classes to upper level network management and security classes that demand a large amount of physical equipment and a great number of student lab contact hours prior to the emulated lab. This emulation solution eliminates our dependencies on expensive proprietary networking equipment. It also reduces the burden of system administration. This remote lab has been used and tested in three networking courses. The feedbacks show that the lab delivers similar or better lab experiences for the DE students while the lab construction and maintenance cost is reduced to the minimum.
Abstract:
The explosive growth of online services powered by data centres (web search, cloud computing, etc.) has motivated intense research into data centre network (DCN) design over the past decade. Computational demands are testing the limits of current engineering capabilities, and a new field in theoretical research has emerged with its roots in the area of traditional interconnection networks; adaptations of well-understood topologies such as generalized hypercubes, fat-trees, random regular graphs, WK-recursive networks, butterfly networks, as well as topologies geared explicitly towards DCNs, have all been proposed as DCN topologies in the last 7 years. Along with these custom-built (graph) topologies comes the need for theoretical analysis in scenarios that can be radically different from those ordinarily expected in traditional networks. I will give an overview of this emerging field.
I will then present a generic method of adapting a suitably chosen graph G to build a server-centric DCN topology G*, in such a way that any good networking properties of G can be preserved. In particular, routing algorithms on G can be used in G*, and when G is regular, the bisection width of G* (a well known throughput metric) can be obtained from the solution to an edge-isoperimetric problem on G.
Efficiency and Effectiveness: Shared services to support STEM subjectsJisc
The VRE-CI project is funded by JISC to provide a framework for researchers and clinicians involved in cancer imaging to share information, images, and algorithms. The project will develop a module for cancer imaging that allows sharing of DICOM images and image processing algorithms through scientific workflows and web services. It will also include an image viewer using DeepZoom technology and functionality for loading images securely while maintaining appropriate levels of anonymity.
Information Technology in Industry(ITII) - November Issue 2018ITIIIndustries
IT Industry publishes original research articles, review articles, and extended versions of conference papers. Articles resulting from research of both theoretical and/or practical natures performed by academics and/or industry practitioners are welcome. IT in Industry aims to become a leading IT journal with a high impact factor.
Efficient Point Cloud Pre-processing using The Point Cloud LibraryCSCJournals
Robotics, video games, environmental mapping and medical are some of the fields that use 3D data processing. In this paper we propose a novel optimization approach for the open source Point Cloud Library (PCL) that is frequently used for processing 3D data. Three main aspects of the PCL are discussed: point cloud creation from disparity of color image pairs; voxel grid downsample filtering to simplify point clouds; and passthrough filtering to adjust the size of the point cloud. Additionally, OpenGL shader based rendering is examined. An optimization technique based on CPU cycle measurement is proposed and applied in order to optimize those parts of the pre-processing chain where measured performance is slowest. Results show that with optimized modules the performance of the pre-processing chain has increased 69 fold.
TexGen: Open Source Software for Modelling of Textile CompositesJisc
TexGen is open source software for modeling textile composites that was developed at the University of Nottingham. It was reimplemented in 2006 as a platform-independent code under an open source license to address limitations of the previous version. Releasing it as open source encouraged its use in research, resulting in over 5,000 downloads and supporting over $1 million in funded projects.
[Thesis] Tangible Collaboration applied in Space Systems Concurrent Engineeri...Christopher Cerqueira
This document discusses applying tangible user interfaces (TUIs) to space systems concurrent engineering (CE) concept studies sessions. It proposes using physical TUI artifacts to represent models that multidisciplinary engineering teams currently manipulate digitally. The document reviews how teams currently collaborate using predominantly screen-based tools, and analyzes collaborative behaviors and model types used in CE sessions. It then presents the concept of TUIs and a taxonomy for classifying their elements and interactions. Finally, the document proposes a proof-of-concept TUI design for CE sessions involving physical artifact representations of models that team members can directly manipulate.
The document discusses using Smalltalk for logic synthesis of circuits from behavioral code. It describes compiling Smalltalk methods into logic graphs by propagating type information to infer the computation being performed. This allows fully automatic synthesis of combinational logic circuits from applicative Smalltalk programs. However, numeric synthesis of data paths and microprocessor meshes has not yet been implemented.
Design of an IT Capstone Subject - Cloud RoboticsITIIIndustries
This paper describes the curriculum of the three year IT undergraduate program at La Trobe University, and the faculty requirements in designing a capstone subject, followed by the ACM’s recommended IT curriculum covering the five pillars of the IT discipline. Cloud robotics, a broad multidisciplinary research area, requiring expertise in all five pillars with mechatronics, is an ideal candidate to offer capstone experiences to IT students. Therefore, in this paper, we propose a long term master project in developing a cloud robotics testbed, with many capstone sub-projects spanning across the five IT pillars, to meet the objectives of capstone experience. This paper also describes the design and implementation of the testbed, and proposes potential capstone projects for students with different interests.
This document discusses performance analysis of cloud computing services. It begins by defining cloud computing and describing its key characteristics like on-demand access to computing resources and pay-per-use models. It then reviews several studies on using virtualization technologies and frameworks for evaluating cloud performance and workload generation. The document concludes that tools are needed for comprehensive performance analysis of large scientific clouds to evaluate metrics like response time, cost and scalability across different cloud vendors.
Cloud computing is a realized wonder. It delights its users by providing applications, platforms and infrastructure without any initial investment. The “pay as you use” strategy comforts the users. The usage can be increased by adding infrastructure, tools or applications to the existing application. The realistic beauty of cloud computing is that there is no need for any sophisticated tool for access, web browser or even smartphone will do. Cloud computing is a windfall for small organizations having less sensitive information. But for large organizations, the risks related to security may be daunting. Necessary steps have to be taken for managing the issues like confidentiality, integrity, privacy, availability and so on. In this paper availability is taken and studied in a multi-dimensional perspective. Availability is taken a key issue and the mechanisms that enable enhancement are analyzed.
A High Throughput Bioinformatics Distributed Computing PlatformHabibur Rahman
A Research Paper review presentation on "A High throughput bioinformatics distribute computing platform", presented by Md. Habibur Rahman, BIT0216, Institute of Information Technology University of Dhaka.
- Acronis is a global leader in cyber protection with over 5.5 million prosumers and $250 million in revenue. It has dual headquarters in Switzerland and Singapore.
- The document discusses future computing technologies like quantum computing, photonic computing, brain-inspired computing and their potential to solve problems beyond the capabilities of classical computers. It also discusses challenges like fundamental physical limits, heat dissipation and the need for new materials and algorithms.
- A new research university called SIT is proposed to address global challenges through technology and innovation in areas like cybersecurity, AI, quantum technologies and new materials. It will be located in Schaffhausen, Switzerland near the Rhein Falls and partner with top universities
On Modeling and Testing When Unpredictability Becomes the Pattern (April 2nd,...Benoit Combemale
Unpredictability is becoming the pattern for modeling and testing complex software systems. Model-driven engineering (MDE) aims to manage variability at both design-time and run-time to support unpredictable changes. Feature models are used to represent variability and manage configurations. The Common Variability Language (CVL) provides a standardized way to describe and resolve variability models. MDE supports adapting systems dynamically in response to changes by automatically re-weaving aspects of models.
IRJET- A Real Time Yolo Human Detection in Flood Affected Areas based on Vide...IRJET Journal
This document proposes a method for real-time human detection in flood-affected areas using video content analysis and the YOLO object detection algorithm. It trains YOLO on the COCO Human dataset to detect and localize humans in video frames from surveillance cameras. The results show that YOLO can accurately detect multiple humans, even with occlusion, and single humans under varying illumination. This approach aims to help rescue operations quickly identify affected areas and prioritize aid.
Weeds detection efficiency through different convolutional neural networks te...IJECEIAES
The preservation of the environment has become a priority and a subject that is receiving more and more attention. This is particularly important in the field of precision agriculture, where pesticide and herbicide use has become more controlled. In this study, we propose to evaluate the ability of the deep learning (DL) and convolutional neural network (CNNs) technology to detect weeds in several types of crops using a perspective and proximity images to enable localized and ultra-localized herbicide spraying in the region of Beni Mellal in Morocco. We studied the detection of weeds through six recent CNN known for their speed and precision, namely, VGGNet (16 and 19), GoogLeNet (Inception V3 and V4) and MobileNet (V1 and V2). The first experiment was performed with the CNNs architectures from scratch and the second experiment with their pre-trained versions. The results showed that Inception V4 achieved the highest precision with a rate of 99.41% and 99.51% on the mixed image sets and for its version from scratch and its pre-trained version respectively, and that MobileNet V2 was the fastest and lightest with its size of 14 MB.
The document discusses the DAME (Data Mining & Exploration) project, which aims to implement data mining applications and services for massive data analysis and exploration using a distributed computing environment. It seeks to standardize data mining methods and make them interoperable within the virtual observatory. The project has developed several web applications and investigates using a plugin architecture and standardized accounting to improve interoperability between applications and minimize data transfer requirements. The goal is to develop a unified data mining application approach for the virtual observatory.
The document discusses a proposed novel many-core architecture called FlexTiles that is based on reconfigurable devices like FPGAs, DSPs, and GPPs. It aims to provide an adaptive technique and autonomous decision making to improve programming efficiency and reduce time to market for applications with time-varying workloads. The project has a budget of 3.67M euros over 3 years and involves several partner organizations.
1. Large scale GPU clusters are increasingly being used for machine learning training as neural network architectures become more complex and distributed training becomes necessary.
2. New trends in machine learning include more complex neural network architectures, diverse data types and applications, automated machine learning, and federated learning which distributes training across decentralized devices.
3. To support these new trends, machine learning platforms need to enable fine-grained customization of hardware and software as well as distributed training across multiple nodes.
Nanotechnology is a broad enabling technology that can transform industries from low-tech to high-tech by controlling matter at the atomic and molecular scales. It will impact many areas including imaging, computing, modeling, and sensors. The author has participated in several nanotechnology projects involving modeling molecular structures, imaging silicon wafers for defects, developing technologies for future combat vehicles, incorporating image processing into transmission electron microscopes, developing materials for space missions and vehicles, inspecting composites, and compressing speech for consumer devices. Nanotechnology is being applied across industries like defense, automotive, aerospace, and computing.
Asymmetric image encryption scheme based on Massey Omura scheme IJECEIAES
The document describes a proposed asymmetric image encryption scheme based on the Massey Omura cryptosystem. The scheme works as follows: the sender and receiver agree on public parameters like a prime number p. The sender then represents the plaintext image as a matrix and encrypts it using their public key to get the first ciphertext image. This is sent to the receiver who encrypts it again with their public key to produce the second ciphertext image, which is sent back to the sender. The sender decrypts this using their private key to produce the third ciphertext image, which is sent to and decrypted by the receiver using their private key to recover the original plaintext image. The document provides an example of implementing the scheme and discusses evaluating its security using
The Grid Observatory Initiative aims to create a digital curation process for monitoring and analyzing data from the EGI grid infrastructure. It preserves operational monitoring data, makes it available through its web portal, and analyzes it to build knowledge about globalized IT systems and e-science computing practices. Research using the data helps improve systems performance while advancing scientific understanding of complex systems and computational behaviors.
UnaCloud is an opportunistic based cloud infrastructure
(IaaS) that allows to access on-demand computing
capabilities using commodity desktops. Although UnaCloud
tried to maximize the use of idle resources to deploy virtual
machines on them, it does not use energy-efficient resource
allocation algorithms. In this paper, we design and implement
different energy-aware techniques to operate in an energyefficient
way and at the same time guarantee the performance
to the users. Performance tests with different algorithms and
scenarios using real trace workloads from UnaCloud, show how
different policies can change the energy consumption patterns
and reduce the energy consumption in opportunistic cloud
infrastructures. The results show that some algorithms can
reduce the energy-consumption power up to 30% over the
percentage earned by opportunistic environment.
El documento describe las diferentes formas de aumentar la capacidad computacional disponible para investigadores, incluyendo el uso de servidores, clústeres y grids computacionales. Explica que un grid computacional permite compartir recursos de cómputo y almacenamiento entre organizaciones para apoyar proyectos de investigación de forma transparente. También describe los esfuerzos en Colombia para establecer una infraestructura grid nacional que permita compartir recursos entre instituciones.
Efficiency and Effectiveness: Shared services to support STEM subjectsJisc
The VRE-CI project is funded by JISC to provide a framework for researchers and clinicians involved in cancer imaging to share information, images, and algorithms. The project will develop a module for cancer imaging that allows sharing of DICOM images and image processing algorithms through scientific workflows and web services. It will also include an image viewer using DeepZoom technology and functionality for loading images securely while maintaining appropriate levels of anonymity.
Information Technology in Industry(ITII) - November Issue 2018ITIIIndustries
IT Industry publishes original research articles, review articles, and extended versions of conference papers. Articles resulting from research of both theoretical and/or practical natures performed by academics and/or industry practitioners are welcome. IT in Industry aims to become a leading IT journal with a high impact factor.
Efficient Point Cloud Pre-processing using The Point Cloud LibraryCSCJournals
Robotics, video games, environmental mapping and medical are some of the fields that use 3D data processing. In this paper we propose a novel optimization approach for the open source Point Cloud Library (PCL) that is frequently used for processing 3D data. Three main aspects of the PCL are discussed: point cloud creation from disparity of color image pairs; voxel grid downsample filtering to simplify point clouds; and passthrough filtering to adjust the size of the point cloud. Additionally, OpenGL shader based rendering is examined. An optimization technique based on CPU cycle measurement is proposed and applied in order to optimize those parts of the pre-processing chain where measured performance is slowest. Results show that with optimized modules the performance of the pre-processing chain has increased 69 fold.
TexGen: Open Source Software for Modelling of Textile CompositesJisc
TexGen is open source software for modeling textile composites that was developed at the University of Nottingham. It was reimplemented in 2006 as a platform-independent code under an open source license to address limitations of the previous version. Releasing it as open source encouraged its use in research, resulting in over 5,000 downloads and supporting over $1 million in funded projects.
[Thesis] Tangible Collaboration applied in Space Systems Concurrent Engineeri...Christopher Cerqueira
This document discusses applying tangible user interfaces (TUIs) to space systems concurrent engineering (CE) concept studies sessions. It proposes using physical TUI artifacts to represent models that multidisciplinary engineering teams currently manipulate digitally. The document reviews how teams currently collaborate using predominantly screen-based tools, and analyzes collaborative behaviors and model types used in CE sessions. It then presents the concept of TUIs and a taxonomy for classifying their elements and interactions. Finally, the document proposes a proof-of-concept TUI design for CE sessions involving physical artifact representations of models that team members can directly manipulate.
The document discusses using Smalltalk for logic synthesis of circuits from behavioral code. It describes compiling Smalltalk methods into logic graphs by propagating type information to infer the computation being performed. This allows fully automatic synthesis of combinational logic circuits from applicative Smalltalk programs. However, numeric synthesis of data paths and microprocessor meshes has not yet been implemented.
Design of an IT Capstone Subject - Cloud RoboticsITIIIndustries
This paper describes the curriculum of the three year IT undergraduate program at La Trobe University, and the faculty requirements in designing a capstone subject, followed by the ACM’s recommended IT curriculum covering the five pillars of the IT discipline. Cloud robotics, a broad multidisciplinary research area, requiring expertise in all five pillars with mechatronics, is an ideal candidate to offer capstone experiences to IT students. Therefore, in this paper, we propose a long term master project in developing a cloud robotics testbed, with many capstone sub-projects spanning across the five IT pillars, to meet the objectives of capstone experience. This paper also describes the design and implementation of the testbed, and proposes potential capstone projects for students with different interests.
This document discusses performance analysis of cloud computing services. It begins by defining cloud computing and describing its key characteristics like on-demand access to computing resources and pay-per-use models. It then reviews several studies on using virtualization technologies and frameworks for evaluating cloud performance and workload generation. The document concludes that tools are needed for comprehensive performance analysis of large scientific clouds to evaluate metrics like response time, cost and scalability across different cloud vendors.
Cloud computing is a realized wonder. It delights its users by providing applications, platforms and infrastructure without any initial investment. The “pay as you use” strategy comforts the users. The usage can be increased by adding infrastructure, tools or applications to the existing application. The realistic beauty of cloud computing is that there is no need for any sophisticated tool for access, web browser or even smartphone will do. Cloud computing is a windfall for small organizations having less sensitive information. But for large organizations, the risks related to security may be daunting. Necessary steps have to be taken for managing the issues like confidentiality, integrity, privacy, availability and so on. In this paper availability is taken and studied in a multi-dimensional perspective. Availability is taken a key issue and the mechanisms that enable enhancement are analyzed.
A High Throughput Bioinformatics Distributed Computing PlatformHabibur Rahman
A Research Paper review presentation on "A High throughput bioinformatics distribute computing platform", presented by Md. Habibur Rahman, BIT0216, Institute of Information Technology University of Dhaka.
- Acronis is a global leader in cyber protection with over 5.5 million prosumers and $250 million in revenue. It has dual headquarters in Switzerland and Singapore.
- The document discusses future computing technologies like quantum computing, photonic computing, brain-inspired computing and their potential to solve problems beyond the capabilities of classical computers. It also discusses challenges like fundamental physical limits, heat dissipation and the need for new materials and algorithms.
- A new research university called SIT is proposed to address global challenges through technology and innovation in areas like cybersecurity, AI, quantum technologies and new materials. It will be located in Schaffhausen, Switzerland near the Rhein Falls and partner with top universities
On Modeling and Testing When Unpredictability Becomes the Pattern (April 2nd,...Benoit Combemale
Unpredictability is becoming the pattern for modeling and testing complex software systems. Model-driven engineering (MDE) aims to manage variability at both design-time and run-time to support unpredictable changes. Feature models are used to represent variability and manage configurations. The Common Variability Language (CVL) provides a standardized way to describe and resolve variability models. MDE supports adapting systems dynamically in response to changes by automatically re-weaving aspects of models.
IRJET- A Real Time Yolo Human Detection in Flood Affected Areas based on Vide...IRJET Journal
This document proposes a method for real-time human detection in flood-affected areas using video content analysis and the YOLO object detection algorithm. It trains YOLO on the COCO Human dataset to detect and localize humans in video frames from surveillance cameras. The results show that YOLO can accurately detect multiple humans, even with occlusion, and single humans under varying illumination. This approach aims to help rescue operations quickly identify affected areas and prioritize aid.
Weeds detection efficiency through different convolutional neural networks te...IJECEIAES
The preservation of the environment has become a priority and a subject that is receiving more and more attention. This is particularly important in the field of precision agriculture, where pesticide and herbicide use has become more controlled. In this study, we propose to evaluate the ability of the deep learning (DL) and convolutional neural network (CNNs) technology to detect weeds in several types of crops using a perspective and proximity images to enable localized and ultra-localized herbicide spraying in the region of Beni Mellal in Morocco. We studied the detection of weeds through six recent CNN known for their speed and precision, namely, VGGNet (16 and 19), GoogLeNet (Inception V3 and V4) and MobileNet (V1 and V2). The first experiment was performed with the CNNs architectures from scratch and the second experiment with their pre-trained versions. The results showed that Inception V4 achieved the highest precision with a rate of 99.41% and 99.51% on the mixed image sets and for its version from scratch and its pre-trained version respectively, and that MobileNet V2 was the fastest and lightest with its size of 14 MB.
The document discusses the DAME (Data Mining & Exploration) project, which aims to implement data mining applications and services for massive data analysis and exploration using a distributed computing environment. It seeks to standardize data mining methods and make them interoperable within the virtual observatory. The project has developed several web applications and investigates using a plugin architecture and standardized accounting to improve interoperability between applications and minimize data transfer requirements. The goal is to develop a unified data mining application approach for the virtual observatory.
The document discusses a proposed novel many-core architecture called FlexTiles that is based on reconfigurable devices like FPGAs, DSPs, and GPPs. It aims to provide an adaptive technique and autonomous decision making to improve programming efficiency and reduce time to market for applications with time-varying workloads. The project has a budget of 3.67M euros over 3 years and involves several partner organizations.
1. Large scale GPU clusters are increasingly being used for machine learning training as neural network architectures become more complex and distributed training becomes necessary.
2. New trends in machine learning include more complex neural network architectures, diverse data types and applications, automated machine learning, and federated learning which distributes training across decentralized devices.
3. To support these new trends, machine learning platforms need to enable fine-grained customization of hardware and software as well as distributed training across multiple nodes.
Nanotechnology is a broad enabling technology that can transform industries from low-tech to high-tech by controlling matter at the atomic and molecular scales. It will impact many areas including imaging, computing, modeling, and sensors. The author has participated in several nanotechnology projects involving modeling molecular structures, imaging silicon wafers for defects, developing technologies for future combat vehicles, incorporating image processing into transmission electron microscopes, developing materials for space missions and vehicles, inspecting composites, and compressing speech for consumer devices. Nanotechnology is being applied across industries like defense, automotive, aerospace, and computing.
Asymmetric image encryption scheme based on Massey Omura scheme IJECEIAES
The document describes a proposed asymmetric image encryption scheme based on the Massey Omura cryptosystem. The scheme works as follows: the sender and receiver agree on public parameters like a prime number p. The sender then represents the plaintext image as a matrix and encrypts it using their public key to get the first ciphertext image. This is sent to the receiver who encrypts it again with their public key to produce the second ciphertext image, which is sent back to the sender. The sender decrypts this using their private key to produce the third ciphertext image, which is sent to and decrypted by the receiver using their private key to recover the original plaintext image. The document provides an example of implementing the scheme and discusses evaluating its security using
The Grid Observatory Initiative aims to create a digital curation process for monitoring and analyzing data from the EGI grid infrastructure. It preserves operational monitoring data, makes it available through its web portal, and analyzes it to build knowledge about globalized IT systems and e-science computing practices. Research using the data helps improve systems performance while advancing scientific understanding of complex systems and computational behaviors.
UnaCloud is an opportunistic based cloud infrastructure
(IaaS) that allows to access on-demand computing
capabilities using commodity desktops. Although UnaCloud
tried to maximize the use of idle resources to deploy virtual
machines on them, it does not use energy-efficient resource
allocation algorithms. In this paper, we design and implement
different energy-aware techniques to operate in an energyefficient
way and at the same time guarantee the performance
to the users. Performance tests with different algorithms and
scenarios using real trace workloads from UnaCloud, show how
different policies can change the energy consumption patterns
and reduce the energy consumption in opportunistic cloud
infrastructures. The results show that some algorithms can
reduce the energy-consumption power up to 30% over the
percentage earned by opportunistic environment.
El documento describe las diferentes formas de aumentar la capacidad computacional disponible para investigadores, incluyendo el uso de servidores, clústeres y grids computacionales. Explica que un grid computacional permite compartir recursos de cómputo y almacenamiento entre organizaciones para apoyar proyectos de investigación de forma transparente. También describe los esfuerzos en Colombia para establecer una infraestructura grid nacional que permita compartir recursos entre instituciones.
Cloud Computing es un modelo que está siendo adoptado por muchas empresas a nivel mundial ya que los más grandes proveedores de software y vendedores de software independiente (ISVs) de todos los tamaños han empezado a diseñar sus aplicaciones para que sean consumidas bajo el modelo cloud computing de Software como Servicio (SaaS). El diseño de soluciones de Software como Servicio involucra grandes retos a nivel técnico incluyendo aspectos como la escalabilidad
de la soluciones, los niveles de personalización, el modelo multi-arrendatario, el acceso desde múltiples dispositivos, la seguridad y el desarrollo sobre servicios PaaS e IaaS son algunos de los temas abordados en esta presentación realizada en el XXXIII Salón de Informática realizado en la ciudad de Bogotá, Colombia (2013).
UnaGrid is an opportunistic virtual grid infrastructure that takes advantage of the idle processing capabilities of conventional desktop machines in computer rooms through the use of Customizable Processing Virtual Clusters (CPVCs), these capabilities are used within the development of e-Science projects. This paper presents the design, implementation and assessment of a virtual storage system, which simultaneously allows UnaGrid to take advantage of the storage and processing capabilities available in tens or hundreds of desktop machines. The first tests have shown that this system allows attaining large storage capabilities, at low cost, and superior performance than a NFS-NAS dedicated solution.
El documento describe el emprendimiento en Internet y cómo ha cambiado el mundo a través de nuevas tecnologías y empresas. Explica que Internet seguirá cambiando todas las industrias a medida que surjan nuevos emprendimientos internacionales y colombianos. También cubre cómo comienza un emprendedor, el enfoque de Lean Startup, la validación de clientes, los tres estadios de una startup, la inversión y el miedo al fracaso.
Cloud computing promises to offer great opportunities for research groups; however when researchers want to execute applications in cloud infrastructures many complex processes must be accomplished. In this presentation we present the e-Clouds project which will allow researchers to easily execute many applications on public Infrastructure as a Service (IaaS) solutions. Designed for being a Software as a Service (SaaS) marketplace for scientific applications, e-Clouds allows researchers to submit jobs which are transparently executed on public IaaS platforms, such as Amazon Web Services (AWS). e-Clouds manages the on-demand provisioning and configuration of computing instances, storage, applications, schedulers, jobs, and data. The architectural design and how a first application has been supported on e-Clouds are presented. e-Clouds will allow researchers to easily share and execute applications in the cloud at low TCO (Total Cost of Ownership) and without the complexities associated with details of IT configurations and management. e-Clouds provides new opportunities for research groups with low or none budget for dedicated cluster or grid solutions, providing on-demand access to ready-to-use applications and accelerating the result generation of e-Science projects.
This document summarizes existing IoT testbed frameworks and identifies gaps. It discusses how existing testbeds are mostly based on wireless sensor networks using proprietary hardware, which causes issues like vendor lock-in, lack of interoperability, and inability to scale. It also notes simulations can provide unrealistic results. The document introduces a proposed new open IoT testbed that aims to address these issues by offering sensors, actuators, and APIs as interoperable services using open source devices and standards.
This document summarizes existing IoT testbed frameworks and identifies gaps. It describes the proposed research which aims to address these gaps by developing an open IoT testbed that offers sensors, actuators, platforms and APIs as interoperable services. The testbed would improve on existing approaches by standardizing data formats, optimizing data collection and handling errors. It explores publishing sensor data using a message broker to reduce dependencies and learning curves.
Grid computing is a distributed computing model that enables transparent sharing and aggregation of computing, storage, and network resources across dynamic and geographically dispersed organizations. Key characteristics include distributing computational resources among multiple and widely separated sources and users, providing a means for using distributed resources to solve large problems, and making resources appear as a single virtual machine with powerful capabilities. Example applications discussed include scientific computing, business applications, and volunteer computing projects.
Conference: IEEE 16th International Conference on
Industrial Informatics (INDIN2018).
Porto, Portugal – July 18-20, 2018
Title of the paper: Towards the Adoption of Cyber-
Physical Systems of Systems Paradigm in Smart
Manufacturing Environments
Authors: Borja Ramis Ferrer, Gerardo Beruvides,
Wael M. Mohammed, Alberto Villalonga, Jose L.
Martinez Lastra, Fernando Castaño, Rodolfo E.
Haber
This document discusses an open IoT testbed and architectural framework. It describes IoT systems as consisting of interconnected devices that can communicate and exchange data. A core component is embedded systems/devices that include sensors to measure the environment and actuators to perform physical actions. Microcontrollers interface with these devices and communicate via various protocols. The document proposes an open IoT testbed with a control plane that can discover resources/services, orchestrate based on user demands, and resolve conflicts through a lock release model. It provides a functional and detailed architecture for the proposed framework.
This document discusses an open IoT testbed and architectural framework. It describes IoT systems as consisting of interconnected devices that can communicate and exchange data. A core component is embedded systems/devices that include sensors to measure the environment and actuators to perform physical actions. Microcontrollers interface with these devices and communicate via various protocols. The document proposes an open IoT testbed with a control plane that can discover resources/services, orchestrate based on user demands, and resolve conflicts through a lock release model. It provides a functional and detailed architecture for the proposed framework.
German Conference on Bioinformatics 2021
https://gcb2021.de/
FAIR Computational Workflows
Computational workflows capture precise descriptions of the steps and data dependencies needed to carry out computational data pipelines, analysis and simulations in many areas of Science, including the Life Sciences. The use of computational workflows to manage these multi-step computational processes has accelerated in the past few years driven by the need for scalable data processing, the exchange of processing know-how, and the desire for more reproducible (or at least transparent) and quality assured processing methods. The SARS-CoV-2 pandemic has significantly highlighted the value of workflows.
This increased interest in workflows has been matched by the number of workflow management systems available to scientists (Galaxy, Snakemake, Nextflow and 270+ more) and the number of workflow services like registries and monitors. There is also recognition that workflows are first class, publishable Research Objects just as data are. They deserve their own FAIR (Findable, Accessible, Interoperable, Reusable) principles and services that cater for their dual roles as explicit method description and software method execution [1]. To promote long-term usability and uptake by the scientific community, workflows (as well as the tools that integrate them) should become FAIR+R(eproducible), and citable so that author’s credit is attributed fairly and accurately.
The work on improving the FAIRness of workflows has already started and a whole ecosystem of tools, guidelines and best practices has been under development to reduce the time needed to adapt, reuse and extend existing scientific workflows. An example is the EOSC-Life Cluster of 13 European Biomedical Research Infrastructures which is developing a FAIR Workflow Collaboratory based on the ELIXIR Research Infrastructure for Life Science Data Tools ecosystem. While there are many tools for addressing different aspects of FAIR workflows, many challenges remain for describing, annotating, and exposing scientific workflows so that they can be found, understood and reused by other scientists.
This keynote will explore the FAIR principles for computational workflows in the Life Science using the EOSC-Life Workflow Collaboratory as an example.
[1] Carole Goble, Sarah Cohen-Boulakia, Stian Soiland-Reyes,Daniel Garijo, Yolanda Gil, Michael R. Crusoe, Kristian Peters, and Daniel Schober FAIR Computational Workflows Data Intelligence 2020 2:1-2, 108-121 https://doi.org/10.1162/dint_a_00033.
Philip Morris International deals with large quantities of scientific data from various sources in different formats. They developed an agile approach using microservices and graph databases to integrate data and provide a single interface for users. A prototype was completed by one scientist and developer in less than 3 months, called the ChemoInformatics KnowledgeBase, which assembles chemical constituent data and associates it with internal and public scientific studies.
Presentation in the "Whole genome sequencing for clinical microbiology:Translation into routine applications" Symposium , Basel , Switzerland, 2 Sep 2017
The document discusses providing actuator and sensor access as a service over the internet. It proposes an algorithm for resource requisition that creates locks on actuator instances to prevent multiple simultaneous requests. This ensures actuators can only respond to one command at a time. The algorithm also analyzes request volume to optimize traffic to unavailable resources. An API is developed to abstract away hardware details and provide platform-independent parameter retrieval and actuation. This allows developers to focus on application logic rather than hardware integration.
This document discusses the need for an open source IoT development environment and testbed to allow software developers to create IoT applications without requiring hardware expertise. It notes that existing IoT testbeds often use proprietary hardware and software, limiting interoperability. The proposed solution aims to provide virtual access to sensors and actuators through an API, as well as a microcontroller platform as a service. This would allow developers to write code without worrying about hardware integration and deployment details. The goal is to make IoT development and testing more accessible through an open testbed that addresses issues like sensor availability and cost.
Using Microservices to Design Patient-facing Research SoftwareMartin Chapman
Using microservices allows patient-facing research software to address challenges in software development, modularity, and processing time. The microservices architecture separates a system into individual communicating services, each providing a single functionality. This allows different languages to be used, logic to be replaced with minimal impact, and long execution times to be isolated. The CONSULT system demonstrates these benefits through its use of microservices to integrate data from various sources and provide decision support for stroke patients.
This document discusses challenges and opportunities around data management for integrated structural biology research. It notes that structural biology projects often use multiple experimental techniques and facilities, generating data in different formats. There is a need for metadata standards and repositories that can integrate diverse structural biology data types. A proposed solution is a virtual research environment portal that provides collaborative data sharing, processing and analysis tools, along with mechanisms for data quality control, citation and recognition. Such a system could help address the challenges of managing multi-technique structural biology data throughout the research lifecycle.
AI for All: Biology is eating the world & AI is eating Biology Intel® Software
Advances in cell biology and creation of an immense amount of data are converging with advances in Machine learning to analyze this data. Biology is experiencing its AI moment and driving the massive computation involved in understanding biological mechanisms and driving interventions. Learn about how cutting edge technologies such as Software Guard Extensions (SGX) in the latest Intel Xeon Processors and Open Federated Learning (OpenFL), an open framework for federated learning developed by Intel, are helping advance AI in gene therapy, drug design, disease identification and more.
This document discusses using the T-BioInfo platform to provide practical education in bioinformatics. It describes how the platform can integrate different types of omics data and analysis into intuitive, visual pipelines. This allows non-experts to analyze and interpret complex datasets. Example projects are provided, such as using RNA-seq data to identify genes involved in a disease. The goal is to teach bioinformatics through collaborative, project-based learning without requiring programming skills. Learners would reconstruct simulated biological processes and contribute to ongoing analysis of real scientific datasets.
This document discusses the challenges of open source biological software projects including community engagement, integration with other tools, and increasing accessibility (democratization). It provides examples of how the Biopython project addresses these challenges such as through the Google Summer of Code program, improving documentation, and leveraging cloud computing resources to more easily distribute and access biological data and tools.
This document discusses the challenges of open source biological software projects including community engagement, integration with other tools, and increasing accessibility (democratization). It provides examples of how the Biopython project addresses these challenges such as through the Google Summer of Code program, improving documentation, and leveraging cloud computing resources to more easily distribute and access data and tools.
The document summarizes several biomedical applications that utilize the EGEE grid computing infrastructure:
- There are 3 pilot applications that have been running since 2004: CDSS for clinical decision support, GATE for radiation therapy planning, and GPS@ for protein structure analysis.
- CDSS classifies medical cases from distributed databases accessed through the grid. GATE uses grid resources to parallelize Monte Carlo simulations for radiation therapy planning. GPS@ provides bioinformaticians access to international databases and algorithms.
- Additional applications described include co-registration of medical images for contrast agent diffusion analysis, and SiMRI3D for magnetic resonance image simulation. These applications utilize the grid's parallel computing resources to process large medical datasets.
IRJET - Development of Cloud System for IoT ApplicationsIRJET Journal
This document discusses the development of cloud systems for IoT applications. It begins with an introduction stating that one major problem IoT faces is storing and managing vast amounts of data generated. It then reviews 6 papers related to IoT cloud platforms, cloud storage systems, developments in cloud and IoT, exploring IoT platform development, minimizing energy consumption and SLA violations in cloud data centers, and IoT data classification. The document concludes that a detailed review of 6 IoT platform development approaches was presented and a framework was proposed to help select approaches based on requirements.
The Microsoft Biology Foundation (MBF) is an open-source library of bioinformatics algorithms and services built on .NET. MBF provides modular and reusable code for tasks like genomics, sequencing, and analysis. It leverages existing Microsoft technologies and allows distribution of computations across platforms from local to cloud. The first version was released in June 2010. MBF is developed openly on CodePlex and aims to benefit both commercial and non-commercial users.
Similar to Bio-UnaGrid: Easing bioinformatics workflow execution (20)
En esta presentación se describe el estado del emprendimiento en Colombia y el mundo actualizado a 2017, incluyendo temáticas como: el incremento en el número de usuarios conectados a Internet a nivel mundial, ejemplos de proyectos de emprendimiento exitosos a nivel nacional (Tappsi, DomiciliosBogota, Rappi, Mensajeros Urbanos, Ticketcode) e internacional, metodologías al momento de emprender (Lean Startup, Business Model Generation, Customer Development), estrategias para crear mínimos productos viables (MVPs), la importancia del equipo, las habilidades del equipo, las etapas de una startup, las alternativas de inversión y buenas prácticas que deberían ser tenidas en cuentas por emprendedores.
La presentación está dirigida y fue realizada a estudiantes de pregrado y posgrado en el curso "Las Mil Caras de Internet" en la Universidad de los Andes.
e-Clouds A Platform and Marketplace to Access and Publish Scientific Applicat...Mario Jose Villamizar Cano
Cloud computing promises to offer great opportunities for research groups; however when researchers want to execute applications in cloud infrastructures many complex processes must be accomplished. In this presentation we present the e-Clouds project which will allow researchers to easily execute many applications on public Infrastructure as a Service (IaaS) solutions. Designed for being a Software as a Service (SaaS) marketplace for scientific applications, e-Clouds allows researchers to submit jobs which are transparently executed on public IaaS platforms, such as Amazon Web Services (AWS). e-Clouds manages the on-demand provisioning and configuration of computing instances, storage, applications, schedulers, jobs, and data. The architectural design and how a first application has been supported on e-Clouds are presented. e-Clouds will allow researchers to easily share and execute applications in the cloud at low TCO (Total Cost of Ownership) and without the complexities associated with details of IT configurations and management. e-Clouds provides new opportunities for research groups with low or none budget for dedicated cluster or grid solutions, providing on-demand access to ready-to-use applications and accelerating the result generation of e-Science projects.
En esta presentación realizada en la Universidad de los Andes, se dan a conocer herramientas, metodologías y buenas prácticas que pueden ser utilizadas por estudiantes y personas que estén interesados en crear o que hagan parte de proyectos de emprendimiento en Internet (Startups). En la presentación se trata principalmente el tema de desarrollo ágil de productos web y móviles, se comparten algunas experiencias al momento de crear empresas en Internet, se dan recomendaciones a tener en cuenta al validar productos con clientes, y se referencian herramientas cloud computing y libros que pueden ser muy útiles para los emprendedores.
This presentation is a about a light and personal overview of success histories, challenges, opportunities, methodologies and recommendations that may be relevant to people/professionals/students/entrepreneurs that are trying to create new ideas, products and business models that heavily use information and communication technologies (ICT) in the new era of information where there are more than two billions of people connected to the Internet and where there are emerging technologies that will continue changing the world in the near future.
Presentación donde se describe el estado actual del modelo Cloud Computing y cómo este modelo está cambiando la manera en que las empresas adquieren infraestructuras, plataformas de desarrollo y aplicaciones. Presentación realizada en el evento Cloud Computing Latin American Conference 2011 organizado por Channel Planet.
UnaCloud: an opportunistic cloud computing Infrastructure as a Service (IaaS) model implementation, which provides at lower cost, fundamental computing resources (processing, storage and networking) to run arbitrary software, including operating systems and applications.
Taxonomía de los modelos de entrega de servicios, despliegue y facturación en...Mario Jose Villamizar Cano
En esta presentación se hace una descripción de los diferentes modelos de entrega, despliegue y facturación, disponibles al momento de querer utilizar soluciones Cloud Computing (Computación en la nube).
BacteriumSimulatorGrid (BSGrid) - Tool for Simulating the Behavior of the Bac...Mario Jose Villamizar Cano
We developed BSGrid, an application to simulate the behavior of bacterial populations using stochastic methods, using high performance computing infrastructures (HPCIs) as cluster and/or grid computing.
Presenta el diseño y la implementación de una infraestructura grid virtual oportunista que permite aprovechar las capacidades de procesamiento existentes en los equipos pertenecientes a los laboratorios de cómputo de un campus universitario, permitiendo a los usuarios locales tener prioridad para el acceso a los recursos, mientras en forma paralela un cluster virtual aprovecha los recursos no utilizados por los usuarios.
Esta es la presentación del Nivel 6 del curso "ISIS-1204 Algorítmia y Programación por Objetos 1" (APO1) de la Universidad de los Andes, Bogotá, Colombia.
Esta es la presentación del Nivel 4 del curso "ISIS-1204 Algorítmia y Programación por Objetos 1" (APO1) de la Universidad de los Andes, Bogotá, Colombia.
Esta es la presentación del Nivel 10 del curso "ISIS-1205 Algorítmia y Programación por Objetos 2" (APO2) de la Universidad de los Andes, Bogotá, Colombia.
Esta es la presentación del Nivel 9 del curso "ISIS-1205 Algorítmia y Programación por Objetos 2" (APO2) de la Universidad de los Andes, Bogotá, Colombia.
Esta es la presentación del Nivel 3 del curso "ISIS-1204 Algorítmia y Programación por Objetos 1" (APO1) de la Universidad de los Andes, Bogotá, Colombia.
Esta es la presentación del Nivel 8 del curso "ISIS-1205 Algorítmia y Programación por Objetos 2" (APO2) de la Universidad de los Andes, Bogotá, Colombia.
El documento presenta información sobre el curso Algorítmica y Programación 1 (APO 1) Nivel 2. Se incluyen los datos del profesor a cargo, los nuevos conceptos que se abordarán en este nivel como tipos de datos, expresiones, instrucciones condicionales y casos de estudio. También se describen ejemplos y hojas de trabajo para aplicar los conocimientos.
Esta es la presentación del Nivel 7 del curso "ISIS-1205 Algorítmia y Programación por Objetos 2" (APO2) de la Universidad de los Andes, Bogotá, Colombia.
Esta es la presentación del Nivel 1 del curso "ISIS-1204 Algorítmia y Programación por Objetos 1" (APO1) de la Universidad de los Andes, Bogotá, Colombia.
In the realm of cybersecurity, offensive security practices act as a critical shield. By simulating real-world attacks in a controlled environment, these techniques expose vulnerabilities before malicious actors can exploit them. This proactive approach allows manufacturers to identify and fix weaknesses, significantly enhancing system security.
This presentation delves into the development of a system designed to mimic Galileo's Open Service signal using software-defined radio (SDR) technology. We'll begin with a foundational overview of both Global Navigation Satellite Systems (GNSS) and the intricacies of digital signal processing.
The presentation culminates in a live demonstration. We'll showcase the manipulation of Galileo's Open Service pilot signal, simulating an attack on various software and hardware systems. This practical demonstration serves to highlight the potential consequences of unaddressed vulnerabilities, emphasizing the importance of offensive security practices in safeguarding critical infrastructure.
The Department of Veteran Affairs (VA) invited Taylor Paschal, Knowledge & Information Management Consultant at Enterprise Knowledge, to speak at a Knowledge Management Lunch and Learn hosted on June 12, 2024. All Office of Administration staff were invited to attend and received professional development credit for participating in the voluntary event.
The objectives of the Lunch and Learn presentation were to:
- Review what KM ‘is’ and ‘isn’t’
- Understand the value of KM and the benefits of engaging
- Define and reflect on your “what’s in it for me?”
- Share actionable ways you can participate in Knowledge - - Capture & Transfer
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
Northern Engraving | Modern Metal Trim, Nameplates and Appliance PanelsNorthern Engraving
What began over 115 years ago as a supplier of precision gauges to the automotive industry has evolved into being an industry leader in the manufacture of product branding, automotive cockpit trim and decorative appliance trim. Value-added services include in-house Design, Engineering, Program Management, Test Lab and Tool Shops.
"Frontline Battles with DDoS: Best practices and Lessons Learned", Igor IvaniukFwdays
At this talk we will discuss DDoS protection tools and best practices, discuss network architectures and what AWS has to offer. Also, we will look into one of the largest DDoS attacks on Ukrainian infrastructure that happened in February 2022. We'll see, what techniques helped to keep the web resources available for Ukrainians and how AWS improved DDoS protection for all customers based on Ukraine experience
High performance Serverless Java on AWS- GoTo Amsterdam 2024Vadym Kazulkin
Java is for many years one of the most popular programming languages, but it used to have hard times in the Serverless community. Java is known for its high cold start times and high memory footprint, comparing to other programming languages like Node.js and Python. In this talk I'll look at the general best practices and techniques we can use to decrease memory consumption, cold start times for Java Serverless development on AWS including GraalVM (Native Image) and AWS own offering SnapStart based on Firecracker microVM snapshot and restore and CRaC (Coordinated Restore at Checkpoint) runtime hooks. I'll also provide a lot of benchmarking on Lambda functions trying out various deployment package sizes, Lambda memory settings, Java compilation options and HTTP (a)synchronous clients and measure their impact on cold and warm start times.
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
Discover top-tier mobile app development services, offering innovative solutions for iOS and Android. Enhance your business with custom, user-friendly mobile applications.
Session 1 - Intro to Robotic Process Automation.pdfUiPathCommunity
👉 Check out our full 'Africa Series - Automation Student Developers (EN)' page to register for the full program:
https://bit.ly/Automation_Student_Kickstart
In this session, we shall introduce you to the world of automation, the UiPath Platform, and guide you on how to install and setup UiPath Studio on your Windows PC.
📕 Detailed agenda:
What is RPA? Benefits of RPA?
RPA Applications
The UiPath End-to-End Automation Platform
UiPath Studio CE Installation and Setup
💻 Extra training through UiPath Academy:
Introduction to Automation
UiPath Business Automation Platform
Explore automation development with UiPath Studio
👉 Register here for our upcoming Session 2 on June 20: Introduction to UiPath Studio Fundamentals: https://community.uipath.com/events/details/uipath-lagos-presents-session-2-introduction-to-uipath-studio-fundamentals/
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving
Manufacturing custom quality metal nameplates and badges involves several standard operations. Processes include sheet prep, lithography, screening, coating, punch press and inspection. All decoration is completed in the flat sheet with adhesive and tooling operations following. The possibilities for creating unique durable nameplates are endless. How will you create your brand identity? We can help!
AppSec PNW: Android and iOS Application Security with MobSFAjin Abraham
Mobile Security Framework - MobSF is a free and open source automated mobile application security testing environment designed to help security engineers, researchers, developers, and penetration testers to identify security vulnerabilities, malicious behaviours and privacy concerns in mobile applications using static and dynamic analysis. It supports all the popular mobile application binaries and source code formats built for Android and iOS devices. In addition to automated security assessment, it also offers an interactive testing environment to build and execute scenario based test/fuzz cases against the application.
This talk covers:
Using MobSF for static analysis of mobile applications.
Interactive dynamic security assessment of Android and iOS applications.
Solving Mobile app CTF challenges.
Reverse engineering and runtime analysis of Mobile malware.
How to shift left and integrate MobSF/mobsfscan SAST and DAST in your build pipeline.
LF Energy Webinar: Carbon Data Specifications: Mechanisms to Improve Data Acc...DanBrown980551
This LF Energy webinar took place June 20, 2024. It featured:
-Alex Thornton, LF Energy
-Hallie Cramer, Google
-Daniel Roesler, UtilityAPI
-Henry Richardson, WattTime
In response to the urgency and scale required to effectively address climate change, open source solutions offer significant potential for driving innovation and progress. Currently, there is a growing demand for standardization and interoperability in energy data and modeling. Open source standards and specifications within the energy sector can also alleviate challenges associated with data fragmentation, transparency, and accessibility. At the same time, it is crucial to consider privacy and security concerns throughout the development of open source platforms.
This webinar will delve into the motivations behind establishing LF Energy’s Carbon Data Specification Consortium. It will provide an overview of the draft specifications and the ongoing progress made by the respective working groups.
Three primary specifications will be discussed:
-Discovery and client registration, emphasizing transparent processes and secure and private access
-Customer data, centering around customer tariffs, bills, energy usage, and full consumption disclosure
-Power systems data, focusing on grid data, inclusive of transmission and distribution networks, generation, intergrid power flows, and market settlement data
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
"Choosing proper type of scaling", Olena SyrotaFwdays
Imagine an IoT processing system that is already quite mature and production-ready and for which client coverage is growing and scaling and performance aspects are life and death questions. The system has Redis, MongoDB, and stream processing based on ksqldb. In this talk, firstly, we will analyze scaling approaches and then select the proper ones for our system.
inQuba Webinar Mastering Customer Journey Management with Dr Graham HillLizaNolte
HERE IS YOUR WEBINAR CONTENT! 'Mastering Customer Journey Management with Dr. Graham Hill'. We hope you find the webinar recording both insightful and enjoyable.
In this webinar, we explored essential aspects of Customer Journey Management and personalization. Here’s a summary of the key insights and topics discussed:
Key Takeaways:
Understanding the Customer Journey: Dr. Hill emphasized the importance of mapping and understanding the complete customer journey to identify touchpoints and opportunities for improvement.
Personalization Strategies: We discussed how to leverage data and insights to create personalized experiences that resonate with customers.
Technology Integration: Insights were shared on how inQuba’s advanced technology can streamline customer interactions and drive operational efficiency.
1. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: Easing Bioinformatics Workflow Execution Using LONI
Pipeline and a Virtual Desktop Grid
Mario Villamizar, Harold Castro, Silvia Restrepo, Luis Rodriguez
David Mendez Department of Biological Sciences
Departments of Systems and Computer Universidad de los Andes
Engineering Bogotá D.C., Colombia
Universidad de los Andes
Bogotá, Colombia
2. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Agenda
The Problem Context
Related Work
Bio‐UnaGrid: Architecture
Bio‐UnaGrid: Implementation
Bio‐UnaGrid: Performance Tests
Conclusions and Future Work
3. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Agenda
The Problem Context
Related Work
Bio‐UnaGrid: Architecture
Bio‐UnaGrid: Implementation
Bio‐UnaGrid: Performance Tests
Conclusions and Future Work
4. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Around the problem
Bioinformatics researchers use applications that require large computational
capabilities.
Cluster and grid computing are common solutions to support these
capabilities.
Cluster computing allows to
aggregate homogeneous
computational resources for an
specific research group or
organization
5. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Around the problem
Grid computing allows to aggregate heterogeneous computational resources
of different organizations to support larger computational capabilities in e-
Science projects.
In grid computing a set of
organization, with a common
objective (VO), share
computational resources.
6. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Around the problem
High performance computing (HPC) infrastructures, such as cluster or grid
computing, provide results in shorter time, however when bioinformatics
researchers want to execute applications, they face several consuming time
problems.
7. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Around the problem
1. There are many application suites to run different analysis, so researchers
must learn command line syntax (options, input and output files, distributed
execution environment, etc.) for hundreds of applications.
8. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Around the problem
2. Researchers must learn commands to manage and use distributed
computing infrastructures.
9. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Around the problem
3. Applications require specific and complex configurations to operate in
distributed environments.
10. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Around the problem
4. Researchers frequently execute a set of applications in a sequential and/or
coordinated manner, called pipelines or workflows.
In many cases the workflow execution is manually done by the researchers
using command line. Execute an application, wait …., execute the next
application, wait …. This process may take a lot of time.
11. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Around the problem
5. Very often researchers want to execute applications that require larger
processing capabilities than those provided by dedicated cluster computing
infrastructures, so they must wait weeks or months before getting their
results.
12. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
The problem
We were having the following problem:
The manual and command-based application and workflow execution require
bioinformatics researchers to spend most of their time in: configuring
application parameters, managing HPC infrastructures, managing scientific
data, and linking partial results from one application to another.
When a new researcher or student want to use applications he/she spends a
lot of time learning all those things.
13. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Agenda
The Problem Context
Related Work
Bio‐UnaGrid: Architecture
Bio‐UnaGrid: Implementation
Bio‐UnaGrid: Performance Tests
Conclusions and Future Work
14. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Related work
Several projects have been developed to facilitate the automatic workflow
execution in bioinformatics and other scientific fields.
15. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Related work
Most of workflow tools have limitations for bioinformatics applications such
as:
They require applications to be recompiled or modified.
It is too complex to modify or recompile every bioinformatics application.
They support internal data sources with specific data structures.
Bioinformatics data have different formats and they can be internal or
external.
They operate with specific platforms, and cluster or grid middlewares.
Institutions regularly have different cluster and grid configurations. In our
case we are part of different Grid projects such as GISELA, EGEE, and
OSG.
16. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Related work
Most of workflow tools have limitations for bioinformatics applications such
as:
They regularly require researchers to use complex commands involving
consuming-time tasks.
Bioinformatics researchers should use easy GUIs to execute workflows.
They operate on dedicated cluster and grid infrastructures, however in a
university or enterprise campus there are tens or hundreds of desktop
computers that are under-utilized.
Opportunistic or DGVCSs can provide large processing capabilities, at low
cost.
17. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
How to solve these problems
Define an infrastructure that allows that
existing bioinformatics applications can be
easily incorporated without modifications
LONI Pipeline
Workflows should be created through
GUIs and executed on different HPC
infrastructures.
The infrastructure should also allow the
use of different data sources (internal and
external), and the incorporation of MPI
applications.
UnaGrid
For getting results faster the infrastructure
Bio-UnaGrid integrates the should operate on dedicated computing
capabilities of LONI Pipeline and infrastructures and on DGVCS.
UnaGrid to provide these features.
18. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
LONI Pipeline
LONI Pipeline is a free framework used to
execute neuroscience workflows.
From a computational perspective, LONI differs
from other workflow tools in several features:
It does not require external application being
recompiled.
It supports external data sources.
It is hardware platform independent.
It can be installed on different cluster or grid
middlewares using the Pipeline plugins.
From a research user perspective, LONI was
designed having in mind features like usability,
portability, intuitiveness, transparency and
abstraction of cluster or grid infrastructures.
19. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
LONI Pipeline Architecture
20. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
UnaGrid
Dedicated infrastructures are an unviable option in organizations or
countries with low financial resources. However, these organizations have
many computer labs which are not fully utilized by employees or university
students.
21. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
UnaGrid: Opportunistic virtual clusters
X
Cores
Linux
Processing
Virtual Machine
Physical Machine of a
Computer Room
b. When there is not an End User
using the physical machine
A virtual cluster is a set of commodity and interconnected desktops
executing virtual machines (VMs) in background and low-priority through
virtualization technologies, these VMs take advantage of the available idle
processing capabilities in computer labs on an university campus.
22. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
UnaGrid: Opportunistic virtual clusters
A virtual machine is executed on each computer of a lab and it supports
the role of a cluster slave and all of these virtual machines on execution
make up a virtual processing cluster. A dedicated node is necessary for a
virtual cluster and it supports the role of the cluster master.
23. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
UnaGrid: Opportunistic virtual clusters
A virtual machine is executed on each computer of a lab and it supports
the role of a cluster slave and all of these virtual machines on execution
make up a virtual processing cluster. A dedicated node is necessary for a
virtual cluster and it supports the role of the cluster master.
24. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
UnaGrid: Opportunistic virtual clusters
A virtual machine is executed on each computer of a lab and it supports
the role of a cluster slave and all of these virtual machines on execution
make up a virtual processing cluster. A dedicated node is necessary for a
virtual cluster and it supports the role of the cluster master.
25. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
UnaGrid: Opportunistic virtual clusters
Each research group can define its own virtual clusters with custom
application environments (middlewares, applications, configurations, etc.)
A grid solution (several virtual clusters) can be deployed for supporting
the processing capabilities required by some applications.
26. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Agenda
The Problem Context
Related Work
Bio‐UnaGrid: Architecture
Bio‐UnaGrid: Implementation
Bio‐UnaGrid: Performance Tests
Conclusions and Future Work
27. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid Architecture
The LONI Pipeline Client is the main entry point of bioinformatics
researchers to the Bio-UnaGrid infrastructure.
28. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid Operation
Each researcher receives a username and a password.
He/she can proceed to create the workflows, using the bioinformatics
LONI Module Library, drag and drop, and other tools, provided by the
LONI Pipeline Client.
For each LONI Module (a bioinformatics application), the researcher
must define the input and output parameters, and how each LONI
Module is connected with other LONI Modules of the workflow.
Once a workflow has been created, researchers can execute it, sending it
to the LONI Pipeline Server.
Researcher can monitor and visualize the status of the workflow, and
download partial results, through GUIs of the LONI Pipeline Client.
The Grid/Cluster infrastructure is transparent for researchers.
29. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: Bioinformatics Applications
For executing workflows, bioinformatics researchers do
not have to worry about applications’ commands or
complex HPC infrastructures.
30. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: Incorporating Bioinformatics Applications
The process to incorporate existing and new bioinformatics application
suites, like NCBI BLAST, into Bio-UnaGrid infrastructure are summarized in
five steps.
1) Installation and configuration of the application suite on a Shared
Storage System.
2) Creation of a LONI Pipeline Module using the LONI Pipeline Client for
each executable of the suite.
3) Individual tests for each LONI Pipeline Module of the suite from the
LONI Pipeline Client.
4) Storage of the LONI Pipeline Module in the LONI Pipeline Server.
5) Using the modules from workflows created by bioinformatics
researchers in the LONI Pipeline Client.
31. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: Incorporating Bioinformatics Applications
32. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: MPI applications on Bio-UnaGrid
At the time of this development, LONI did not support MPI applications.
To allow the execution of MPI applications, we developed a wrapper, called
mpiJobManager, which uses the Distributed Resource Management
Application API (DRMAA).
To execute an MPI Application, such as mpiBLAST, a researcher uses the
MPI LONI Module of the MPI application to specify the number of MPI
processes to be used.
The LONI Pipeline Server receives the MPI job and proceeds to execute it
on the DRM slaves.
33. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Agenda
The Problem Context
Related Work
Bio‐UnaGrid: Architecture
Bio‐UnaGrid: Implementation
Bio‐UnaGrid: Performance Tests
Conclusions and Future Work
34. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: Implementation
Bioinformatics Infrastructure
6 Dedicated OGE Slaves – 24 Cores
1 Gbps S&C Infrastructure
LONI Client GUMA Client
10
Gbps NFS-NAS Server
10 Gbps
LONI Pipeline Server
10 1 OGE Master
Bioinformatics Gbps Gbps
Researcher GUMA Portal
10 Gbps 10 Gbps
155
LAN/ Mbps 10
Internet Gbps User Database
1 Gbps
Linux CVC with 21 OGE Slaves – 21 Cores
VM VM VM VM ... VM
Windows 7 Lab
UnaGrid Infrastructure
Bio-UnaGrid was implemented to support several genomic projects related
to coffee, potato and cassava, focused in increasing their production.
35. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: Implementation
The LONI Pipeline Server version 5.0.2 and the DRM Master of OGE
version 6.2 update 5.
6 dedicated servers - Intel Xeon X5560 Quad Core processor of 2.8 GHz
and 4 GB of RAM.
A computer lab (Windows 7) - Intel i5 processor of 3.46 GHz and 8 GB of
RAM. 21 VMs (UnaGrid) was deployed and configured with a CPU core
and 4 GB of RAM.
The Database User was installed on a dedicated server using MySQL.
We use an NFSv3-NAS.
All nodes are interconnected through 1 GbE and 10GbE links.
36. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: Implementation
Fault tolerance mechanisms were configured for BoT applications using the
OGE capabilities.
Four bioinformatics application suites have been installed on the Bio-
UnaGrid implementation and 21 LONI Modules were created.
NCBI BLAST version 2.2.20: BLASTn, BLASTp, BLASTx, tBLASTn,
tBLASTx and MEGABLAST.
HMMER version 2.3.2: Build, Search and Calibrate
InterProScan version 4.6: BlastProdom, Coils, Gene3D, PIR, Panther,
Pfam, SEG, SMART, SuperFamily, TIGRfam and fPrintScan
mpiBLAST version 1.6.0: mpiBLAST
Because mpiBLAST requires the use of an MPI implementation, the MPICH-
2 implementation version 1.2.7 was installed.
37. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: A Simple Workflow Example
38. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Bio-UnaGrid: An Example of A More Complex Workflow
39. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Agenda
The Problem Context
Related Work
Bio‐UnaGrid: Architecture
Bio‐UnaGrid: Implementation
Bio‐UnaGrid: Performance Tests
Conclusions and Future Work
40. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Performance Tests - BLASTn
We executed performance tests in both environments, in the dedicated
cluster and in the opportunistic UnaGrid CVC.
We selected the high popular BLAST algorithm to execute these tests.
For testing the performance of Bio-UnaGrid for executing BoT applications
we used the NCBI BLASTn application, varying the number of CPU cores
and the size of the sequence set.
We executed BLASTn searches between nucleotide sets of 480 (487 KB)
and 960 (954 KB) sequences, and the nt database (28 GB).
The searches were executed using 5, 10, 15 and 20, BLASTn independent
processes (BoT BLASTn), using query segmentation.
Each search was executed using a single CPU core.
41. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Performance Tests - BLASTn
Execution time of BoT BLASTn searches between sets of 480 and 960
sequences, and the nt database.
480 seq. - Dedicated Cluster (20 Cores) 10.9x – UnaGrid (20 Cores) 13.1x
960 seq. - Dedicated Cluster (15 Cores) 6.1x – UnaGrid (20 Cores) 10x
42. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Performance Tests - mpiBLAST
For testing the execution of MPI applications on both environments, we
executed the same tests using mpiBLAST.
In these tests we executed MPI processes coordinated on 5, 10, 15 and 20
CPU cores.
In each test, the additional process called mpiJobManager was executed in
another CPU core.
We used database segmentation to divide the database in 5, 10, 15 and
20 partitions using the mpiformatdb installed with mpiBLAST.
43. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Performance Tests - mpiBLAST
Execution time of mpiBLAST searches between sets of 480 and 960
sequences, and the nt database.
480 seq. - Dedicated Cluster (15 Cores) 2.3x – UnaGrid (20 Cores) 4.7x
960 seq. - Dedicated Cluster (20 Cores) 3.0x – UnaGrid (20 Cores) 3.7x
44. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Agenda
The Problem Context
Related Work
Bio‐UnaGrid: Architecture
Bio‐UnaGrid: Implementation
Bio‐UnaGrid: Performance Tests
Conclusions and Future Work
45. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Conclusions
Using the idle processing capabilities available in
computer labs using Windows, Linux or Mac
desktops, Bio-UnaGrid provides additional
processing capabilities to those provided by
dedicated computational clusters.
Bio-UnaGrid allows bioinformatics researchers to
easily define workflows using GUIs and execute
them on cluster and grid computing infrastructures
with a simple click.
Bio-UnaGrid is highly extensible as existing
bioinformatics applications can be incorporated
without modifications.
With Bio-UnaGrid, researchers focus on analysis
of application results, not on technical issues of
distributed computing infrastructures.
46. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Future Work
We will incorporate more bioinformatics
applications suites into Bio-UnaGrid.
We will execute new performance tests with
other applications such as HMMER, MPI-
HMMER, ClustalW and ClustalW-MPI.
We will execute new performance tests in a
larger grid deployment involving hundreds of
heterogeneous desktop computers in different
administrative domains.
We will implement a shared storage system
more scalable than NFSv3.
We also plan to implement a fault tolerance
mechanism for MPI applications.
47. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Thanks for your attention !!!
48. The Third International Conference on Bioinformatics,
Biocomputational Systems and Biotechnologies (BIOTECHNO 2011)
Contact Information
mj.villamizar24@uniandes.edu.co mjvc007@hotmail.com
http://twitter.com/mjvc007
http://linkedin.com/in/mariojosevillamizarcano
Mario José Villamizar Cano