SLALOM Webinar Final Technical Outcomes Explanined "Using the SLALOM Technica...Oliver Barreto Rodríguez
SLALOM organized two live sessions to present the final versions of our legal terms and technical specifications for #Cloud #SLAs. The sessions provide examples showing how to practically apply SLALOM to improve current practice in the industry for # Cloud #SLAs and support development of cloud computing metrics.
The first webinar covered SLALOM Technical track "Using metrics to improve Cloud SLAs".
User Behavior Hashing for Audience ExpansionDatabricks
Learning to hash has been widely adopted as a solution to approximate nearest neighbor search for large-scale data retrieval in many applications. Applying deep architectures to learning to hash has recently gained increasing attention due to its computational efficiency and retrieval quality.
This document discusses techniques for optimizing the performance of PeopleSoft applications. It covers tuning several aspects within a PeopleSoft environment, including server performance, web server performance, Tuxedo performance management, application performance, and database performance. Some key recommendations include implementing a methodology to monitor resource consumption without utilizing critical resources, ensuring load balancing strategies are sound, measuring historical patterns of server resource utilization, capturing key performance metrics for Tuxedo, and focusing on tuning high-resource consuming SQL statements and indexes.
This document provides an overview of resource aware scheduling in Apache Storm. It discusses the challenges of scheduling Storm topologies at Yahoo scale, including increasing heterogeneous clusters, low cluster utilization, and unbalanced resource usage. It then introduces the Resource Aware Scheduler (RAS) built for Storm, which allows fine-grained resource control and isolation for topologies through APIs and cgroups. Key features of RAS include pluggable scheduling strategies, per user resource guarantees, and topology priorities. Experimental results from Yahoo Storm clusters show significant improvements to throughput and resource utilization with RAS. The talk concludes with future work on improved scheduling strategies and real-time resource monitoring.
This document provides an overview of resource aware scheduling in Apache Storm. It discusses the challenges of scheduling Storm topologies at Yahoo scale, including increasing heterogeneous clusters, low cluster utilization, and unbalanced resource usage. It then introduces the Resource Aware Scheduler (RAS) built for Storm, which allows fine-grained resource control and isolation for topologies through APIs and cgroups. Key features of RAS include pluggable scheduling strategies, per user resource guarantees, and topology priorities. Experimental results from Yahoo Storm clusters show significant improvements to throughput and resource utilization with RAS. Future work may include improved scheduling strategies and real-time resource monitoring.
Learning Software Performance Models for Dynamic and Uncertain EnvironmentsPooyan Jamshidi
This document provides background on Pooyan Jamshidi's research related to learning software performance models for dynamic and uncertain environments. It summarizes his past work developing techniques for modeling and optimizing performance across different systems and environments, including using transfer learning to reuse performance data from related sources to build more accurate models with fewer measurements. It also outlines opportunities for using transfer learning to adapt performance models to new environments and systems.
In this video from SC17 in Denver, Dan Reed moderates a panel discussion on HPC Software for Energy Efficiency.
"We have already achieved major gains in energy-efficiency for both the datacenter and HPC equipment. For example, the PUE of the Swiss Supercomputer (CSCS) datacenter prior to 2012 was 1.8, but the current PUE is about 1.25; a factor of ~1.5 improvement. HPC system improvements have also been very strong, as evidenced by FLOPS/Watt performance on the Green500 List. While we have seen gains from data center and HPC system efficiency, there are also energy-efficiency gains to be had from software- application performance improvements, for example. This panel will explore what HPC software capabilities were most helpful over the past years in improving HPC system energy efficiency? It will then look forward; asking in what layers of the software stack should a priority be put on introducing energy-awareness; e.g., runtime, scheduling, applications? What is needed moving forward? Who is responsible for that forward momentum?"
Watch the video: https://wp.me/p3RLHQ-hHQ
Learn more: https://sc17.supercomputing.org/presentation/?id=pan103&sess=sess245
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
SLALOM Webinar Final Technical Outcomes Explanined "Using the SLALOM Technica...Oliver Barreto Rodríguez
SLALOM organized two live sessions to present the final versions of our legal terms and technical specifications for #Cloud #SLAs. The sessions provide examples showing how to practically apply SLALOM to improve current practice in the industry for # Cloud #SLAs and support development of cloud computing metrics.
The first webinar covered SLALOM Technical track "Using metrics to improve Cloud SLAs".
User Behavior Hashing for Audience ExpansionDatabricks
Learning to hash has been widely adopted as a solution to approximate nearest neighbor search for large-scale data retrieval in many applications. Applying deep architectures to learning to hash has recently gained increasing attention due to its computational efficiency and retrieval quality.
This document discusses techniques for optimizing the performance of PeopleSoft applications. It covers tuning several aspects within a PeopleSoft environment, including server performance, web server performance, Tuxedo performance management, application performance, and database performance. Some key recommendations include implementing a methodology to monitor resource consumption without utilizing critical resources, ensuring load balancing strategies are sound, measuring historical patterns of server resource utilization, capturing key performance metrics for Tuxedo, and focusing on tuning high-resource consuming SQL statements and indexes.
This document provides an overview of resource aware scheduling in Apache Storm. It discusses the challenges of scheduling Storm topologies at Yahoo scale, including increasing heterogeneous clusters, low cluster utilization, and unbalanced resource usage. It then introduces the Resource Aware Scheduler (RAS) built for Storm, which allows fine-grained resource control and isolation for topologies through APIs and cgroups. Key features of RAS include pluggable scheduling strategies, per user resource guarantees, and topology priorities. Experimental results from Yahoo Storm clusters show significant improvements to throughput and resource utilization with RAS. The talk concludes with future work on improved scheduling strategies and real-time resource monitoring.
This document provides an overview of resource aware scheduling in Apache Storm. It discusses the challenges of scheduling Storm topologies at Yahoo scale, including increasing heterogeneous clusters, low cluster utilization, and unbalanced resource usage. It then introduces the Resource Aware Scheduler (RAS) built for Storm, which allows fine-grained resource control and isolation for topologies through APIs and cgroups. Key features of RAS include pluggable scheduling strategies, per user resource guarantees, and topology priorities. Experimental results from Yahoo Storm clusters show significant improvements to throughput and resource utilization with RAS. Future work may include improved scheduling strategies and real-time resource monitoring.
Learning Software Performance Models for Dynamic and Uncertain EnvironmentsPooyan Jamshidi
This document provides background on Pooyan Jamshidi's research related to learning software performance models for dynamic and uncertain environments. It summarizes his past work developing techniques for modeling and optimizing performance across different systems and environments, including using transfer learning to reuse performance data from related sources to build more accurate models with fewer measurements. It also outlines opportunities for using transfer learning to adapt performance models to new environments and systems.
In this video from SC17 in Denver, Dan Reed moderates a panel discussion on HPC Software for Energy Efficiency.
"We have already achieved major gains in energy-efficiency for both the datacenter and HPC equipment. For example, the PUE of the Swiss Supercomputer (CSCS) datacenter prior to 2012 was 1.8, but the current PUE is about 1.25; a factor of ~1.5 improvement. HPC system improvements have also been very strong, as evidenced by FLOPS/Watt performance on the Green500 List. While we have seen gains from data center and HPC system efficiency, there are also energy-efficiency gains to be had from software- application performance improvements, for example. This panel will explore what HPC software capabilities were most helpful over the past years in improving HPC system energy efficiency? It will then look forward; asking in what layers of the software stack should a priority be put on introducing energy-awareness; e.g., runtime, scheduling, applications? What is needed moving forward? Who is responsible for that forward momentum?"
Watch the video: https://wp.me/p3RLHQ-hHQ
Learn more: https://sc17.supercomputing.org/presentation/?id=pan103&sess=sess245
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobilDatabricks
ExxonMobil leveraged machine learning at scale using Databricks to extract insights from equipment maintenance logs and improve operations. The logs contained both structured and unstructured text data across a global fleet maintained in legacy systems, limiting traditional analysis. By ingesting and enriching over 60 million records using natural language processing, the system identified outliers, enabled capacity planning, and prioritized maintenance tasks, projected to save millions annually through more effective reliability and maintenance guidance.
FAULT TOLERANCE OF RESOURCES IN COMPUTATIONAL GRIDSMaurvi04
This document discusses fault tolerance techniques for computational grids. It begins with an introduction to grid computing and defines some key terms related to faults and failures. It then discusses different types of faults that can occur in grids, including physical faults, network faults, and process faults. It outlines several fault tolerance techniques used in grids, including job and data replication, checkpointing, scheduling approaches, and load balancing strategies. The document concludes with suggestions for future work, such as optimizing checkpoint storage and granularity.
1. Discuss the structured system analysis and design methodologies
2. What is DSS? Discuss the components and capabilities of DSS.
3. Narrate the stages of SDLC
4. Define OOP. What are the applications of it?
The document outlines a three-phase approach to developing an intelligent monitoring platform:
Phase 1 involves interviewing dev and ops teams to understand current monitoring practices.
Phase 2 focuses on improving the postmortem process and outage understanding.
Phase 3 aims to reduce the time to identify and resolve outages through expanded data collection, correlation analysis, and predictive capabilities.
The document outlines a resilient system design approach for PayPal that aims to isolate functionality and enable independent availability patterns. It proposes using asynchronous reconciliation to resolve failures without affecting the customer experience. The key elements are:
1) Isolating functionality into independent functional components (FCs) that can be developed, deployed and executed separately for increased flexibility.
2) Using a service container (SC) to orchestrate FCs and handle failures by consolidating responses and initiating reconciliation.
3) Implementing circuit breakers and fallback behaviors to protect clients, services, and business from failures in FCs or dependencies.
4) Employing an eventual consistency model with automated reconciliation to resolve inconsistencies due to network partitions
This document provides guidance on evaluating low-code platforms (LCPs) and tools (LCTs) for suitability. It outlines several factors to consider:
1. Verify critical features through prototyping to ensure an LCP supports must-have functionality. Use critical features to filter LCP options.
2. Evaluate security, including data security practices, data ownership policies, user permissions, audit logs, breach response, and disaster recovery.
3. Assess scale, including table size limits, business logic processing ability, transaction capacity, data crunching/analysis capabilities, and user experience with large data volumes.
4. Consider interfaces to other tools, including common data formats, traffic volumes
Documented Requirements are not Useless After All!Lionel Briand
The document discusses challenges related to requirements in agile software development projects, including dealing with natural language requirements, domain knowledge, frequent changes, and configuring requirements for product families. It outlines research conducted to address these challenges through natural language processing techniques like analyzing compliance with boilerplate templates, extracting domain models from requirements, and analyzing change impact when requirements change. The talk aims to discuss when documented requirements are important in practice and how they can be supported to effectively handle common challenges.
The document discusses Enterprise Resource Planning (ERP) systems. It describes the ERP architecture as using a client-server model with a relational database to store and process data. The ERP lifecycle involves definition, construction, implementation, and operation phases. Core ERP components manage accounting, production, human resources and other internal functions, while extended components provide external capabilities like CRM, SCM, and e-business. Proper implementation requires screening software, evaluating packages, analyzing process gaps, reengineering workflows, training staff, testing, and post-implementation support.
A sdn based application aware and network provisioningStanley Wang
The document discusses application aware SDN network provisioning. It begins with an overview of YARN architecture in Hadoop, including its benefits over earlier Hadoop architectures like improved scalability and utilization. It then discusses how SDN can be integrated with big data and cloud computing workloads by optimizing network topology and routing based on traffic patterns. Two approaches are proposed - reactive, where the SDN controller learns patterns from job logs/endpoints and modifies paths, and proactive where applications directly inform the network of intent. Finally, it proposes a service profile based SDN platform that uses network profiles and APIs to declaratively define logical topologies and provide network services and abstractions to applications.
1. The document discusses research activities related to reducing energy consumption by at least 30% through the development of core source technologies for universal operating systems.
2. It describes four papers being presented, including ones on system and device latency modeling, power management frameworks for embedded systems, and automatic selection of power policies for operating systems.
3. It also summarizes four research topics from the National University, including performance evaluation of parallel applications using a power-aware paging method on next-generation memory architectures.
Break-out session by Bernhard Becker, Deltares, at the Delft-FEWS International User Days 2020, during Delft Software Days - Edition 2020. Thursday, 5 November 2020.
The document describes the system architecture for an augmented cognition system. It includes sensors that assess cognitive state, a cognitive state assessor that analyzes sensor data, and various components that select and sequence tasks or switch modalities based on the assessed cognitive state. The goal is to maximize operator performance by detecting and mitigating cognitive overload. It provides details on the sensors used, cognitive state gauges, mitigation strategies like pacing and task sequencing, and how tasks are selected and routed through the system.
AlphaPy is a Python framework for building machine learning pipelines. It contains two main pipelines: the Model Pipeline for generating models and the Domain Pipeline for preparing training and test data. The Model Pipeline uses scikit-learn and other packages to build predictive models for classification and regression. The Domain Pipeline transforms raw data into canonical form suitable for modeling. The output is a persistent Model Object. The document provides examples of using AlphaPy to build stock market prediction models that identify predictive technical indicators and validate results on training and test data.
AlphaPy: A Data Science Pipeline in PythonMark Conway
AlphaPy is a Python framework for building machine learning pipelines. It contains two main pipelines: the Model Pipeline for generating models and the Domain Pipeline for preparing training and test data. The Model Pipeline uses scikit-learn and other packages to build predictive models for classification and regression. The Domain Pipeline transforms raw data into canonical form suitable for modeling. The output is a persistent Model Object. The document provides examples of using AlphaPy to build stock market prediction models that identify predictive technical indicators and validate results on training and test data.
Overview of DuraMat software tool developmentAnubhav Jain
The document discusses software tools being developed by researchers for photovoltaic (PV) applications. It summarizes several software projects funded by DuraMat that address different aspects of PV including: (1) PV system modeling and analysis, (2) operation and degradation modeling, and (3) planning and reducing levelized cost of energy. The software aims to solve a range of PV problems, are open source, and developed collaboratively on GitHub to be reusable and sustainable resources for the community.
The document outlines requirements for an MSO network operations center (NOC). Key requirements include monitoring a complex network supporting various services, having full visibility and management capabilities across all technologies, and providing a single pane of view to minimize screen space. The NOC should focus on service uptime through fault detection, resolution, and performance monitoring. People, processes, and tools are needed to provide a quality NOC. A suggested organization structure includes first line engineers, support engineers, shift leads, managers, and specialized roles. Considerations for NOC systems include flexibility, scalability, cost containment, lean architecture, and simplicity.
The document discusses performance evaluation of computer and telecommunication systems. Performance evaluation aims to quantitatively predict a system's behavior and is used to compare designs, plan for capacity, and debug performance issues. It involves modeling, simulation, and testing approaches of varying cost and accuracy. Key metrics include counts, times, sizes, productivity, response time, and reliability measures. Workload characterization analyzes how systems are used, while benchmarks compare performance across systems running standardized tests.
What Is Machine Learning?
Where do we deploy machine learning and what software and cloud services are out there to support it?
What are the trends in deploying these systems and what are the benefits for IT?
Do you have a IoT Machine Learning Case Study in the Cloud?
Keep Calm and React with Foresight: Strategies for Low-Latency and Energy-Eff...Tiziano De Matteis
This talk has been given at PPoPP 2016 (Barcelona)
The paper addresses the problem of designing control strategies for elastic stream processing applications. Elasticity allows applications to rapidly change their configuration (e.g. the number of used resources) on-the-fly, in response to fluctuations of their workload. In this work we face this problem by adopting the Model Predictive Control technique, a control-theoretic method aimed at finding the optimal application configuration along a limited prediction horizon by solving an online optimization problem. Our control strategies are designed to address latency constraints, by using Queueing Theory models, and energy consumption by changing the number of used cores and the CPU frequency through the Dynamic Voltage and Frequency Scaling (DVFS) function of modern multi-core CPUs. The proactive capabilities, in addition to the latency- and energy-awareness, represent the novel features of our approach. Experiments performed using a high-frequency trading application show the effectiveness compared with state-of-the-art techniques.
A full version of the slides (with transitions) is available at: https://docs.google.com/presentation/d/1VZ3y3RQDLFi_xA7Rl0Vj1iqBdoerxCMG4y53uMz9Ziw/edit?usp=sharing
NLP-Focused Applied ML at Scale for Global Fleet Analytics at ExxonMobilDatabricks
ExxonMobil leveraged machine learning at scale using Databricks to extract insights from equipment maintenance logs and improve operations. The logs contained both structured and unstructured text data across a global fleet maintained in legacy systems, limiting traditional analysis. By ingesting and enriching over 60 million records using natural language processing, the system identified outliers, enabled capacity planning, and prioritized maintenance tasks, projected to save millions annually through more effective reliability and maintenance guidance.
FAULT TOLERANCE OF RESOURCES IN COMPUTATIONAL GRIDSMaurvi04
This document discusses fault tolerance techniques for computational grids. It begins with an introduction to grid computing and defines some key terms related to faults and failures. It then discusses different types of faults that can occur in grids, including physical faults, network faults, and process faults. It outlines several fault tolerance techniques used in grids, including job and data replication, checkpointing, scheduling approaches, and load balancing strategies. The document concludes with suggestions for future work, such as optimizing checkpoint storage and granularity.
1. Discuss the structured system analysis and design methodologies
2. What is DSS? Discuss the components and capabilities of DSS.
3. Narrate the stages of SDLC
4. Define OOP. What are the applications of it?
The document outlines a three-phase approach to developing an intelligent monitoring platform:
Phase 1 involves interviewing dev and ops teams to understand current monitoring practices.
Phase 2 focuses on improving the postmortem process and outage understanding.
Phase 3 aims to reduce the time to identify and resolve outages through expanded data collection, correlation analysis, and predictive capabilities.
The document outlines a resilient system design approach for PayPal that aims to isolate functionality and enable independent availability patterns. It proposes using asynchronous reconciliation to resolve failures without affecting the customer experience. The key elements are:
1) Isolating functionality into independent functional components (FCs) that can be developed, deployed and executed separately for increased flexibility.
2) Using a service container (SC) to orchestrate FCs and handle failures by consolidating responses and initiating reconciliation.
3) Implementing circuit breakers and fallback behaviors to protect clients, services, and business from failures in FCs or dependencies.
4) Employing an eventual consistency model with automated reconciliation to resolve inconsistencies due to network partitions
This document provides guidance on evaluating low-code platforms (LCPs) and tools (LCTs) for suitability. It outlines several factors to consider:
1. Verify critical features through prototyping to ensure an LCP supports must-have functionality. Use critical features to filter LCP options.
2. Evaluate security, including data security practices, data ownership policies, user permissions, audit logs, breach response, and disaster recovery.
3. Assess scale, including table size limits, business logic processing ability, transaction capacity, data crunching/analysis capabilities, and user experience with large data volumes.
4. Consider interfaces to other tools, including common data formats, traffic volumes
Documented Requirements are not Useless After All!Lionel Briand
The document discusses challenges related to requirements in agile software development projects, including dealing with natural language requirements, domain knowledge, frequent changes, and configuring requirements for product families. It outlines research conducted to address these challenges through natural language processing techniques like analyzing compliance with boilerplate templates, extracting domain models from requirements, and analyzing change impact when requirements change. The talk aims to discuss when documented requirements are important in practice and how they can be supported to effectively handle common challenges.
The document discusses Enterprise Resource Planning (ERP) systems. It describes the ERP architecture as using a client-server model with a relational database to store and process data. The ERP lifecycle involves definition, construction, implementation, and operation phases. Core ERP components manage accounting, production, human resources and other internal functions, while extended components provide external capabilities like CRM, SCM, and e-business. Proper implementation requires screening software, evaluating packages, analyzing process gaps, reengineering workflows, training staff, testing, and post-implementation support.
A sdn based application aware and network provisioningStanley Wang
The document discusses application aware SDN network provisioning. It begins with an overview of YARN architecture in Hadoop, including its benefits over earlier Hadoop architectures like improved scalability and utilization. It then discusses how SDN can be integrated with big data and cloud computing workloads by optimizing network topology and routing based on traffic patterns. Two approaches are proposed - reactive, where the SDN controller learns patterns from job logs/endpoints and modifies paths, and proactive where applications directly inform the network of intent. Finally, it proposes a service profile based SDN platform that uses network profiles and APIs to declaratively define logical topologies and provide network services and abstractions to applications.
1. The document discusses research activities related to reducing energy consumption by at least 30% through the development of core source technologies for universal operating systems.
2. It describes four papers being presented, including ones on system and device latency modeling, power management frameworks for embedded systems, and automatic selection of power policies for operating systems.
3. It also summarizes four research topics from the National University, including performance evaluation of parallel applications using a power-aware paging method on next-generation memory architectures.
Break-out session by Bernhard Becker, Deltares, at the Delft-FEWS International User Days 2020, during Delft Software Days - Edition 2020. Thursday, 5 November 2020.
The document describes the system architecture for an augmented cognition system. It includes sensors that assess cognitive state, a cognitive state assessor that analyzes sensor data, and various components that select and sequence tasks or switch modalities based on the assessed cognitive state. The goal is to maximize operator performance by detecting and mitigating cognitive overload. It provides details on the sensors used, cognitive state gauges, mitigation strategies like pacing and task sequencing, and how tasks are selected and routed through the system.
AlphaPy is a Python framework for building machine learning pipelines. It contains two main pipelines: the Model Pipeline for generating models and the Domain Pipeline for preparing training and test data. The Model Pipeline uses scikit-learn and other packages to build predictive models for classification and regression. The Domain Pipeline transforms raw data into canonical form suitable for modeling. The output is a persistent Model Object. The document provides examples of using AlphaPy to build stock market prediction models that identify predictive technical indicators and validate results on training and test data.
AlphaPy: A Data Science Pipeline in PythonMark Conway
AlphaPy is a Python framework for building machine learning pipelines. It contains two main pipelines: the Model Pipeline for generating models and the Domain Pipeline for preparing training and test data. The Model Pipeline uses scikit-learn and other packages to build predictive models for classification and regression. The Domain Pipeline transforms raw data into canonical form suitable for modeling. The output is a persistent Model Object. The document provides examples of using AlphaPy to build stock market prediction models that identify predictive technical indicators and validate results on training and test data.
Overview of DuraMat software tool developmentAnubhav Jain
The document discusses software tools being developed by researchers for photovoltaic (PV) applications. It summarizes several software projects funded by DuraMat that address different aspects of PV including: (1) PV system modeling and analysis, (2) operation and degradation modeling, and (3) planning and reducing levelized cost of energy. The software aims to solve a range of PV problems, are open source, and developed collaboratively on GitHub to be reusable and sustainable resources for the community.
The document outlines requirements for an MSO network operations center (NOC). Key requirements include monitoring a complex network supporting various services, having full visibility and management capabilities across all technologies, and providing a single pane of view to minimize screen space. The NOC should focus on service uptime through fault detection, resolution, and performance monitoring. People, processes, and tools are needed to provide a quality NOC. A suggested organization structure includes first line engineers, support engineers, shift leads, managers, and specialized roles. Considerations for NOC systems include flexibility, scalability, cost containment, lean architecture, and simplicity.
The document discusses performance evaluation of computer and telecommunication systems. Performance evaluation aims to quantitatively predict a system's behavior and is used to compare designs, plan for capacity, and debug performance issues. It involves modeling, simulation, and testing approaches of varying cost and accuracy. Key metrics include counts, times, sizes, productivity, response time, and reliability measures. Workload characterization analyzes how systems are used, while benchmarks compare performance across systems running standardized tests.
What Is Machine Learning?
Where do we deploy machine learning and what software and cloud services are out there to support it?
What are the trends in deploying these systems and what are the benefits for IT?
Do you have a IoT Machine Learning Case Study in the Cloud?
Keep Calm and React with Foresight: Strategies for Low-Latency and Energy-Eff...Tiziano De Matteis
This talk has been given at PPoPP 2016 (Barcelona)
The paper addresses the problem of designing control strategies for elastic stream processing applications. Elasticity allows applications to rapidly change their configuration (e.g. the number of used resources) on-the-fly, in response to fluctuations of their workload. In this work we face this problem by adopting the Model Predictive Control technique, a control-theoretic method aimed at finding the optimal application configuration along a limited prediction horizon by solving an online optimization problem. Our control strategies are designed to address latency constraints, by using Queueing Theory models, and energy consumption by changing the number of used cores and the CPU frequency through the Dynamic Voltage and Frequency Scaling (DVFS) function of modern multi-core CPUs. The proactive capabilities, in addition to the latency- and energy-awareness, represent the novel features of our approach. Experiments performed using a high-frequency trading application show the effectiveness compared with state-of-the-art techniques.
A full version of the slides (with transitions) is available at: https://docs.google.com/presentation/d/1VZ3y3RQDLFi_xA7Rl0Vj1iqBdoerxCMG4y53uMz9Ziw/edit?usp=sharing
Similar to History-Aware Explanations: Towards Enabling Human-in-the-Loop in Self-Adaptive Systems (20)
This is a talk I gave to potential applicants about our pilot of AutoFeedback for term 2 of a first-year Java programming module. AutoFeedback is an automated code feedback platform that tries to lower the barriers to entry as much as possible for submitting code from an IDE, and receiving feedback about it.
MODELS 2019: Querying and annotating model histories with time-aware patternsAntonio García-Domínguez
30 minute slides for our talk at the IEEE / ACM 22nd International Conference on Model Driven Engineering Languages and Systems conference, on our Eclipse Hawk model indexing tool.
These are slides and resources for a 1h talk on resources for high-quality figures and tables in presentations. In the related Gitlab project [1], "samples" includes some samples from the mentioned technologies, from my past papers, my PhD dissertation, and various resources on the web.
[1]: https://gitlab.com/a.garcia-dominguez/2019-talk-presentations
Eclipse Hawk provides scalable querying of models by indexing them into graph databases. It addresses challenges of collaborative modeling on large systems by distributed teams. The Hawk API is designed for flexibility, performance, and scalability through features like multiple communication styles, efficient encodings, and paged results.
Transparencias para una conferencia invitada en la Escuela Superior de Ingeniería (Cádiz) - 2 de abril de 2018.
Tratan sobre Hawk (https://github.com/mondo-project/mondo-hawk), el proyecto MONDO y mi experiencia dando clase e investigando en UK.
These are the slides for a second-year 2-hour lecture in the CS2010 "Group Project" module explaining software quality through the ISO 25010 standard and giving some basics of software testing. The talk illustrates software quality concepts through relevant videogames, in line with the "strategy game" theme chosen for this year's group coursework.
MoDELS'16 presentation: Integration of a Graph-Based Model Indexer in Commerc...Antonio García-Domínguez
Modelio M2T with Jython
HT: Hawk M2T with EGL
A García-Domínguez et al. Integration of a Graph-Based Model Indexer in Commercial Modelling Tools 23 / 27
Intro Integration Evaluation Conclusions
M2T with Hawk: conclusions
Indexing time grows linearly with model size
Generation time grows logarithmically
Hawk+EGL is faster than Modelio+Jython for large models
Hawk scales better due to indexing and querying optimizations
Modelio+Jython has higher constant overhead
Break-even point is ~10k elements
For large projects, Hawk is preferable
SOFTEAM now uses Hawk for all M2
This is a slightly revised version of the slides I used for my talk at BMSD 2015, titled "Domain-Specific Language for Generating Administrative Process Applications". In these slides I present our current version of the AdminDSL language, from which we can produce Django web apps.
Este documento proporciona orientación sobre cómo elaborar un buen póster científico. Explica que un póster debe resumir la investigación de manera concisa utilizando principalmente gráficos e imágenes en lugar de texto extenso. Además, discute consideraciones de diseño como la distribución visual de los elementos, el uso de colores, tipos de letra y formatos de archivo apropiados para imágenes. También ofrece consejos sobre software de diseño y sobre cómo defender efectivamente el póster durante la sesión.
Software libre para la integración de información en la Universidad de CádizAntonio García-Domínguez
Este documento presenta las herramientas y procesos utilizados para integrar información en la Universidad de Cádiz. Describe cómo se localizan, cargan y explotan los datos para generar informes e información de alto nivel. También introduce varias herramientas de código abierto como Kettle, Pentaho, Mondrian y CKAN que se usan para extraer, transformar, cargar y analizar datos de forma integrada.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELgerogepatton
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
KuberTENes Birthday Bash Guadalajara - K8sGPT first impressionsVictor Morales
K8sGPT is a tool that analyzes and diagnoses Kubernetes clusters. This presentation was used to share the requirements and dependencies to deploy K8sGPT in a local environment.
International Conference on NLP, Artificial Intelligence, Machine Learning an...gerogepatton
International Conference on NLP, Artificial Intelligence, Machine Learning and Applications (NLAIM 2024) offers a premier global platform for exchanging insights and findings in the theory, methodology, and applications of NLP, Artificial Intelligence, Machine Learning, and their applications. The conference seeks substantial contributions across all key domains of NLP, Artificial Intelligence, Machine Learning, and their practical applications, aiming to foster both theoretical advancements and real-world implementations. With a focus on facilitating collaboration between researchers and practitioners from academia and industry, the conference serves as a nexus for sharing the latest developments in the field.
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Sinan KOZAK
Sinan from the Delivery Hero mobile infrastructure engineering team shares a deep dive into performance acceleration with Gradle build cache optimizations. Sinan shares their journey into solving complex build-cache problems that affect Gradle builds. By understanding the challenges and solutions found in our journey, we aim to demonstrate the possibilities for faster builds. The case study reveals how overlapping outputs and cache misconfigurations led to significant increases in build times, especially as the project scaled up with numerous modules using Paparazzi tests. The journey from diagnosing to defeating cache issues offers invaluable lessons on maintaining cache integrity without sacrificing functionality.
Understanding Inductive Bias in Machine LearningSUTEJAS
This presentation explores the concept of inductive bias in machine learning. It explains how algorithms come with built-in assumptions and preferences that guide the learning process. You'll learn about the different types of inductive bias and how they can impact the performance and generalizability of machine learning models.
The presentation also covers the positive and negative aspects of inductive bias, along with strategies for mitigating potential drawbacks. We'll explore examples of how bias manifests in algorithms like neural networks and decision trees.
By understanding inductive bias, you can gain valuable insights into how machine learning models work and make informed decisions when building and deploying them.
ACEP Magazine edition 4th launched on 05.06.2024Rahul
This document provides information about the third edition of the magazine "Sthapatya" published by the Association of Civil Engineers (Practicing) Aurangabad. It includes messages from current and past presidents of ACEP, memories and photos from past ACEP events, information on life time achievement awards given by ACEP, and a technical article on concrete maintenance, repairs and strengthening. The document highlights activities of ACEP and provides a technical educational article for members.
The CBC machine is a common diagnostic tool used by doctors to measure a patient's red blood cell count, white blood cell count and platelet count. The machine uses a small sample of the patient's blood, which is then placed into special tubes and analyzed. The results of the analysis are then displayed on a screen for the doctor to review. The CBC machine is an important tool for diagnosing various conditions, such as anemia, infection and leukemia. It can also help to monitor a patient's response to treatment.
Redefining brain tumor segmentation: a cutting-edge convolutional neural netw...IJECEIAES
Medical image analysis has witnessed significant advancements with deep learning techniques. In the domain of brain tumor segmentation, the ability to
precisely delineate tumor boundaries from magnetic resonance imaging (MRI)
scans holds profound implications for diagnosis. This study presents an ensemble convolutional neural network (CNN) with transfer learning, integrating
the state-of-the-art Deeplabv3+ architecture with the ResNet18 backbone. The
model is rigorously trained and evaluated, exhibiting remarkable performance
metrics, including an impressive global accuracy of 99.286%, a high-class accuracy of 82.191%, a mean intersection over union (IoU) of 79.900%, a weighted
IoU of 98.620%, and a Boundary F1 (BF) score of 83.303%. Notably, a detailed comparative analysis with existing methods showcases the superiority of
our proposed model. These findings underscore the model’s competence in precise brain tumor localization, underscoring its potential to revolutionize medical
image analysis and enhance healthcare outcomes. This research paves the way
for future exploration and optimization of advanced CNN models in medical
imaging, emphasizing addressing false positives and resource efficiency.
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
Comparative analysis between traditional aquaponics and reconstructed aquapon...bijceesjournal
The aquaponic system of planting is a method that does not require soil usage. It is a method that only needs water, fish, lava rocks (a substitute for soil), and plants. Aquaponic systems are sustainable and environmentally friendly. Its use not only helps to plant in small spaces but also helps reduce artificial chemical use and minimizes excess water use, as aquaponics consumes 90% less water than soil-based gardening. The study applied a descriptive and experimental design to assess and compare conventional and reconstructed aquaponic methods for reproducing tomatoes. The researchers created an observation checklist to determine the significant factors of the study. The study aims to determine the significant difference between traditional aquaponics and reconstructed aquaponics systems propagating tomatoes in terms of height, weight, girth, and number of fruits. The reconstructed aquaponics system’s higher growth yield results in a much more nourished crop than the traditional aquaponics system. It is superior in its number of fruits, height, weight, and girth measurement. Moreover, the reconstructed aquaponics system is proven to eliminate all the hindrances present in the traditional aquaponics system, which are overcrowding of fish, algae growth, pest problems, contaminated water, and dead fish.
2008 BUILDING CONSTRUCTION Illustrated - Ching Chapter 02 The Building.pdf
History-Aware Explanations: Towards Enabling Human-in-the-Loop in Self-Adaptive Systems
1. History-Aware Explanations: Towards
Enabling Human-in-the-Loop in
Self-Adaptive Systems
J.M. Parra-Ullauri, A. García-Domínguez, N. Bencomo, L.H. García-Paucar
SAM 2022, 24 October 2022
3. Explainability for trustworthy self-adaptive systems
Software working in difficult environments
• Fixed behaviour cannot handle complex and uncertain situations
• Instead, a self-adaptive system changes its behaviour to meet its
goals as needed
• Consider self-driving cars, complex cloud deployments, data/power
networks...
Emergent behaviour needs to be explained
• Lack of trust on SAS is hindering their adoption
• Trust can be gained by allowing users to understand why the SAS
made its decisions, and to influence the decisions as desired
1
4. Are humans integrated in decision-making loops?
• Many SAS follow feedback loops: MAPE-K is a common
architecture
• How does the human get involved there?
• Can the human observe the loop?
• Can the human pitch in with their own input, e.g. driving preferences
for a self-driving car, or what to clean for a robot vaccuum? 2
5. Context: roadmap for history-aware self-adaptive systems
• This work is part of our
roadmap for history-aware SAS
• Level 1: explain decisions after
the fact
• Level 2: explain behaviour on
the fly
• Level 3 (this paper): external
agent (human) uses the
explanations to influence the
system via “effectors”
(adaptation controls)
3
7. Extending MAPE-K: explanatory and feedback layer
• We propose adding an layer to
MAPE-K to integrate the
human
• Filter: collect relevant history
of the system
• Explain: use history to describe
system behaviour
• Feedback: human uses relatable
“effectors” to influence
behaviour
4
8. Extending MAPE-K: the Filter component
Log
timesliceID: EString
Agent
name: EString
Decision
name: EString
Observation
description: EString
probability: EDouble
Action
name: EString
NFR
name: EString
0..*
0..*
0..*
0..*
0..*
decisions
0..1
0..*
observations
0..1
0..*
actionTaken
0..1
observation
0..1
• The Filter component collects
information from the Monitor,
Analyze, and Plan stages: for
instance, raw sensor / decision
logs
• This information is reshaped
according to a trace
metamodel, divided into a
algorithm-independent half and
an algorithm-centric half
• Model versions are indexed by
Hawk into a temporal graph DB
5
9. Extending MAPE-K: the Explain component
Explanation construction: done in this paper
• Query the TGDB for the info to create explanations
• Time-aware EOL dialect in Hawk for formulating questions
Explanation presentation: done in this paper
• Plots (e.g. time series of key performance metrics)
• Yes/no answers (e.g. “was X always/never true?”)
• Examples of matches of a given situation
Explanation reception: future work
• Collect info on how the user reacted to the explanations
• Track what the user knows and how they perceive the system
6
10. Extending MAPE-K: the Feedback component
Abstracting away influences into “effectors”
• Users should not have to be familiar with the underlying algorithm
• The system should include effectors to allow the user to influence
the system, expressed in their terms
• User input should be recorded in system history (for accountability)
Possible effectors at Plan/Execute stages
• A SAS manages tradeoffs between competing goals: users can
influence the relative priority of those goals (e.g. performance vs
efficiency)
• Users can suggest specific actions to the SAS at the Execute stage,
triggering a reconfiguration to meet its new preference
7
12. Case study: Remote Data Mirroring (RDM)
• SAS manages data servers and
network links
• Two actions: switch between
minimal/redundant topologies
• Handles
cost/reliability/performance
tradeoffs, while meeting SLAs
• SLA satisfaction partially
observable over monitoring
variables (RBC, TTW, ANL)
• Uses Requirements-aware
Model POMDP for
decision-making
8
13. RDM: Filter
Filter component collects into a temporal graph DB:
• Initial stakeholder preferences about the NFRs and SLAs
• Adaptation strategies selected by SAS based on preferences, and
their impact on the observed satisfaction levels
• Situations detected at runtime, where initial preferences may drive
SAS to unsuitable adaptation strategies
9
14. RDM: Explain (construction)
1 var result : Sequence;
2 var nfrs = NFRBelief.latest.all;
3 /∗ ... ∗/
4 for (nfr in nfrs) {
5 var currentNFR = nfr.latest;
6 result.add(Sequence {
7 currentNFR.eContainer.eContainer.timesliceID,
8 currentNFR.nfr.name,
9 currentNFR.satisfied,
10 currentNFR.estimatedProbability,
11 currentNFR.eContainer.actionTaken.name,
12 aveMEC, aveMR, aveMP
13 });
14 }
15 return result;
• An EOL query is run after each
timeslice
• For each NFR, we know:
• Timeslice ID
• Name of NFR
• Considered satisfied? (Y/N)
• Satisfaction level
• Taken action (topology)
• Average MEC/MR/MP
satisfaction over the history
of the system
10
15. RDM: Explain (presentation)
• Results are fed to a custom GUI, with historic/current values
• User can track satisfaction levels over time
11
16. RDM: Feedback
• +/- buttons allows for changing relative weights for Plan stage
• Simple description: “make the algorithm focus less/more on this”
• Interactions are recorded, and algorithm still tries to meet all SLAs 12
17. RDM: example - slices 1–323
Initially, the system is working as expected by the user.
13
18. RDM: example - slices 324–645
System suffers connectivity issues, but relative weights of
reliability/cost/performance keep it on the minimal spanning topology.
14
19. RDM: example - use of effector
User decides to put more focus onto reliability, clicking on “+” under MR:
GUI runs an EOL query, and shows a dialog with a quick summary of the
current situation before asking for confirmation.
User confirms the action, and MR weight is increased.
15
20. RDM: example - slices 646+
System switches to RT after putting more weight on reliability, which
does impact cost/performance but stays within SLAs.
16
21. RDM: example - impact of change
Before update of
preferences
After update of
preferences
• Before the preferences were
updated, average satisfaction of
MR was below SLA threshold
• After the update, MR
satisfaction improves at the
expense of the others, but all
SLAs are still met
17
22. What we have done so far
Extension to MAPE-K
• We proposed involving humans in the MAPE-K feedback loop, by
adding an explanatory & feedback layer
• Layer made up of Filter, Explain, and Feedback components
Implementation of E&F layer
• Filter: reshape to trace model + index into temporal graph DB
• Explain: query temporal graph + generate plots/answers
• Feedback: effectors for users to influence Plan/Execute
Case study: RDM
• Applied E&F layer to the RDM SAS
• Custom GUI with system-specific effectors
• Simulated scenario of preference readjustment
23. What’s next?
Explanation receptions
• Explanations currently targeted SAS developers
• SAS users will need a different style of explanations
• Follow-up study on explanation efficacy and appropriateness
(Opportunity-Willigness-Capability), and effectors’ impact on
trustworthiness
Further lines of work
• Currently ongoing: non-human consumers of explanations (e.g.
external system optimising AI/ML hyper-parameters)
• Additional case studies on other SAS
• Other explanations besides factual ones, e.g. formulating hypotheses
and producing evidence supporting/rejecting them
• Distributed SAS (→ distributed trace models)