In this session we’ll leave the need for performance a foregone conclusion and take a whirlwind tour through the complexity of modern Internet architectures. The complexities lead to evil optimization problems and significant challenges troubleshooting production issues to a speedy and successful end.
Starting with the simple facts that you can’t fix what you can’t see and you can’t improve what you can’t measure, we’ll discuss what needs monitoring and why. We’ll talk about unlikely allies in the fight for time and budget to instrument systems, applications and processes for observability.
You’ll leave the session with a better understanding of what it looks like to troubleshoot the storm of a malfunctioning large architecture and some tools and techniques you can use to not be swallowed by the Kraken.
What is observability and how is it different from traditional monitoring? How do we effectively monitor and debug complex, elastic microservice architectures? In this interactive discussion, we’ll answer these questions. We’ll also introduce the idea of an “observability pipeline” as a way to empower teams following DevOps practices. Lastly, we’ll demo cloud-native observability tools that fit this “observability pipeline” model, including Fluentd, OpenTracing, and Jaeger.
Observability has emerged as one of the hottest topics on the DevOps landscape. Organizations seek to improve visibility into their cloud infrastructure and applications and identify production issues that may negatively impact #customerexperience.
➡️ But what are some of the best practices for scaling observability for modernapplications?
➡️ What challenges are #cloudplatforms facing?
Explore how to overcome the challenges and unlock speed, observability, and automation across your DevOps lifecycle.
How to Move from Monitoring to Observability, On-Premises and in a Multi-Clou...Splunk
With the acceleration of customer and business demands, site reliability engineers and IT Ops analysts now require operational visibility into their entire architecture, something that traditional APM tools, dev logging tools, and SRE tools aren’t equipped to provide. Observability enables you to inspect and understand your IT stack on premises and in the cloud(s); It’s no longer about whether your system works (monitoring), but being able to task why it is not working? (Observability). This presentation will outline key steps to take to move from monitoring to observability.
Whether you’re an enterprise migrating to cloud native or born in the cloud, most of today’s APM and Observability tools don’t support how your engineers and DevOps teams need to develop, deploy, and support their software. Observability needs to shift left and reflect the modern way companies organize their development teams and their vital interdependencies.
Chronosphere is the only vendor addressing the unique requirements for observability in a cloud native world. Join this webinar to learn:
- What cloud native observability is and how it is different from the promises made by traditional cloud APM and observability vendors
- How to use cloud native observability to do more “Dev” and less “Ops” so you can dramatically improve developer and engineer workflows and productivity
- How to make on-call shifts less stressful so your engineers aren’t getting burned out
How deeply can you understand what is happening inside your application? In modern, microservices-based applications, it’s critical to have end-to-end observability of each component and the communications between them in order to quickly identify and debug issues. In this session, we show how to have the necessary instrumentation and how to use the data you collect to have a better grasp of your production environment. On AWS, CloudWatch collects monitoring and operational data in the form of logs, metrics, and events, providing you with a unified view of AWS resources, applications, and services. With AWS X-Ray, you can understand how your application and its underlying services are performing to identify and troubleshoot the root cause of performance issues and errors. X-Ray provides an end-to-end view of requests as they travel through your application, and shows a map of your application’s underlying components. AWS App Mesh standardizes how your microservices communicate, giving you end-to-end visibility and helping to ensure high-availability for your applications.
What is observability and how is it different from traditional monitoring? How do we effectively monitor and debug complex, elastic microservice architectures? In this interactive discussion, we’ll answer these questions. We’ll also introduce the idea of an “observability pipeline” as a way to empower teams following DevOps practices. Lastly, we’ll demo cloud-native observability tools that fit this “observability pipeline” model, including Fluentd, OpenTracing, and Jaeger.
Observability has emerged as one of the hottest topics on the DevOps landscape. Organizations seek to improve visibility into their cloud infrastructure and applications and identify production issues that may negatively impact #customerexperience.
➡️ But what are some of the best practices for scaling observability for modernapplications?
➡️ What challenges are #cloudplatforms facing?
Explore how to overcome the challenges and unlock speed, observability, and automation across your DevOps lifecycle.
How to Move from Monitoring to Observability, On-Premises and in a Multi-Clou...Splunk
With the acceleration of customer and business demands, site reliability engineers and IT Ops analysts now require operational visibility into their entire architecture, something that traditional APM tools, dev logging tools, and SRE tools aren’t equipped to provide. Observability enables you to inspect and understand your IT stack on premises and in the cloud(s); It’s no longer about whether your system works (monitoring), but being able to task why it is not working? (Observability). This presentation will outline key steps to take to move from monitoring to observability.
Whether you’re an enterprise migrating to cloud native or born in the cloud, most of today’s APM and Observability tools don’t support how your engineers and DevOps teams need to develop, deploy, and support their software. Observability needs to shift left and reflect the modern way companies organize their development teams and their vital interdependencies.
Chronosphere is the only vendor addressing the unique requirements for observability in a cloud native world. Join this webinar to learn:
- What cloud native observability is and how it is different from the promises made by traditional cloud APM and observability vendors
- How to use cloud native observability to do more “Dev” and less “Ops” so you can dramatically improve developer and engineer workflows and productivity
- How to make on-call shifts less stressful so your engineers aren’t getting burned out
How deeply can you understand what is happening inside your application? In modern, microservices-based applications, it’s critical to have end-to-end observability of each component and the communications between them in order to quickly identify and debug issues. In this session, we show how to have the necessary instrumentation and how to use the data you collect to have a better grasp of your production environment. On AWS, CloudWatch collects monitoring and operational data in the form of logs, metrics, and events, providing you with a unified view of AWS resources, applications, and services. With AWS X-Ray, you can understand how your application and its underlying services are performing to identify and troubleshoot the root cause of performance issues and errors. X-Ray provides an end-to-end view of requests as they travel through your application, and shows a map of your application’s underlying components. AWS App Mesh standardizes how your microservices communicate, giving you end-to-end visibility and helping to ensure high-availability for your applications.
Is your company built on software? How do you know if your customer's experience is slow and sucks? How do you debug slowness or troubleshoot an incident? Observability! David Mitchell, VP of Engineering at Datadog will talk to use about Observability, why it's important, what it is and how Datadog helps reduce toil in your environment.
GDG Cloud Southlake #13
As engineers we spend much of our time getting stuff to production and making sure our infrastructure doesn’t burn down out right. We however spend very little time learning to understand and respond to outages. Does our platform degrade in a graceful way or what does a high cpu load really mean? What can we learn from level 1 outages to be able to run our platforms more reliably.
Plenty of people are jumping on the new hype, Observability, lots of them are replacing their “legacy” monitoring stack. Not all of them achieve the goals they set. But observability is not a tool — it is a property of a system. Moving from many small black boxes to a more holistic view of your system.
In this talk we ll talk about how to prepare teams to tweak their testing and monitoring setup and work instructions to quickly observe, react to and resolve problems. We look at improving your monitoring by adapting your culture and then maybe your tooling. Where we as engineers not only write, maintain and operate our software platforms but actively pursue ways to learn and predict its (non-functional) behavior.
Furthermore we ll discuss the need for and the options of not only monitoring our platforms and it's envitable outages, but also their (potential) length and impact. We ll look at tools like at using Service Level Objects for ways to prepare teams to tweak their testing and monitoring setup and runbooks to quickly observe, react to and resolve problems.
With Instana the "Classic" Observability is not the end of the line. Find out what Observability means and how it can help DevOps, Developers, SREs day-by-day.
Slides from my session at MeasureWorks' Performance Labs... Topic is Observability, the new buzzword in Web Performance & DevOps, trying to explain what it is and why it matters for your operations...
Can you understand what’s happening inside your code and system, simply by asking questions using your tools? Can you answer any new question you think of, or only the ones you prepared for?
Slides for the Observability-101 session held on 16th May 2021.
Link for the event: https://community.cncf.io/events/details/cncf-mumbai-presents-observability-101/
More Than Monitoring: How Observability Takes You From Firefighting to Fire P...DevOps.com
For some, observability is just a hollow rebranding of monitoring, for others it’s monitoring on steroids. But what if we told you observability is the new way to find out why—not just if—your distributed system or application isn’t working as expected? Today, we see that traditional monitoring approaches can fall short if a system or application doesn’t adequately externalize its state.
This is truer as workloads move into the cloud and leverage ephemeral technologies, such as microservices and containers. To reach observability, IT and DevOps teams need to correlate different sources from logs, metrics, traces, events and more. This becomes even more challenging when defining the online revenue impact of a failed container—after all, this is what really matters to the business.
This webinar will cover:
The differences between observability and monitoring
Why it is a bigger challenge in a multicloud and containerized world
How observability results in less firefighting and more fire prevention
How new platforms can help gain observability (on premises and in the cloud) for containers, microservices and even SAP or mainframes
Monitoring with Dynatrace Presentation.pptxKnoldus Inc.
Dynatrace is an APM solution designed to automatically measure the performance of an application or micro-service, discover the topology and dependencies of your architecture, and determine if a problem is in the code or in the software and hardware infrastructure supporting your application. It is a one-end solution to diagnosing all stacks of your application with logs, metrics and traces, and that too, in real-time.
Getting started with Site Reliability Engineering (SRE)Abeer R
"Getting started with Site Reliability Engineering (SRE): A guide to improving systems reliability at production"
This is an intro guide to share some of the common concepts of SRE to a non-technical audience. We will look at both technical and organizational changes that should be adopted to increase operational efficiency, ultimately benefiting for global optimizations - such as minimize downtime, improve systems architecture & infrastructure:
- improving incident response
- Defining error budgets
- Better monitoring of systems
- Getting the best out of systems alerting
- Eliminating manual, repetitive actions (toils) by automation
- Designing better on-call shifts/rotations
How to design the role of the Site Reliability Engineer (who effectively works between application development teams and operations support teams)
Design of observers for nonlinear systems using the Frobenius theorem. Presentation for the defense of my MSc Thesis at the School of Applied Mathematics, NTU Athens.
Is your company built on software? How do you know if your customer's experience is slow and sucks? How do you debug slowness or troubleshoot an incident? Observability! David Mitchell, VP of Engineering at Datadog will talk to use about Observability, why it's important, what it is and how Datadog helps reduce toil in your environment.
GDG Cloud Southlake #13
As engineers we spend much of our time getting stuff to production and making sure our infrastructure doesn’t burn down out right. We however spend very little time learning to understand and respond to outages. Does our platform degrade in a graceful way or what does a high cpu load really mean? What can we learn from level 1 outages to be able to run our platforms more reliably.
Plenty of people are jumping on the new hype, Observability, lots of them are replacing their “legacy” monitoring stack. Not all of them achieve the goals they set. But observability is not a tool — it is a property of a system. Moving from many small black boxes to a more holistic view of your system.
In this talk we ll talk about how to prepare teams to tweak their testing and monitoring setup and work instructions to quickly observe, react to and resolve problems. We look at improving your monitoring by adapting your culture and then maybe your tooling. Where we as engineers not only write, maintain and operate our software platforms but actively pursue ways to learn and predict its (non-functional) behavior.
Furthermore we ll discuss the need for and the options of not only monitoring our platforms and it's envitable outages, but also their (potential) length and impact. We ll look at tools like at using Service Level Objects for ways to prepare teams to tweak their testing and monitoring setup and runbooks to quickly observe, react to and resolve problems.
With Instana the "Classic" Observability is not the end of the line. Find out what Observability means and how it can help DevOps, Developers, SREs day-by-day.
Slides from my session at MeasureWorks' Performance Labs... Topic is Observability, the new buzzword in Web Performance & DevOps, trying to explain what it is and why it matters for your operations...
Can you understand what’s happening inside your code and system, simply by asking questions using your tools? Can you answer any new question you think of, or only the ones you prepared for?
Slides for the Observability-101 session held on 16th May 2021.
Link for the event: https://community.cncf.io/events/details/cncf-mumbai-presents-observability-101/
More Than Monitoring: How Observability Takes You From Firefighting to Fire P...DevOps.com
For some, observability is just a hollow rebranding of monitoring, for others it’s monitoring on steroids. But what if we told you observability is the new way to find out why—not just if—your distributed system or application isn’t working as expected? Today, we see that traditional monitoring approaches can fall short if a system or application doesn’t adequately externalize its state.
This is truer as workloads move into the cloud and leverage ephemeral technologies, such as microservices and containers. To reach observability, IT and DevOps teams need to correlate different sources from logs, metrics, traces, events and more. This becomes even more challenging when defining the online revenue impact of a failed container—after all, this is what really matters to the business.
This webinar will cover:
The differences between observability and monitoring
Why it is a bigger challenge in a multicloud and containerized world
How observability results in less firefighting and more fire prevention
How new platforms can help gain observability (on premises and in the cloud) for containers, microservices and even SAP or mainframes
Monitoring with Dynatrace Presentation.pptxKnoldus Inc.
Dynatrace is an APM solution designed to automatically measure the performance of an application or micro-service, discover the topology and dependencies of your architecture, and determine if a problem is in the code or in the software and hardware infrastructure supporting your application. It is a one-end solution to diagnosing all stacks of your application with logs, metrics and traces, and that too, in real-time.
Getting started with Site Reliability Engineering (SRE)Abeer R
"Getting started with Site Reliability Engineering (SRE): A guide to improving systems reliability at production"
This is an intro guide to share some of the common concepts of SRE to a non-technical audience. We will look at both technical and organizational changes that should be adopted to increase operational efficiency, ultimately benefiting for global optimizations - such as minimize downtime, improve systems architecture & infrastructure:
- improving incident response
- Defining error budgets
- Better monitoring of systems
- Getting the best out of systems alerting
- Eliminating manual, repetitive actions (toils) by automation
- Designing better on-call shifts/rotations
How to design the role of the Site Reliability Engineer (who effectively works between application development teams and operations support teams)
Design of observers for nonlinear systems using the Frobenius theorem. Presentation for the defense of my MSc Thesis at the School of Applied Mathematics, NTU Athens.
Velocity EU 2013 What is the velocity of an unladen swallow?pdyball
Seatwave was growing fast, success was unabated, and industry awards were landing on their doormat. Infrastructure had been revamped, load patterns were understood. Everything was going just great…
Until…
The marketing team planned Seatwave’s first UK TV campaign – all regions – simultaneously, but only told the engineering team the day before the first advert was due to run!
10 seconds into the advert the site melted and there was a collective thud as heads hit desks.
It was expensive lesson to learn but also the wake up call that forced everyone in Seatwave to focus on the performance of their site.
In this session we’ll share that pain we experienced, and how we improved performance so that when all our competitors crashed during the UKs largest concert ticket sale, we were able to take 20 days revenue in just 2 hours!
However, maintaining performance is a challenge, product owners want new features, the site starts to put on weight and slowly performance starts to degrade once more.
Will it take another disaster to focus everyone on performance or is there another way to avoid “boom and bust”?
We’ll talk about the steps we’re taking to avoid “boom and bust” by making both performance and the impact performance has on our customers visible to everyone across Seatwave including:
Our Adobe Site Catalyst installation with a custom implementation of the W3C Navigation Timing API allowing us to segment our business KPI’s by speed.
How we’re using a WebPageTest within continuous integration for our QA and production builds.
How we constantly review our performance against competitors using our own installation of the HTTPArchive.
Join us on our quest in search of the Holy Grail of truly understanding how web site performance affects our business, and the processes and systems we are putting in place to ensure we keep speed at the heart of our product development roadmap.
Why Page Speed Isn't Enough - Tim Morrow - Velocity Europe 2012Tim Morrow
Betfair has completed a ground-up re-write of our Sports betting web platform. We focused on the fundamentals of resiliency and performance. We measured and improved and measured again and in the process, saw our page load times drop by a factor of 5; on average our full page load is under 3 seconds. We met our promise to our customers.
But some of our customers kept telling us our new site was slow. Many sophisticated customers have chosen to stay using our previous platform; there is a big opportunity cost while those customers have not chosen to experience the new products and features we’ve delivered on the new platform.
All our indicators say our new site is fast and our customers should be happy. We all know a faster site means more money, but we’re not getting the benefits we expect.
In this talk, hear Tim Morrow tell the story of how making Betfair’s site faster for most customers actually made it 3x slower for their biggest customers. Truly, you get what you measure.
In-kernel Analytics and Tracing with eBPF for OpenStack CloudsPLUMgrid
As the movement of applications from bare metal to the cloud continues, considerations around analytics and tracing are becoming more prevalent for security, monitoring, and accounting. As an open source project under the Linux Foundation, the IO Visor Project is working with the kernel community on extending BPF (eBPF) and is being used by many companies for security, tracing, and analytics. This talk will describe how an OpenStack micro-segmentation framework using eBPF can be utilized for analytics and tracing to secure application workloads. Use cases around application security, intrusion detection using service insertion, identity will be described. While networking is one piece of the solution, sandboxing applications to avoid attacks is also important. We will also touch upon how eBPF technology and a unified policy framework can secure application workloads in areas beyond networking.
This talk was given at Velocity '13 in Santa Clara by Abe Stanway and Jon Cowie. It talks about how Etsy make sense of the 250k metrics they gather, using their new Kale stack.
MeasureWorks - Velocity Conference Europe 2012 - a Web Performance dashboard ...MeasureWorks
For the Velocity Conference Europe 2012 workshop day this presentation is about the essentials for creation and building a Web Performance dashboard. This with ultimate goal of providing the audience a framework for designing and building a web performance dashboard. The session will cover the following 3 items:
Design guidelines: What defines a web performance dashboard? How to make sure it’s actionable and for people to actually use it on day to day basis?
Data collection: Why performance data? The various ways there are to collect data (e.g. synthetic versus RUM data, Webpagetest, Mobile) and how to correlate the different types of data and tools
Building the dashboard: How to build the actual dashboard, providing an overview of the tools/techniques used
At the end of the workshop you will be able to design and build your own dashboard based on the framework provided, or to optimize the current dashboards within your organization.
Awesome! Traffic to your site is really picking up and everything is lookin’ good. Well, except for that database back in the corner, but it will hold… right? No one really wants to deal with scaling the database tier, but hopefully your customers will drag you (perhaps kicking and screaming) to some sort of distributed database architecture.
This talk is all about scaling MySQL through hardware optimizations and sharding from a Site Engineering perspective. This includes real world examples of finding pain points, identifying risks, and evaluating cloud vs hardware scaling. I’ll also discus distributed database management, dealing with data purging, making consistent backups, and how to keep the site up when things go bad.
Be Mean to Your Code with Gauntlt and the Rugged Way // Velocity EU 2013 Work...James Wickett
This is a hands-on workshop for working with Gauntlt. The first half is philosophy, theory and social commentary. The second half is the hands on workshop.
There are two options for working through the workshop. The recommended way is to use the virtual box image as there are a couple of security tools (arachni, nmap, ...) that we will be using. It is not required for you to use it though and you can just clone the repo if you have ruby 1.9.3 and bundler.
If you want to use the vagrant box setup for the workshop, please follow the instructions in 02_Using Vagrant Box.md and if you want to just use our own box, follow the directions in 03_Using Repo Only.md
This has been tested to work on linux and OS X. You can follow along using the instructions > https://gist.github.com/wickett/25d90a462706639446cc
One of the dying skill sets in today’s engineering teams is the multi-disciplinary analyst that can truly dissect dysfunction in the radically complex architectures of today. As tools emerge that connect the dots, it might be faster to collect the data needed to analysis and decision making, but the knowledge and techniques to actually make the assessments needed are hard to come by.
In this session, we’ll walk through a complex architecture and discuss what an engineer in this role really needs to understand. We’ll analyze a few anecdotal problems and see why this world of magical automation and elastic deployments will never really displace the need for root on a production box, a debugger, and the ability to move fast, take risks and destroy performance problems.
User generated data is an old problem. Systems and network telemetry, page analytics and application state combine to form an ever growing mountain of data collected by today's tools. Collecting and storing this data requires more than just a single application, having no single point where the user touches the system and gets an answer makes debugging a nightmare and reproducing the error intractable. Distributed systems require a clear perspective on production systems and access to data in real time to have any hope of solving complex problems related to state, all while not impacting user experience.
We will explain the problem, the pains and how we solved them. Develop in production; push code to development.
Linux Tracing Superpowers by Eugene PirogovPivorak MeetUp
For a long time Linux was far behind operating systems of Unix family from the perspective of debuggability, specifically in a live production systems.
However, over the course of 2016 Linux saw a series of patches that brought it on par with Unix world: an old Linux tool called BPF has risen and extended into powerful new one – eBPF. Some say that eBPF marks the begining of true DTrace for Linux.
During the presentation I'm going to talk about tracing basics, cover a series of events that led to the development of eBPF and will compare eBPF with DTrace from Unix world. Current state of affairs of Linux tracing tools will be explored. Finally, together we'll look at some of the exciting examples of eBPF application.
***
Eugene is well known in our Ruby (and Elixir) communities. Last time when he was at #pivorak he made a very light and interesting intro to the Elixir. You can check his speech out here - http://bit.ly/2evCd9R
Building Data Driven Products With Ruby - RubyConf 2012Ryan Weald
Description
Slides from RubyConf 2012 talk:
"Big data and data science have become hot topics in the developer community during the past year. This talk will show how ruby is used to build real data driven products at scale.
Data scientist Ryan Weald walks through the building of data driven products at Sharethrough, from exploratory analysis to production systems, with an emphasis on the role Ruby plays in each phase of the data driven product cycle.
He discusses how Ruby interacts with other data analysis tools -- such as Hadoop, Cascading, Python, and Javascript -- with a constructive look at Ruby's weaknesses, and presents suggestions on how Ruby can contribute more to data science in the areas of visualization and machine learning."
This is a massive slide deck I used as the starting point for a 1.5 hour talk at the 2012 www.nerlscd.org conference. Mixture of old and (some) new slides from my usual stuff.
As we move further into the future of digital design, web design is no longer just about creating a single great desktop browsing experience. The interactive design industry is at a crossroads; mobile and tablet devices continue to propagate and fuel new interactions, and the web is now found on more devices than ever.
So, how do we adjust accordingly? More often than not, designers and programmers use old methodologies to tackle new problems. The real tool kit of a great web designer starts off-line and off-screen. This lecture will showcase important skills that will prepare flexible designs for future facing web projects. We will outline a set of new philosophies, collaborative processes and custom tools that enable productivity in this ever-changing world. We'll also cover the importance of creating your own tools and adapting to new needs, so you can stay ahead of the game.
By demonstrating the right workflow, the right tools and a future facing philosophy, this talk aims to help anyone who has thought to themselves: 'there has to be a better way'. The future isn't 12-column grids and pixel perfect PSDs. It's a flexible thinking model that relies on your understanding of development and a strong design philosophy.
This presentation talks about three topics related to monitoring. The first is a brief history and future forecast of monitoring trends. The second is a second look at the inputs, outputs, and techniques for setting SLOs. The third sets some basic tenets one should always follow when monitoring systems.
A tour of challenges today's software engineers will fast (and material they should familiarize themselves with) to cope with the issues that arise due to the distributed nature of today's applications.
Craftsmanship in software tends to erode as team sizes increase. This can be due to a large variety of reasons, but is often dependent on code base size, team size, and autonomy. In this session I'll talk about some of the challenges companies face as these things change and how to manipulate teams, architectures and how people work to maintain software craftsmanship will still delivering product.
Technology changes and process changes in how people build and manage Internet systems have driven a need for a new approach to monitoring. We talk about why, what and how.
There are two common tenets of operations: "hell is other people's software," and "better software is produced by those forced to operate it." In this session I'll take a fly-by-tour of two pieces of software that were built from the ground up for operability from the hard-earned teachings of their inoperable predecessors: a distributed datastore replacing PostgreSQL, and a message queue replacing RabbitMQ.
We'll discuss specific design aspects that increase resiliency in the event of failure and observability at all times.
There are many modern techniques for identifying anomalies in datasets. There are fewer that work as online algorithms suitable for application to real-time streaming data. What’s worse? Most of these methodologies require a deep understanding of the data itself. In this talk, we tour what the options are for identifying anomalies in real-time data and discuss how much we really need to know before hand to guess at the ever-useful question: is this normal?
What you should think about putting in webops dashboards. There's a lot of discussion that isn't annotated in the slide stack -- so you're missing a lot without audio.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
2. Hi! I’m @postwait
I founded @OmniTI
and @MessageSystems
and @Circonus
Tuesday, October 2, 12
3. Hi! I’m @postwait
I am very active in @TheOfficialACM
participating in @ACMQueue
and the practitioners board.
Tuesday, October 2, 12
4. Hi! I’m @postwait
I (regrettably) build complex systems.
Tuesday, October 2, 12
5. Why we are here
We’re here to talk about
coping with breakage
Tuesday, October 2, 12
6. Rule #1
Direct observation of failure
leads to quicker rectification.
Tuesday, October 2, 12
7. Rule #2
You cannot correct
what you cannot measure.
Tuesday, October 2, 12
8. Solution Approach #1
Debugging failures requires either
visibility into the
precipitating state
Tuesday, October 2, 12
9. Precipitating State
Single threaded applications
✓ Easy
Tuesday, October 2, 12
10. Precipitating State
Multi-threaded applications
✓ Challenging
Tuesday, October 2, 12
11. Precipitating State
Distributed applications
here there be dragons
Tuesday, October 2, 12
12. Solution Approach #2
or
direct observation of a
(and likely very many)
failing transaction
Tuesday, October 2, 12
13. Direct Observation
Observing something fail...
is priceless.
Tuesday, October 2, 12
14. Direct Observation
Observation leads to
intelligent questioning.
Tuesday, October 2, 12
15. Direct Observation
Questioning leads to answers...
but only through more observation.
Tuesday, October 2, 12
16. Direct Observation
Questioning leads to answers...
but only through more observation.
and herein lies the rub.
Tuesday, October 2, 12
17. Leaning Towards Scientific Process
In production you don’t have
• repeatability
• control groups
• external verification
Tuesday, October 2, 12
18. Leaning Towards Scientific Process
In production you don’t have
• repeatability
• control groups
• external verification
... or do you?
Tuesday, October 2, 12
19. What’s monitoring got to do with it?
Monitoring is all about the
passive observation of
telemetry data.
Tuesday, October 2, 12
20. Monitoring Telemetry
cannot pinpoint problems
can provides evidence of
the existence of a problem
Tuesday, October 2, 12
21. Monitoring
Gives you evidence that
there is a problem
Tuesday, October 2, 12
22. Monitoring
Gives you evidence that
you have fixed a problem
(or at least the symptoms)
Tuesday, October 2, 12
23. Monitoring Tactically
If it could be of interest,
measure it and
expose the measurement
Tuesday, October 2, 12
31. Repeatability is a Pipe Dream
You production problem is a
(hopefully pathological)
outcome of circumstance.
A circumstance which often
cannot be repeated.
Tuesday, October 2, 12
32. Control Groups
Control groups can
compensate for the
inability to
precisely repeat an experiment.
Tuesday, October 2, 12
33. Control Groups
Most architectures have redundancy.
Tuesday, October 2, 12
34. Control Groups
With the right design,
you can turn that redundancy
into a debugging environment.
[1] http://omniti.com/surge/2012/sessions/xtreme-deployment
Tuesday, October 2, 12
35. Control Groups: Simple Example
I have 10 web servers
I fix 1
I verify 1 is fixed
I verify 9 are still broken
Tuesday, October 2, 12
36. Control Groups: Seems Easy
Web servers tend to be:
• homogeneous
• share-(nothing|little)
• independent
Tuesday, October 2, 12
37. Control Groups: Not So Easy
Most other services aren’t so
homogeneous and equal:
databases, batch processes (think
billings), orchestration middleware,
message queues
Tuesday, October 2, 12
38. Observability
Some might claim that
seeing telemetry data is
observation...
It is doubly indirect at best.
Tuesday, October 2, 12
39. Observability
I want to
directly see
the
errant behaviour
Tuesday, October 2, 12
40. Observability is forgiving
In complex, multi-component
architectures, errors can be
observed as errant behaviour in
many junction points.
Tuesday, October 2, 12
42. Observing the network
Looking at just the
arrival of new connections
tcpdump -nnq -tttt -s384
'tcp port 80 and (tcp[13] & (2|16) == 2)'
Tuesday, October 2, 12
43. Observing the network
Looking at just the data
arrival and departure times
tcpdump -nnq -tt
-s 384 'tcp port 80 and (((ip[2:2] - ((ip[0]&0xf)*4)) - ((tcp[12]&0xf0)/4)) != 0)'
snoop -rq -ta
-s 384 'tcp port 80 and (((ip[2:2] - ((ip[0]&0xf)*4)) - ((tcp[12]&0xf0)/4)) != 0)'
Tuesday, October 2, 12
44. Observing the network
Finding the difference between
a client’s question and
a server’s answer
(tcpdump | awk filter).
{
gsub(".[0-9]+(: | >)"," & ");
gsub("[:=]"," ");
EP=sprintf("%s%s", ($4==".80")?$6:$3, ($4==".80")?$7:$4);
if(S[EP] == "C" && $4 == ".80") { printf("%f %sn", $1 - L[EP], EP); }
S[EP]= ($4==".80")?"S":"C";
L[EP]= $1;
}
Tuesday, October 2, 12