The recording in https://eviltester.com/talks has:
- longer practice session recording
- live recording - local recording better quality
- 8 bonus recordings with an extra hour of material
- will automation take over
- impact of buzzwords
- how to cope with trends
- contextual problem solving
- information about the references
- exercises
- behind the scenes look at how the talk was prepared and tools used
- transcripts
- subtitles
My aim here is to tell you that I learned to work with Agility rather than work with the Agile Rituals and Definitions. And I learned to trust that working with Agility trumps Rituals and Definitions the hard way. Because sticking to rituals and definitions led to rigidity, rather than agility.
And then "What does testing look like when you adopt that mindset?"
In this presentation you will short cut your learning on the topic of Agility, so you understand "What does testing look like when you adopt an Agility mindset?". Applying this mind set naturally leads to incorporating exploratory testing, technical testing, automated execution, end to end testing and risk. Adopting this mindset allows you to fit into any Agile Software Development project and create a customized testing approach that works.
Keynote at the internal Rabobank Testing Conference on Feb 15th 2018 in Utrecht.
https://www.compendiumdev.co.uk/page/rabobank201802
Your Automated Execution Does Not Have to be FlakyAlan Richardson
This webinar is for anybody who has accepted 'flaky' test automation. Alan believes that to describe and accept your test execution as flaky is merely an excuse. In this webinar he will explore the myths of flakiness, so that you never use those excuses again!
Categories of common problems with suggested solutions.
For more information visit http://eviltester.com/flaky
Test Automation Day 2015 Keynote Alan Richardson - Practical Lessons Learned ...Alan Richardson
Practical Lessons Learned Automating in Testing.
Let us forget theory for a moment, and concentrate on the practice of automation. Alan will describe lessons learned from both success and failure;
as a tester, an automator, and a manager. But more importantly, you will discover how to apply these lessons and improve your automation.
Learn how to stay focused, how to experiment and still add value, how to manage even if you cannot code, and more...
Test Bash Netherlands Alan Richardson "How to misuse 'Automation' for testing...Alan Richardson
We often hear about how ‘test automation’ can go wrong, which is all fine and dandy for the pessimists in the audience, and balancing feel good positive case studies exist for the optimists. But what about the anarchists? What about the rule breakers? What about the pragmatists? In this talk Alan will explain how to ‘misuse’ the ‘automation’ tools you’ve heard so much about, because you need to get things done. You’ve no doubt heard that ‘Cucumber is not a test tool’, and you’ve no doubt noticed that people use Cucumber during their testing. It’s misuse cases like this that we will celebrate, and as a bonus, you’ll learn what Cucumber ‘really' is. We’ll look at other tools; to find out their true nature and how you can turn it to your personal advantage. To further groom you for success, we’ll explain the mental models which give you guilt free flexibility in your approach. If you’ve ever wanted an ‘expert’ to quote to give you permission to use the tools how you want, this is the talk for you.
I'm going to be talking about finding the 'essence' of the tool, rather than what everyone 'says' about it, and that can lead to a radical overhaul in your beliefs and usage of the specific tool.
I blogged about my slide creation process for the conference, and there are some sneak peaks of some possible content in there are well.
http://blog.eviltester.com/2016/10/a-case-study-in-creating-conference.html
Joy of Coding Conference 2019 slides - Alan RichardsonAlan Richardson
Adventures in Testing, Programming, Teaching, Automating and Marketing
When you already know how to code, it's easy to forget how hard some of that learning was... until you have to teach people. And if all you've ever built are applications, you don't know really know the nuances of writing code to automate them. And if you've written the code but never had to market the applications then you've not really experienced the full joy of coding.
In this presentation Alan will revisit many of his past projects to identify lessons learned. Lessons from: writing commercial and open source tools, multi-user adventure games, REST APIs, test automation, automating applications to make them do things they are not supposed to do, and coding for technical marketing.
Some lessons we will learn:
* The 'install' is the hardest part
* Writing frameworks is too much fun and should be banned
* Applications are just "code calling other libraries"
* Writing a Text Adventure s the most fun and educational thing you'll ever code
* The Dangers of knowing how to code
We will also learn the dangers of knowing how to code and discover how our coding skills can give us an edge, in business and online live in general, if we choose to harness our skills to improve our daily experiences.
Add More Security To Your Testing and Automating - Saucecon 2021Alan Richardson
Presented at SauceCon 2021, April.
More details: https://www.eviltester.com/conference/saucecon2021_conference/
Security Testing is a highly technical set of skills, covering a wide domain of knowledge that can take a long time to learn and gain proficiency. We already have enough to learn with Software Testing and even more when we add in Automating. So are there any simple ways to increase the scope of what we already do, that provide more insight into the security of our application? Answer: Yes. And in this talk we will cover practical steps, dos and don’ts to add some Security focus fast, without spending years learning how to Hack applications.
My aim here is to tell you that I learned to work with Agility rather than work with the Agile Rituals and Definitions. And I learned to trust that working with Agility trumps Rituals and Definitions the hard way. Because sticking to rituals and definitions led to rigidity, rather than agility.
And then "What does testing look like when you adopt that mindset?"
In this presentation you will short cut your learning on the topic of Agility, so you understand "What does testing look like when you adopt an Agility mindset?". Applying this mind set naturally leads to incorporating exploratory testing, technical testing, automated execution, end to end testing and risk. Adopting this mindset allows you to fit into any Agile Software Development project and create a customized testing approach that works.
Keynote at the internal Rabobank Testing Conference on Feb 15th 2018 in Utrecht.
https://www.compendiumdev.co.uk/page/rabobank201802
Your Automated Execution Does Not Have to be FlakyAlan Richardson
This webinar is for anybody who has accepted 'flaky' test automation. Alan believes that to describe and accept your test execution as flaky is merely an excuse. In this webinar he will explore the myths of flakiness, so that you never use those excuses again!
Categories of common problems with suggested solutions.
For more information visit http://eviltester.com/flaky
Test Automation Day 2015 Keynote Alan Richardson - Practical Lessons Learned ...Alan Richardson
Practical Lessons Learned Automating in Testing.
Let us forget theory for a moment, and concentrate on the practice of automation. Alan will describe lessons learned from both success and failure;
as a tester, an automator, and a manager. But more importantly, you will discover how to apply these lessons and improve your automation.
Learn how to stay focused, how to experiment and still add value, how to manage even if you cannot code, and more...
Test Bash Netherlands Alan Richardson "How to misuse 'Automation' for testing...Alan Richardson
We often hear about how ‘test automation’ can go wrong, which is all fine and dandy for the pessimists in the audience, and balancing feel good positive case studies exist for the optimists. But what about the anarchists? What about the rule breakers? What about the pragmatists? In this talk Alan will explain how to ‘misuse’ the ‘automation’ tools you’ve heard so much about, because you need to get things done. You’ve no doubt heard that ‘Cucumber is not a test tool’, and you’ve no doubt noticed that people use Cucumber during their testing. It’s misuse cases like this that we will celebrate, and as a bonus, you’ll learn what Cucumber ‘really' is. We’ll look at other tools; to find out their true nature and how you can turn it to your personal advantage. To further groom you for success, we’ll explain the mental models which give you guilt free flexibility in your approach. If you’ve ever wanted an ‘expert’ to quote to give you permission to use the tools how you want, this is the talk for you.
I'm going to be talking about finding the 'essence' of the tool, rather than what everyone 'says' about it, and that can lead to a radical overhaul in your beliefs and usage of the specific tool.
I blogged about my slide creation process for the conference, and there are some sneak peaks of some possible content in there are well.
http://blog.eviltester.com/2016/10/a-case-study-in-creating-conference.html
Joy of Coding Conference 2019 slides - Alan RichardsonAlan Richardson
Adventures in Testing, Programming, Teaching, Automating and Marketing
When you already know how to code, it's easy to forget how hard some of that learning was... until you have to teach people. And if all you've ever built are applications, you don't know really know the nuances of writing code to automate them. And if you've written the code but never had to market the applications then you've not really experienced the full joy of coding.
In this presentation Alan will revisit many of his past projects to identify lessons learned. Lessons from: writing commercial and open source tools, multi-user adventure games, REST APIs, test automation, automating applications to make them do things they are not supposed to do, and coding for technical marketing.
Some lessons we will learn:
* The 'install' is the hardest part
* Writing frameworks is too much fun and should be banned
* Applications are just "code calling other libraries"
* Writing a Text Adventure s the most fun and educational thing you'll ever code
* The Dangers of knowing how to code
We will also learn the dangers of knowing how to code and discover how our coding skills can give us an edge, in business and online live in general, if we choose to harness our skills to improve our daily experiences.
Add More Security To Your Testing and Automating - Saucecon 2021Alan Richardson
Presented at SauceCon 2021, April.
More details: https://www.eviltester.com/conference/saucecon2021_conference/
Security Testing is a highly technical set of skills, covering a wide domain of knowledge that can take a long time to learn and gain proficiency. We already have enough to learn with Software Testing and even more when we add in Automating. So are there any simple ways to increase the scope of what we already do, that provide more insight into the security of our application? Answer: Yes. And in this talk we will cover practical steps, dos and don’ts to add some Security focus fast, without spending years learning how to Hack applications.
# Automating Pragmatically
Testival Meetup 20190604
## Alan Richardson
- EvilTester.com
- @EvilTester
- compendiumdev.co.uk
- digitalonlinetactics.com
---
~~~~~~~~
Title: Automating Pragmatically
The online discussions of automating can leave me confused.
- Should you automate through the GUI?
- Should GUI automating be banned?
- Do all testers need to code? Is automating part of testing
or not?
- Do we need to automate to get a job?
In this short session Alan will discuss automating
from a pragmatic and contextual position and
share how he thinks about automating.
~~~~~~~~
Risk Mitigation Using Exploratory and Technical Testing - QASymphony Webinar ...Alan Richardson
A Webinar on Risk Analysis and Management, Exploratory Testing, and Technical Testing.
I want to get across the model that I have for risks, which is that risks are “beliefs” and a result of our beliefs. We believe some things will go wrong more than others. And because our beliefs are limited but the range of risks is not, we need to somehow go beyond our beliefs and look at tools and processes for doing that.
Also we know that risk is important for testing. What I want to do in this talk is present risk as the underpinning and driving force behind everything we do in testing.
You can use risk to justify the stuff that you do as a tester. And you can use risk to derive your test scope as well as your test process.
Software Testing Terms Defined. Answering the FAQ "What is Regression Testing?"
- What is Regression Testing?
- How to do Regression Testing?
- Why do we do Regression Testing?
- How to re-think Regression Testing in terms of Risk?
Secrets and Mysteries of Automated Execution Keynote slidesAlan Richardson
Test Automation, Programming Automation, Automated Execution. This presentations contains some high level models, abstractions and approaches for effective, non-flakey and maintainable automation.
https://www.eviltester.com
What is Shift Left Testing? Do you need to use that term to improve your Software Testing and Development process? I don't think so.
- why I don't use the term Shift Left
- Explanation of what Shift Left means when people use it
- Explanation of what Shift Left might mean when people hear it
- How to Shift Left incorrectly
- How to improve your test process without using the phrase Shift Left.
Hire me for consultancy and buy my online books and training at:
- https://compendiumdev.co.uk
- http://eviltester.com
- http://seleniumsimplified.com
- http://javafortesters.com
DevFest 14th Dec 2019 Bishkek
- Alan Richardson
https://www.eviltester.com/conference/devfestbishkek2019_conference
- EvilTester.com
- @EvilTester
- CompendiumDev.co.uk
---
Have you ever wondered how other people test applications? Not in theory, but in practice? What thought processes are used? How did they model the application? What tools were used? How did they track the testing? That's what this talk is all about. This talk will be based on a short Case Study of testing an open source web application. Why open source? Because then there is no commercial confidentiality about the process, tools or thought processes.
---
Alan will explain his thought processes, coverage, approaches, tools used, risks identified and results found. And generalise from this into reusable models and principles that can be applied to your testing. This covers the What?, and the Why? of practical exploratory web testing.
Technical and Testing Challenges: Using the "Protect The Square" GameAlan Richardson
How good are your Technical Testing in the Browser and JavaScript skills? Put them to the test with the "Protect The Square" game.
https://www.compendiumdev.co.uk/games/buggygames/protect_the_square/protect_the_square.html
Much of the automating we do to support testing involves detecting change. Once our tests pass, they fail when the system changes and the automated execution alerts us to the change. There are other ways that automating can help us.
Automation vs. intelligence - "follow me if you want to live"Viktor Slavchev
Have you ever heard the story that your job is automatable, that all the human testers will be replaced by machines or automated tests and you will lose your job? Or even worse, that machines and artificial intelligence will take over our craft and our life and we will be totally useless. Do you buy these? Are you afraid?
“Come with me, if you want to live” – this was the famous line that many members of the Human resistance in the Terminator franchise used, when offering their help in the war against Skynet.
So, come with me (and John Connor), and join the testing resistance to fight on the side of intellect against the evil machine army. I am willing to challenge the I part in AI on contest by focusing on few key topics:
Can we translate testing into machine language? Polymorphic and mimeomorphic actions – what are these?
Do we really know what are the benefits of human testing? What are human testers irreplaceable for?
Do we really have empirical evidence that computers are capable of doing professional testing? Do we have evidence of “intelligence” at all?
Last year at RTC ‘17 I was asked – “Is AI the answer to all test automation problems?”. My answer is “No, it’s not!”. And this talk is my explanation why.
Re-thinking Test Automation and Test Process Modelling (in pictures)Alan Richardson
- Why do we talk about Test Automation the way we do?
- Why do we talk about 100% Test Automation?
- How do we model automation as part of our Test Process?
- How does Testing provide information?
- Why was a Waterfall Test Process Different from an Agile Process?
- Why, in reality, both processes are fundamentally the same.
- How we modelled "Test Automation" incorrectly, and an alternative way to model it.
Read the associated blog post at http://blog.eviltester.com/2017/09/rethinking-test-process-automation-modelling.html
Slides for Automation Guild 2016 Conference
If you want to automate, you learn to code, and you learn to code well.
“Automate” doesn’t mean “Automate Testing” it means “Automate part of your test process”.
You need to learn to code to do that with the most options open to you.
We’ll look at some ‘we do this alot’ and ‘we want to automate’ activities which we can use tools for. But we’ll also see that we are limited by the tools.
When we code, we can do a lot with minimum code, and gain a lot more flexibility.
Then we’ll cover how to think about learning to code.
solve a problem quickly (automate tactically)
solve a problem for the long term (automate strategically)
To work strategically we need to learn:
to code well,
understand refactoring,
libraries vs frameworks,
abstractions,
etc.
This talk isn’t just for beginners, we’ll cover stuff that should make it useful for the experts in the audience.
We’ll cover a lot in 45 mins, with code examples and tool examples, and I’ll make it all pretty practical.
For more details visit:
https://www.compendiumdev.co.uk/page/tag2017
The act of testing is a scientific exploration of a system done in three phases - RECONNAISSANCE to understand and plan, SEARCH to look for issues, REST&RECOVER to analyse and course correct. To enable the various activities in each phase to be done quickly and effectively, is where the SEVEN Thinking Tools outlined in this presentation. How to apply these tools in a session-based approach is also briefly outlined.
This article version of this SlideShare is available at http://bit.ly/7ThinkTools.
How to Improve Your Technical Test Ability - AADays 2015 KeynoteAlan Richardson
We often work on improving the testability of an application to better support our testing. And what if, in addition to this, we actively improved our "Test Ability"? Because then we can take advantage of the new and existing application features during our testing. Alan will describe the steps he has taken to improve his Test Ability. The main examples will be drawn from his experience of testing web and HTTP based applications. Alan will explain how you can use the inbuilt browser features to help you, and describe add-ons you can use. Also, how you can chain external tools like sniffers and proxies, and why you would want to. Because, and this is more important than the individual tool examples, Alan will describe how he models an application to identify gaps in his knowledge and tooling, and then improves his Technical Test Ability by filling those gaps.
Slides for Agile Testers Conference 2018
Technology Based Testing by Alan Richardson
What do you learn if you want to test 'beyond the acceptance criteria'? Technical risk based testing can help. In this case I'm going to use the phrase Technical Testing to cover: "identify technology based risks to drive testing". This thought process can help us make informed decisions about the scope of exploratory testing we will carry out. It also helps focus your studies on the technical knowledge appropriate for the project you are testing.
## Blurb
This requires:
- understanding of the technology
- risk identification
- tools applicable to the technology
This presentation will use a simple example to demonstrate that:
- Even simple technology can pose risk
- Combining simple technology can increase risk
- Understanding technology allows us to evaluate risk
* http://www.eviltester.com
* http://www.compendiumdev.co.uk
* https://twitter.com/eviltester
Black Ops Testing Workshop from Agile Testing Days 2014Alan Richardson
At Agile Testing Days 2014. Steve Green, Tony Bruce and Alan Richardson hosted a double track Black Ops Testing workshop, where Redmine was the target application.
Find out more about the Black Ops Testing Team: http://blackopstesting.com/page/about.html
What does Technical Testing mean? For Alan, it means going beyond requirements and using Technical Information about the implementation and an understanding of the technologies used in the building of the system to add to the risk profile and use to help derive test approaches. Using Web Testing as an example we explain how approaching testing from a technical perspective changes how you view the system and how you test. Also explained, how a technical understanding leads to a different use of tooling an automation. This webinar presented 1st April 2015 to Tabara De Testare
Harry Collins - Testing Machines as Social Prostheses - EuroSTAR 2013TEST Huddle
EuroSTAR Software Testing Conference 2013 presentation on Testing Machines as Social Prostheses by Harry Collins.
See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
# Automating Pragmatically
Testival Meetup 20190604
## Alan Richardson
- EvilTester.com
- @EvilTester
- compendiumdev.co.uk
- digitalonlinetactics.com
---
~~~~~~~~
Title: Automating Pragmatically
The online discussions of automating can leave me confused.
- Should you automate through the GUI?
- Should GUI automating be banned?
- Do all testers need to code? Is automating part of testing
or not?
- Do we need to automate to get a job?
In this short session Alan will discuss automating
from a pragmatic and contextual position and
share how he thinks about automating.
~~~~~~~~
Risk Mitigation Using Exploratory and Technical Testing - QASymphony Webinar ...Alan Richardson
A Webinar on Risk Analysis and Management, Exploratory Testing, and Technical Testing.
I want to get across the model that I have for risks, which is that risks are “beliefs” and a result of our beliefs. We believe some things will go wrong more than others. And because our beliefs are limited but the range of risks is not, we need to somehow go beyond our beliefs and look at tools and processes for doing that.
Also we know that risk is important for testing. What I want to do in this talk is present risk as the underpinning and driving force behind everything we do in testing.
You can use risk to justify the stuff that you do as a tester. And you can use risk to derive your test scope as well as your test process.
Software Testing Terms Defined. Answering the FAQ "What is Regression Testing?"
- What is Regression Testing?
- How to do Regression Testing?
- Why do we do Regression Testing?
- How to re-think Regression Testing in terms of Risk?
Secrets and Mysteries of Automated Execution Keynote slidesAlan Richardson
Test Automation, Programming Automation, Automated Execution. This presentations contains some high level models, abstractions and approaches for effective, non-flakey and maintainable automation.
https://www.eviltester.com
What is Shift Left Testing? Do you need to use that term to improve your Software Testing and Development process? I don't think so.
- why I don't use the term Shift Left
- Explanation of what Shift Left means when people use it
- Explanation of what Shift Left might mean when people hear it
- How to Shift Left incorrectly
- How to improve your test process without using the phrase Shift Left.
Hire me for consultancy and buy my online books and training at:
- https://compendiumdev.co.uk
- http://eviltester.com
- http://seleniumsimplified.com
- http://javafortesters.com
DevFest 14th Dec 2019 Bishkek
- Alan Richardson
https://www.eviltester.com/conference/devfestbishkek2019_conference
- EvilTester.com
- @EvilTester
- CompendiumDev.co.uk
---
Have you ever wondered how other people test applications? Not in theory, but in practice? What thought processes are used? How did they model the application? What tools were used? How did they track the testing? That's what this talk is all about. This talk will be based on a short Case Study of testing an open source web application. Why open source? Because then there is no commercial confidentiality about the process, tools or thought processes.
---
Alan will explain his thought processes, coverage, approaches, tools used, risks identified and results found. And generalise from this into reusable models and principles that can be applied to your testing. This covers the What?, and the Why? of practical exploratory web testing.
Technical and Testing Challenges: Using the "Protect The Square" GameAlan Richardson
How good are your Technical Testing in the Browser and JavaScript skills? Put them to the test with the "Protect The Square" game.
https://www.compendiumdev.co.uk/games/buggygames/protect_the_square/protect_the_square.html
Much of the automating we do to support testing involves detecting change. Once our tests pass, they fail when the system changes and the automated execution alerts us to the change. There are other ways that automating can help us.
Automation vs. intelligence - "follow me if you want to live"Viktor Slavchev
Have you ever heard the story that your job is automatable, that all the human testers will be replaced by machines or automated tests and you will lose your job? Or even worse, that machines and artificial intelligence will take over our craft and our life and we will be totally useless. Do you buy these? Are you afraid?
“Come with me, if you want to live” – this was the famous line that many members of the Human resistance in the Terminator franchise used, when offering their help in the war against Skynet.
So, come with me (and John Connor), and join the testing resistance to fight on the side of intellect against the evil machine army. I am willing to challenge the I part in AI on contest by focusing on few key topics:
Can we translate testing into machine language? Polymorphic and mimeomorphic actions – what are these?
Do we really know what are the benefits of human testing? What are human testers irreplaceable for?
Do we really have empirical evidence that computers are capable of doing professional testing? Do we have evidence of “intelligence” at all?
Last year at RTC ‘17 I was asked – “Is AI the answer to all test automation problems?”. My answer is “No, it’s not!”. And this talk is my explanation why.
Re-thinking Test Automation and Test Process Modelling (in pictures)Alan Richardson
- Why do we talk about Test Automation the way we do?
- Why do we talk about 100% Test Automation?
- How do we model automation as part of our Test Process?
- How does Testing provide information?
- Why was a Waterfall Test Process Different from an Agile Process?
- Why, in reality, both processes are fundamentally the same.
- How we modelled "Test Automation" incorrectly, and an alternative way to model it.
Read the associated blog post at http://blog.eviltester.com/2017/09/rethinking-test-process-automation-modelling.html
Slides for Automation Guild 2016 Conference
If you want to automate, you learn to code, and you learn to code well.
“Automate” doesn’t mean “Automate Testing” it means “Automate part of your test process”.
You need to learn to code to do that with the most options open to you.
We’ll look at some ‘we do this alot’ and ‘we want to automate’ activities which we can use tools for. But we’ll also see that we are limited by the tools.
When we code, we can do a lot with minimum code, and gain a lot more flexibility.
Then we’ll cover how to think about learning to code.
solve a problem quickly (automate tactically)
solve a problem for the long term (automate strategically)
To work strategically we need to learn:
to code well,
understand refactoring,
libraries vs frameworks,
abstractions,
etc.
This talk isn’t just for beginners, we’ll cover stuff that should make it useful for the experts in the audience.
We’ll cover a lot in 45 mins, with code examples and tool examples, and I’ll make it all pretty practical.
For more details visit:
https://www.compendiumdev.co.uk/page/tag2017
The act of testing is a scientific exploration of a system done in three phases - RECONNAISSANCE to understand and plan, SEARCH to look for issues, REST&RECOVER to analyse and course correct. To enable the various activities in each phase to be done quickly and effectively, is where the SEVEN Thinking Tools outlined in this presentation. How to apply these tools in a session-based approach is also briefly outlined.
This article version of this SlideShare is available at http://bit.ly/7ThinkTools.
How to Improve Your Technical Test Ability - AADays 2015 KeynoteAlan Richardson
We often work on improving the testability of an application to better support our testing. And what if, in addition to this, we actively improved our "Test Ability"? Because then we can take advantage of the new and existing application features during our testing. Alan will describe the steps he has taken to improve his Test Ability. The main examples will be drawn from his experience of testing web and HTTP based applications. Alan will explain how you can use the inbuilt browser features to help you, and describe add-ons you can use. Also, how you can chain external tools like sniffers and proxies, and why you would want to. Because, and this is more important than the individual tool examples, Alan will describe how he models an application to identify gaps in his knowledge and tooling, and then improves his Technical Test Ability by filling those gaps.
Slides for Agile Testers Conference 2018
Technology Based Testing by Alan Richardson
What do you learn if you want to test 'beyond the acceptance criteria'? Technical risk based testing can help. In this case I'm going to use the phrase Technical Testing to cover: "identify technology based risks to drive testing". This thought process can help us make informed decisions about the scope of exploratory testing we will carry out. It also helps focus your studies on the technical knowledge appropriate for the project you are testing.
## Blurb
This requires:
- understanding of the technology
- risk identification
- tools applicable to the technology
This presentation will use a simple example to demonstrate that:
- Even simple technology can pose risk
- Combining simple technology can increase risk
- Understanding technology allows us to evaluate risk
* http://www.eviltester.com
* http://www.compendiumdev.co.uk
* https://twitter.com/eviltester
Black Ops Testing Workshop from Agile Testing Days 2014Alan Richardson
At Agile Testing Days 2014. Steve Green, Tony Bruce and Alan Richardson hosted a double track Black Ops Testing workshop, where Redmine was the target application.
Find out more about the Black Ops Testing Team: http://blackopstesting.com/page/about.html
What does Technical Testing mean? For Alan, it means going beyond requirements and using Technical Information about the implementation and an understanding of the technologies used in the building of the system to add to the risk profile and use to help derive test approaches. Using Web Testing as an example we explain how approaching testing from a technical perspective changes how you view the system and how you test. Also explained, how a technical understanding leads to a different use of tooling an automation. This webinar presented 1st April 2015 to Tabara De Testare
Harry Collins - Testing Machines as Social Prostheses - EuroSTAR 2013TEST Huddle
EuroSTAR Software Testing Conference 2013 presentation on Testing Machines as Social Prostheses by Harry Collins.
See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Technology should be a beneficial force in our lives, taking the world in exciting new directions and making us better humans. To ensure this, we need to facilitate a conversation between data technology and the human experience. Keeping social responsibility and ethical behavior in mind when designing AI systems enables us to put the right systems in place to contribute to the society we want, fostering higher levels of cognitive and emotional skills.
Jivan Virdee and Hollie Lubbock explore how to address fairness, accountability, and the long-term effects on our society when designing with data, focusing on four key areas for consideration in this space:
— Responsibility and accountability for machine learning systems
— Fair and transparent data science
— Trust and human-machine collaboration
—Automation and changes in the way we work
Along the way, they cover key issues in creating ethical AI systems and detail how we might go about tackling them and outline further questions that will need to be addressed going forward.
Big thanks to @fjord and @accenturedock for their help and support
Talk by:
https://www.linkedin.com/in/hollie-lubbock-703b77b/
https://www.linkedin.com/in/jivanvirdee/
Data Reliability Challenges with Spark by Henning Kropp (Spark & Hadoop User ...Comsysto Reply GmbH
Current Data Lake projects are facing enormous issues over generating business value. According to Gartner, more than 65% of the projects are failing. The most common reasons for projects to fail are centered around data reliability and performance issues resulting in delays, complexity, and errors.
Delta is the next-generation analytics engine as part of the Databricks Runtime tackling some of the most challenging issues with Spark today. Delta provides ACID, Data Versioning, and Schema Enforcement on top of Apache Parquet. In this talk, we will discuss the current challenges and give a live demo of Delta.
The law and ethics of data-driven artificial intelligencePyData
By Aileen Nielsen
PyData New York City 2017
This talk is a completely non-technical discussion of how the law currently regulates artificial intelligence (if it does at all) and what is likely to change in the near future. The talk is geared towards technically-minded practitioners of data driven intelligence with the aim of increasing discussion of the social and ethical impact of data driven AI and how to code responsibly.
Ethical AI summit Dec 2023 notes from HB keynoteHelen Beetham
Somewhat extended and tidied up text of HB keynote at the ALT winter summit on AI and Ethics, December 2023. Slides draft quality for navigation only - a better quality set of slides is also available.
Ethics Ex Machina – Designing the Future With a ConscienceBrandon Ward
Volkswagen designed an unethical solution to a global and customer problem. It cost them $30 BILLION literal dollars, and who knows how much more in PR.
Cambridge Analytica is now a shuttered entity after unethical, if not illegal usage of our personal data.
Facebook's CEO had to testify before Congress, and Microsoft's AI bot "Tay" became a slanderous racist in less than 16 hours.
Forbes found that "Millennials prefer to do business with corporations and brands with pro-social messages, sustainable manufacturing methods and ethical business standards."
In this talk we'll dive into how we can begin to think about our work with the lens of what Alan Cooper calls being a good ancestor. How we can inject ethics and standards into our daily efforts to ensure we're designing the future we want, and our children deserve.
Unlocking the Potential: Data as a Medium for Design & JusticeJess Freaner
As a designer and data scientist, I work with data in service of meeting people’s needs. This data is inherently subjective, which is what makes it both an excellent medium for design and vulnerable to misuse. I’ll share what it means to design with data and how data science can contribute to and augment the design process. Once we see what’s exciting and newly possible, we’ll delve into why now, more than ever, human-centered design matters as we discuss ethics and the impact of AI designs on individuals, communities, and societies.
Talk given by Jess Freaner (IDEO) at UX Strategy Meetup in Chicago - November 2019
Automating Tactically vs Strategically SauceCon 2020Alan Richardson
One of the biggest concepts that has made a difference to my programming and automating in recent years is the concept of “Tactical vs. Strategic.” Automating tactically might be for a specific purpose, possibly small, possibly a bit rough around the edges, not necessarily completely robust for everyone, etc. And Strategic automation is more critical to long-term aims, maintained and maintainable, etc.
In this talk, Alan Richardson will provide examples of automating both Strategically and Tactically for activities as diverse as supporting testing, marketing and general life. We will also consider how and when to move from automating tactically to strategically, and how the concept has helped me change my programming style and how to write better code.
Programming katas for Software Testers - CounterStringsAlan Richardson
What would be suitable Code Katas for people wanting to learn how to code to support their testing?
CounterStrings
- `*3*5*7*9*12*15*`
A CounterString is a string like this `*3*5*7*9*12*15*` where the `*` represent the position in the string of the number immediately proceeding it. This is a 15 character CounterString.
These are useful because if you paste them into a field, and are truncated then it is easy to see what they were truncated to, it is as James Bach describes it, self documenting test data.
https://www.eviltester.com/blog/eviltester/2019-02-27-programming-katas-for-testers/
Have you ever wished that you had a worked example of how to test a REST API?
Not just automate the API, but how to interact with it with command line tools, and GUI tools to support your manual interactive testing. And then take your testing forward into automating the API?
That's what this book provides.
Read the 74 page sample and find out more information on the book page.
https://www.compendiumdev.co.uk/page/tracksrestapibook
The full book has over 200 pages of actual hands on case study information that can improve your testing and automating of REST API based applications.
TDD - Test Driven Development - Java JUnit FizzBuzzAlan Richardson
A short example Test Driven Development session where I code FizzBuzz.
FizzBuzz is often used as a programming interview question and as a Kata for practicing your coding.
The GitHub code repository with the Java code for this exercise is available at:
https://github.com/eviltester/fizzbuzz
Read the blog post for the video:
http://blog.eviltester.com/2018/03/tdd-test-driven-development-java-junit.html
What is Testability vs Automatability? How to improve your Software Testing.Alan Richardson
Testability is different from Automatability.
- Testability - does the application have features that make it easier for a human to test?
- Automatizability (Automatability) - does the application have features that make it easier to control and interrogate by another application.
You will learn:
- What is Testability?
- What is automatability?
- What is automatizability?
- Adding testability features can introduce risk.
- Features that aid automated execution, can overlap with features that aid testing, but they are not the same.
A Common Sense Guide to Agile Development and Testing that might just change your Agile approach forever.
Answering the 9 most common questions asked about Agile Testing:
- What is Agile Testing?
- Do we still need testers in Agile?
- What is an Agile Tester?
- What does a Software Tester Actually Do?
- Should we automate our testing?
- What tools should we use for our Agile Testing?
- How Much Should we Automate?
- How can we automate and still finish the sprint?
- How can we finish all our testing in the sprint?
A high quality download of the 9 points as a free "Print out and Keep" Poster is available at http://eviltester.com/agile
The Evil Tester Show - Episode 001 Halloween 2017Alan Richardson
The Evil Tester Show - Episode 001 Halloween 2017
## Halloween Special 2017
## Alan Richardson
- Houdini
- Charles Fort
- Ghost Hunting
- Unconventional Influences
http://eviltester.com/show/001-halloween-2017/
---
# _TLDR; The world needs a new Testing Podcast, so I created one_
---
# We are in the Uncertainty Business.
We find and investigate anomalous Phenomena
## Anomalous - "deviating from what is standard, normal, or expected."
We are part of a long tradition of Anomalous Phenomena seekers.
---
# The Podcast
- [Audio]
https://eviltester.podbean.com/e/the-evil-tester-show-episode-001-halloween-special-2017/
- [Video]
https://youtu.be/TLMtOM0FXRA
- [Show Notes]
http://eviltester.com/show/001-halloween-2017/
Simple ways to add and work with a `.jar` file in your local maven setupAlan Richardson
TL;DR Hack - add as a library in IntelliJ project. Tactic - add as system scope in maven. Tactic/Strategic - install locally to .m2. Strategic - use a repository management tool, publish to maven central
Sometimes you want to work with a jar file that isn't hosted in maven central.
It might be a 3rd party jar, it might be one that you have written.
Regardless.
You have a lot of options for this. The approaches that I have used:
- add .jar files as an IntelliJ project dependency
- install it locally to your .m2 repository
- add it to your project as a system scoped file
- use a repository management tool like Nexus or Archiva
- publish the dependency to maven central
Automating Strategically or Tactically when TestingAlan Richardson
"Test Automation" can be viewed as strategic or tactical.
This presentation describes reasons for making this distinction and how you know if you are working strategically or tactically when you automate as part of your test approach.
Learning in Public - A How to Speak in Public WorkshopAlan Richardson
Glossophobia, the fear of public speaking, usually ranks pretty high on surveys of 'what people fear'. And for good reason. We've all attended conferences where the keynote speakers were seriously injured after being hit by a torrent of rolled up feedback forms, or speakers were left bleeding from a rain of plastic name badges thrown Shuriken-like by the Ninja trained attendees.
You can learn to avoid these outcomes, and when you do, you gain a skill that will win you recognition, improve your job prospects and allow you to travel the world talking to fellow testers.
In this workshop Alan will provide hints and tips for improving your public speaking. Sharing, from experience, what works for him, and discuss some conventional wisdom on public speaking. Alan will also share a few secrets, and unconventional exercises that he uses to prepare.
Public speaking is a skill we have to learn in public, but it is a skill, it is learn-able, and it is a skill that you can learn.
Read more in the supporting blog post:
http://blog.eviltester.com/2017/09/overcome-imposter-syndrome-public-speaking.html
How to Practise to Remove Fear of Public SpeakingAlan Richardson
Tips on how to overcome fear of public speaking:
- the 'fear' is a learned response, it is not innate
- recognise that it is not fear, it is excitement
- channel the excitement into energy to boost your talk
- practice with different styles of presentation
- record yourself practicing
- practice out loud, as well as in your head.
Speaking in public is a skill, that you can develop if you care enough about the message that you want to deliver. It is simply practice, and you can do that.
FAQ - why does my code throw a null pointer exception - common reason #1 Rede...Alan Richardson
A common reason for Null Pointer Exceptions in Java is a variable redeclaration instead of instantiation. Learn what that means, how to avoid it, and how to spot it, in this presentation.
Read the full blog post: http://testerhq.com/post/blogs/javafortesters/2017-08-29-faq-null-pointer-exception/
Visit my Java Web Site: http://javafortesters.com
---
# FAQ - why does my code throw a null pointer exception - common reason #1 Redeclaration
- Using `@BeforeClass` or `@Before` can setup data for use in tests
- Any 'variables' we instantiate need to be 'fields' rather than variables
- We want to instantiate them in the setup method rather than redeclare them
---
# Example of the Problem
I know I will use an `Adder` in my test so I create it as a field:
~~~~~~~~
public class WhyCodeThrowsNullPointerExceptionTest {
Adder adder;
~~~~~~~~
I don't want to re-instantiate it each time so I make an `@BeforeClass` method to instantiate it:
~~~~~~~~
@BeforeClass
public static void setupAdder(){
Adder adder = new Adder();
}
~~~~~~~~
**Warning: Error in the above code**
---
# Semantic Error
I just made a Semantic coding error. This won't be caught by a compiler, but it will cause my `@Test` to fail with a Null Pointer Exception.
In the setup method I really wanted to assign a value to the field, instead I created an new variable with the same name.
# In General
- Try to write one test at a time so that if you have a problem it is easier to identify where the problem is
- Try to write working isolated tests and then refactor to a more general solution when you need it - that way, you know it was working, so you just have to work backwards to find out what went wrong
- Try to use automated IDE refactoring rather than move code around manually
- Use the IDE syntax highlighting to help spot any issues
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
When stars align: studies in data quality, knowledge graphs, and machine lear...
The Future of Testing Webinar
1. Exploring How Software Testing Is Going
to Change
Alan Richardson
@EvilTester
EvilTester.com
compendiumdev.co.uk
digitalonlinetactics.com
@EvilTester 1
3. What we want to happen
reveals our biases
reveals our experience
@EvilTester 3
4. What we think will happen
reveals our model of the world
reveals our perception of the industry
@EvilTester 4
5. The Future
Is driven by individual vision
backed by individuals who believe
Is crafted by individuals taking action
Is allowed by individuals taking no action
Individuals are key.
@EvilTester 5
6. Warning: As a tester, risk of perceived
negativity ahead
@EvilTester 6
7. Machiavelli Says...
"he who would forecast what is about to happen should look to what
has been ; since all human events, whether present or to come, have
their exact counterpart in the past. And this, because these events are
brought about by men, whose passions and dispositions remaining in
all ages the same naturally give rise to the same effects"
Chapter XLIII of "Discourses on the first decade of Titus Livius"
@EvilTester 7
8. Predict More of the Same
"Predicting the future is much too easy, anyway. You look at the
people around you, the street you stand on, the visible air you
breathe, and predict more of the same..."
Ray Bradbury
@EvilTester 8
9. What I think ‑ based on the present
We will pursue the 'new', seeking the 'one' true tool
rather than find full potential in the 'old'
Market over‑saturated with 'new' and 'better' tools
no such thing as a bad tool, but hard to differentiate
Technology will be over‑hyped
seeds fear, and disappointment
Lack of confidence from Testing
@EvilTester 9
10. We Need to Have Confidence in Testing
Lack of confidence from Testing leads to
New Trends will continue to unnerve Testers
Testing will be defined by people who have not studied Testing
Attempting to replace Testers with Tools
AI, Automation
Testers will not be confident about their core
Testers will continue to reframe
@EvilTester 10
11. Individuals, Organisation, Industry
"I don’t try to predict the future. I try to prevent it."
Ray Bradbury
The future is not yet written.
@EvilTester 11
13. “A good science fiction story should be able to
predict not the automobile but the traffic jam."
– Frederik Pohl
@EvilTester 13
14. "he who looks carefully into the matter will find,
that in all human affairs, we cannot rid ourselves of
one inconvenience without running into another."
Niccolo Machiavelli
@EvilTester 14
16. Q: What process is used to detect the
manifestation of risk?
A: Testing
@EvilTester 16
17. What I Want
So tell me what you want, what you really, really want
I wanna, (ha) I wanna, (ha) I wanna, (ha) I wanna, (ha)
I wanna really, really, really wanna zigazig ah
Spice Girls
@EvilTester 17
18. "I want better"
"Predicting the future is much too easy, anyway. You look at the
people around you, the street you stand on, the visible air you
breathe, and predict more of the same.
To hell with more. I want better."
Ray Bradbury
@EvilTester 18
19. What I Want ‑ Better Understanding of
Testing
More effective communication and understanding of Testing
People experienced in Testing, describe "Testing"
@EvilTester 19
20. Go Meta ‑ the core of testing remains
untouched
Six: What do you want?
Two: Information.
Six: Whose side are you on?
Two: That would be telling. We want information...information...
information!!!
Six: You won't get it!
Two: By hook or by crook, we will.
The Prisoner, title sequence
@EvilTester 20
21. What I Want ‑ Better Team Interaction
Teams constructed based on skill mix and utilisation
rather than role based
Separation between Automating and Testing
@EvilTester 21
22. What I Want ‑ Better and Flexible Tools
Tools support more Open and Standard APIs
User expandable tools ‑ APIs and Scriptable
Fewer frameworks and more libraries
Less risk in tool adoption
migration paths and open standards
Eco systems of interoperable tools and libraries
@EvilTester 22
23. What I want ‑ supporting tools
Understanding that 'people' are at the center of testing
Tools which observe more
Tools which help me interrogate faster
Tools which help me manipulate more easily
@EvilTester 23
24. What I Want ‑ Effective Processes
Algorithmic and Model Based Approaches
so much we haven't built on
More tools suitable for more skill ranges
Easier for people to decide which tool to adopt
@EvilTester 24
25. What I Want ‑ AI Augmentation
Not AI that replaces algorithms
AI that monitors and flags weak signals
AI that understands its limits
points out 'things'
and brings in people
AI that works in combination with people
@EvilTester 25
26. What I Want ‑ Gaps in tooling filled
visualising and communicating our work
deterministic algorithmic model coverage
spin off side projects
test data
...
@EvilTester 26
27. What I want ‑ Optimism
Testers Taking Testing Forward
Bringing in Lessons from other disciplines
Testers Defining Testing
Testers Demonstrating Value
Testers Harnessing Technology
@EvilTester 27
28. Whichever Future we pick ‑ Model it
Whichever Future we pick: Speculative, Want, Think
What would prevent it happening?
What would prevent us being part of it?
What would be the impact?
How do we get there?
@EvilTester 28
29. How to Handle the Future Now
Identify end point
Identify how you fit
identify what skills and attitudes you need
Work backwards from there
to create a plan for evolving
@EvilTester 29
30. Build the Future
Do not just let the future happen to you
Work to be part of the future you want
@EvilTester 30
31. Work to create the future you want
Alan Richardson ‑ EvilTester.com
@EvilTester 31