EuroSTAR Software Testing Conference 2010 presentation on Weekend Testing, Skilled Software Testing Unleashed by Ajay Balamnrugadas. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Erkki Poyhonen - Software Testing - A Users GuideTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Software Testing - A Users Guide by Erkki Poyhonen. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Gustav Olsson - Agile - Common Sense with a New Name Tag revisedTEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Agile - Common Sense with a New Name Tag revised by Gustav Olsson. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Tafline Murnane - The Carrot or The Whip-What Motivates Testers? - EuroSTAR 2010TEST Huddle
EuroSTAR Software Testing Conference 2010 presentation on The Carrot or The Whip-What Motivates Testers? by Tafline Murnane. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Ken Johnston - Big Bugs That Got Away - EuroSTAR 2010TEST Huddle
EuroSTAR Software Testing Conference 2010 presentation on Big Bugs That Got Away by Ken Johnston . See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Michael Bolton - Heuristics: Solving Problems RapidlyTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Heuristics: Solving Problems Rapidly by Michael Bolton. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Herman- Pieter Nijhof - Where Do Old Testers Go?TEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Where Do Old Testers Go? by Herman- Pieter Nijhof. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Isabel Evans - A Statement for the Future TEST Huddle
This document outlines an agenda for a workshop aimed at drafting statements for a future software testing manifesto. The workshop will have delegates discuss and edit manifesto statements in small groups. They will choose topics and draft statements for two sessions. The output of the workshop will be conclusions, manifesto statements for a final conference presentation, and documentation of the workshop to be published online along with the statements.
Rikard Edgren - Testing is an Island - A Software Testing DystopiaTEST Huddle
This document summarizes trends in software testing that could diminish its effectiveness and enjoyment. It notes an increasing focus on verification over validation, precise measurement over subjective judgement, and short-term metrics over long-term quality. This narrowing scope risks making testers isolated and limiting their creativity, motivation and ability to consider the full context of a project. The document advocates a holistic and subjective approach that considers people and intangible factors, not just short-term quantifiable results. Subjectivity and considering the whole system, not just parts, are presented as useful for testing.
Erkki Poyhonen - Software Testing - A Users GuideTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Software Testing - A Users Guide by Erkki Poyhonen. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Gustav Olsson - Agile - Common Sense with a New Name Tag revisedTEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Agile - Common Sense with a New Name Tag revised by Gustav Olsson. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Tafline Murnane - The Carrot or The Whip-What Motivates Testers? - EuroSTAR 2010TEST Huddle
EuroSTAR Software Testing Conference 2010 presentation on The Carrot or The Whip-What Motivates Testers? by Tafline Murnane. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Ken Johnston - Big Bugs That Got Away - EuroSTAR 2010TEST Huddle
EuroSTAR Software Testing Conference 2010 presentation on Big Bugs That Got Away by Ken Johnston . See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Michael Bolton - Heuristics: Solving Problems RapidlyTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Heuristics: Solving Problems Rapidly by Michael Bolton. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Herman- Pieter Nijhof - Where Do Old Testers Go?TEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Where Do Old Testers Go? by Herman- Pieter Nijhof. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Isabel Evans - A Statement for the Future TEST Huddle
This document outlines an agenda for a workshop aimed at drafting statements for a future software testing manifesto. The workshop will have delegates discuss and edit manifesto statements in small groups. They will choose topics and draft statements for two sessions. The output of the workshop will be conclusions, manifesto statements for a final conference presentation, and documentation of the workshop to be published online along with the statements.
Rikard Edgren - Testing is an Island - A Software Testing DystopiaTEST Huddle
This document summarizes trends in software testing that could diminish its effectiveness and enjoyment. It notes an increasing focus on verification over validation, precise measurement over subjective judgement, and short-term metrics over long-term quality. This narrowing scope risks making testers isolated and limiting their creativity, motivation and ability to consider the full context of a project. The document advocates a holistic and subjective approach that considers people and intangible factors, not just short-term quantifiable results. Subjectivity and considering the whole system, not just parts, are presented as useful for testing.
Shrini Kulkarni - Software Metrics - So Simple, Yet So Dangerous TEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Software Metrics - So Simple, Yet So Dangerous by Shrini Kulkarni. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
This talk suggests how we might make sense of the tools landscape of the near future, where the pressure to modernise processes and automate is greatest, and what a new test process supported by tools might look like.
Takeaways:
- We need to take machine learning in testing seriously, but it won’t be taking our jobs just yet
- We don’t need more test automation tools; today we need tools that capture tester knowledge
- Tools that that learn and think can’t work for testers until we solve the knowledge capture challenge.
View On-Demand Webinar: https://youtu.be/EzyUdJFuzlE
About Joseph Ours' Presentation – “Bad Metric – Bad!”
Metrics have always been used in corporate sectors, primarily as a way to gain insight into what is an otherwise invisible world. Organizations blindly adopt a set of metrics as a way of satisfying some process transparency requirement, rarely applying any statistical or scientific thought behind the measures and metrics they establish and interpret. Many metrics do not represent what people believe they do and as a result can lead to erroneous decisions. Joseph looks at some of the common and some of the humorous testing metrics and determines why they are failures. He further discusses the real purpose of metrics, metrics programs and finishes with pitfalls into which you fall.
Stefaan Lukermans & Dominic Maes - Testers And Garbage Men - EuroSTAR 2011TEST Huddle
EuroSTAR Software Testing Conference 2013 presentation on Testers And Garbage Men by Stefaan Lukermans & Dominic Maes. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Fabian Scarano - Preparing Your Team for the FutureTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Preparing Your Team for the Future by Fabian Scarano. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Julian Harty - Alternatives To Testing - EuroSTAR 2010TEST Huddle
EuroSTAR Software Testing Conference 2010 presentation on "Presentation Title" by "Speaker Name". See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
- The speaker proposes 16 "test axioms" that are intended to provide a framework for testing approaches and represent principles that are context-insensitive and self-evidently true.
- The axioms are grouped into three categories: stakeholders, design, and delivery. The speaker argues the axioms can help testers think critically about testing and identify flaws in arguments.
- It is argued that process improvement models are not effective for improving testing because there is no consensus on best practices and processes must be tailored to context. True improvement requires understanding why current approaches are used given the context.
The document provides guidance for managing a team of junior testers. It discusses challenges such as lack of skills and experience in junior testers. It recommends setting clear expectations, providing frequent communication and feedback, ensuring knowledge sharing, and protecting the team to help them succeed. Patience and structure are important, as is repeating key messages, to help junior testers learn and improve. The goal is for the team to work cooperatively toward a common objective.
Ruud Teunissen - Test Process Improvement on a Shoestring TEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Test Process Improvement on a Shoestring by Ruud Teunissen. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
A Rapid Introduction to Rapid Software TestingTechWell
You're under tight time pressure and have barely enough information to proceed with testing. How do you test quickly and inexpensively, yet still produce informative, credible, and accountable results? Rapid Software Testing, adopted by context-driven testers worldwide, offers a field-proven answer to this all-too-common dilemma. In this one-day sampler of the approach, Michael Bolton introduces you to the skills and practice of Rapid Software Testing through stories, discussions, and "minds-on" exercises that simulate important aspects of real testing problems. The rapid approach isn't just testing with speed or a sense of urgency; it's mission-focused testing that eliminates unnecessary work, assures that the most important things get done, and constantly asks how testers can help speed up the successful completion of the project. Join Michael to see how rapid testing focuses on both the mind set and skill set of the individual tester who uses tight loops of exploration and critical thinking skills to help continuously re-optimize testing to match clients' needs and expectations.
A test strategy is the set of ideas that guides your test design. It's what explains why you test this instead of that, and why you test this way instead of that way. Strategic thinking matters because testers must make quick decisions about what needs testing right now and what can be left alone. You must be able to work through major threads without being overwhelmed by tiny details. James Bach describes how test strategy is organized around risk but is not defined before testing begins. Rather, it evolves alongside testing as we learn more about the product. We start with a vague idea of our strategy, organize it quickly, and document as needed in a concise way. In the end, the strategy can be as formal and detailed as you want it to be. In the beginning, though, we start small. If you want to focus on testing and not paperwork, this approach is for you.
A Rapid Introduction to Rapid Software TestingTechWell
This document provides a summary of a presentation on Rapid Software Testing. The presentation was given by Michael Bolton of DevelopSense and covered the methodology and mindset of rapid software testing. It emphasizes testing software expertly under uncertainty and time pressure. The presentation defines rapid testing as testing more quickly and less expensively while still achieving excellent results. It compares rapid testing to other approaches like exhaustive, ponderous, and slapdash testing. The presentation also discusses principles of rapid testing, how to recognize problems quickly using heuristics, and testing rapidly to fulfill the mission of testing.
These slides provide an introduction to usability testing. This well-known method in user-centred design is used to improve products, by having participants interact with these products and by measuring their performances and responses.
I presented this topic as a guest lecturer to first-year Psychology students at the University of Twente at February 6th, 2017. Providing examples and best practices from Dutch digital design agency Mirabeau, I explained to them the required steps for the preparation, the moderation, and the analysis of usability tests. Moreover, I highlighted the importance of psychologists’ knowledge, (research) methods and skills for design, which I believe to be invaluable.
The document discusses 10 signs that an organization's software testing may not be enough. These include having excessive production bugs, bugs found during user acceptance testing, growing bug counts over test cycles, not investing in testing compared to competitors, lacking clear criteria for what constitutes "enough" testing, testers advising against releasing software, weak prevention efforts like code reviews, lack of developer unit testing, frequently reduced testing periods causing deadline problems, and high tester turnover. The document advocates treating testing as risk management, increasing test reuse and automation, and addresses common challenges and questions around software testing.
This document discusses the need for leadership in the testing community to drive innovation and change. It provides examples of challenges facing testers at different companies and how they are addressing them through approaches like shifting testing left into development, adopting agile practices, and using analytics. It argues that testing is no longer just an end phase but must be integrated into continuous delivery. For change to happen, testers will need to embrace new approaches, challenge old ways of thinking, and stand up as leaders to define the future of testing.
This document discusses why checklists are better than test cases for documentation in quality assurance. It argues that test cases become overcrowded and focus too much on documentation rather than core functions. Checklists are more time-saving and easy to update. An example compares a test case to a checklist for login/registration flows. The author's company Hipo uses a test pad and robot framework integrated with checklists to share with clients and team members.
The document discusses how test axioms can be used to advance testing practices. It introduces 16 proposed test axioms grouped into stakeholder, design, and delivery axioms. The axioms represent critical thinking processes for testing any system. The document discusses how the axioms can help testers design test strategies, assess improvement opportunities, and define needed skills. It also proposes a "first equation of testing" that separates axioms, context, values, and thinking to allow for different valid approaches. Additionally, the concept of "quantum testing" is introduced to discuss assigning significance to tests rather than defining their value, which can only be determined by stakeholders.
Creating Agile Test Strategies for Larger EnterprisesTEST Huddle
Having difficulty creating an agile test strategy for your company? Let Testing Excellence Award winner, Derk-Jan de Grood, show you how it’s done
View webinar recording here - http://huddle.eurostarsoftwaretesting.com/resource/agile-testing/creating-agile-test-strategies-larger-enterprises/
Test reporting is something few testers take time to practice. Nevertheless, it's a fundamental skill—vital for your professional credibility and your own self management. Many people think management judges testing by bugs found or test cases executed. Actually, testing is judged by the story it tells. If your story sounds good, you win. A test report is the story of your testing. It begins as the story we tell ourselves, each moment we are testing, about what we are doing and why. We use the test story within our own minds, to guide our work. James Bach explores the skill of test reporting and examines some of the many different forms a test report might take. As in other areas of testing, context drives good reporting. Sometimes we make an oral report; occasionally we need to write it down. Join James for an in-depth look at the art of the reporting.
James Whittaker - Pursuing Quality-You Won't Get There - EuroSTAR 2011TEST Huddle
EuroSTAR Software Testing Conference 2011 presentation on Pursuing Quality-You Won't Get There by James Whittaker. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
'Houston We Have A Problem' by Rien van Vugt & Maurice SiteurTEST Huddle
Prevent the surprise, become a pro-active test manager. Too often projects suddenly seem to spin out of control. Challenges and risks keep stacking up and the defect count grows exponentially. At the same time, management can put pressure on you, asking when testing will be completed.
A surprise? Not really, defects only paint half the picture. The test effort, after all, is primarily determined by the number of tests that need to be completed. For an on the spot status of testing and accurate view on the quality and risks of the entire project we need to organize the test process to provide flexible, up-to-date metrics and trends on a daily basis. E.g. we need a view on baseline vs. actuals and ETC’s on test cases. Advanced metrics will provide answers on what needs to be done tomorrow to stay on track, the location and root cause of issues and who is required to take action. Also the test effort remaining for an acceptable product (or a specific risk level) can be estimated fairly accurately.
In addition early involvement and preparation in the development life cycle, performing test intakes rather than reviews, will help you bridge the gap between different development teams and allows you to verify consistency between business requirements, the integration model, functional specifications and technical specifications. It facilitates knowledge transfer and provides you with the “story” behind the specifications. This will help prevent structural issues in an early stage and avoid blocking issues during test execution.
This presentation combines daily test metrics and trends with test process dynamics and shows you how to become a “pro-active” test manager. Even better you can apply it tomorrow and take your test process to a distinct higher maturity level.
Michael Bolton - Two Futures of Software TestingTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Two Futures of Software Testing by Michael Bolton. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Shrini Kulkarni - Software Metrics - So Simple, Yet So Dangerous TEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Software Metrics - So Simple, Yet So Dangerous by Shrini Kulkarni. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
This talk suggests how we might make sense of the tools landscape of the near future, where the pressure to modernise processes and automate is greatest, and what a new test process supported by tools might look like.
Takeaways:
- We need to take machine learning in testing seriously, but it won’t be taking our jobs just yet
- We don’t need more test automation tools; today we need tools that capture tester knowledge
- Tools that that learn and think can’t work for testers until we solve the knowledge capture challenge.
View On-Demand Webinar: https://youtu.be/EzyUdJFuzlE
About Joseph Ours' Presentation – “Bad Metric – Bad!”
Metrics have always been used in corporate sectors, primarily as a way to gain insight into what is an otherwise invisible world. Organizations blindly adopt a set of metrics as a way of satisfying some process transparency requirement, rarely applying any statistical or scientific thought behind the measures and metrics they establish and interpret. Many metrics do not represent what people believe they do and as a result can lead to erroneous decisions. Joseph looks at some of the common and some of the humorous testing metrics and determines why they are failures. He further discusses the real purpose of metrics, metrics programs and finishes with pitfalls into which you fall.
Stefaan Lukermans & Dominic Maes - Testers And Garbage Men - EuroSTAR 2011TEST Huddle
EuroSTAR Software Testing Conference 2013 presentation on Testers And Garbage Men by Stefaan Lukermans & Dominic Maes. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Fabian Scarano - Preparing Your Team for the FutureTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Preparing Your Team for the Future by Fabian Scarano. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Julian Harty - Alternatives To Testing - EuroSTAR 2010TEST Huddle
EuroSTAR Software Testing Conference 2010 presentation on "Presentation Title" by "Speaker Name". See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
- The speaker proposes 16 "test axioms" that are intended to provide a framework for testing approaches and represent principles that are context-insensitive and self-evidently true.
- The axioms are grouped into three categories: stakeholders, design, and delivery. The speaker argues the axioms can help testers think critically about testing and identify flaws in arguments.
- It is argued that process improvement models are not effective for improving testing because there is no consensus on best practices and processes must be tailored to context. True improvement requires understanding why current approaches are used given the context.
The document provides guidance for managing a team of junior testers. It discusses challenges such as lack of skills and experience in junior testers. It recommends setting clear expectations, providing frequent communication and feedback, ensuring knowledge sharing, and protecting the team to help them succeed. Patience and structure are important, as is repeating key messages, to help junior testers learn and improve. The goal is for the team to work cooperatively toward a common objective.
Ruud Teunissen - Test Process Improvement on a Shoestring TEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Test Process Improvement on a Shoestring by Ruud Teunissen. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
A Rapid Introduction to Rapid Software TestingTechWell
You're under tight time pressure and have barely enough information to proceed with testing. How do you test quickly and inexpensively, yet still produce informative, credible, and accountable results? Rapid Software Testing, adopted by context-driven testers worldwide, offers a field-proven answer to this all-too-common dilemma. In this one-day sampler of the approach, Michael Bolton introduces you to the skills and practice of Rapid Software Testing through stories, discussions, and "minds-on" exercises that simulate important aspects of real testing problems. The rapid approach isn't just testing with speed or a sense of urgency; it's mission-focused testing that eliminates unnecessary work, assures that the most important things get done, and constantly asks how testers can help speed up the successful completion of the project. Join Michael to see how rapid testing focuses on both the mind set and skill set of the individual tester who uses tight loops of exploration and critical thinking skills to help continuously re-optimize testing to match clients' needs and expectations.
A test strategy is the set of ideas that guides your test design. It's what explains why you test this instead of that, and why you test this way instead of that way. Strategic thinking matters because testers must make quick decisions about what needs testing right now and what can be left alone. You must be able to work through major threads without being overwhelmed by tiny details. James Bach describes how test strategy is organized around risk but is not defined before testing begins. Rather, it evolves alongside testing as we learn more about the product. We start with a vague idea of our strategy, organize it quickly, and document as needed in a concise way. In the end, the strategy can be as formal and detailed as you want it to be. In the beginning, though, we start small. If you want to focus on testing and not paperwork, this approach is for you.
A Rapid Introduction to Rapid Software TestingTechWell
This document provides a summary of a presentation on Rapid Software Testing. The presentation was given by Michael Bolton of DevelopSense and covered the methodology and mindset of rapid software testing. It emphasizes testing software expertly under uncertainty and time pressure. The presentation defines rapid testing as testing more quickly and less expensively while still achieving excellent results. It compares rapid testing to other approaches like exhaustive, ponderous, and slapdash testing. The presentation also discusses principles of rapid testing, how to recognize problems quickly using heuristics, and testing rapidly to fulfill the mission of testing.
These slides provide an introduction to usability testing. This well-known method in user-centred design is used to improve products, by having participants interact with these products and by measuring their performances and responses.
I presented this topic as a guest lecturer to first-year Psychology students at the University of Twente at February 6th, 2017. Providing examples and best practices from Dutch digital design agency Mirabeau, I explained to them the required steps for the preparation, the moderation, and the analysis of usability tests. Moreover, I highlighted the importance of psychologists’ knowledge, (research) methods and skills for design, which I believe to be invaluable.
The document discusses 10 signs that an organization's software testing may not be enough. These include having excessive production bugs, bugs found during user acceptance testing, growing bug counts over test cycles, not investing in testing compared to competitors, lacking clear criteria for what constitutes "enough" testing, testers advising against releasing software, weak prevention efforts like code reviews, lack of developer unit testing, frequently reduced testing periods causing deadline problems, and high tester turnover. The document advocates treating testing as risk management, increasing test reuse and automation, and addresses common challenges and questions around software testing.
This document discusses the need for leadership in the testing community to drive innovation and change. It provides examples of challenges facing testers at different companies and how they are addressing them through approaches like shifting testing left into development, adopting agile practices, and using analytics. It argues that testing is no longer just an end phase but must be integrated into continuous delivery. For change to happen, testers will need to embrace new approaches, challenge old ways of thinking, and stand up as leaders to define the future of testing.
This document discusses why checklists are better than test cases for documentation in quality assurance. It argues that test cases become overcrowded and focus too much on documentation rather than core functions. Checklists are more time-saving and easy to update. An example compares a test case to a checklist for login/registration flows. The author's company Hipo uses a test pad and robot framework integrated with checklists to share with clients and team members.
The document discusses how test axioms can be used to advance testing practices. It introduces 16 proposed test axioms grouped into stakeholder, design, and delivery axioms. The axioms represent critical thinking processes for testing any system. The document discusses how the axioms can help testers design test strategies, assess improvement opportunities, and define needed skills. It also proposes a "first equation of testing" that separates axioms, context, values, and thinking to allow for different valid approaches. Additionally, the concept of "quantum testing" is introduced to discuss assigning significance to tests rather than defining their value, which can only be determined by stakeholders.
Creating Agile Test Strategies for Larger EnterprisesTEST Huddle
Having difficulty creating an agile test strategy for your company? Let Testing Excellence Award winner, Derk-Jan de Grood, show you how it’s done
View webinar recording here - http://huddle.eurostarsoftwaretesting.com/resource/agile-testing/creating-agile-test-strategies-larger-enterprises/
Test reporting is something few testers take time to practice. Nevertheless, it's a fundamental skill—vital for your professional credibility and your own self management. Many people think management judges testing by bugs found or test cases executed. Actually, testing is judged by the story it tells. If your story sounds good, you win. A test report is the story of your testing. It begins as the story we tell ourselves, each moment we are testing, about what we are doing and why. We use the test story within our own minds, to guide our work. James Bach explores the skill of test reporting and examines some of the many different forms a test report might take. As in other areas of testing, context drives good reporting. Sometimes we make an oral report; occasionally we need to write it down. Join James for an in-depth look at the art of the reporting.
James Whittaker - Pursuing Quality-You Won't Get There - EuroSTAR 2011TEST Huddle
EuroSTAR Software Testing Conference 2011 presentation on Pursuing Quality-You Won't Get There by James Whittaker. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
'Houston We Have A Problem' by Rien van Vugt & Maurice SiteurTEST Huddle
Prevent the surprise, become a pro-active test manager. Too often projects suddenly seem to spin out of control. Challenges and risks keep stacking up and the defect count grows exponentially. At the same time, management can put pressure on you, asking when testing will be completed.
A surprise? Not really, defects only paint half the picture. The test effort, after all, is primarily determined by the number of tests that need to be completed. For an on the spot status of testing and accurate view on the quality and risks of the entire project we need to organize the test process to provide flexible, up-to-date metrics and trends on a daily basis. E.g. we need a view on baseline vs. actuals and ETC’s on test cases. Advanced metrics will provide answers on what needs to be done tomorrow to stay on track, the location and root cause of issues and who is required to take action. Also the test effort remaining for an acceptable product (or a specific risk level) can be estimated fairly accurately.
In addition early involvement and preparation in the development life cycle, performing test intakes rather than reviews, will help you bridge the gap between different development teams and allows you to verify consistency between business requirements, the integration model, functional specifications and technical specifications. It facilitates knowledge transfer and provides you with the “story” behind the specifications. This will help prevent structural issues in an early stage and avoid blocking issues during test execution.
This presentation combines daily test metrics and trends with test process dynamics and shows you how to become a “pro-active” test manager. Even better you can apply it tomorrow and take your test process to a distinct higher maturity level.
Michael Bolton - Two Futures of Software TestingTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Two Futures of Software Testing by Michael Bolton. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Scott Andress - Collaboration not Competition updatedTEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Collaboration not Competition updated by Scott Andress. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Martin Gijsen - Effective Test Automation a la Carte TEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Effective Test Automation a la Carte by Martin Gijsen. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Ian Smith - Mobile Software Testing - Facing Future ChallengesTEST Huddle
This document discusses challenges in testing mobile software systems. It notes the increasing capabilities of mobile devices and complexity of mobile applications. Key challenges include high variability in cellular networks and devices, changing platform landscapes, and ensuring security of sensitive data on devices. The document recommends approaches like managing complexity through architectural partitioning, maximizing code reuse across platforms, and combining emulation with automated GUI testing. It provides an example case study of developing an automated mobile call generation system and discusses lessons learned.
EuroSTAR Software Testing Conference 2008 presentation on The Damage Zone by Julien Bensaid. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Jonas Skjoldan - Automatic GUI test with Ruby and WatirTEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Automatic GUI test with Ruby and Watir by Jonas Skjoldan . See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Dietmar Strasser - Traditional QA meets Agile DevelopmentTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Traditional QA meets Agile Development by Dietmar Strasser. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Ruud Teunissen - Personal Test Improvement - Dealing with the FutureTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Personal Test Improvement - Dealing with the Future by Ruud Teunissen. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Otto Vinter - Analysing Your Defect Data for Improvement PotentialTEST Huddle
EuroSTAR Software Testing Conference 2008 presentation on Analysing Your Defect Data for Improvement Potential by Otto Vinter. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
EuroSTAR Software Testing Conference 2011 presentation on Implementing Change by Torben Hoelgaard. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
EuroSTAR Software Testing Conference 2009 presentation on The Power of Risk by Erik Beolen. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Markus Gartner - Alternative Paths for Self-Education in Software Testing - E...TEST Huddle
EuroSTAR Software Testing Conference 2013 presentation on Alternative Paths for Self-Education in Software Testing by Markus Gartner . See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Jelle Calsbeek - Stay Agile with Model Based Testing revisedTEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Evolution of New Feature Verification in 3G Networks by Michael Monaghan. See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Doron Reuveni - The Mobile App Quality Challenge - EuroSTAR 2010TEST Huddle
EuroSTAR Software Testing Conference 2010 presentation on The Mobile App Quality Challenge by Doron Reuveni. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Michael Roar Borlund & Christian Carlsen - Real Exploratory Testing, Now With...TEST Huddle
Exploratory testing approaches like "hotspot" and "coffee break" were presented as ways to optimize time spent testing and find more defects when performing exploratory testing in a service-oriented architecture (SOA). The "hotspot" approach resulted in finding more defects on average but took more time per defect. The "coffee break" approach found fewer defects but in less time. Both approaches provided broader test coverage and additional knowledge of the system compared to traditional testing. The presentation concluded that using a customized mix of both exploratory testing methods can minimize wasted time and add value to a project.
Johan Jonasson - Introducing Exploratory Testing to Save the ProjectTEST Huddle
EuroSTAR Software Testing Conference 2009 presentation on Introducing Exploratory Testing to Save the Project by Johan Jonasson . See more at conferences.eurostarsoftwaretesting.com/past-presentations/
Martin Koojj - Testers in the Board of DirectorsTEST Huddle
Martin Kooij discusses the evolution and future of software testing. He traces his experience in testing telecom equipment from the 1960s to present day. Testing has matured from a technical focus on defects to a more strategic, risk-based approach. Metrics now consider risks rather than just defects. Kooij believes that by 2018, testers will report directly to boards of directors on product risks translated into business risks. Testers will broaden their skills and take responsibility for cost-effective testing to estimate and mitigate risks. For testing to continue evolving, testers must be brave, independent, and politically skilled while focusing on business risks over personal agendas.
Henrik Andersson - Exploratory Testing Champions - EuroSTAR 2010TEST Huddle
EuroSTAR Software Testing Conference 2010 presentation on Henrik Andersson by Exploratory Testing Champions. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
Fredrik Rydberg - Can Exploratory Testing Save Lives - EuroSTAR 2010TEST Huddle
EuroSTAR Software Testing Conference 2010 presentation on Can Exploratory Testing Save Lives by Fredrik Rydberg. See more at: http://conference.eurostarsoftwaretesting.com/past-presentations/
If you’ve never requested a usability study bid before or you want to see how our process differs from others you have worked with in the past — this deck is for you. Here is the 5-step process June UX uses to plan and conduct moderated usability studies.
Usability Testing Basics: What's it All About? at Web SIG ClevelandCarol Smith
Presented to Web SIG Cleveland on May 21, 2011 at Notre Dame College in South Euclid (Cleveland), Ohio.
Learn all you need to get started:
- Where you can conduct studies (does it have to be in a lab?)
- Types of studies (RITE, think aloud, etc.)
- Tips for recruiting participants
- Tips for Interacting with participants without biasing the study
- Preparing for the study (materials needed, forms, etc.)
- Guidance for analyzing the study
Heather Vancura and Bruno Souza have a new book out on the topic of how to have a successful and rewarding career as a software developer. The title of this session is the title of the book. Ed Burns wrote a book on the same topic nearly fifteen years ago. In this 45-minute session, you will learn the most important parts of the the new book from the author of the old book. Ed brings the insight he gained from writing his book to the task of presenting Heather and Bruno's book which contains the lessons learned in interviewing a diverse selection of 26 successful developers. Between the authors of the two books and the interviews they conducted, there is over a century of developer career experience in this talk! This informative and fun session will give you some practical tips to improve your own career.
The document discusses changes in the role of testers and quality assurance. It notes that companies like Facebook do not have dedicated testing teams, but instead rely on automation, user reports, and privileged beta testers to find bugs. However, as software increases in importance and complexity, more structured testing approaches are needed. The document advocates for testers to take a broader view, collaborate more across teams, and adapt their skills while maintaining independence. Testers are encouraged to find new areas to contribute and ways to help ensure overall product quality.
The article discusses test coverage and how to avoid missing things during testing. It notes that complete test coverage is not possible due to differences between test and customer environments. The author provides tips for improving coverage, including testing all requirements, using different test techniques, automating tests, testing edge cases, and getting feedback from real users. While not every bug can be found, following best practices can help improve test coverage.
How To: Developers' Community-driven Career GrowthC4Media
Video and slides synchronized, mp3 and slide download available at URL https://bit.ly/2G4YPN4.
Georgiy Mogelashvili talks about the “Game of Roles” that Booking.com uses to grow their own developers into senior or leadership positions. He talks about how the framework came about, what it means in details, how they are using it, and, most important, how to apply the same principles at another organization without much effort but with high outcome. Filmed at qconsf.com.
Georgiy Mogelashvili is working at Booking.com as a Senior Developer and Team Lead. He is responsible for company products development as well as people management.
Michael Larsen discusses lessons learned from his involvement with Weekend Testing, a volunteer organization that holds short testing sessions online to help testers learn and practice their skills. Some key lessons include: considering participants' varying access to applications and internet speeds based on geography; the value of focused session-based testing to thoroughly test applications; using a "divide and conquer" approach to split testing tasks among groups; ensuring testing goals and missions are clearly communicated; and being aware of potential cultural differences that could impact requirements or date conventions. Overall, Weekend Testing is an effective way for passionate testers globally to collaborate, train others, and have fun advancing their testing skills.
What drives innovation? Is it business? Is it opportunity? Is it need? Is it challenges? I believe Architects love to solve problems, provide a vision for growth, mitigate issues, drive business change, rethink legacy and create a new. PASSION to create is what drives innovation. ARCHON is an initiative created by me, to try and ensure all the enthusiastic architects live their passion during this event. First EVER global Architecture Hackathon event
Formative Evaluation for Educational Product DevelopmentVanessa Gennarelli
This document discusses formative evaluation for educational product development. Formative evaluation involves testing an educational product with users during development to inform the product's direction. It can be conducted at any time during development. Some key methods discussed include interviews, think-aloud protocols, focus groups, questionnaires, and click-testing. Conducting formative evaluation with target users for around a week can help identify usability issues, measure user appeal and engagement, and test user comprehension to improve the educational product.
The document provides information about conducting usability testing. It discusses what usability testing involves, including setting tasks for test participants and noting any problems they encounter. It provides tips for testing, such as teaming up with a partner, selecting 3-5 test participants, having them complete 2-3 tasks in 30-50 minutes, and one person acting as note-taker and moderator. The document also discusses how to find participants, what to tell them, questions to ask as moderator, common testing errors to avoid, and metrics to capture like completion rates, time on task, errors and satisfaction.
Guided by Optimizely's training team, participants will leave this session armed with the tools and resources to enable their organization with the technical and strategic skills needed to build a successful optimization program. Through modeling of best-practices and facilitation techniques, participants will gain a strong foundation in the three key Optimizely training modules: Optimizely X - Platform Training, Strategy Ideation, and Hypothesis Creation, and Result Review and Analysis.
This workshop focuses on how to facilitate and what to expect before, during, and after a training. We will discuss the key learning objectives, common questions and takeaways, and follow-up resources from each module.
Join us and learn how to deliver powerful Optimizely trainings to enable your team.
Using Testing Marathons to Support a Culture of QualityRodrigo Cursino
The document describes how test marathons called CESAR.thons are used at the Brazilian Innovation Institute to promote a culture of quality. CESAR.thons bring together multidisciplinary teams to test a product over the course of a day using a structured framework. The framework includes planning, instructions, execution, debriefing, closing, and reporting phases. CESAR.thons have helped train new QA employees, identify over 300 bugs and ideas for improvement, and support business goals while engaging employees.
The document introduces design methods categorized into three phases: Discover, Define, and Develop. The Discover phase focuses on research methods like observation, interviews, and brainstorming. The Define phase is used to analyze research findings and prioritize ideas through methods like focus groups and customer journey mapping. Finally, the Develop phase employs creative techniques such as scenarios, prototyping, and role playing to refine ideas into viable solutions.
The document introduces design methods categorized into three phases: Discover, Define, and Develop. The Discover phase focuses on research methods like observation, user diaries, and surveys. The Define phase is used to analyze findings and prioritize ideas through methods like assessment criteria and focus groups. Finally, the Develop phase develops solutions using profiles, scenarios, prototyping, and other techniques. Overall, the document outlines various qualitative and quantitative design research methods that can be applied at different stages of the design process.
Testing the unknown: the art and science of working with hypothesisArdita Karaj
Testing what we know, or have a clear understanding of, is relatively straight forward, as is making decisions based on the expected result. But today’s world is presenting us with the Unknown and the Ambiguous, which can only be approached by hypothesizing and experimenting - a lot! This requires intentional thinking, and a different strategy to observe in context.
This session will uncover how testers are helping their teams and product owners, by basing their testing on the science behind creating hypotheses and running experiments. A testing mindset and probing the context around use cases are some of the most valuable competencies testers bring to the team in order to enable decisions based on data.
Pairing: The Secret Sauce of Agile TestingTechWell
Finding time to learn test techniques, mentor other testers, grow application knowledge, and cross-train your team members is a daunting task with a complicated recipe. What if you could do these things while testing and finding bugs? Enter Pair Testing. What’s that? Well, maybe you’ve heard of pair programming. It’s like that—only you’re testing rather than programming. And it’s the secret sauce of agile testing because it makes your routine, bland testing so much more fun! Testers on Jess Lancaster’s team use pair testing not only to make better software but also to foster better relationships along the way. Jess explores why pairing works, how to run an effective pairing session, and just how easy it is to get started with pairing. Armed with Jess’ easy-to-use Pair Testing recipe card, plan your first pairing encounter so you are ready to roll when you get back to the office. This sounds easy enough, but you know there will be mistakes when you try it. Jess has you covered there, too. Learn his team’s pairing mistakes and the things he did to improve their pairing sessions.
This document provides guidance on conducting a usability study. It outlines the schedule, expectations for participants, and guidelines for running test sessions. Participants will be divided into groups of 3, with one person acting as the user and two observing. As a user, they will complete tasks while thinking aloud, and observers will take notes on successes, obstacles, and areas of confusion. After each task the group will switch roles. The goal is to evaluate the design and identify ways to improve the user experience.
Are you looking to gather insights from your potential customers? When it comes to your prospects, do you really know what they want? Many startup teams tell us they are missing the key information they need to get into their users' mind. Without this information, the products often fall short of delighting users.
There are those that believe that user research and usability testing must be a complex and scientific process that takes lots of time, money, and resources. However, in the real world, most startups don't have the luxury to spend weeks or months on their user research. That's where guerrilla research techniques come into play.
Sourcing The Right Participants For Your UX Research & TestingUserZoom
This document discusses sourcing the right participants for UX research and testing. It begins by noting that sourcing participants is often the #1 challenge and outlines an agenda to cover various sourcing methods. The methods are divided into two categories: doing it yourself using your own customers or visitors, and using external companies. Doing it yourself involves techniques like guerrilla usability testing, using customer databases, social media sites, websites/apps, and in-store locations. Using external companies involves online panels, traditional recruiting firms, new recruiting startups, specialized panels, and a new sourcing engine. It provides tips on screeners, study quality, incentives, and special considerations for regulated industries, B2B, research with kids
Building Delightful Products: A Customer-Centric Approach to Product Strategy...Perfetti Media
In this presentation, Carbonite's User Experience Director, Christine Perfetti, will share proven techniques for quickly exploring new design approaches based on solid customer data. She will share approaches for gathering customer insights, generating new product concepts, and evaluating designs.
Similar to Ajay Balamnrugadas - Weekend Testing, Skilled Software Testing Unleashed - EuroSTAR 2010 (20)
Why We Need Diversity in Testing- AccentureTEST Huddle
In this webinar Rasa (Testing capability lead for Denmark) and Matthias (EALA Testing capability lead) will share some of their own experiences why diversity matters, give insights into how Accenture as a global firm is promoting diversity and how we are in the process of changing our attitudes and processes to make all of this sustainable
Keys to continuous testing for faster delivery euro star webinar TEST Huddle
Your business needs to deliver faster. To accommodate, Development needs to introduce fewer changes but in a much more frequent cadence. This creates a challenge for test teams to keep up with the rapid pace of change without compromising on quality. Automation is paramount to the success or failure of Continuous Delivery, and Continuous Testing enables early and frequent quality feedback throughout the CI/CD pipeline.
In this webinar, Eran & Ayal will explore how to implement Continuous Testing to ensure high quality releases in a Continuous Delivery environment; including what to test and when to automate new functionality in order to optimize your efforts.
Why you Shouldnt Automated But You Will Anyway TEST Huddle
The document discusses automation in software testing. It begins by outlining common claims made about the benefits of automation, such as saving time and improving quality, but argues that these claims often don't hold true. Automation does not inherently save time, guarantee quality, or reduce resources needed. It also does not always save money when development, maintenance, and infrastructure costs are considered. The document provides a formula for determining when automation is worthwhile based on how many times a test case would need to be rerun manually. It concludes by acknowledging that, despite these drawbacks, organizations will still automate testing because it is exciting, managers demand it, and it benefits careers.
In this webinar Carsten will explore the role of the tester in a Scrum team. He will examine where the tester play an important role in Scrum and how you can contribute to a teams performance.
Leveraging Visual Testing with Your Functional TestsTEST Huddle
Designing and implementing (or selecting) the right automation strategy, for functional testing, with visual testing, can help your project with greater test coverage while improving test scalability
Big Data: The Magic to Attain New HeightsTEST Huddle
This document discusses how big data and data science can be used to attain new heights, likening it to magic. It provides an overview of Ken Johnston's background and experiences in data science. It then discusses six keys to a "big" magic show with big data: trying multiple times, addressing issues with over-counting, experimentation techniques like A/B testing, infrastructure for big data, tools and skills, and security, privacy and fraud protection. The document emphasizes the importance of an assistant to help the data scientist or data engineer with various tasks.
The document discusses Test Driven Development (TDD) and Test Driven Design. It uses the analogy of building a lightsaber and later a Death Star to illustrate the TDD process and benefits. Some benefits mentioned are better test coverage, less debugging, and better design. The document provides tips for practicing TDD including planning ahead, defining boundaries, taking small steps to pass each test, and maintaining discipline. It emphasizes trying TDD in a team and considering Behavior Driven Development (BDD) as well.
Scaling Agile with LeSS (Large Scale Scrum)TEST Huddle
In this webinar, Elad will cover the principles that the #LeSS framework has to offer in order to enable bug organisations to become agile.
View webinar recording - https://huddle.eurostarsoftwaretesting.com/resource/agile-testing/scaling-agile-less-large-scale-scrum/
3 key takeaways
- Do you know the meaning of your organisation, system, product?
- Can you deliver the important risks right away?
- How can you communicate about the (process and product) risks your dealing with?
View Webinar recording: https://huddle.eurostarsoftwaretesting.com/resource/test-management/is-there-a-risk/
Are Your Tests Well-Travelled? Thoughts About Test CoverageTEST Huddle
This document summarizes a presentation on test coverage given by Dorothy Graham. It uses an analogy of travel to different locations to explain what test coverage means and some caveats. Coverage refers to the relationship between tests and the parts of a system being tested, but achieving 100% coverage does not mean everything is tested. There are four caveats discussed: coverage only measures one aspect of testing, a single test can achieve coverage, coverage does not indicate quality, and it only applies to the existing system not missing pieces. The key recommendation is to ask "coverage of what?" when the term is used rather than assuming more coverage is always better.
Growing a Company Test Community: Roles and Paths for TestersTEST Huddle
Over the past three years, our company’s test team has grown from three lonesome testers to a community of nine – with more planned. Since we don’t see testers as “click monkeys”, but as valuable and integrated project members who bring a specific skill set to the table, it’s important for us to choose testers well and to train them in various areas so that they can contribute, grow and see their own career path within testing.
To structure to our internal tester training program, we have been developing role descriptions, education paths and career options for our testers, which I’d like to share with you in this webinar.
View webinar - https://huddle.eurostarsoftwaretesting.com/resource/webinar/growing-company-test-community-roles-paths-testers/
It’s the same argument again and again. One side says “team members should all be able to do everything, and the programmers should do their testing and all testers should be writing code”. The other side says “No, that can’t possibly work – programmers don’t know how to test, they don’t have the right mindset”. And on and on it goes.
http://huddle.eurostarsoftwaretesting.com/resource/webinar/need-testers-agile-teams/
In this webinar, Dave Haeffner (Elemental Selenium, USA) discusses how to:
- Build an integrated feedback loop to automate test runs and find issues fast
- Setup your own infrastructure or connect to a cloud provider
-Dramatically improve test times with parallelization
https://huddle.eurostarsoftwaretesting.com/resource/webinar/use-selenium-successfully/
Testers & Teams on the Agile Fluency™ Journey TEST Huddle
The document discusses the Agile Fluency model, which aims to help teams and testers improve their agile skills and practices over time. It describes a pathway with increasing levels of fluency that provide more benefits, including delivering value, optimizing value, and innovating. Reaching higher levels requires investments in training, coaching, and changing team structures and roles. The model can help organizations determine what level of fluency they need and what investments are required for testing teams to operate at that level.
Practical Test Strategy Using HeuristicsTEST Huddle
Key Takeaways
- See what makes a good test strategy
- Learn how to make a thorough test strategy
- Identify what is the ‘Heuristic Test Strategy Model’ is
- Develop a solid test strategy that fits fast
- Discover how diversification can help you to create a test strategy
Key Takeaways:
- A diagramming method that helps discuss roles
- A one page analysis heuristic for roles
- Why roles matter on projects
https://huddle.eurostarsoftwaretesting.com/resource/people-skills/thinking-through-your-role/
Key Takeaways:
- What will this release contain
- What impact will it have on your test runs
- How can you preserve your existing investment in tests using the Selenium WebDriver APIs, and your even older RC tests
- Looking forward, when will the W3C spec be complete
- What can we expect from Selenium 4
https://huddle.eurostarsoftwaretesting.com/
New Model Testing: A New Test Process and ToolTEST Huddle
Paul Gerrard presented a new test process and tool called Cervaya that combines elements of structured and exploratory testing. The process involves testers surveying features using Cervaya to iteratively build system models and test plans. This shifts testing earlier in the development process. Cervaya logs tester activity, supports real-time collaboration, and could generate documentation. The goal is to make testing more aligned with agile and continuous delivery approaches. Gerrard invited collaboration on further developing Cervaya.
Five Digital Age Trends That Will Dramatically Impact Testing And Quality Sk...TEST Huddle
Key Takeaways:
- Understand the key digital age trends that will disrupt large enterprises
- Learn what impact and opportunities these trends present for testing and quality engineering skills
- Discover how a comprehensive digital testing strategy integrated with high velocity intelligent automation enables success for the high performers of the future
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
[OReilly Superstream] Occupy the Space: A grassroots guide to engineering (an...Jason Yip
The typical problem in product engineering is not bad strategy, so much as “no strategy”. This leads to confusion, lack of motivation, and incoherent action. The next time you look for a strategy and find an empty space, instead of waiting for it to be filled, I will show you how to fill it in yourself. If you’re wrong, it forces a correction. If you’re right, it helps create focus. I’ll share how I’ve approached this in the past, both what works and lessons for what didn’t work so well.
What is an RPA CoE? Session 1 – CoE VisionDianaGray10
In the first session, we will review the organization's vision and how this has an impact on the COE Structure.
Topics covered:
• The role of a steering committee
• How do the organization’s priorities determine CoE Structure?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
AppSec PNW: Android and iOS Application Security with MobSFAjin Abraham
Mobile Security Framework - MobSF is a free and open source automated mobile application security testing environment designed to help security engineers, researchers, developers, and penetration testers to identify security vulnerabilities, malicious behaviours and privacy concerns in mobile applications using static and dynamic analysis. It supports all the popular mobile application binaries and source code formats built for Android and iOS devices. In addition to automated security assessment, it also offers an interactive testing environment to build and execute scenario based test/fuzz cases against the application.
This talk covers:
Using MobSF for static analysis of mobile applications.
Interactive dynamic security assessment of Android and iOS applications.
Solving Mobile app CTF challenges.
Reverse engineering and runtime analysis of Mobile malware.
How to shift left and integrate MobSF/mobsfscan SAST and DAST in your build pipeline.
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/temporal-event-neural-networks-a-more-efficient-alternative-to-the-transformer-a-presentation-from-brainchip/
Chris Jones, Director of Product Management at BrainChip , presents the “Temporal Event Neural Networks: A More Efficient Alternative to the Transformer” tutorial at the May 2024 Embedded Vision Summit.
The expansion of AI services necessitates enhanced computational capabilities on edge devices. Temporal Event Neural Networks (TENNs), developed by BrainChip, represent a novel and highly efficient state-space network. TENNs demonstrate exceptional proficiency in handling multi-dimensional streaming data, facilitating advancements in object detection, action recognition, speech enhancement and language model/sequence generation. Through the utilization of polynomial-based continuous convolutions, TENNs streamline models, expedite training processes and significantly diminish memory requirements, achieving notable reductions of up to 50x in parameters and 5,000x in energy consumption compared to prevailing methodologies like transformers.
Integration with BrainChip’s Akida neuromorphic hardware IP further enhances TENNs’ capabilities, enabling the realization of highly capable, portable and passively cooled edge devices. This presentation delves into the technical innovations underlying TENNs, presents real-world benchmarks, and elucidates how this cutting-edge approach is positioned to revolutionize edge AI across diverse applications.
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsDianaGray10
Join us to learn how UiPath Apps can directly and easily interact with prebuilt connectors via Integration Service--including Salesforce, ServiceNow, Open GenAI, and more.
The best part is you can achieve this without building a custom workflow! Say goodbye to the hassle of using separate automations to call APIs. By seamlessly integrating within App Studio, you can now easily streamline your workflow, while gaining direct access to our Connector Catalog of popular applications.
We’ll discuss and demo the benefits of UiPath Apps and connectors including:
Creating a compelling user experience for any software, without the limitations of APIs.
Accelerating the app creation process, saving time and effort
Enjoying high-performance CRUD (create, read, update, delete) operations, for
seamless data management.
Speakers:
Russell Alfeche, Technology Leader, RPA at qBotic and UiPath MVP
Charlie Greenberg, host
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Digital Banking in the Cloud: How Citizens Bank Unlocked Their MainframePrecisely
Inconsistent user experience and siloed data, high costs, and changing customer expectations – Citizens Bank was experiencing these challenges while it was attempting to deliver a superior digital banking experience for its clients. Its core banking applications run on the mainframe and Citizens was using legacy utilities to get the critical mainframe data to feed customer-facing channels, like call centers, web, and mobile. Ultimately, this led to higher operating costs (MIPS), delayed response times, and longer time to market.
Ever-changing customer expectations demand more modern digital experiences, and the bank needed to find a solution that could provide real-time data to its customer channels with low latency and operating costs. Join this session to learn how Citizens is leveraging Precisely to replicate mainframe data to its customer channels and deliver on their “modern digital bank” experiences.
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Freshworks Rethinks NoSQL for Rapid Scaling & Cost-EfficiencyScyllaDB
Freshworks creates AI-boosted business software that helps employees work more efficiently and effectively. Managing data across multiple RDBMS and NoSQL databases was already a challenge at their current scale. To prepare for 10X growth, they knew it was time to rethink their database strategy. Learn how they architected a solution that would simplify scaling while keeping costs under control.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
1. Ajay Balamurugadas
on behalf of
Weekend Testing Team
Weekend Testing Skilled Software Testing Unleashed
2. My sincere thanks to…
Amit Kulkarni
Anna Baik
Geir Gulbrandsen
Lisa Crispin
Markus Gärtner
Matthew Heusser
Michael Bolton
Richard Hill
Rosie Sherry
Sumit Kalra
Tiffany Fodor
Tony Bruce
2
3. Agenda
Brief Introduction
How do we conduct sessions
Journey so far
Chapter Initiators
Challenges faced
Benefits to community
Achievements
Future Plans3
4. Brief Introduction
What is Weekend Testing?
An initiative started by four passionate software testers to improve their testing skills by practicing software testing.
Why Weekends
First Session
A group with a difference
Cost vs. Value4
5. Mission and Vision
What is your Mission statement?
A platform for software testers to collaborate, test various kinds of software, foster hope, gain peer recognition, and be of value to the community.
What is your Vision statement?
To find people with similar synergy –towards testing software and learning from it and improve the craft of software testing.
5
6. How Do We Conduct Sessions
Registration:
Emails, Events added in Forums.
Announcements on Website, Twitter, Blogs, SMS.
Facilitation:
Skype Group chat, Give mission, Set time limit.
Provide adequate information.
Testing Session:
Make sure testers do not divert from mission.
Time boxed session.
Discussion Session:
Encourage questions, clear assumptions, de-brief. 6
8. Journey So Far (Part 1)
More than 100 sessions.
Over 1000 members.
Chapters in Asia, Europe, Australia.
Skills focused:
Observation , Note Taking, Learning, Clearing Traps, Questioning, Bug Hunting, Bug Investigation, Modeling,
Automation, Bug Advocacy, Presentation, Group Discussion, Collaboration. 8
9. Journey So Far (Part 2)
Over 10 Guest Facilitators.
New repository of bugs and tester reports.
Variety of products tested.
Experience sharing, new test ideas.
Fun weekends, New bloggers. 9
14. Benefits To Software Community
Tester community
More testers’ interactions, Tester clubs, Practice and self- critique, independent and community based skill development
Programmer community
Testers’ opinion, better open source software
Software community
Better testers, better software, happy customers14
15. Achievements (Part 1)
Testimonials from programmers.
Emails and Comments from Testers
I have improved a lot in my assignments at office after taking this session. Thanks everyone for making this activity a success.
–Rajesh N Iyer
‘Better Software Magazine’ and ‘The Testing Planet’
Nov 2009 STC Conference, Bangalore.
Nov 2010 EuroSTAR Conference, Copenhagen.
15
16. Achievements (Part 2)
Future of Skilled Testing
–Michael Bolton
It’s a sapient testing insurgency!
–James Bach
I liked this a lot, as normally I work alone, so I guess this is the closest I get to working in a team environment.
–Anne-Marie Charrett16
17. What have we learnt
Fail faster and safer
Appreciate diversity
Uncover hidden talent
Aware of traps
Varied experience
Contribute to open source
Build unity in our field
17
18. Want to Initiate a Chapter?
Traps
Members
Assistance
Guidelines
Access to website
Finally, facilitation is also a skill
18
21. The Initiators
Parimala Shankaraiah
Author of Curious Tester blog.
Sharath Byregowda
Author of Test to Tester blog
Ajay Balamurugadas
Author of Enjoy Testingblog
Manoj M V
Author of Testing Redefinedblog
Santhosh Tuppad
Author of Tuppad.comblog 21