As good engineers, we use metrics to make decisions on quality, testing. My view is that measurements have inherent assumptions that we have to be cognizant to, otherwise decisions can be dangerous.
Testing As A Bottleneck - How Testing Slows Down Modern Development Processes...TEST Huddle
We often claim the purpose of testing is to verify that software meets a desired level of quality. Frequently, the term “testing” is associated with checking for functional correctness. However, in large, complex software systems with an established user-base, it is also important to verify system constraints such as backward compatibility, reliability, security, accessibility, usability. Kim Herzig from Microsoft explores these issues with the latest webinar on test Huddle.
A case study trying to answering the question "Are there statistical correlations between statement coverage and the number of failures detected?" and running a comparison between different reliability growth models
Testing As A Bottleneck - How Testing Slows Down Modern Development Processes...TEST Huddle
We often claim the purpose of testing is to verify that software meets a desired level of quality. Frequently, the term “testing” is associated with checking for functional correctness. However, in large, complex software systems with an established user-base, it is also important to verify system constraints such as backward compatibility, reliability, security, accessibility, usability. Kim Herzig from Microsoft explores these issues with the latest webinar on test Huddle.
A case study trying to answering the question "Are there statistical correlations between statement coverage and the number of failures detected?" and running a comparison between different reliability growth models
Using language workbenches and domain-specific languages for safety-critical ...Markus Voelter
Language workbenches support the efficient creation, integration, and use of domain-specific languages. Typically, they execute models by code generation to programming language code. This can lead to increased productivity and higher quality. However, in safety-/mission-critical environments, generated code may not be considered trustworthy, because of the lack of trust in the generation mechanisms. This makes it harder to justify the use of language workbenches in such an environment. In this paper, we demonstrate an approach to use such tools in critical environments. We argue that models created with domain-specific languages are easier to validate and that the additional risk resulting from the transformation to code can be mitigated by a suitably designed transformation and verification architecture. We validate the approach with an industrial case study from the healthcare domain. We also discuss the degree to which the approach is appropriate for critical software in space, automotive, and robotics systems.
Why Automated Testing Matters To DevOpsdpaulmerrill
“Automated testing is a pain in my ear! Why can’t QA get it right? Why do the tests keep breaking? And for Pete’s sake, stop blaming the infrastructure!”
…Ok, maybe you chose a different word than “ear”.
How often do you have thoughts like this? Daily?
Let’s talk about these frustrations, why they exist and how we can use them to improve our systems!
In this talk, Paul Merrill, founder and Principal Automation Engineer at Beaufort Fairmont explores why automated testing matters to DevOps. Join us to learn how automated testing can be a useful tool in the creation and release of your systems!
Effective Software Test Case Design Approach highlights typical wrong approaches to software test case design and focuses on an effective methodology in test case design from a collaborative approach.
Through the use of an example requirement/user story, this presentation highlights the "interactions" between the stakeholders, i.e. Product Owner, Developer, and Test Engineer in the development of user story acceptance criteria, details, test scope, and effective, consistent and valid test cases.
It's usually not enough time for improving comfort of code writing and product monitoring. But this is an important thing for making software products with high quality. IT society even made awesome tools for making our life easier and culture of software engineering continue growing.
4 Nisan 2015 tarihinde Kadir Has Üniversitesi'nde yapılan 9. Yazılım Teknolojileri Seminer etkinliğinde Eralp Erat'ın yaptığı TDD (Test Driven Design) sunumu
Fantastic Tests - The Crimes of Bad Test DesignWinston Laoh
Presentation given at the February 2019 Gridspace Talks event by Winston Laoh, Software Engineer in Test. The main purpose of the presentation was to inform the audience of the importance of good test design and factors that contribute to bad test design.
in this presentation i demonstrated all of the testings that can be done to improve the continuous delivery system in the development process of software in the end i have demonstrated what TDD is and what are it's benefits after the slides i demonstrated the TDD methodology by building a small project(a simple stack) with three rules of TDD.
Using language workbenches and domain-specific languages for safety-critical ...Markus Voelter
Language workbenches support the efficient creation, integration, and use of domain-specific languages. Typically, they execute models by code generation to programming language code. This can lead to increased productivity and higher quality. However, in safety-/mission-critical environments, generated code may not be considered trustworthy, because of the lack of trust in the generation mechanisms. This makes it harder to justify the use of language workbenches in such an environment. In this paper, we demonstrate an approach to use such tools in critical environments. We argue that models created with domain-specific languages are easier to validate and that the additional risk resulting from the transformation to code can be mitigated by a suitably designed transformation and verification architecture. We validate the approach with an industrial case study from the healthcare domain. We also discuss the degree to which the approach is appropriate for critical software in space, automotive, and robotics systems.
Why Automated Testing Matters To DevOpsdpaulmerrill
“Automated testing is a pain in my ear! Why can’t QA get it right? Why do the tests keep breaking? And for Pete’s sake, stop blaming the infrastructure!”
…Ok, maybe you chose a different word than “ear”.
How often do you have thoughts like this? Daily?
Let’s talk about these frustrations, why they exist and how we can use them to improve our systems!
In this talk, Paul Merrill, founder and Principal Automation Engineer at Beaufort Fairmont explores why automated testing matters to DevOps. Join us to learn how automated testing can be a useful tool in the creation and release of your systems!
Effective Software Test Case Design Approach highlights typical wrong approaches to software test case design and focuses on an effective methodology in test case design from a collaborative approach.
Through the use of an example requirement/user story, this presentation highlights the "interactions" between the stakeholders, i.e. Product Owner, Developer, and Test Engineer in the development of user story acceptance criteria, details, test scope, and effective, consistent and valid test cases.
It's usually not enough time for improving comfort of code writing and product monitoring. But this is an important thing for making software products with high quality. IT society even made awesome tools for making our life easier and culture of software engineering continue growing.
4 Nisan 2015 tarihinde Kadir Has Üniversitesi'nde yapılan 9. Yazılım Teknolojileri Seminer etkinliğinde Eralp Erat'ın yaptığı TDD (Test Driven Design) sunumu
Fantastic Tests - The Crimes of Bad Test DesignWinston Laoh
Presentation given at the February 2019 Gridspace Talks event by Winston Laoh, Software Engineer in Test. The main purpose of the presentation was to inform the audience of the importance of good test design and factors that contribute to bad test design.
in this presentation i demonstrated all of the testings that can be done to improve the continuous delivery system in the development process of software in the end i have demonstrated what TDD is and what are it's benefits after the slides i demonstrated the TDD methodology by building a small project(a simple stack) with three rules of TDD.
The anonymised slides from an old (but hopefully still relevant) talk on the case for placing a strategic focus on design testability. The material covers the technical, process and organisational considerations arising from such a strategy and is predominantly a summary of the ideas presented in Brett Pettichord's 2001 "Design For Testability' paper available here. The presentation makes a case for why a high level of design testability can be seen as a critical success factor in achieving sustained agility.
This outlines FIVE key application scenarios of validation using doSmartQA, a smart probing assistant to test deeply & rapidly.
It facilitates rapid testing in short sessions of Recon, Explore & Recoup, based on HyBIST -
‘Hypothesis Based Immersive Session Testing’, an intellectual practice of probing.
“Despite all the testing we do, field issues do not seem to abate. Sometimes it is a few serious issues that cause us to react intensely, sometimes it is a bunch of simple issues that make us consume bandwidth. Clearly the backlog is building up, with debts to be serviced, straining capacity to deliver new ideas.”
This is what I hear from senior engineering managers of product companies. How do you go about fixing this? Well, I have seen a flurry of activity to identify root cause(s) and address them. They help to set focus, but fizzle out.
Analysing 'quality of technical debt’ to understand types of issues that leak enables practical actions, rather than jumping into the ‘reason of why’ (root cause). Smart QA it is, to do failure analytics differently, to ‘tighten the purse’.
Technical debt is indeed a serious drain on engineering capacity, forcing one to fix issues at the expense of building revenue yielding new features. Smart failure analytics visualises problems well, enabling clear actions to strengthen practice and reduce debt significantly.
If you are “choked by technical debt”, then you may find our SmartQA consulting (stagsoftware.com/smartqa) interesting, where we unshackle your practice so that you can exploit technology.
"We track a lot of metrics related to progress of development and quality every sprint, like backlogs, technical debt, velocity, task status etc. What is not very evident is the 'quality of movement' i.e. how well done, so that we create less debt as we move. How can I get a better insight of the quality of tests done and a more objective measure of product quality?"
Extrinsic metrics are easier to measure and give visibility of direction, progress, speed and external feel of product quality. Intrinsic metrics are deeper, harder to measure but can give greater insight into the quality of work. Measuring this requires a good structure and organisation of test artefacts. The benefit - a greater insight into effectiveness of outcome and therefore lower technical debt & greater acceleration, don't you think?
Metrics can be classified as measuring work progress, work quality, product quality and practice quality. Except for the first one on work progress where we have a lot of measures facilitated by project and test management tools, the others depend on test organisation and clarity of types of issues to uncover. 'Quality Levels' based on HBT (Hypothesis Based Testing) provides a strong foundation for these, enabling you to assess potential test effectiveness, judge product quality objectively and fine tune practice quality .
If you are keen on "insightful quality metrics", then you may find our SmartQA consulting (stagsoftware.com/smartqa) interesting, where we unshackle your practice so that you can see clearly and do far better.
“As we embrace faster release cycles, testing has become a bottleneck. Yes, we have embraced automation as the way forward. We have a huge regression suite and therefore a big backlog for automation, a tough balance to speed up and yet maintain the fast paced release rhythm. What can I do?”
Automated tests are great to monitor a system’s health. Rather than just use regression as the candidate for automation, key flows that signify the pulse of a system's health are superior, don’t you think? And, this won’t create a huge backlog for automation, right?
Most often I have seen automation embraced as the solution to speed up testing. Conceptually correct it is, the problem is - what makes it worth the while to automate? Automated tests have to be in sync with the product and are therefore not a one time effort.
Choosing the right ones implies, it needs to be at the level of user flow, and be a clear indicator of health. Unless test scenarios are well structured and organised, choosing the right ones will turn out to be difficult, and ultimately weigh you down. It then becomes a pursuit of catching up with automation rather than making it work for you.
The goal is not 100% automation, it really is no leakage of defects. Automated tests are really ‘checks’ that assess key paths for good health (correctness) while intelligent human tests are focused on finding issues(robustness). A harmonious balance between these two enables clean code to be delivered without being weighed down by automation.
If you are “weighed down by automation“, then you may find our SmartQA consulting (stagsoftware.com/smartqa) interesting, where we unshackle your practice so that you can exploit technology.
Inspired by how the world is handling Covid19, this slideshare lists actions taken and criteria met to contain the pandemic and correlate this to how we can deliver clean code for large scale software systems. This article focuses on the process flow and criteria for delivering clean code.
The act of testing is a scientific exploration of a system done in three phases - RECONNAISSANCE to understand and plan, SEARCH to look for issues, REST&RECOVER to analyse and course correct. To enable the various activities in each phase to be done quickly and effectively, is where the SEVEN Thinking Tools outlined in this presentation. How to apply these tools in a session-based approach is also briefly outlined.
This article version of this SlideShare is available at http://bit.ly/7ThinkTools.
Agile and automation have been great enablers to doing tests faster. How we can accelerate further to accomplish more by doing less is the objective of this webinar.
“Left-shifting” by smart decomposition of dev testing aided by smart lightweight aids to perform rapid dev testing will be the takeaways of this webinar.
Three ideas to regression test smarter and outline THREE AIDS to do this.
AID #1: Fault propagation analyser - Figure out how what-to-retest by doing a smarter impact analysis using a scientific approach to understanding fault propagation due to change.
AID #2 : Automation analyser - Ensure scenarios are fit-to-automate so that they are easily scriptable and easily maintainable
AID #3 : Yield analyser : Figure out how much not to regress by analysing defect yields over time to understand what parts of the system have been hardened
Well, automation is an obvious choice, ensure that the scenarios are “fit enough for automation” so that you don’t end spend much effort maintaining the scripts to be in sync with every change.
Drawing inspiration from Atul Gawande's book "The checklist manifesto", T Ashok, CEO, STAG Software, explores at how we can exploit the power of checklist to delivering good quality code.
Drawing inspiration from Atul Gawande's book "The checklist manifesto", T Ashok, CEO, STAG Software, explores at how we can exploit the power of checklist to delivering good quality code.
This is the webinar recording on the topic ‘Test Case Immunity’- Optimize testing. In this webinar we have conveyed an interesting idea of measuring “Test Case Immunity” to logically assess what test cases to drop by so that we can 'do none'
This slide share contains the webinar, slides and the transcribed audio. The discussion outlines the entities to be considered for design, level based design, the optimal approach (think & prove/execute & evaluate) and finally design techniques.
Part#2 of Tri-webinar series consisting of three webinars commencing with 'How-to question to understand an user story and identify gaps', moving onto 'How-to set clear baseline' to ensure an effective strategy, and finally culminating with 'How-to design test scenarios/cases' using a scientific and disciplined approach
Part1 of Tri-webinar series consisting of three webinars commencing with 'How-to question to understand an user story and identify gaps', moving onto 'How-to set clear baseline' to ensure an effective strategy, and finally culminating with 'How-to design test scenarios/cases' using a scientific and disciplined approach
"Language shapes the way you think" was the topic of the talk presented by T Ashok, CEO STAG Software, to a group of test professionals at a Pune-based IT services and solutions provider on June 16, 2014.
HBT Innovation Series webinar presented by T Ashok, Architect-HBT and Founder & CEO, STAG Software on the topic - Deliver Superior Outcomes Using HBT Visualization Tool - on Feb 26, 2014.
This presentation on Hypothesis Based Testing (HBT) was delivered by Mr Satvik Kini, Associate Quality Manager, Suite Test Centre, SAP Labs India Pvt. Ltd at STeP-IN Forum webinar on Dec 19, 2013.
This presentation was part of the talk delivered by T Ashok Founder & CEO STAG Software at the HSTC 2013: "Think Testing" Conference on Nov 21 & 22 at Hyderabad.
STAG Software presented a webinar on Aug 21, 2013 on the topic - Improving Defect Yield - a three step approach". The webinar was hosted by T Ashok, Founder & CEO, STAG Software and Architect of HBT.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
2. Abstract
We collect and analyze metrics to make decisions, without understanding
the underlying assumptions and therefore make potentially dangerous
decisions. Some of the aspects that we track are (1) Quality (2) Progress
and (3) Adequacy/Effectiveness.
Some of the typical measurements are defect rates, distributions, density,
coverage, distribution of test cases by test entity, test cases executed etc.
Though each of these seem logical, my contention is that we may not
remember the assumptions underlying these.
For example, does a low defect rate mean a good product or poor testing?
Does high code coverage mean great testing? How do we know what the
desired density is ? Also some of the measurements are focused on the low
level activity (for example test case execution progress), which could be
useless as we lose sight the overall goal ( i.e. how is quality progressing?).
Measurements like test case immunity, test case growth, quality growth are
possibly better indicators ensuring that we stay focussed on the goal of
effective testing. The intent is to understand the assumptions and ensure we
do not forget these and also look at interesting measurements that could be
better indicators.
Copyright 2011 STAG Software Private Limited www.stagsoftware.com
3. Some typical measurements
Code coverage
Requirement coverage Design
#Test cases(TC) categorized by attributes
Test productivity
Execution
TC execution rate
Pass/Fail rate
Defect arrival rate
Defect distribution Assessment
Defect density
Copyright 2011 STAG Software Private Limited www.stagsoftware.com
4. Assumptions/ Dangers
Code coverage
Requirement coverage Design
#Test cases(TC) categorized by attributes
# Intent is to ensure that behaviors of all LOC have indeed been examined
>>>> Only functional behaviors are examined. It is assumed that *all* behaviors
have been coded.
# Intent is to ensure that all requirements will indeed be examined
>>>> What if we have just *One TC *per requirement? Necessary condition but
not sufficient.
# Intent is to have just enough test cases that will ensure that our examination
will be successful. We categorize these in multiple ways : (1) by features (2)by
priority etc..
>>>> Assumption is that *many* is good, how do we measure quality of TC i.e.
*defect-type yielding* ability?
Copyright 2011 STAG Software Private Limited www.stagsoftware.com
5. Assumptions/ Dangers
Test productivity
Execution
TC execution rate
#We focus on the test activity here, how are we doing with respect to the intended goal?
>>>>Assumption - more quickly we execute, more closer we are to the goal.
#We focus on the test activity here, how are we doing with respect to the intended goal?
>>>>Assumption - more quickly we execute, more closer we are to the goal. But what is the
*yield* ?
>>>>We need to execute *all* TC, gives us comfort , but is it effective?But what is the *yield* ?
>>>>We need to execute *all* TC, gives us comfort , but is it effective?
Copyright 2011 STAG Software Private Limited www.stagsoftware.com
6. Assumptions/ Dangers
Pass/Fail rate
Defect arrival rate
Assessment
Defect distribution
Defect density
#Assumption that *more failures* means high risk to delivery
>>>>We are assuming that * #failures* is kinda indicative of risk. Should we not analyze from a
defect type perspective?
# Higher the rate of arrival of defects, the worse it seems; lower it is , better it seems...
>>>>Assumption that *defect acceleration* is important, need to sure that we have the right
goal i.e. test cases need to be complete- how sure are we about this?
# Distribution by severity - More severe the defect, the worse it is..
>>>Do we know what a *good* distribution is?
# More dense the defect *clump* is, higher the risk - seems logical
>>>> Do we know what kind of clump it is ? Simple ‘fungal infection’ or ‘cancerous’
>>>> Distribution/Generalized metric (e.g density) need a clear UCL/LCL to infer
Copyright 2011 STAG Software Private Limited www.stagsoftware.com
7. Think...
• Shift from measuring pure activities to outcome/goals
• Do we measure “Test Case Immunity” i.e. over time, which of the test
cases have lost the power to uncover defects?
• Would test case growth(needs to be qualified) be an interesting metric
that indicates that we are constantly expanding the net?
• Rather than using pure defect metrics to serve as indicators of quality,
should we not shift to defect type metrics to give us a objective sense of
quality?
Copyright 2011 STAG Software Private Limited www.stagsoftware.com