The document provides testing criteria for evaluating the accessibility of Android applications for users with disabilities. It includes tests in the following categories: usage with limited vision, usage without color perception, usage with limited hearing, usage without vocal capability, usage with limited dexterity, and usage with limited cognition. Each category contains tests related to navigation, control, feedback, display, adjustments/settings, and external devices. The tests are marked as either critical, which an application must pass, or warning, where it can accumulate points for minor issues. An application passes overall if it passes all critical tests and accumulates 3 or fewer warning points.
The burgeoning use of mobile devices has created enormous opportunities for organizations to leverage mobile to increase sales, advertise products, and collaborate with internal and external resources. However, with increasing usage, the need to perform testing on these devices is increasing significantly. This is not an easy task considering the number of devices, device operating systems, and operating system versions. To manage the number of variations, organizations rely on mobile testing tools to support their testing efforts. David Dang shares his experiences analyzing numerous mobile testing tool platforms for a prominent shopping network. Learn how identifying the "right" mobile testing tool depends on multiple factors such as supported devices, level of testing, resources, and required integration with other tools. Take back to share with your team a review of common tools on the market and the pros and cons of each.
Software Testing Interview Questions For Experiencedzynofustechnology
The document discusses various topics related to software testing interviews for experienced testers. It covers reliability testing, handling bugs, challenges of thorough testing, testing without complete requirements, differences between retesting and regression testing, challenges of software testing, types of functional testing, and more. Key points made include that it is impossible to thoroughly test a program due to subjective requirements and too many inputs/paths, the importance of regression testing when modules are updated, and differences between bugs, defects, and errors.
Mr. Avinash Birnale VP - Technology, Quality @ Endeavour Software Technologies delivered a session on Smart Skills that every developer must have at Mobile Monday Ahmendabd on 22nd March 2014.
Assess with your blinkers off widen your scope for tool evaluationIndium Software
This document discusses software quality assurance and testing tool evaluation. It presents a refined process for evaluating testing tools that involves understanding the application's technologies, architecture, and test cases before selecting the top 4 tools and conducting proofs of concept. Two case studies demonstrate benefits like increased test coverage, faster defect detection, and cost savings when using the refined evaluation approach. Overall, the document advocates evaluating tools with an open mind and comprehensive understanding of the application to select the most effective automation solution.
Mobile App Testing Strategy by RapidValue SolutionsRapidValue
There has been an increase in the adoption of smartphones, tablets and several mobile devices with the passage of time. And it has led to an enormous growth of mobile applications in recent years. Mobile device is considered to be the primary medium of interaction for the customers and also, businesses worldwide. And mobile applications are, actually, driving the communication.
People, generally, do not give much importance to mobile application testing because of its expensive nature. But it is very much essential to ensure that the consumers have a great experience, every time they use the application.
The aim of mobile application testing should be to acquire knowledge about the quality of the service that you are offering. Does it work properly or not? Will it provide services as per the expectation of the customer? These questions need to be answered to ensure that the customer comes back to you, for your service again. Mobile testing is becoming more and more complicated and complex with each passing day. Strategies are invented and used to simplify the mobile application testing.
This presentation primarily addresses, the strategy to be adopted in mobile app testing, the types of mobile app testing and the stages to be followed, before the application is set live.
The burgeoning use of mobile devices has created enormous opportunities for organizations to leverage mobile to increase sales, advertise products, and collaborate with internal and external resources. However, with increasing usage, the need to perform testing on these devices is increasing significantly. This is not an easy task considering the number of devices, device operating systems, and operating system versions. To manage the number of variations, organizations rely on mobile testing tools to support their testing efforts. David Dang shares his experiences analyzing numerous mobile testing tool platforms for a prominent shopping network. Learn how identifying the "right" mobile testing tool depends on multiple factors such as supported devices, level of testing, resources, and required integration with other tools. Take back to share with your team a review of common tools on the market and the pros and cons of each.
Software Testing Interview Questions For Experiencedzynofustechnology
The document discusses various topics related to software testing interviews for experienced testers. It covers reliability testing, handling bugs, challenges of thorough testing, testing without complete requirements, differences between retesting and regression testing, challenges of software testing, types of functional testing, and more. Key points made include that it is impossible to thoroughly test a program due to subjective requirements and too many inputs/paths, the importance of regression testing when modules are updated, and differences between bugs, defects, and errors.
Mr. Avinash Birnale VP - Technology, Quality @ Endeavour Software Technologies delivered a session on Smart Skills that every developer must have at Mobile Monday Ahmendabd on 22nd March 2014.
Assess with your blinkers off widen your scope for tool evaluationIndium Software
This document discusses software quality assurance and testing tool evaluation. It presents a refined process for evaluating testing tools that involves understanding the application's technologies, architecture, and test cases before selecting the top 4 tools and conducting proofs of concept. Two case studies demonstrate benefits like increased test coverage, faster defect detection, and cost savings when using the refined evaluation approach. Overall, the document advocates evaluating tools with an open mind and comprehensive understanding of the application to select the most effective automation solution.
Mobile App Testing Strategy by RapidValue SolutionsRapidValue
There has been an increase in the adoption of smartphones, tablets and several mobile devices with the passage of time. And it has led to an enormous growth of mobile applications in recent years. Mobile device is considered to be the primary medium of interaction for the customers and also, businesses worldwide. And mobile applications are, actually, driving the communication.
People, generally, do not give much importance to mobile application testing because of its expensive nature. But it is very much essential to ensure that the consumers have a great experience, every time they use the application.
The aim of mobile application testing should be to acquire knowledge about the quality of the service that you are offering. Does it work properly or not? Will it provide services as per the expectation of the customer? These questions need to be answered to ensure that the customer comes back to you, for your service again. Mobile testing is becoming more and more complicated and complex with each passing day. Strategies are invented and used to simplify the mobile application testing.
This presentation primarily addresses, the strategy to be adopted in mobile app testing, the types of mobile app testing and the stages to be followed, before the application is set live.
Mobile Application Testing - White PaperJade Global
Mobile applications are the sole reason for the rise in popularity of smartphones. The ease and convenience of mobile applications usage has created a huge dependency on it. Over the years, the number and variety of consumer and enterprise mobile applications has grown phenomenally.
Usability testing for mobile apps - Touch Tour ChennaiRIA RUI Society
This document discusses usability testing for mobile applications. It begins with an agenda and overview of the mobile app market. It then defines usability and explains why usability testing is important for mobile apps. Guidelines are provided for conducting usability tests, including testing 5-8 users, measuring success rates and task times, and identifying errors. Both formative testing during development and summative testing after release are described. Examples are given of evaluating different aspects of usability.
The document discusses mobile application testing, including the types of mobile applications like web, native, and hybrid applications; different mobile platforms like Android and iOS; how to test aspects like installation, features, and network connectivity; challenges in mobile application testing like lack of devices; and how to overcome challenges using emulators and automation tools.
The Essentials of Mobile App Testing and MonitoringMobilePundits
Mobile Technology is transforming the way people use their cell phones. Although demand is highest for consumer apps, enterprise applications are evolving too, allowing businesses to work more productively In this document we look at how the testing of mobile applications helps to achieve quality. Here we explore a typical way that an app is developed, look at the testing stages involved, answer some of the frequent questions concerning testing, and provide a definition of the common testing terms.
Training Opportunity: Certified Mobile App Professional (CMAP) TestingITpreneurs
Downloaded an app onto your phone lately? With app downloads mounting into the billions each year, there is a need for general standards in the creation and usage of mobile technology – and this provides a great opportunity for training companies: Meet Certified Mobile App Professional (CMAP). CMAP is a certification from the International Software Quality Institute (ISQI) specifically designed for anyone involved in Mobile App Testing. A great next step for those with Foundation Level certificate as “ISTQB Certified Tester” (CTFL).
Agenda:
Introduction: What will you learn today?
About CMAP
The Challenges of Mobile App Testing
Business Needs
Go-To-Market Strategies for CMAP
How to Get Started
Questions & Answers
The document outlines different types of mobile testing including usability testing, functional testing, performance testing, memory leakage testing, interrupt testing, laboratory testing, installation testing, localization testing, certification testing, cross platform testing, and automation testing. It describes the mobile application development lifecycle of analyzing requirements, designing interfaces, implementing features, accepting feedback, and testing the application. The types of mobile testing are defined to ensure the application meets requirements, has a usable interface, functions properly on different devices and networks, and can be installed, updated, and localized across platforms.
Mobile application testing faces several key challenges, including the variety of mobile devices with different hardware capabilities, diverse mobile platforms and operating systems, and different mobile networks. Effective mobile application testing requires testing across multiple devices, platforms, and networks. It also requires the use of emulators and cloud-based testing to efficiently test applications. The document outlines various types of mobile application testing including compatibility, interface, services, and performance testing. It discusses test strategies, common issues encountered in testing, and the importance of automation to address the complexities of mobile application testing.
BStriker's multi-language ALTM test management tool seamlessly integrates into software ecosystems and collaborates with existing tools. It allows centralized test cycle management for all platforms independent of the software development lifecycle used. With recommended international testing standards, it improves quality assurance. Innovative features enable teams to perform better using risk-based approaches and receive real-time reports. The automated issue detection and fixing will disrupt the market.
The document discusses building an enterprise-grade mobile testing strategy. It recommends starting with tools that fit your current workflow and then transforming your strategy over time to improve coverage, reliability, and governance. A balanced strategy considers factors like time to market, device and network coverage, and existing tools and processes. The document describes using HP UFT Mobile for test automation across real mobile devices and emulators at scale.
Your mobile test plan can't be adhoc. Watch this webinar and learn how to start testing more methodically to provide test coverage that reflects your users and ensures a great digital experience.
We'll cover the most important insights from the latest issue of the Digital Mobile Test Coverage Index, including benchmarks, geographies and the latest devices and OS's that app developers and testers need to know.
This white paper discusses an approach to shift mobile accessibility testing left in the development cycle by implementing unit testing. It proposes testing each unit of code for accessibility before integration to find issues earlier. The approach includes developing automated unit test cases for accessibility and integrating them into a continuous accessibility certification framework to regularly check for issues during development. This allows issues to be fixed sooner, reducing rework and helping release accessible applications on time and on budget.
A 5 minute guide to delivering Flawless Mobile AppsCygnet Infotech
Scale up the performance and functioning of mobile applications by implementing dedicated mobile test automation practices with thoughtful planning and strategizing.
This document outlines the design and implementation of a bar-code scanning price comparison mobile application. The app is intended to help elderly, disabled, and unemployed individuals easily compare product prices by scanning barcodes. It will allow users to scan barcodes to get product information, prices, and nutritional facts from a database. The implementation section discusses screenshots of the system, the database design, and testing the app through white box, black box, and user evaluation methods to identify defects and ensure requirements are met.
Software development involves planning, implementing, testing, and maintaining software. The main stages are gathering requirements, designing and coding the software, testing it, and deploying it to users. Once deployed, software often requires maintenance to fix bugs and add new features. Effective development requires clearly defining requirements, managing scope, and providing user training and support during and after deployment.
12 considerations for mobile testing (march 2017)Antoine Aymer
The document is a brochure that outlines 12 key considerations for choosing a mobile application testing solution. It discusses the importance of testing apps on real devices and emulators, enabling remote access to devices, supporting both manual and automated testing, testing under realistic network conditions, simulating common user interruptions, using object ID recognition, and testing the functional, performance, and security aspects of apps. It positions HPE's mobile testing solutions as addressing all 12 considerations by supporting testing on devices/emulators, remote access, manual/automated testing, network simulation, interruption simulation, object ID recognition, and functional, performance, and security testing. It emphasizes the importance of an end-to-end solution and expertise in mobile testing.
Evaluation is an important process for designers to determine if their work is useful and desirable to users. It can be done at different stages of product development through various techniques, including quick and dirty evaluations, usability testing, field studies, and predictive evaluations. These techniques involve observing users, asking for user and expert opinions, testing user performance, and modeling task performance. Common evaluation methods are questionnaires, interviews, observation, and measuring time and errors during usability testing. Likert scales are often used to analyze results. Evaluation allows designers to improve their products and ensure they meet user needs.
This document contains the resume of R. Balakrishnan. It summarizes his objective, skills, experience, education and projects. His objective is to utilize his skills and be creatively involved in work that contributes to organizational growth. He has over 4 years of experience in software testing and skills like manual testing, test management tools, bug tracking and mobile testing. He has worked on projects testing printer drivers and software for Xerox, and a textile management system for Tejas Fashion. He is proficient in tools like ClearQuest, Aptest and Magen Test. He holds an MCA degree from PSNA College of Engineering & Technology.
In this session you will learn:
Testing Concepts and Manual Testing
Overview of Testing Life Cycle
Testing Methodologies
Dynamic Testing
Black Box Testing
White Box Testing
Gray Box Testing
Unit Testing
Integration Testing
System Testing
Regression Testing
User Acceptance Testing (UAT)
Stop guessing, start testing – mobile testing done right - Timo EuteneuerJAXLondon_Conference
The document discusses the importance of mobile app testing and some of the challenges involved. It notes that 48% of users won't use an app again after encountering issues. Various testing approaches like manual testing and automation are presented, and it is argued that automation is key for high test coverage and reducing time to market. Challenges like device, OS, and screen size fragmentation are also covered.
Mobile Application Testing Training PresentationMobiGnosis
Mobile Application Testing Training Presentation in Bangalore by experienced Professionals in Industry. Get a FREE Demo Now. Visit http://www.mobignosis.com/mobile-testing-training/
JIMS Vasant KunjII is the Top institute for BCA. JIMS is one of the Best BCA Colleges in Delhi which offers best placements in Top IT Companies in Delhi NCR. It is amongst the top A+ Category highest ranked colleges in Delhi, provides 3 years Regular Degree from UGC Approved University.
This unit of Software Testing is a part of BCA 5th sem syllabi.
User Acceptance Testing (UAT) involves actual software users testing a system in real-world scenarios to ensure it meets requirements and solves users' problems. UAT occurs after functional, system, and regression testing and validates that a system works for users as intended. During UAT, users document any issues found and developers fix bugs to prepare the system for release. UAT focuses on validating user experience rather than just finding defects and is important for demonstrating a system supports business needs in a real world environment.
Mobile Application Testing - White PaperJade Global
Mobile applications are the sole reason for the rise in popularity of smartphones. The ease and convenience of mobile applications usage has created a huge dependency on it. Over the years, the number and variety of consumer and enterprise mobile applications has grown phenomenally.
Usability testing for mobile apps - Touch Tour ChennaiRIA RUI Society
This document discusses usability testing for mobile applications. It begins with an agenda and overview of the mobile app market. It then defines usability and explains why usability testing is important for mobile apps. Guidelines are provided for conducting usability tests, including testing 5-8 users, measuring success rates and task times, and identifying errors. Both formative testing during development and summative testing after release are described. Examples are given of evaluating different aspects of usability.
The document discusses mobile application testing, including the types of mobile applications like web, native, and hybrid applications; different mobile platforms like Android and iOS; how to test aspects like installation, features, and network connectivity; challenges in mobile application testing like lack of devices; and how to overcome challenges using emulators and automation tools.
The Essentials of Mobile App Testing and MonitoringMobilePundits
Mobile Technology is transforming the way people use their cell phones. Although demand is highest for consumer apps, enterprise applications are evolving too, allowing businesses to work more productively In this document we look at how the testing of mobile applications helps to achieve quality. Here we explore a typical way that an app is developed, look at the testing stages involved, answer some of the frequent questions concerning testing, and provide a definition of the common testing terms.
Training Opportunity: Certified Mobile App Professional (CMAP) TestingITpreneurs
Downloaded an app onto your phone lately? With app downloads mounting into the billions each year, there is a need for general standards in the creation and usage of mobile technology – and this provides a great opportunity for training companies: Meet Certified Mobile App Professional (CMAP). CMAP is a certification from the International Software Quality Institute (ISQI) specifically designed for anyone involved in Mobile App Testing. A great next step for those with Foundation Level certificate as “ISTQB Certified Tester” (CTFL).
Agenda:
Introduction: What will you learn today?
About CMAP
The Challenges of Mobile App Testing
Business Needs
Go-To-Market Strategies for CMAP
How to Get Started
Questions & Answers
The document outlines different types of mobile testing including usability testing, functional testing, performance testing, memory leakage testing, interrupt testing, laboratory testing, installation testing, localization testing, certification testing, cross platform testing, and automation testing. It describes the mobile application development lifecycle of analyzing requirements, designing interfaces, implementing features, accepting feedback, and testing the application. The types of mobile testing are defined to ensure the application meets requirements, has a usable interface, functions properly on different devices and networks, and can be installed, updated, and localized across platforms.
Mobile application testing faces several key challenges, including the variety of mobile devices with different hardware capabilities, diverse mobile platforms and operating systems, and different mobile networks. Effective mobile application testing requires testing across multiple devices, platforms, and networks. It also requires the use of emulators and cloud-based testing to efficiently test applications. The document outlines various types of mobile application testing including compatibility, interface, services, and performance testing. It discusses test strategies, common issues encountered in testing, and the importance of automation to address the complexities of mobile application testing.
BStriker's multi-language ALTM test management tool seamlessly integrates into software ecosystems and collaborates with existing tools. It allows centralized test cycle management for all platforms independent of the software development lifecycle used. With recommended international testing standards, it improves quality assurance. Innovative features enable teams to perform better using risk-based approaches and receive real-time reports. The automated issue detection and fixing will disrupt the market.
The document discusses building an enterprise-grade mobile testing strategy. It recommends starting with tools that fit your current workflow and then transforming your strategy over time to improve coverage, reliability, and governance. A balanced strategy considers factors like time to market, device and network coverage, and existing tools and processes. The document describes using HP UFT Mobile for test automation across real mobile devices and emulators at scale.
Your mobile test plan can't be adhoc. Watch this webinar and learn how to start testing more methodically to provide test coverage that reflects your users and ensures a great digital experience.
We'll cover the most important insights from the latest issue of the Digital Mobile Test Coverage Index, including benchmarks, geographies and the latest devices and OS's that app developers and testers need to know.
This white paper discusses an approach to shift mobile accessibility testing left in the development cycle by implementing unit testing. It proposes testing each unit of code for accessibility before integration to find issues earlier. The approach includes developing automated unit test cases for accessibility and integrating them into a continuous accessibility certification framework to regularly check for issues during development. This allows issues to be fixed sooner, reducing rework and helping release accessible applications on time and on budget.
A 5 minute guide to delivering Flawless Mobile AppsCygnet Infotech
Scale up the performance and functioning of mobile applications by implementing dedicated mobile test automation practices with thoughtful planning and strategizing.
This document outlines the design and implementation of a bar-code scanning price comparison mobile application. The app is intended to help elderly, disabled, and unemployed individuals easily compare product prices by scanning barcodes. It will allow users to scan barcodes to get product information, prices, and nutritional facts from a database. The implementation section discusses screenshots of the system, the database design, and testing the app through white box, black box, and user evaluation methods to identify defects and ensure requirements are met.
Software development involves planning, implementing, testing, and maintaining software. The main stages are gathering requirements, designing and coding the software, testing it, and deploying it to users. Once deployed, software often requires maintenance to fix bugs and add new features. Effective development requires clearly defining requirements, managing scope, and providing user training and support during and after deployment.
12 considerations for mobile testing (march 2017)Antoine Aymer
The document is a brochure that outlines 12 key considerations for choosing a mobile application testing solution. It discusses the importance of testing apps on real devices and emulators, enabling remote access to devices, supporting both manual and automated testing, testing under realistic network conditions, simulating common user interruptions, using object ID recognition, and testing the functional, performance, and security aspects of apps. It positions HPE's mobile testing solutions as addressing all 12 considerations by supporting testing on devices/emulators, remote access, manual/automated testing, network simulation, interruption simulation, object ID recognition, and functional, performance, and security testing. It emphasizes the importance of an end-to-end solution and expertise in mobile testing.
Evaluation is an important process for designers to determine if their work is useful and desirable to users. It can be done at different stages of product development through various techniques, including quick and dirty evaluations, usability testing, field studies, and predictive evaluations. These techniques involve observing users, asking for user and expert opinions, testing user performance, and modeling task performance. Common evaluation methods are questionnaires, interviews, observation, and measuring time and errors during usability testing. Likert scales are often used to analyze results. Evaluation allows designers to improve their products and ensure they meet user needs.
This document contains the resume of R. Balakrishnan. It summarizes his objective, skills, experience, education and projects. His objective is to utilize his skills and be creatively involved in work that contributes to organizational growth. He has over 4 years of experience in software testing and skills like manual testing, test management tools, bug tracking and mobile testing. He has worked on projects testing printer drivers and software for Xerox, and a textile management system for Tejas Fashion. He is proficient in tools like ClearQuest, Aptest and Magen Test. He holds an MCA degree from PSNA College of Engineering & Technology.
In this session you will learn:
Testing Concepts and Manual Testing
Overview of Testing Life Cycle
Testing Methodologies
Dynamic Testing
Black Box Testing
White Box Testing
Gray Box Testing
Unit Testing
Integration Testing
System Testing
Regression Testing
User Acceptance Testing (UAT)
Stop guessing, start testing – mobile testing done right - Timo EuteneuerJAXLondon_Conference
The document discusses the importance of mobile app testing and some of the challenges involved. It notes that 48% of users won't use an app again after encountering issues. Various testing approaches like manual testing and automation are presented, and it is argued that automation is key for high test coverage and reducing time to market. Challenges like device, OS, and screen size fragmentation are also covered.
Mobile Application Testing Training PresentationMobiGnosis
Mobile Application Testing Training Presentation in Bangalore by experienced Professionals in Industry. Get a FREE Demo Now. Visit http://www.mobignosis.com/mobile-testing-training/
JIMS Vasant KunjII is the Top institute for BCA. JIMS is one of the Best BCA Colleges in Delhi which offers best placements in Top IT Companies in Delhi NCR. It is amongst the top A+ Category highest ranked colleges in Delhi, provides 3 years Regular Degree from UGC Approved University.
This unit of Software Testing is a part of BCA 5th sem syllabi.
User Acceptance Testing (UAT) involves actual software users testing a system in real-world scenarios to ensure it meets requirements and solves users' problems. UAT occurs after functional, system, and regression testing and validates that a system works for users as intended. During UAT, users document any issues found and developers fix bugs to prepare the system for release. UAT focuses on validating user experience rather than just finding defects and is important for demonstrating a system supports business needs in a real world environment.
The document discusses various aspects of software testing including the need for testing, types of testing, testing tools, the testing life cycle, and determining when to stop testing. It notes that software testing is a planned process used to identify correctness, completeness, security and quality of software. The testing life cycle involves requirements analysis, test planning, writing and reviewing test cases, bug logging and tracking, and closing and reopening bugs.
Testing and Debugging Flutter Apps: A Comprehensive Approach QSS Technosoft Inc.
QSS Technosoft Inc is an experienced provider of comprehensive software development services. We are experts in application development and testing & debugging for Flutter apps. Our in-depth knowledge of the Flutter framework provides us with the proficiency to create specialised apps that align perfectly with your business goals.
In today's digital age, where mobile applications have become an integral part of our daily lives, ensuring the quality and reliability of these apps is of utmost importance.
The document provides an overview of software testing and quality assurance concepts. It defines key terms like software, software testing, quality assurance, and validation and verification. It describes the objectives and types of testing like static and dynamic analysis. It also discusses quality factors, metrics, statistical quality control methods like control charts, and total quality management approaches.
The document discusses testing and distribution of mobile apps. It provides an overview of:
1) A mobile maturity model that organizations can use to assess their mobile strategy and capabilities across different areas including testing.
2) The importance of testing throughout the app development lifecycle from definition to development to acceptance. It describes various testing types like unit, integration, and usability testing.
3) How automated testing can help with frequent verification but still requires manual testing. It provides examples of unit and functional automated tests.
4) The different phases of testing in a project including definition to set testing requirements, development where testing is integrated, and acceptance testing by the customer.
UAT involves developing a test strategy, scenarios, and scripts. The test strategy outlines the testing approach, including people, tools, procedures, and support. Test scenarios describe situations to test. Test scripts define actual inputs and expected results. An effective test strategy is specific, practical, and justified, clarifying major tasks and challenges. It identifies the type and timing of testing, critical success factors, and tradeoffs.
In this session you will learn:
Overview of Testing Life Cycle
Testing Methodologies
Black Box Testing
White Box Testing
Gray Box Testing
Integration Testing
System Testing
Regression Testing
User Acceptance Testing (UAT)
For more information, click here:
https://www.mindsmapped.com/courses/quality-assurance/software-testing-tutorial/
Find out more about quality assurance training and specifically about Testing concepts and manual testing. Topics covered in this session are:
Overview of Testing Life Cycle
Testing Methodologies
Black Box Testing
White Box Testing
Gray Box Testing
Integration Testing
System Testing
Regression Testing
User Acceptance Testing (UAT)
For more information, visit: https://www.mindsmapped.com/courses/quality-assurance/quality-assurance-training-learn-manual-and-automation-testing/
In this quality assurance training, you will learn Testing Concepts and Manual Testing. Topics covered in this session are:
• Overview of Testing Life Cycle
• Testing Methodologies
• Black Box Testing
• White Box Testing
• Gray Box Testing
• Integration Testing
• System Testing
• Regression Testing
• User Acceptance Testing (UAT)
For more information, visit this link: https://www.mindsmapped.com/courses/quality-assurance/software-testing-training-beginners-and-intermediate-level/
In this session you will learn:
Overview of Testing Life Cycle
Testing Methodologies
Black Box Testing
White Box Testing
Gray Box Testing
Integration Testing
System Testing
Regression Testing
User Acceptance Testing (UAT)
For more information: https://www.mindsmapped.com/courses/quality-assurance/qa-software-testing-training-for-beginners/
In this quality assurance training session, you will learn Testing Concepts and Manual Testing. Topics covered in this course are:
• Overview of Testing Life Cycle
• Testing Methodologies
• Black Box Testing
• White Box Testing
• Gray Box Testing
• Integration Testing
• System Testing
• Regression Testing
• User Acceptance Testing (UAT)
To know more, visit this link: https://www.mindsmapped.com/courses/quality-assurance/software-testing-quality-assurance-qa-training-with-hands-on-exercises/
GUI testing focuses on testing the graphical user interface of an application to ensure functionality, usability, and consistency. It allows comprehensive testing of user interactions and can be automated. Installation testing verifies correct installation and setup on various platforms to identify issues. Error handling testing checks that applications appropriately guide users through potential errors.
The document discusses software testing concepts including the importance of testing, the testing life cycle, types of testing, quality assurance and control, and bug reporting. It provides definitions and descriptions of key testing terms like errors, bugs, faults, failures, test plans, test cases, unit testing, integration testing, system testing, and regression testing. Testing roles like testers, QA leads, and test analysts are also outlined.
A regression testing approach is needed to evaluate the overall working of the application after it has undergone a recent change.
- Examines the functionality of the upgrade
- Ensures that the change does not interfere with existing features
- Identifies any errors or bugs in the application's existing functionalities
- It helps determine bugs in the newly pushed code
## Strategy for regression testing - what to test, how often, and more
Regression testing strategy depends on several key factors, like how often developers upgrade the application, how significant the new change is, and what existing sections it could affect.
Here are some tried and tested proven strategies that you could follow during regression testing:
- Measure or change the scale of the upgrade to determine how likely it would affect the application.
- When introducing automation testing, outline the test cases and scenarios to know which should be automated and manually tested.
- The regression testing approach must cover all the possible test cases and impacted functionalities.
- Focus on the testing process, technology, and roles when automating regression testing
- Perform risk analysis based on the size of your business/project and its complexity, along with its importance
## Tips for managing regression risks
- Proactively identify and assess regression risks before starting the regression testing process. You can then focus all your efforts on the most critical areas.
- Use risk mitigation strategies to reduce the impact of identified risks. For example, if a particular threat could result in data loss, you could create backups to mitigate the risk.
- Use a structured approach for managing regression risks, such as a risk registry or risk management plan; this will help ensure that all threats are captured and tracked.
- Communicate any potential impacts of regression risks to stakeholders to make informed decisions about the release schedule.
## Steps to perform regression testing:
- Prepare for manual and automated tests: This involves getting the required tools and resources ready, such as test data, test cases, test scripts, and more.
- Identify which changes or upgrades on existing modules of the application will impact its functionalities: You need to specifically identify which areas of the application will be affected by the changes or upgrades to focus your testing efforts during regression testing on those areas.
- Use manual and automated tests accordingly: Once you have identified the impacted functionalities, you can use both manual and automation tests to validate that the changes or upgrades have not adversely affected those functionalities.
Some of the most common regressions that need testing include functionalities such as login, search, and checkout. To detect these regressions, you can use different methods such as checking the application's output against expected results, performing functional tests, and using automated tools.
The document discusses challenges with Voluntary Product Accessibility Templates (VPATs). It provides an overview of what a VPAT is, including that it is a tool to document a product's conformance with Section 508 accessibility standards. However, it notes that VPATs can be difficult for consumers to interpret and compare products. Specifically, VPATs may lack detail, have subjective assessments, and it is hard to verify claims without the product. The document provides tips for evaluating a VPAT, such as reviewing the language, using accessibility checkers, testing with a screen reader, meeting with the company, and getting user feedback.
What is the Difference between Functional Testing and UAT?Serena Gray
User experience, usability, functionality and performance are the key parameters that are thoroughly checked in UAT. In this article, you will get to know about the differences between functional testing and UAT.
Top 7 reasons why software testing is crucial in SDLCSLAJobs Chennai
Software Testing is an important thing in Software Development Life Cycle (SDLC) because errors can occur anywhere in the application and some of them remain undiscovered until deployment. The final product with bugs brings many difficulties such as redevelopment, wastage of cost and time, poor functionality, and many more. Software testing proves the quality of the product that helps to gain more customer satisfaction by meeting their requirements in an application or a product.
Similar to Accessibility Testing Criteria for Android Applications (20)
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Infrastructure Challenges in Scaling RAG with Custom AI modelsZilliz
Building Retrieval-Augmented Generation (RAG) systems with open-source and custom AI models is a complex task. This talk explores the challenges in productionizing RAG systems, including retrieval performance, response synthesis, and evaluation. We’ll discuss how to leverage open-source models like text embeddings, language models, and custom fine-tuned models to enhance RAG performance. Additionally, we’ll cover how BentoML can help orchestrate and scale these AI components efficiently, ensuring seamless deployment and management of RAG systems in the cloud.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Accessibility Testing Criteria for Android Applications
1. Page 1 of 61
Accessibility Testing Criteria for Android Applications
version 1.2: July 2015
__________________________________________________
Introduction
Context and Background
This set of Accessibility Testing Criteria is intended to be used to check the
accessibility of an application for users with impairments in one or more categories of
vision, colour perception, hearing, speech, dexterity, cognition on an Android Device.
The basis for this is built on input from a number of sources including:
o Mobile Manufacturers Forum GARI website,
o W3C WCAG 2.0 recommendations on accessibility,
o Android developer site recommendations on accessibility,
o AT&T recommendation on website accessibility,
o AQuA members experience and their accessibility teams.
More details and links to sources are at the end of this document.
Accessibility needs to be designed into applications from the outset, therefore use of
this set of Testing Criteria is to be encouraged at the prototype development stage,
before delivery of a finished product, and when further updates or amendments are
made to the application, to ensure that accessibility is not broken by subsequent
changes. As the tests only look at the Application’s accessibility, the Application
should also be tested against the AQuA Testing Criteria for Android applications
(which focuses on the general usability of the Application).
Not all applications will be needed to be tested for all categories, and requirements
across the different categories may be contradictory (e.g. enhancing the sound
based feedback is great for improving accessibility for usage with limited vision, but
may load too much information to audio for usage with limited hearing). Application
providers will need to consider how to approach accessibility for different categories
of impairment, as appropriate for their application and their audience. Therefore,
rather than simply running all tests on any application, it will be necessary to
understand which usage conditions are appropriate to an application, and only run
those sections of the tests that are relevant to its intended use and audience.
The Application will be considered to have passed the testing if it has passed for one
category. All passes will be categorized appropriately. (e.g. Pass for Usage With
Limited Vision). Note that a judgement call will need to be made on whether to test
for usage with a particular limitation, i.e. it may not always be appropriate for a
sound-based application to be tested for usage with limited hearing – although,
conversely, users with limited hearing may still be able to enjoy part of the frequency
range of a music player application.
General points to note
Applications intended to meet the needs of users with specific impairments should
not be developed in isolation from the users they intend to support; rather, wherever
possible the developer should maintain contact with appropriate organizations or
communities, and should seek to involve representatives in design and testing.
2. Page 2 of 61
These tests are written so they can be performed by a tester without impairments.
However, we would advise that, if possible, there should be at least one cycle of
testing carried out by people with the specific impairment or impairments which the
tests address.
When using this set of Testing Criteria, please bear in mind that it is not designed
purely as a set of prescriptive step-by-step tests. Rather, it is a guide to questioning
assumptions that may be made in the design and development stages, and of finding
ways to check that those assumptions do not reduce usability. Such a process
normally enhances usability.
No specific distinction is made between accessibility features provided by an
application and those provided by the device’s operating system, as the focus is on
the end user experience as a whole, which should ideally be seamless and
consistent regardless of the source of the function. Within that context, it should be
borne in mind that these tests are only intended for evaluating the behavior of
applications or an update to them.
The TalkBack accessibility service will need to be installed and enabled on the
Android Device. This is usually pre-installed on devices with Android 4.0 or higher,
where it can be found as an option under Accessibility in the device Settings menu.
For devices without the option, it can be installed from Google Play
(https://play.google.com/store/apps/details?id=com.google.android.marvin.talkback).
General advice on Android accessibility can be found at the Android Accessibility
Help pages - https://support.google.com/accessibility/android#topic=6007234.
Help on using TalkBack can be found at the Enable TalkBack page -
https://support.google.com/accessibility/android/answer/6007100?hl=en-
GB&ref_topic=3529932.
It is recommended that the tester goes through the tutorial for TalkBack if they are
unfamiliar with it. Injudicious use can render the phone hard to use as familiar touch
controls no longer function as expected – the tester should particularly note that
single-tap only selects an item while TalkBack is in use, and double-tap is always
required to action a control or item.
Greg Jotham
Chief Quality Auditor, AQuA
3. Page 3 of 61
Structure of the Criteria
The accessibility tests fall into basic categories (with a few sub-categories) for users
with restrictions in that area.
The categories or sub-categories are set out in this document as sections of tests.
The Application may be tested against each section, or against several sections
below:
• Usage with limited vision (includes usage without vision for the purposes of
this document)
• Usage without perception of colour / minimising photosensitive seizure
triggers
• Usage with limited hearing (includes usage without hearing for the
purposes of this document)
• Usage without vocal capability
• Usage with limited manipulation or strength (includes usage with limited
reach for the purposes of this document)
• Usage with limited cognition
Each category is broken down into a set of functional areas:
• Navigation
• Control (execution of actions)
• Feedback
• Display
• Adjustments / Settings
• External devices
External devices include but are not limited to:
• Keypads
• D-pads
• Joystick
• Braille displays
• Induction loop
• Hearing aids
• Headsets
• Switch control
• Sip and puff control systems
• Location beacons
4. Page 4 of 61
Critical and Warning levels of Tests
We recognise that many of the tests that are performed do not produce a binary
result. They are often subjective leaving the interpretation to the tester. It is unfair
therefore to fail an application for one minor error that may be down to a tester’s
opinion.
To account for this, the individual tests are each marked as either Critical or Warning.
Critical Level Tests
As the name suggests, a Critical level test must be passed.
If the Application fails the test then the Application has an overall fail.
Warning Level Tests
For a test which is considered Warning level, we have allowed for four different
results; pass, annoying, difficult and impossible.
These warning levels are described as follows;
Pass = the Application has passed the test. There are no issues
Annoying = a minor error has occurred with the Application - e.g. one or two typos
that would make the Application not perfect but still very useable
Difficult = a more serious issue has occurred with the Application e.g. multiple typos
making the Application difficult to use but not impossible
Impossible = a very serious issue has occurred with the Application - the errors are
so bad as to make the Application unusable.
Once all appropriate tests have been carried out, points should be attributed
according to the following scale.
Warning levels:
Annoying = 1 points
Difficult = 2 points
Impossible = 4 points
For the Application to pass, the errors must not add up to more than 3 points.
4 points or more is a failure.
Severity of error Warning test type Critical test type
No error 0 points 0 points
Annoying error 1 point
Difficult error 2 points
Impossible error 4 points
Fail test 5 points
(As an example, the Application could have 3 annoying results, or 1 difficult and 1
annoying and still pass.)
5. Page 5 of 61
Contents
1. Usage with limited vision .........................................................................................7
1.1. Navigation ..................................................................................................... 7
1.1.1. Install ..................................................................................................... 7
1.1.2. Audio prompts for all content.................................................................. 8
1.2. Control (execution of actions) ........................................................................ 9
1.2.1. Explore By Touch................................................................................... 9
1.2.2. Gesture control interaction ................................................................... 10
1.3. Feedback..................................................................................................... 11
1.3.1. Feedback - sufficient............................................................................ 11
1.3.2. Feedback – audio elements differentiated............................................ 12
1.3.3. Feedback – audio prompt overloading / underloading.......................... 13
1.4. Display......................................................................................................... 14
1.4.1. Display element sizing / portrayal......................................................... 14
1.4.2. Display orientation ............................................................................... 15
1.5. Adjustments / Settings................................................................................. 16
1.5.1. Contrast Control................................................................................... 16
1.5.2. Volume control..................................................................................... 17
1.5.3. Expanded font sizes............................................................................. 18
1.6. External Devices.......................................................................................... 19
1.6.1. Alert on connection / disconnection...................................................... 19
1.6.2. Navigation with external directional control device ............................... 20
1.6.3. Operation with external switch control.................................................. 21
1.6.4. Interaction with location beacons ......................................................... 22
2. Usage without perception of colour / Minimising photosensitive seizure triggers....23
2.1. Display......................................................................................................... 23
2.1.1. Adjust colour scheme........................................................................... 23
2.1.2. Monochrome presentation ................................................................... 24
2.1.3. Photosensitive seizure triggers ............................................................ 25
2.2. Feedback..................................................................................................... 26
2.2.1. Feedback elements – colour schemes................................................. 26
2.3. Adjustments / Settings................................................................................. 27
3. Usage with limited hearing.....................................................................................28
3.1. Navigation ................................................................................................... 28
3.1.1. Visual navigation.................................................................................. 28
3.2. Control (execution of actions) ...................................................................... 29
3.2.1. Visual notification of actions................................................................. 29
3.3. Feedback..................................................................................................... 30
3.3.1. No audio-only feedback ....................................................................... 30
3.3.2. Visual alerts ......................................................................................... 31
3.4. Display......................................................................................................... 32
3.4.1. Notification bar..................................................................................... 32
3.5. Adjustments / Settings................................................................................. 33
3.5.1. Replacement of audible alerts with visual ones.................................... 33
3.6. External Devices.......................................................................................... 34
3.6.1. Support for external volume control with hearing assistance devices... 34
3.6.2. Detection of external hearing aids and visual notification ..................... 35
4. Usage without vocal capability...............................................................................36
4.1. Navigation ................................................................................................... 36
4.1.1. Navigation not limited to speech input.................................................. 36
6. Page 6 of 61
4.2. Control (execution of actions) ...................................................................... 37
4.2.1. Control actions not limited to speech input........................................... 37
4.3. Feedback..................................................................................................... 37
4.4. Display......................................................................................................... 38
4.4.1. Text display.......................................................................................... 38
4.5. Adjustments / Settings................................................................................. 39
4.5.1. Text-to-speech configuration & quality ................................................. 39
4.6. External Devices.......................................................................................... 40
4.6.1. External Text-To-Speech devices ........................................................ 40
4.6.2. External text displays ........................................................................... 41
5. Usage with limited manipulation or strength..........................................................42
5.1. Navigation ................................................................................................... 42
5.1.1. Alternative inputs for navigation ........................................................... 42
5.2. Control (execution of actions) ...................................................................... 43
5.2.1. Assistive options for control.................................................................. 43
5.2.2. Pressure-related input options ............................................................. 44
5.2.3. Multi-finger control options ................................................................... 45
5.3. Feedback..................................................................................................... 46
5.4. Display......................................................................................................... 46
5.4.1. Dialogue boxes & timeouts................................................................... 46
5.4.2. Displayed information – cognitive overload .......................................... 47
5.5. Adjustments / Settings................................................................................. 48
5.5.1. Touch-related settings.......................................................................... 48
5.6. External Devices.......................................................................................... 49
5.6.1. Connection and operation with external devices .................................. 49
6. Usage with limited cognition..................................................................................50
6.1. Navigation ................................................................................................... 50
6.1.1. Help information................................................................................... 50
6.1.2. Clarity of single action.......................................................................... 51
6.1.3. Limiting the number of options / choices presented to the user............ 52
6.1.4. Language complexity ........................................................................... 53
6.2. Control (execution of actions) ...................................................................... 54
6.2.1. Presentation of control elements.......................................................... 54
6.3. Feedback..................................................................................................... 55
6.3.1. Clarity of feedback ............................................................................... 55
6.4. Display......................................................................................................... 56
6.4.1. Text fonts and sizes ............................................................................. 56
6.4.2. Colours for reading comprehension ..................................................... 57
6.4.3. Dialogue boxes & timeouts................................................................... 58
6.5. Adjustments / Settings................................................................................. 59
6.5.1. Reading level options........................................................................... 59
6.6. External Devices.......................................................................................... 59
7. USAGE WITH LIMITED VISION
Page 7 of 61
Testing Criteria
1. Usage with limited vision
1.1. Navigation
1.1.1. Install
Test ID
1.1.1
Test Title
Install
Critical
Test Description
Check that the Application installs with the correct information for
TalkBack and screen readers.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled when testing compatibility with TalkBack.
Audio prompting should also be checked with a screen reader if this is able to read
installation prompts, and during screen reader testing TalkBack should be turned off
unless specifically required by the screen reader.
Ensure that speech output pitch and speed settings, where provided, are set to values
that produce optimum results for the tester.
Testing Steps
1. Install the Application.
2. Navigate the focus to the Application icon on the screen.
3. Listen to the name of the Application.
RESULT:
The Application name should be clearly and understandably spoken aloud and
should be unambiguous in the target language or languages of the device.
Result of Test
PASS FAIL
8. USAGE WITH LIMITED VISION
Page 8 of 61
1.1.2. Audio prompts for all content
Test ID
1.1.2
Test Title
Audio prompts for all content
Critical
Test Description
Verify that user interface controls that provide information (graphics or
text) or allow user action, have clear and accurate audio descriptions
when TalkBack is enabled or a screen reader used, and controls are
focused.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled when testing compatibility with TalkBack.
Audio prompting should also be checked with a screen reader where practical, and
during screen reader testing TalkBack should be turned off unless specifically required
by the screen reader.
Ensure that speech output pitch and speed settings, where provided, are set to values
that produce optimum results for the tester.
Testing Steps
1. Launch the Application.
2. Use directional controls to move the focus between Application layout
elements.
3. Check that all controls, images, text and other elements that present
information visually have meaningful audio descriptions, which present the
same amount of information that is available visually.
4. Check that all labels and tags have meaningful names.
5. Check that non-informational elements (e.g. UI decoration) do not generate
audio feedback, as this could be confusing.
RESULT:
A single audio description of displayed elements is given, which enables
understanding of the Application without use of the visual interface.
Result of Test
PASS FAIL
9. USAGE WITH LIMITED VISION
Page 9 of 61
1.2. Control (execution of actions)
1.2.1. Explore By Touch
Test ID
1.2.1
Test Title
Explore By Touch – Android
Accessibility
Critical
Test Description
Verify that Explore By Touch prompts and controls are correctly
presented and actioned.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled.
Ensure that Explore By Touch is enabled in TalkBack settings.
If the Application is intended to be used with TalkBack and a screen reader in
combination, this use case should also be checked.
Ensure that speech output pitch and speed settings, where provided, are set to values
that produce optimum results for the tester.
Testing Steps
1. Launch the Application.
2. Move a finger all over the screen to hear audio descriptions that identify
screen elements and controls.
3. Check that tags and labels have meaningful names.
4. Double-tap to open applications, menus and other controls.
5. Navigate within the Application to ensure that all navigation prompts and
selections are valid, menu structures are correctly presented and usable,
and exiting the Application or putting it into background and restoring it can
still be properly executed.
6. If Help information is provided, access it and check that the content is
presented correctly through audio.
7. Swipe with two fingers to change screens and scroll within lists.
8. Check that application-specific gestures requiring swiping, zooming or
scrolling continue to work properly after TalkBack, or TalkBack and Explore
By Touch are enabled, or that an alternative interface for these functions is
provided.
9. Where speech input and control are provided, check that these operate
correctly and the level of mis-recognition of input is no greater than would
be found in general use of speech recognition outside of the Application
(e.g. in existing facilities like Google Search or Navigation).
RESULT:
All controls should be correctly identified, and it should be possible for a user
to navigate and use the Application without reference to the visual content.
Speech control and input where provided should be sufficiently functional for
normal usage.
Result of Test
PASS FAIL
10. USAGE WITH LIMITED VISION
Page 10 of 61
1.2.2. Gesture control interaction
Test ID
1.2.2
Test Title
Gesture control interaction
Critical
Test Description
Actions controlled by gestures should be usable when TalkBack is
enabled.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled.
Testing Steps
1. Launch the Application.
2. Explore all the functionality of the Application.
3. Check that application-specific gestures, such as zooming images,
scrolling lists, swiping between pages or navigating carousel controls all
continue to work when TalkBack is enabled. If these gestures do not
function, then an alternative interface for these actions must be provided.
RESULT:
All actions normally controlled by gestures should still be provided by a
suitable interface when TalkBack is enabled.
Result of Test
PASS FAIL
11. USAGE WITH LIMITED VISION
Page 11 of 61
1.3. Feedback
1.3.1. Feedback - sufficient
Test ID
1.3.1
Test Title
Feedback - sufficient
Critical
Test Description
Verify that audio / haptic feedback is sufficient for usage conditions.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled when testing compatibility with TalkBack.
Audio prompting should also be checked with a screen reader if practical, and during
screen reader testing, TalkBack should be turned off unless specifically required by
the screen reader.
Testing Steps
1. Launch the Application.
2. Explore navigation and use of the Application’s functions.
3. Check that auditory prompts are provided before use of a function.
4. Check that haptic and audio confirmation of a function being used is
available.
5. Check that spinning wheels, progress bars and other progress indicators
have a suitably informative non-visual equivalent.
6. Check that dialog boxes have full audio description.
7. Check that when it is possible to scroll to data above or below the
displayed screen, rising or falling tones are used to give feedback on
current position when scrolling
RESULT:
All audio / haptic feedback should be sufficiently informative, and should be
adequate to make the Application usable without reference to the visual
content.
Result of Test
PASS FAIL
12. USAGE WITH LIMITED VISION
Page 12 of 61
1.3.2. Feedback – audio elements differentiated
Test ID
1.3.2
Test Title
Feedback – audio elements differentiated
Critical
Test Description
Verify that audio feedback of multiple elements is not confusingly
similar.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled when testing compatibility with TalkBack.
Audio prompting should also be checked with a screen reader, and during screen
reader testing TalkBack should be turned off unless specifically required by the screen
reader.
Testing Steps
1. Launch the Application.
2. Explore any elements that present elements in grouped areas, such as lists
or contact details.
3. Check that closely related elements are given sufficiently different audio
description such that it is possible to easily distinguish between them without
reference to visual content. E.g. in a contacts list entry, the contact picture,
telephone, email etc should not be simply labeled with the contact’s name
identically for each element.
RESULT:
Audio feedback should be sufficiently differentiated that all elements within
groups and list can be identified correctly without reference to visual content.
Result of Test
PASS FAIL
13. USAGE WITH LIMITED VISION
Page 13 of 61
1.3.3. Feedback – audio prompt overloading / underloading
Test ID
1.3.3
Test Title
Feedback – audio prompt overloading /
underloading
Critical
Test Description
Verify that audio prompting is neither too little nor too great for clarity.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled when testing compatibility with TalkBack.
Audio prompting should also be checked with a screen reader, and during screen
reader testing TalkBack should be turned off unless specifically required by the screen
reader.
Testing Steps
1. Launch the Application.
2. Explore grouped controls in each part of the Application.
3. Check that closely related controls provide an appropriate level of audio
information that enables users to understand and act on a screen element.
Too much or too little prompting can make it difficult to understand and use a
control.
RESULT:
The level of audio prompting should be sufficient for easy use of all controls,
but not so great that confusion can arise through inability to retain / recall
multiple complex prompts.
Result of Test
PASS FAIL
14. USAGE WITH LIMITED VISION
Page 14 of 61
1.4. Display
1.4.1. Display element sizing / portrayal
Test ID
1.4.1
Test Title
Display element sizing / portrayal
Critical
Test Description
Verify that display elements are correctly sized and depicted that they
allow users with limited vision to distinguish between them.
Required for:
All applications.
Testing Note
http://www.afb.org/info/living-with-vision-loss/reading-and-writing/making-print-more-
readable/235 provides some context as to what is acceptable in printed formats, and
from which it should be possible to make a judgement about the displayed size on
electronic device screens.
Testing Steps
1. Launch the Application.
2. Examine all display elements.
3. Check that these elements are of sufficient size, spacing and distinctive
design as to allow users with limited vision to distinguish between them,
understand their purpose, and operate them in a touchscreen environment.
RESULT:
All visual elements should be sized, spaced and designed so that they are
practical for use with restricted vision.
Result of Test
PASS FAIL
15. USAGE WITH LIMITED VISION
Page 15 of 61
1.4.2. Display orientation
Test ID
1.4.2
Test Title
Display orientation
Critical
Test Description
The Application must be usable in supported orientations and make
limitations clear before use.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. If the Application is restricted to a single orientation (portrait / landscape) of
the device, it should give an audio announcement of this at launch. It is
permissible for there to be a user setting to turn this announcement off, as
long as it is on by default at installation.
3. Check that switching between orientations does not cause display
elements to fail test 1.4.1 (Display element sizing / portrayal).
4. Check that audio prompts continue to function correctly after change of
orientation.
RESULT:
The Application should remain usable regardless of any changes caused by
device reorientation.
Result of Test
PASS FAIL
16. USAGE WITH LIMITED VISION
Page 16 of 61
1.5. Adjustments / Settings
1.5.1. Contrast Control
Test ID
1.5.1
Test Title
Contrast Control
Critical
Test Description
The Application should offer different display contrast levels.
Required for:
All applications.
Testing Note
To pass this test, the Application should offer a minimum contrast ratio between
background and text / elements of 4.5:1 to meet the WCAG2.0 level AA standard.
See http://www.w3.org/TR/WCAG20-TECHS/G18.html for details of the W3C
recommendations.
For the purposes of this test, it is not necessary to formally measure the luminance of
elements. It will be sufficient if the contrast appears, when judged by eye, to be not
less than the 4.5:1 samples provided at http://trace.wisc.edu/contrast-ratio-examples/.
The set of examples in the section headed “Text Samples that just pass at several
Contrast Ratios” may provide the easiest comparison for judging by eye.
Testing Steps
1. Launch the Application.
2. If the Application does not offer a high-contrast display by default, open its
Settings and confirm that there is at least one high-contrast display option.
3. Check that the either the provided default or the high-contrast display
option provides a contrast level that meets the recommended standard.
RESULT:
The Application should offer a high-contrast display option that gives a
significant contrast increase when used.
Result of Test
PASS FAIL
17. USAGE WITH LIMITED VISION
Page 17 of 61
1.5.2. Volume control
Test ID
1.5.2
Test Title
Volume control
Critical
Test Description
Audio prompt volume should be sufficient and remain responsive to
device controls.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled when testing compatibility with TalkBack.
Audio prompting should also be checked with a screen reader, and during screen
reader testing TalkBack should be turned off unless specifically required by the screen
reader.
If the Application provides audio prompts without TalkBack, the test should be
performed twice, once with TalkBack on, and once with it off.
Testing Steps
1. Launch the Application.
2. Examine the Application’s Settings and audio controls for evidence of any
option to boost the maximum volume above the maximum normally
offered, e.g. for speaker announcement of prompts in noisy situations. If
such an option exists, the test should be performed once with it off, and
once with it at the maximum setting.
3. Check that it is possible to achieve an adequate volume level for prompts
to be clearly heard. Volume level without any boost should be at least
sufficient for normal usage indoors, but where the Application would
normally be expected to be used outdoors at least part of the time, the
volume level achieved – with boost if offered – should be sufficient for this
usage. If the Application is expressly designed for use only with
headphones and makes this clear to the user, adequate headphone
volume will be sufficient.
4. Check that the volume delivered responds promptly to the operation of the
device volume keys.
5. Where a boosted volume is such that there is a risk of hearing damage if
the device is held close to the ear, check that “press and hold” of the
device Volume Down key can reduce the volume to safe levels within a
second or less.
RESULT:
Audio prompt volume should be sufficient for the intended use and should
respond quickly to device controls.
Result of Test
PASS FAIL
18. USAGE WITH LIMITED VISION
Page 18 of 61
1.5.3. Expanded font sizes
Test ID
1.5.3
Test Title
Expanded font sizes
Critical
Test Description
Where large fonts are used, all elements should be clearly presented.
Required for:
All applications.
Testing Note
Select the largest font size / clearest font face available in the device’s Settings –
Display – Screen Display menu for this test.
Testing Steps
1. Launch the Application.
2. Open the Application’s Settings and identify whether larger fonts are
offered. If they are, select the largest one available. If not, continue with the
default font provided by the device / application.
3. Check that the largest font available in the App is suitable for users with
visual impairment. Use of screen zoom is permissible so long as this does
not render the Application difficult to use correctly.
4. Check that when an extra large font is used, that neither individual
characters nor blocks of text are misaligned, distorted, overlaid or
otherwise presented unreadably.
RESULT:
The Application should be able to correctly present text in a font size suitable
for users with visual impairment.
Result of Test
PASS FAIL
19. USAGE WITH LIMITED VISION
Page 19 of 61
1.6. External Devices
1.6.1. Alert on connection / disconnection
Test ID
1.6.1
Test Title
Alert on connection / disconnection
Critical
Test Description
Connection or disconnection of external devices should generate an
informative alert.
Required for:
All applications.
Testing Note
1. Ensure that the Android accessibility service ‘TalkBack’ is installed on the device,
and enabled.
2. If a sufficiently informative alert is generated by the device OS rather than the
Application, this will be equally acceptable and constitute a Pass.
Testing Steps
1. Launch the Application.
2. Connect, and then disconnect any external devices which can be used with
the Application (e.g. D-pad, joystick, switch control, Braille display, puff and
sip controls).
3. Check that a clearly understandable confirmation of connection /
disconnection is given, whether by spoken audio, distinct tones that clearly
identify the two states, or other feedback (e.g. haptic), and that these
confirmations are not easily confused with other confirmations received
during use of the device in general, and the Application in particular.
RESULT:
Confirmation of the connection / disconnection of external devices should
leave the user in no doubt as to the status of each device, without needing to
reference any visual confirmations that may be produced.
Result of Test
PASS FAIL
20. USAGE WITH LIMITED VISION
Page 20 of 61
1.6.2. Navigation with external directional control device
Test ID
1.6.2
Test Title
Audio prompts with external directional
control device
Critical
Test Description
Audio prompts should be satisfactory when using external directional
control devices instead of touchscreen control.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled.
If the Application provides audio prompts without TalkBack, the test should be
performed twice, once with TalkBack on, and once with it off.
Testing Steps 1 & 2 can be reversed if the Application is designed to be used with a
specific device that has to be connected while it is running.
Testing Steps
1. Connect an external directional control device such as a D-pad or joystick.
2. Launch the Application.
3. Navigate through the Application’s functions and menus, interact with
controls, and attempt to understand the Application’s layout and operation.
4. Check that the audio prompts and confirmations generated by the
Application while using the external directional control device are sufficient
for the Application to be usable.
RESULT:
When using an external directional control device, audio prompts and
confirmations should be sufficient that the Application can be used without
reference to the visual context for navigation.
Result of Test
PASS FAIL
21. USAGE WITH LIMITED VISION
Page 21 of 61
1.6.3. Operation with external switch control
Test ID
1.6.3
Test Title
Audio prompts with external switch
control
Critical
Test Description
Audio prompts should be satisfactory when using external switch
control instead of touchscreen control.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled.
If the Application provides audio prompts without TalkBack, the test should be
performed twice, once with TalkBack on, and once with it off.
Testing Steps 1 & 2 can be reversed if the Application is designed to be used with a
specific device that has to be connected while it is running.
Testing Steps
1. Connect an external switch control.
2. Launch the Application.
3. Navigate through the Application’s functions and menus, interact with
controls, and attempt to understand the Application’s layout and operation.
4. Check that the audio prompts and confirmations generated by the
Application while using the external switch control are sufficient for the
Application to be usable.
RESULT:
When using an external switch control, audio prompts and confirmations
should be sufficient that the Application can be used without reference to the
visual context for the function or functions that the switch control provides.
Result of Test
PASS FAIL
22. USAGE WITH LIMITED VISION
Page 22 of 61
1.6.4. Interaction with location beacons
Test ID
1.6.4
Test Title
Interaction with location beacons
Critical
Test Description
Non-visual prompts, messages or tones should be presented correctly
and in a timely manner when interacting with location beacons.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled.
If the Application provides audio prompts without TalkBack, the test should be
performed twice, once with TalkBack on, and once with it off.
Testing Steps
1. Launch the Application.
2. Ensure the Application is correctly set up to make connection with external
beacons, whether by Bluetooth or other means, and prompts are set to be
triggered by the interaction.
3. Check that the non-visual prompts, messages or tones generated by the
Application while interacting with location beacons are sufficient for the
Application to be usable.
4. Check that information is never absent or presented incorrectly in any way
that would be likely to put the user at risk.
5. Check that disconnection or exiting from the area of beacon coverage is
unambiguously signaled to the user without excessive delay.
RESULT:
When interacting with external beacons, non-visual prompts and confirmations
should be sufficient that the Application can be used safely and without
reference to any visual context.
Result of Test
PASS FAIL
23. USAGE WITHOUT PERCEPTION OF COLUR / MINIMISING PHOTOSENSTIVE
SEIZURE TRIGGERS
Page 23 of 61
2. Usage without perception of colour / Minimising
photosensitive seizure triggers
2.1. Display
2.1.1. Adjust colour scheme
Test ID
2.1.1
Test Title
Adjust colour scheme
Critical
Test Description
Display options should be offered to suit differing colour perception.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Open the available Settings.
3. Check that a range of display schemes are offered that will suit users with
differing colour perception, e.g. red/green distinction.
RESULT:
The Application should offer display schemes that are suited to users with
differing colour perception, and which will render all displayed elements in
acceptably distinct shades.
Result of Test
PASS FAIL
24. USAGE WITHOUT PERCEPTION OF COLUR / MINIMISING PHOTOSENSTIVE
SEIZURE TRIGGERS
Page 24 of 61
2.1.2. Monochrome presentation
Test ID
2.1.2
Test Title
Monochrome presentation
Critical
Test Description
Display options should include a monochrome format.
Required for:
All applications.
Testing Note
Note that when using monochrome formats care should be taken to avoid strongly
contrasting geometric patterns, which can act as a trigger for photosensitive epilepsy
(see also the test covering these issues).
Testing Steps
1. Launch the Application.
2. Open the available Settings.
3. Check that a monochrome display option is available, and the contrast
levels chosen are suitable for users with all kinds of colour perception
limitation.
RESULT:
The Application should offer a monochrome display option that delivers
acceptable levels of contrast for users with all types of colour perception
limitation.
Result of Test
PASS FAIL
25. USAGE WITHOUT PERCEPTION OF COLUR / MINIMISING PHOTOSENSTIVE
SEIZURE TRIGGERS
Page 25 of 61
2.1.3. Photosensitive seizure triggers
Test ID
2.1.3
Test Title
Photosensitive seizure triggers
Critical
Test Description
Display schemes and content should avoid using known photosensitive
seizure triggers.
Required for:
All applications.
Testing Note
1. Where multiple display schemes are available in the Application, presentation
should be checked in each scheme.
2. Known triggers include (but are not limited to) lights or images which flash at rates
between 3 Hz (flashes per second) and 60 Hz, or contrasting light and dark
geometric patterns, such as black/white stripes or checks.
3. Embedded video should not include multiple flash photography.
Testing Steps
1. Launch the Application.
2. Check all displayed pages and content for photosensitive seizure triggers.
RESULT:
Display schemes and content should avoid using known photosensitive
seizure triggers.
Result of Test
PASS FAIL
26. USAGE WITHOUT PERCEPTION OF COLUR / MINIMISING PHOTOSENSTIVE
SEIZURE TRIGGERS
Page 26 of 61
2.2. Feedback
2.2.1. Feedback elements – colour schemes
Test ID
2.2.1
Test Title
Feedback elements – colour schemes
Critical
Test Description
Colour choices for all displayed feedback elements should be suitable
for users with colour perception limitations.
Required for:
All applications.
Testing Note
Where multiple colour display schemes are available in the Application, presentation
of feedback elements should be checked in each scheme.
Testing Steps
1. Launch the Application.
2. Check the presentation of feedback elements such as spinning wheels,
progress bars, shadows, error messages and action confirmations.
3. Check that colour is never used alone to convey information (for example,
using red text to show a message is an error). Text or icons must convey
the full message content without needing perception of a particular colour.
RESULT:
Available display colour schemes for feedback elements should cater for the
needs of users with all common limitations of colour perception.
Result of Test
PASS FAIL
27. USAGE WITHOUT PERCEPTION OF COLUR / MINIMISING PHOTOSENSTIVE
SEIZURE TRIGGERS
Page 27 of 61
2.3. Adjustments / Settings
Not Applicable
28. USAGE WITH LIMITED HEARING
Page 28 of 61
3. Usage with limited hearing
3.1. Navigation
3.1.1. Visual navigation
Test ID
3.1.1
Test Title
Visual navigation
Critical
Test Description
Navigation within the Application should not rely on audio prompts.
Required for:
All applications.
Testing Note
Before performing this test, set the device to silent mode; or if this setting is not
available, mute all sounds and notifications within the Settings menu.
Testing Steps
1. Launch the Application.
2. Explore the Settings and functionality of the Application.
3. Check that it is possible to navigate all elements of the Application without
use of audio prompts or confirmations.
RESULT:
Visual and tactile (e.g. haptic) feedback within the Application should be
sufficient to enable navigation of all features without use of audio responses.
Result of Test
PASS FAIL
29. USAGE WITH LIMITED HEARING
Page 29 of 61
3.2. Control (execution of actions)
3.2.1. Visual notification of actions
Test ID
3.2.1
Test Title
Visual notification of actions
Critical
Test Description
Actions within the Application should be notified visually.
Required for:
All applications.
Testing Note
Before performing this test, set the device to silent mode; or if this setting is not
available, mute all sounds and notifications within the Settings menu.
If the handset and the Application both support multi-application windowing, have
another application open at the same time as the Application under test, to observe
whether onscreen messages appear within the Application’s display area.
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that all pending and completed actions are notified visually by
onscreen messages within the Application display area.
RESULT:
All actions within the Application should be notified visually and within the
screen area used for display of the Application.
Result of Test
PASS FAIL
30. USAGE WITH LIMITED HEARING
Page 30 of 61
3.3. Feedback
3.3.1. No audio-only feedback
Test ID
3.3.1
Test Title
No audio-only feedback
Critical
Test Description
The Application should not rely on audio-only feedback at any point.
Required for:
All applications.
Testing Note
Before testing, ensure that the device is not set to silent mode, and that sounds and
audible notifications are turned on in the Settings menu (both in the device and
Application where present), so that audible and visual feedback can be compared.
Testing Steps
1. Launch the Application.
2. Explore the functions of the Application.
3. Check that visual feedback is always offered at any time that audio
feedback is present.
RESULT:
The Application should always present the user with visual equivalents to any
audio feedback given.
Result of Test
PASS FAIL
31. USAGE WITH LIMITED HEARING
Page 31 of 61
3.3.2. Visual alerts
Test ID
3.3.2
Test Title
Visual alerts
Critical
Test Description
All alerts issued by the Application should have a visual component.
Required for:
All applications.
Testing Note
For this test, alerts are treated as being different from feedback in that they are either
unanticipated or asynchronous – that is, they may occur an indeterminate amount of
time after the action that will eventually trigger them.
Before testing, ensure that the device is not set to silent mode, and that sounds and
audible notifications are turned on in Settings, so that audible and visual alerts can be
compared.
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that all generated alerts have a sufficiently informative visual
component so that they can be properly understood without reference to
the audible component.
RESULT:
All alerts issued by the Application should be capable of drawing attention to
themselves and being understood by the user, without that user having
reference to any audio component of the alert.
Result of Test
PASS FAIL
32. USAGE WITH LIMITED HEARING
Page 32 of 61
3.4. Display
3.4.1. Notification bar
Test ID
3.4.1
Test Title
Notification bar
Critical
Test Description
The Notification Bar should be used for persistent feedback or alerts.
Required for:
All applications.
Testing Note
Before performing this test, set the device to silent mode; or if this setting is not
available, mute all sounds and notifications within the Settings menu.
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that all feedback and alerts which should require user interaction to
dismiss make proper use of notifications posted to the Notification Bar. The
Application should never rely solely on constant or repeating audio alarms
to solicit action from the user.
RESULT:
All persistent feedback or alarms / alerts that require user dismissal should
have a visual component posted to the Notification Bar.
Result of Test
PASS FAIL
33. USAGE WITH LIMITED HEARING
Page 33 of 61
3.5. Adjustments / Settings
3.5.1. Replacement of audible alerts with visual ones
Test ID
3.5.1
Test Title
Replacement of audible alerts with visual
ones
Critical
Test Description
The Application should offer a simple one-step option for replacing all
audio alerts with visual ones.
Required for:
All applications.
Testing Note
1. Before testing, ensure that the device is not set to silent mode, and that sounds
and audible notifications are turned on in the device Settings.
2. Standard alerts and notifications generated outside the Application are not
included in the scope of this test.
Testing Steps
1. Launch the Application.
2. Check that the Application’s Settings menu or other functionality offers an
easily-understood single-step process to replace all audio alerts with visual
ones.
3. Check that when this option is selected, the Application does not generate
any audible alerts.
RESULT:
The Application should offer a simple way for all of its audio alerts to be
replaced with visual ones, so that a user with hearing limitation does not
inadvertently generate audio alerts in situations where they may be
unwelcome or inconvenient.
Result of Test
PASS FAIL
34. USAGE WITH LIMITED HEARING
Page 34 of 61
3.6. External Devices
3.6.1. Support for external volume control with hearing assistance devices
Test ID
3.6.1
Test Title
Support for external volume control with
hearing assistance devices
Critical
Test Description
The Application should respect and not counteract the operation of
external volume controls used with hearing assistance devices.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Use the functionality of the Application whilst making adjustments of
volume in both the Application and the external device.
3. The Application must never interact uncontrollably with the external device
as far as volume control is concerned, and where the Application is
intended to interact with volume settings on an external device it should
always respond to them correctly and obey any mute / un-mute command
when issued.
RESULT:
The Application should interact correctly with any volume control used on an
external hearing assistance device (e.g. amplifier, filter or hearing aid).
Result of Test
PASS FAIL
35. USAGE WITH LIMITED HEARING
Page 35 of 61
3.6.2. Detection of external hearing aids and visual notification
Test ID
3.6.2
Test Title
Detection of external hearing aids and
visual notification
Critical
Test Description
The Application should correctly detect and connect to external hearing
aids and assistance devices, and provide visual confirmation of this.
Required for:
All applications.
Testing Note
Set the device into silent mode, or if this is not available, turn off or mute sounds and
audio notifications in the device’s Settings.
Testing Steps
1. Launch the Application.
2. Bring a suitable external device (hearing aid, microphone pickup or other
hearing assistance device with which the Application is designed to work)
within range, and set it into a state suitable for connecting with the
Application.
3. Set the Application into the appropriate state for connecting to an external
device if this is not done automatically at startup.
4. Check that the Application correctly detects and connects to the assistance
device, and that detection and connection are confirmed visually, or by
haptic feedback if this is sufficiently distinct from other feedback as to be
completely unambiguous.
5. Check the handling of error situations during connection by turning the
external device off during connection, and likewise by taking it out of range
in the same state, and confirm that the feedback (whether visual or haptic)
is clear and unambiguous and does not need an audio component to be
understood.
RESULT:
Connection and disconnection to the external aid should be reliable and easily
understood without any audio feedback.
Result of Test
PASS FAIL
36. USAGE WITHOUT VOCAL CAPABILITY
Page 36 of 61
4. Usage without vocal capability
4.1. Navigation
4.1.1. Navigation not limited to speech input
Test ID
4.1.1
Test Title
Navigation not limited to speech input
Critical
Test Description
Non-visual navigation should have an alternative to speech input.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled.
Ensure that Explore By Touch is enabled in TalkBack settings.
Testing Steps
1. Launch the Application.
2. Explore navigation of the Application’s functions.
3. Check that it is possible to navigate the Application satisfactorily without
being required to use speech input.
4. Where the Application has optional Settings to assist with other
accessibility issues (e.g. sight impairment) check that using those options
does not render the Application unusable for someone unable to use a
speech-driven interface.
5. Check that proper use is made of TalkBack and Explore By Touch to make
the Application accessible to a user who cannot fully use visual and
speech-driven interfaces.
RESULT:
Navigation within the Application should not be limited to solely speech-driven
input, even when the Application is intended for users with other accessibility
issues such as sight impairment.
Result of Test
PASS FAIL
37. USAGE WITHOUT VOCAL CAPABILITY
Page 37 of 61
4.2. Control (execution of actions)
4.2.1. Control actions not limited to speech input
Test ID
4.2.1
Test Title
Control actions not limited to speech
input.
Critical
Test Description
Non-visual control actions should have an alternative to speech input.
Required for:
All applications.
Testing Note
Ensure that the Android accessibility service ‘TalkBack’ is installed on the device, and
enabled.
Ensure that Explore By Touch is enabled in TalkBack Settings.
Testing Steps
1. Launch the Application.
2. Explore use of the Application’s functions.
3. Check that it is possible to operate the Application satisfactorily without
being required to use speech input.
4. Where the Application has optional Settings to assist with other
accessibility issues (e.g. sight impairment) check that using those options
does not render the Application unusable for someone unable to use a
speech-driven interface.
5. Check that proper use is made of TalkBack and Explore By Touch to make
the Application accessible to a user who cannot fully use visual and
speech-driven interfaces.
RESULT:
Control of the Application’s functionality should not be limited to solely speech-
driven input, even when the Application is intended for users with other
accessibility issues such as sight impairment.
Result of Test
PASS FAIL
4.3. Feedback
Not Applicable
38. USAGE WITHOUT VOCAL CAPABILITY
Page 38 of 61
4.4. Display
4.4.1. Text display
Test ID
4.4.1
Test Title
Text displays
Critical
Test Description
The Application should produce acceptable assistive text on the device
display where that is part of its intended function.
Required for:
Applications which produce assistive text which can be displayed on the
device.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application while making use of the facility
to display assistive text to third parties in place of speech.
3. Check that the resulting output is of acceptable quality and layout such that
it can be easily understood by a viewer without prior training or experience.
RESULT:
Useful assistive text is provided wherever the user is likely to need it.
Result of Test
PASS FAIL
39. USAGE WITHOUT VOCAL CAPABILITY
Page 39 of 61
4.5. Adjustments / Settings
4.5.1. Text-to-speech configuration & quality
Test ID
4.5.1
Test Title
Text-to-speech configuration & quality
Critical
Test Description
Text-to-speech (TTS) assistance should be configurable to produce
acceptable quality output.
Required for:
All applications.
Testing Note
If additional TTS libraries or voice files are required for full functionality or optimum
quality, these should be installed before testing, and device Settings for TTS should
be selected for best output quality outside of the Application.
Testing Steps
1. Launch the Application.
2. Using the Settings menu of the Application (if necessary, in conjunction
with the device’s TTS Settings), select options to produce assistive text-to-
speech (unless this is already enabled by default), and to vary the speed
and / or pitch of output if available.
3. Check that it is easy for the user to obtain good quality output that would
be understood by a listener without training or experience.
4. Check that where different “voices” are offered, each one offers easily
understood output of acceptable quality.
RESULT:
Configuration options on assistive text-to-speech should make it possible to
achieve acceptable output for general use.
Result of Test
PASS FAIL
40. USAGE WITHOUT VOCAL CAPABILITY
Page 40 of 61
4.6. External Devices
4.6.1. External Text-To-Speech devices
Test ID
4.6.1
Test Title
External Text-To-Speech devices
Critical
Test Description
The Application should operate correctly with external text-to-speech
(TTS) devices where that is part of its intended function.
Required for:
Applications which produce assistive text-to-speech in conjunction with
an external audio output device.
Testing Note
If additional TTS libraries or voice files are required for full functionality or optimum
quality, these should be installed before testing, and device Settings for TTS should
be selected for best output quality outside of the Application.
Testing Steps
1. Launch the Application.
2. Using the Settings menu of the Application (if necessary, in conjunction
with the device’s TTS Settings), select options to produce assistive text-to-
speech on the external device, and to vary the speed and / or pitch of
output if available.
3. Check that it is easy for the user to obtain good quality output that would
be understood by a listener without prior training or experience.
RESULT:
Text-to-speech functionality using an external device should be of acceptable
quality.
Result of Test
PASS FAIL
41. USAGE WITHOUT VOCAL CAPABILITY
Page 41 of 61
4.6.2. External text displays
Test ID
4.6.2
Test Title
External text displays
Critical
Test Description
The Application should operate correctly with external text displays
where that is part of its intended function.
Required for:
Applications which produce assistive text which can be output through
an external display.
Testing Note
The device should be set up to obtain optimum quality output on the external display.
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application while making use of the facility
to display assistive text to third parties in place of speech.
3. Check that the resulting output is of acceptable quality and layout such that
it can be easily understood by a viewer without prior training or experience.
RESULT:
Assistive text output to external displays should be of acceptable quality for
general use.
Result of Test
PASS FAIL
42. USAGE WITH LIMITED MANIPULATION OR STRENGTH
Page 42 of 61
5. Usage with limited manipulation or strength
5.1. Navigation
5.1.1. Alternative inputs for navigation
Test ID
5.1.1
Test Title
Alternative inputs for navigation
Critical
Test Description
Where navigation inputs require specific dexterity abilities, alternatives
should be offered.
Required for:
All applications.
Testing Note
This test should be made with the assumption that the user will have limited finger or
joint mobility and control, therefore operations which call for:
• finger pinch movements
• twisting of the hand or device
• rotation of the device
will be unacceptable unless alternative navigation methods are available.
Testing Steps
1. Launch the Application.
2. Where the Settings of the Application offer options relevant to users with
limited dexterity, ensure those options are selected.
3. Check that the Application can be navigated throughout its normal range of
usage by a user with dexterity limitations.
RESULT:
The Application must be navigable by a user with limited dexterity.
Result of Test
PASS FAIL
43. USAGE WITH LIMITED MANIPULATION OR STRENGTH
Page 43 of 61
5.2. Control (execution of actions)
5.2.1. Assistive options for control
Test ID
5.2.1
Test Title
Assistive options for control
Critical
Test Description
Control options should exist for combining actions requiring dexterity.
Required for:
All applications.
Testing Note
This test should be made with the assumption that the user will have limited finger or
joint mobility and control. Operations which call for simultaneous touch inputs or a
repeated sequence of inputs should offer options to simplify the actions required, e.g.
the “sticky keys” options used to allow some sequential control actions to be treated
as if they were simultaneous.
Testing Steps
1. Launch the Application.
2. Where the Settings of the Application offer options relevant to users with
limited dexterity, ensure those options are selected.
3. Check that any control simplifying / combining options would be useful to a
user with limited dexterity, and no commonly-used combinations have been
omitted.
RESULT:
The Application should offer assistive options for control that will assist a user
with dexterity impairment, and commonly-used options are not omitted.
Result of Test
PASS FAIL
44. USAGE WITH LIMITED MANIPULATION OR STRENGTH
Page 44 of 61
5.2.2. Pressure-related input options
Test ID
5.2.2
Test Title
Pressure-related input options
Critical
Test Description
Applications should offer input methods for users with limited ability to
control touch input.
Required for:
All applications.
Testing Note
This test should concentrate on the needs of users with tremor or muscle control
limitations that can create difficulty in maintaining consistent pressure or accuracy for
touch / press / hold operations.
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that the Application can be used without requiring distinction
between single-tap, double-tap, long-press or press-and-drag actions (i.e.
the Application responds to all these actions as if they are single-tap).
4. Check that actions like play / pause functions are not limited to single-touch
operation of the same control (e.g. separate controls rather than the same
control changing function with each touch).
5. Check that any popup messages that contains confirm / cancel or other
buttons are not positioned over controls in the background UI that could
cause an undesired or irreversible action. The user may be unable to avoid
touching twice in the same location, which could cause unintended
operation of an underlying control when the popup is dismissed.
RESULT:
Options offered should be of genuine value to users with limitations in touch
consistency or accuracy, and UI layout should avoid control positioning that
risks unintended operation through multiple touches.
Result of Test
PASS FAIL
45. USAGE WITH LIMITED MANIPULATION OR STRENGTH
Page 45 of 61
5.2.3. Multi-finger control options
Test ID
5.2.3
Test Title
Multi-finger control options
Critical
Test Description
Applications should offer an alternative to multi-finger gestures.
Required for:
All applications.
Testing Note
This test should ensure that the needs of users who cannot use multiple fingers in
gesture controls are met.
Where Settings offer options that provide alternatives to multi-finger gestures, these
should be selected.
Note that for the purposes of this test, “pinch-to-zoom” would be counted as a multi-
finger gesture as it requires two fingers in contact with the screen to execute.
Testing Steps
1. Launch the Application.
2. Explore the Application by navigating through all of its screens and using
its functions, using only one finger in contact with the screen at any time.
3. Check that all of its functionality can be used without needing to use multi-
finger gestures (i.e. that where multi-finger gestures exist, single-point
equivalents using tap or press are available).
RESULT:
The Application should be fully usable without the use of multi-finger gestures.
Result of Test
PASS FAIL
46. USAGE WITH LIMITED MANIPULATION OR STRENGTH
Page 46 of 61
5.3. Feedback
Not Applicable
5.4. Display
5.4.1. Dialogue boxes & timeouts
Test ID
5.4.1
Test Title
Dialogue boxes & timeouts
Critical
Test Description
Interactive elements should not pressure users to respond quickly.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that all dialogue boxes & controls allow the user unlimited time to
read information, provide responses or operate controls.
4. Where auto-scrolling or auto-refreshing text is used, it must be possible for
the user to cancel or pause these actions without restriction, to allow
unlimited time for information to be understood and responded to if
required.
5. Where timeouts are in use by default, the user should be provided with
adequate warning of an imminent timeout, and the ability to prevent that
timeout from acting.
RESULT:
The Application should always wait for as long as the user needs to interact
with it and should not dismiss informative displays or move to another function
if the user cannot respond in a set time.
Result of Test
PASS FAIL
47. USAGE WITH LIMITED MANIPULATION OR STRENGTH
Page 47 of 61
5.4.2. Displayed information – cognitive overload
Test ID
5.4.2
Test Title
Displayed information – cognitive
overload
Critical
Test Description
Applications should not display successive interactive or informational
messages until each preceding one has been actioned by the user.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that the language used is plain and simple to understand, with use
of icons where possible, to simplify the presentation of information.
4. Check that whenever information is presented to the user or interaction
requested, it is always in manageable quantities, and if successive
messages need to be displayed, the Application waits for user input /
acknowledgement before proceeding to the next message.
RESULT:
The Application should always present information in manageable quantities
and wait for user input before stepping to the next message.
Result of Test
PASS FAIL
48. USAGE WITH LIMITED MANIPULATION OR STRENGTH
Page 48 of 61
5.5. Adjustments / Settings
5.5.1. Touch-related settings
Test ID
5.5.1
Test Title
Touch-related settings
Critical
Test Description
Applications should offer adjustment to suit users with difficulty in
maintaining optimum touch pressure, consistency or accuracy.
Required for:
All applications.
Testing Note
This test should concentrate on settings that address the needs of users with tremor
or muscle control limitations that can create difficulty in maintaining consistent
pressure or accuracy for touch / press / hold operations.
Testing Steps
1. Launch the Application.
2. Explore use of Settings touch / pressure / key repetition rate options.
3. Check that the options available implement real-world benefits for users
with these limitations, e.g.
a. Haptic pulsing (for tremor),
b. Ability to vary touch / hold pressure thresholds.
c. Options for replacing augmenting long-press / hold actions with
other single-touch controls.
d. Options for enlarging the size of control sensing areas (not just the
size of the displayed control element).
e. Actions like long press to trigger a key-repetition function can be
disabled.
RESULT:
Options offered should be of genuine value to users with limitations in touch
consistency or accuracy.
Result of Test
PASS FAIL
49. USAGE WITH LIMITED MANIPULATION OR STRENGTH
Page 49 of 61
5.6. External Devices
5.6.1. Connection and operation with external devices
Test ID
5.6.1
Test Title
Connection and operation with external
devices
Critical
Test Description
The Application should operate correctly with external devices and
controls that provide dexterity assistance.
Required for:
All applications.
Testing Note
This test addresses the use of external switches, keyboards and other controls which
provide dexterity-related improvements in the accessibility of the device and
Application, such as joysticks, puff and sip controls, Braille keyboards and audio aid
devices.
To the best extent possible, the external device or devices should be set up and
confirmed as working correctly before launching the Application (unless such devices
are only operable within the Application).
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that all external devices relevant to operation of the Application
have been correctly detected & connected, and that they provide the
expected functionality in a usable manner.
RESULT:
The Application should operate correctly with all relevant external devices.
Result of Test
PASS FAIL
50. USAGE WITH LIMITED COGNITION
Page 50 of 61
6. Usage with limited cognition
6.1. Navigation
6.1.1. Help information
Test ID
6.1.1
Test Title
Help information
Critical
Test Description
Help information to be complete, understandable and not over-complex.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Check that Help information is available for all functions of the Application.
3. Check that each item of information is neither too long nor too detailed to
be usable.
4. Each item should cover a single prompt or function that the user can act
upon before returning to the Help for assistance with the next step.
RESULT:
Help should be provided for all functions of the Application. Help information
should always be in manageable quantities.
Result of Test
PASS FAIL
51. USAGE WITH LIMITED COGNITION
Page 51 of 61
6.1.2. Clarity of single action
Test ID
6.1.2
Test Title
Clarity of single action
Critical
Test Description
Application functions should be expressed as single clear actions.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that each function within the Application is presented as a single,
understandable action or concept.
4. Where complex ideas or actions have to be introduced, they should be
broken down into single-concept elements and presented successively to
prevent cognitive overload.
5. Check that good use of header – sub header – paragraph structures
breaks the information down into manageable sized pieces.
6. Check that links have a descriptive label/text showing what the link
accesses. They should not use generic phrases like “Click Here” or “More”.
RESULT:
Everything the Application does should be simply and directly presented as
single actions.
Result of Test
PASS FAIL
52. USAGE WITH LIMITED COGNITION
Page 52 of 61
6.1.3. Limiting the number of options / choices presented to the user
Test ID
6.1.3
Test Title
Limiting the number of options / choices
presented to the user
Critical
Test Description
The Application should not present long list of options / choices.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore the Settings (where present) and functionality of the Application.
3. Check that choices are given as a limited number of simple high level
concepts:
a. Each entry should break down into successive layers.
b. Each layer should introduce only one idea or action.
c. At any point in the structure, only a small number of choices should
be presented.
RESULT:
Only small sets of choices should be shown in a single step or screen.
Result of Test
PASS FAIL
53. USAGE WITH LIMITED COGNITION
Page 53 of 61
6.1.4. Language complexity
Test ID
6.1.4
Test Title
Language complexity
Critical
Test Description
Language used in the Application should be simple and direct.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore its functionality.
3. Check that:
a. Prompts, labels and informative texts are not too long or detailed to
be easily understood.
b. Single ideas or actions are presented that the user can act upon
before proceeding to the next step.
c. The language used throughout is consistent, simple and
straightforward.
d. Multiple possible actions are not introduced in the same sentence.
e. The user should not be required to keep multiple ideas or concepts
in mind at the same time to successfully use the Application.
RESULT:
Information should always be presented in simple, direct language. Complex
sentence structures should not be used.
Result of Test
PASS FAIL
54. USAGE WITH LIMITED COGNITION
Page 54 of 61
6.2. Control (execution of actions)
6.2.1. Presentation of control elements
Test ID
6.2.1
Test Title
Presentation of control elements
Critical
Test Description
Control elements should be presented in a way that suggests the
outcome and maximises clarity.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that:
a. Controls are simply identified.
b. Controls with similar appearance or labeling are not presented at
the same time.
c. Appearance and labeling of controls is easy to associate with the
outcome.
d. Understanding the purpose or action of controls is possible without
having to hold a context in mind over multiple screens.
RESULT:
Control elements should be clear and easily understood.
Result of Test
PASS FAIL
55. USAGE WITH LIMITED COGNITION
Page 55 of 61
6.3. Feedback
6.3.1. Clarity of feedback
Test ID
6.3.1
Test Title
Clarity of feedback
Critical
Test Description
All feedback produced by the Application should be expressed simply
and clearly.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that:
a. Feedback is expressed in simple, clear terms.
b. Feedback is not excessively lengthy or composed of multiple
elements.
c. Feedback explains its purpose well (especially if its appearance
could be unexpected by the user).
d. Notification of errors is short, unambiguous and easily understood.
e. Error messages state the action to recover from the error.
f. Avoid passive sentence structures (“x should be done”). Use active
sentences (“do x”).
RESULT:
Feedback should be easily understood and acted on by the user.
Result of Test
PASS FAIL
56. USAGE WITH LIMITED COGNITION
Page 56 of 61
6.4. Display
6.4.1. Text fonts and sizes
Test ID
6.4.1
Test Title
Text fonts and sizes
Critical
Test Description
Text should be displayed in font faces and sizes that favour clarity and
logical relationships.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore all parts of the Application – functionality, Settings, Help etc.
3. Check that:
a. Font faces and sizes used maximise readability and clarity.
b. Where font sizes change between blocks of text, the reason for the
change is obvious and logical for the user.
c. Headings are logical and consistent and make good use of font
sizes for clarity.
RESULT:
Font faces and sizes should be chosen for clarity and easy understanding.
Result of Test
PASS FAIL
57. USAGE WITH LIMITED COGNITION
Page 57 of 61
6.4.2. Colours for reading comprehension
Test ID
6.4.2
Test Title
Colours for reading comprehension
Warning
Test Description
The user should be able to choose from a range of colours for text and
backgrounds that may aid reading comprehension.
Required for:
All applications which address reading comprehension issues.
Testing Note
1. Research (http://www.dyslexic.org.uk/research/vision-coloured-filters) has shown
that use of coloured filters can help reading comprehension. Display options
which mimic this should be offered where possible.
2. This test is only appropriate if the Application is intended to directly aid reading
comprehension issues such as dyslexia.
Testing Steps
1. Launch the Application.
2. Explore the functionality and Settings of the Application.
3. Check it is possible to display text in a colour arrangement that aids
reading comprehension in different light levels. Possible options would be:
a. Black, white, grey, yellow or blue backgrounds.
b. Black, white, grey, yellow or blue text.
4. The colour combinations offered should be clear with acceptable contrast
ranges.
RESULT:
The Application should offer ways of presenting text for best reading
comprehension.
Result of Test
Pass Annoying Difficult Impossible
This test is not applicable where...
The Application is not intended to aid reading comprehension issues.
58. USAGE WITH LIMITED COGNITION
Page 58 of 61
6.4.3. Dialogue boxes & timeouts
Test ID
6.4.3
Test Title
Dialogue boxes & timeouts
Critical
Test Description
Interactive elements should not pressure users to respond quickly.
Required for:
All applications.
Testing Note
Testing Steps
1. Launch the Application.
2. Explore the functionality of the Application.
3. Check that all dialogue boxes & controls allow the user unlimited time to
read information, provide responses or operate controls.
4. Where auto-scrolling or auto-refreshing text is used, it must be possible for
the user to cancel or pause these actions without restriction, to allow
unlimited time for information to be understood and responded to if
required.
5. Where timeouts are in use by default, the user should be provided with
adequate warning of an imminent timeout, and the ability to prevent that
timeout from acting.
RESULT:
The Application should always wait for as long as the user needs to interact
with it and should not dismiss informative displays or move to another function
if the user cannot respond in a set time.
Result of Test
PASS FAIL
59. USAGE WITH LIMITED COGNITION
Page 59 of 61
6.5. Adjustments / Settings
6.5.1. Reading level options
Test ID
6.5.1
Test Title
Reading level options
Critical
Test Description
Applications presenting complex information should have an option to
simplify what is presented to the user.
Required for:
All applications.
Testing Note
This test is only intended to apply to an Application whose purpose is to present
complex information to the user. The intention behind the test is to establish whether
the Application offers display options that reduce the amount of information on a
single screen to be easier to understand (for example “Simple” versus “Advanced”
display options).
Testing Steps
1. Launch the Application.
2. Explore the Settings of the Application.
3. Check:
a. Whether the Application defaults to a complex information display.
b. Whether controls or their labels are complex.
c. Whether an option is offered to simplify this presentation to easily-
understood basics.
d. Whether the simplified presentation maximises clarity and
understanding.
RESULT:
If an Application normally shows complex information or controls, it should
offer a simpler, clearer presentation that is restricted to key elements only.
Result of Test
PASS FAIL
6.6. External Devices
Not Applicable
60. Page 60 of 61
Attributions and References
Portions of this document are modifications based on work created and shared by
the Android Open Source Project and used according to terms described in
the Creative Commons 2.5 Attribution License.
Accessibility Testing Checklist - Android Developers -
http://developer.android.com/tools/testing/testing_accessibility.html
W3C WCAG 2.0 : Mobile Accessibility: How WCAG 2.0 and other W3C/WAI
Guidelines apply to mobile – Feb 2015 - http://www.w3.org/TR/mobile-accessibility-
mapping/
Understanding WCAG 2.0 - A guide to understanding and implementing Web
Content Accessibility Guidelines 2.0 - http://www.w3.org/TR/2015/NOTE-
UNDERSTANDING-WCAG20-20150226/Overview.html
Understanding Techniques for WCAG Success Criteria -
http://www.w3.org/TR/2015/NOTE-UNDERSTANDING-WCAG20-
20150226/understanding-techniques.html
Index of Color Contrast Samples - http://trace.wisc.edu/contrast-ratio-examples/
ETSI EN 301 549 v1.1.2 (2015-04): Accessibility requirements suitable for public
procurement of ICT products and services in Europe -
http://www.etsi.org/deliver/etsi_en/301500_301599/301549/01.01.02_60/en_301549
v010102p.pdf
61. Page 61 of 61
Version control
Version Date Changes made
v0.8 Apr 2015 Draft version for circulation & discussion.
v1.0 May 2015 Updated after detailed comments & feedback.
v1.1 Jun 2015 Updated after wider consultation.
v1.2 Jul 2015 First release version after further feedback.
DISCLAIMER. THIS TESTING CRITERIA DOCUMENT ("DOCUMENT") IS FOR INFORMATIONAL
PURPOSES ONLY. YOUR USE OF THIS DOCUMENT AND THE INFORMATION PROVIDED HEREIN IS AT
YOUR OWN RISK. THE DOCUMENT IS PROVIDED ON AN "AS IS" AND "WITH ALL FAULTS" BASIS. THE
APPLICATION QUALITY ALLIANCE (AQuA) INCLUDING THE MEMBERS IT IS COMPRISED THEREOF
DISCLAIM ALL EXPRESS OR IMPLIED CONDITIONS, REPRESENTATIONS, AND WARRANTIES OF ANY
KIND, INCLUDING ANY IMPLIED WARRANTY OR CONDITION OF MERCHANTABILITY, SATISFACTORY
QUALITY, FITNESS FOR A PARTICULAR PURPOSE, OR NON-INFRINGEMENT. THE APP QUALITY
ALLIANCE INCLUDING THE MEMBERS IT IS COMPRISED THEREOF MAKE NO REPRESENTATIONS,
WARRANTIES, CONDITIONS OR GUARANTEES AS TO THE USEFULNESS, QUALITY, SUITABILITY,
TRUTH, ACCURACY OR COMPLETENESS OF THIS DOCUMENT AND MAY CHANGE THIS DOCUMENT AT
ANY TIME WITHOUT NOTICE.