Software testing is a critical part of the software and product development cycle. For years, quality assurance (QA) has been the final gate before a release, a functional validation that the code is ready to go to market. Today, however, despite a growing need for agile and efficient testing, legacy QA models are struggling to keep up with cloud-native environments, fragmented teams and rapid release cycles. This introduces a new era of testing using AI and automation to prioritise tests and analyse vast quantities of data. It also introduces new quality, security and governance risks that must be addressed.
The benefits of AI in testing automation and QA are clear to me. It allows teams to move incredibly rapidly. Companies benefit from faster automation processes, and their productivity increases exponentially. AI can also analyse data to detect defects or coverage gaps and provide teams with high-risk scenarios or recommend additional tests.
The Capgemini World Quality Report 2024-2025 highlighted the impact of AI on the industry, emphasising its ability to optimise test coverage, reduce human error, and introduce intelligent automation[1]. The study found that an impressive 68% of companies are using AI, with 72% reporting faster processes as a result. To me, this signals a clear shift in how testing is perceived. The technology is changing the testing story, moving it away from an after-the-fact process that discovers unexpected errors and frustrates teams and deadlines alike. AI enables testing in near-real-time, as models continuously analyse code throughout the development process, and it can be integrated into development and operations from the outset.
AI models can learn, and they can predict errors that crop up regularly or where they’re most likely to occur. They turn testing into a smoother part of the process, making it proactive and immediate instead of reactive and defined by ticking boxes. They also allow for self-healing, where the automation can detect when a test will break, find a resolution, and then apply the update so the test still passes if the business behaviour remains valid.
Self-healing has the potential to minimise failures caused by minor changes and reduce the burden of repetitive testing for teams. Talented humans now have more time to prioritise exploratory testing or more complex tasks. The technology gives people the space to become high-quality architects, defining risk models, guiding AI, and interpreting patterns, rather than just running tests.
There are agentic platforms that can now take on much of the heavy testing grunt work with minimal human input, acting almost like testing interns that fortunately don’t get tired or frustrated. However, if it sounds too good to be true, it unfortunately can be. While AI has immense value in the QA environment, it also introduces risks that must be considered.
Just as AI in the workplace tends to hallucinate or overcompensate, the same risks apply in testing. Some of the most common problems include false positives and false negatives, which add to testing noise rather than minimising it, over-reliance on automation that impacts skills and awareness, and data privacy and security. Then there’s the reality that AI models are still black boxes – teams lack visibility into the AI decision-making process and don’t know why some tests are prioritised while others aren’t.
Bias, security, private information, limited guardrails, and governance that’s struggling to keep up with the pace of change are very real concerns when introducing AI into the testing environment. Right now, as teams learn to use and optimise AI in testing, it’s important to remember that it’s a tool designed to augment the process. People are needed to validate AI outputs, make release decisions, and, very importantly, serve as guardrails for data privacy and explainability.
This doesn’t mean AI is too risky, just as it doesn’t mean AI is the perfect solution to every testing problem. In my view, AI is as much in its infancy as its use cases, which means testing with AI models and tools should be considered and balanced. Humans are essential, but so are the tools that lift the burden of complexity and deadlines.
