Back to Knowledge Base
Penetration Testing

AI-Powered Penetration Testing: What's Changed and What It Means for Your Defences

8 April 202610 min read
AI-Powered Penetration Testing: What's Changed and What It Means for Your Defences

Penetration testing used to be straightforward -- at least conceptually. You hired skilled testers, they probed your systems for weaknesses, and you fixed what they found. The tools were well-understood, the attack surface was mostly traditional infrastructure and web applications, and the engagement followed a predictable rhythm: scope, test, report, remediate, repeat next year.

That model is breaking down on two fronts.

First, attackers are using AI to find and exploit vulnerabilities faster than ever. Second, the tools available to penetration testers have fundamentally shifted, with AI-powered platforms now capable of autonomous reconnaissance, exploit chaining, and continuous validation at a speed and scale that manual testing cannot match.

For Australian businesses investing in cybersecurity -- whether you're pursuing ISO 27001, SOC 2, or simply trying to protect your customers' data -- understanding this shift is no longer optional. It changes what you should expect from a pentest engagement, what questions you should ask your providers, and how you think about security validation as a whole.


The Two Sides of AI in Penetration Testing

There's an important distinction that often gets lost in the conversation. AI is changing penetration testing in two fundamentally different ways:

  1. AI as a weapon -- attackers (and ethical testers simulating them) using AI to conduct faster, more sophisticated attacks
  2. AI as a testing tool -- penetration testing platforms using AI to automate vulnerability discovery, exploit validation, and continuous security assessment
Both matter. Let's look at each.

How Attackers Are Using AI Right Now

This isn't a future scenario. AI-powered offensive capabilities are being used in production today, and they're reshaping the threat landscape that penetration testing is designed to simulate.

AI-Generated Phishing at Scale

The days of spotting phishing emails by their broken grammar are over. According to KnowBe4's 2025 Phishing Trends Report, approximately 83% of phishing emails are now AI-generated. These messages are grammatically flawless, contextually relevant, and often reference real projects, colleagues, and internal processes scraped from publicly available data.

The Hoxhunt Cyber Threat Intelligence Report observed a 14x surge in AI-generated phishing campaigns since December 2025, with these now representing roughly half of all attacks reported by users. The campaigns are not just more polished -- they're multi-channel. Attackers now coordinate email, SMS, voice calls, and even video conferencing to create a layered web of legitimacy that bypasses traditional single-channel detection.

Deepfake Social Engineering

Deepfake technology has moved from novelty to operational tool. In one of the most cited incidents, a finance worker at engineering firm Arup transferred US$25 million after attending what appeared to be a legitimate video call with the company's CFO and senior leadership -- every face and voice on the screen was AI-generated.

Deepfake-as-a-service (DaaS) platforms emerged as one of the fastest-growing tools for cybercriminals in 2025. According to Cyble's Executive Threat Monitoring report, AI-powered deepfakes were involved in over 30% of high-impact corporate impersonation attacks. These services are now accessible to attackers with minimal technical skill, offering ready-to-use tools for voice cloning, video generation, and persona simulation.

Automated Vulnerability Discovery

AI is accelerating the reconnaissance and exploitation phases of attacks. Threat actors can now deploy AI agents that continuously probe external attack surfaces, identify misconfigurations, and chain exploits -- all at a speed that manual operators cannot match. As one security researcher noted, cybercriminals are becoming highly effective at using AI to find and exploit unpatched vulnerabilities at scale, turning what used to take weeks into hours.

This matters for penetration testing because it raises a fundamental question: if your annual pentest simulates a methodical human attacker working over a two-week engagement, but the real threat is an AI-powered system probing your perimeter 24/7, how representative is that test of your actual risk?


How Penetration Testing Tools Have Changed

The penetration testing industry has responded to this shift with a wave of AI-powered platforms that change what testing looks like in practice.

From Periodic Assessments to Continuous Validation

The traditional model -- an annual or biannual pentest engagement -- is increasingly insufficient. The market is shifting towards Penetration Testing as a Service (PTaaS) and continuous security validation, where AI-powered platforms run ongoing assessments that adapt as your environment changes.

AWS launched its Security Agent in March 2026, deploying specialised AI agents to discover, validate, and report security vulnerabilities through multi-step attack scenarios customised for each application. This signals that continuous, AI-driven security testing is becoming a mainstream expectation, not a niche offering.

Agentic Penetration Testing

The most significant development in 2025-2026 has been the emergence of "agentic" penetration testing -- platforms that use multiple AI agents working together, each specialising in different aspects of an attack. Rather than a single scanning engine, these systems deploy coordinated teams of AI agents: one for reconnaissance, one for exploit development, one for lateral movement, and one for reporting.

These platforms can chain exploits, reason about application behaviour, and discover business logic vulnerabilities that traditional scanners consistently miss -- issues like broken object-level authorisation (BOLA), insecure direct object references (IDOR), and privilege escalation through workflow bypasses.

Research from USENIX demonstrated a 228% improvement in task completion when LLMs were structured into modular roles rather than used as monolithic assistants. The key insight is that AI penetration testing works best when the problem is decomposed into specialised sub-tasks, not when a single model tries to do everything.

What AI Pentesting Does Well -- and Where It Falls Short

It's important to be realistic about capabilities. AI pentesting tools excel at:
  • Speed and coverage -- scanning hundreds of endpoints, APIs, and services simultaneously
  • Pattern recognition -- identifying known vulnerability classes across large codebases
  • Continuous validation -- retesting every time code changes, integrated into CI/CD pipelines
  • Reducing false positives -- validating findings through actual exploitation rather than signature matching
But they still struggle with:
  • Business logic flaws -- understanding the intent behind an application's design and finding ways to abuse it in context
  • Creative attack scenarios -- the kind of lateral thinking that experienced human testers bring to complex environments
  • Physical and social engineering -- testing human processes, not just technical controls
  • Contextual risk assessment -- understanding which findings actually matter for your specific business
The PentestEval research, published in late 2025, provides a useful reality check: it breaks the penetration testing workflow into six stages and shows that most stages still come in under 50% success across evaluated AI models. The technology is genuinely impressive, but it's not a replacement for experienced human testers -- it's a force multiplier.

The New Attack Surface: Testing Your AI Systems

There's a third dimension to this conversation that many organisations haven't considered: if you're deploying AI in your own products or operations, those AI systems need to be tested too.

The OWASP Top 10 for LLM Applications (2025 edition) provides the most widely referenced framework for understanding AI-specific security risks. The top threats include:

LLM01: Prompt Injection remains the number one risk for the second consecutive year. Attackers craft inputs that cause the model to ignore its original instructions and follow new ones -- either directly through user prompts or indirectly through external content the model processes (documents, web pages, emails). The challenge is that you cannot simply patch your way out of this; it exploits how LLMs fundamentally process language.

LLM02: Sensitive Information Disclosure -- LLMs can memorise and reproduce fragments of their training data, including personally identifiable information, proprietary business data, and credentials. Poorly configured models can inadvertently expose confidential information through their outputs.

LLM06: Excessive Agency -- as AI agents gain the ability to send emails, query databases, call APIs, and make decisions, the model's permissions become the attack surface. Give an agent more access than it needs, and you've created an exploitable pathway.

LLM07: System Prompt Leakage -- attackers extracting the system instructions that define your AI's behaviour, potentially revealing business logic, access controls, and internal processes.

LLM08: Vector and Embedding Weaknesses -- a new entry for 2025, targeting vulnerabilities in Retrieval-Augmented Generation (RAG) systems. Attackers can poison vector databases, manipulate embedding models, or exploit weak access controls in vector stores.

If your organisation is deploying chatbots, AI assistants, automated decision-making systems, or any LLM-powered tooling, these vulnerabilities need to be part of your penetration testing scope. Traditional web application testing won't catch them.


What This Means for Australian Businesses

Your Pentest Expectations Should Change

If your penetration testing engagement still looks like a two-week manual assessment once a year, you're likely testing against yesterday's threat model. The shift towards AI-powered attacks means you should be asking:
  • Is your testing provider using AI-augmented tools? The best engagements now combine human expertise with AI-powered platforms for broader coverage.
  • How often are you testing? Annual pentests leave 11 months of unvalidated changes. Consider continuous or quarterly validation for critical systems.
  • Does your scope include AI-specific risks? If you've deployed LLMs, chatbots, or AI-driven automation, prompt injection and related risks need to be in scope.
  • Are social engineering tests keeping pace? With AI-generated phishing and deepfake capabilities, your security awareness training and testing should reflect the current threat, not last year's.

Regulatory and Compliance Implications

For organisations pursuing ISO 27001 certification, penetration testing is a key component of demonstrating that your controls are effective. ISO 27001:2022 Annex A includes specific controls relevant to this discussion:
  • A.8.8 -- Management of technical vulnerabilities requires organisations to obtain information about technical vulnerabilities in a timely fashion and take appropriate measures
  • A.5.36 -- Compliance with policies, rules and standards for information security requires verification that security practices are implemented as designed
  • A.8.25 -- Secure development life cycle requires security testing as part of the development process
The evolving threat landscape means that demonstrating compliance with these controls increasingly requires testing methodologies that reflect how attacks actually happen -- including AI-powered techniques.

For organisations subject to APRA CPS 234, the requirement to maintain information security capability "commensurate with the size and extent of threats" is directly relevant. As AI-powered threats scale up, your testing methodology should scale with them.

Practical Steps You Can Take Now

  1. Review your penetration testing scope and methodology. Ask your provider what AI-powered tools they use and whether they test for AI-specific vulnerabilities.
  2. Inventory your AI deployments. Document every AI and ML system in use across your organisation -- including shadow AI tools your staff may be using without approval. This inventory is also a requirement under ISO 42001 if you're considering AI governance certification. Take our ISO 42001 Readiness Quiz to see where you stand.
  3. Update your security awareness training. Your staff need to understand AI-generated phishing and deepfake impersonation. Traditional "spot the typo" training is no longer sufficient.
  4. Consider continuous security validation. If annual pentesting is your only security testing, explore PTaaS or AI-augmented continuous testing to close the gap between assessments.
  5. Include AI in your risk register. Both the risks from AI-powered attacks and the risks of your own AI deployments should be formally assessed and managed within your ISMS.
---

The Bottom Line

AI hasn't made penetration testing obsolete -- it's made it more important than ever. But it has changed what good testing looks like.

The organisations that will be best positioned are those that combine experienced human testers with AI-powered tools, test continuously rather than annually, and expand their scope to include the AI systems they're deploying internally. The threat landscape has shifted. Your defences need to shift with it.


Further Reading

Penetration TestingAICybersecurityCompliance

Need Help With Your Security?

Our team of experts can guide you through implementation and certification. Start with a free assessment.

Start Free Assessment