Cyber Security

AI phishing: How to defend AI-generated attacks

Phishing has been transformed with the advent of AI-driven automation and sophisticated psychological techniques. Here are our tips and strategies to help you defend against AI-generated attacks.

The rise of ever-dangerous AI phishing attacks

Phishing attacks have always been dangerous, but AI is taking the level of sophistication and their scale and precision. Large language models can now generate highly tailored phishing emails with just minimal data points, making it easier and faster to create convincing deceptions. The advanced use of psychological tools continues to exploit our inherent human tendency like trust in authority, fear of missing out, and or the desire to respond.

Research published this year has shown that when AI-generated emails are combined with these psychological techniques, the success rates are very high. It shows us that AI-generated phishing emails can achieve click-through rates as high as 44%, and when combined with psychological models, success rates soar to upwards of 80%. Phishing campaigns that once took significant time and skill are now automated, lowering the bar for non-technical attackers to launch targeted phishing at scale, increasing their impact while reducing the cost, time and effort.

Phishing works because it preys on common human biases, which we all have, but it’s more nuanced than that. What one person falls for, may raise an instinctive concern for another. Some are reeled in by a sense of urgency, secrecy, or simply just trusting the sender or source. Our complex diversity of potential human responses is why phishing attacks continue to be so effective despite growing awareness, especially financial scams, that exploits the same human condition.

Using AI as a Defence Against Phishing

The enhanced AI driven tools that attackers use to craft those sophisticated phishing emails can also be used to detect them, faster and more effectively intercepting the attack. Whilst the AI-enabled security tech is improving, its performance is not always perfect, and phishing detection still needs a ‘human in the loop’ security and IT professionals continue to provide essential human oversight. Combining AI detection with human verification and advanced behavioural analytics, provides a robust approach. Of course, on-going education to help recognise subtle and advanced phishing attempts, sharing and celebrating experiences.

Whilst human error continues to make up the vast percentage of data breach incidents, as reported by the ICO, I recommend avoiding calling out your workforce as the weakest link, share poor and celebrate good behaviours elevating everyone as your strongest asset. Go beyond one-and-done annual interventions, they are ineffective against this threat, only continuous supportive education on this topic genuinely supports positive security culture outcomes.

Changing behaviour requires making it easy for people to do something. It's not enough to simply tell your staff or write it in a policy. Simplifying steps to help them respond quickly and effectively to potential risks. Nudge theory, which has been around for some time, and used in society to help change or influence behaviour. Not to be confused with notification, nudging focuses on subtle behaviours to help people make better choices naturally, without force, remember when sweets and snacks were in the check-out aisle.

In our world of cyber security, instead of relying on fear tactics that can cause resistance, nudging offers a smoother approach. To try and build better security habits by making the right decision feel like the easiest one, steering people towards better outcomes without overwhelming or pressuring them.

Nudging can be effective but may face limitations in combating AI-driven phishing. As AI generates more and more sophisticated and personalised attacks, subtle nudges might not always be sufficient to counter such tailored threats. Attackers can exploit behavioural tendencies with precision, making it harder for general nudges to guide users effectively. In this fast-evolving threat landscape, stronger interventions, such as multi-layered defences, real-time AI detection, and advanced AI-enabled behavioural analysis, are necessary to maintain pace with the sophistication of AI-enhanced phishing.

Building Resilience Against AI Phishing

To help with the sophistication of AI-enhanced phishing, your multi-layered security strategy must use both tech and human insight, a powerful collaborative force. Having security teams that are skilled in understanding and collaborating with AI, will give you a competitive advantage.

Four strategies to combat AI phishing:

1. AI-Driven Behavioural Analytics

Security systems must use AI to analyse user behaviour in real-time, flagging anomalies such as unusual communication patterns or login activity. These systems need to adapt dynamically to new phishing tactics, offering proactive defence against evolving threats.

2. Advanced Email Filtering

Organisations should adopt AI-powered email filters that go beyond traditional keyword detection, focusing on behavioural signals and language patterns that may indicate phishing intent. These tools should be integrated with machine learning models that evolve with new attack vectors.

3. Personalised Staff Training

Just as attackers use AI to personalise phishing emails, organisations should consider using this approach to tailor security training. AI-driven phishing simulations should adapt to the behaviour of individual employees, with integrated nudging techniques.

4. Adopt Zero Trust Principles

As AI-enabled phishing attacks become more personalised, a Zero Trust approach assumes you are in a hostile environment, and will be compromised, while ensures that no action is trusted without verification. This includes verifying identities, permissions, and intent at every stage of interaction, helping to reduce the impact of a breach.

Evolve your approach with sophisticated defences

The convergence of AI, security and psychology in phishing has fundamentally altered the threat landscape. Attackers are no longer limited by time or skill; they can now launch highly targeted, effective campaigns at low costs and scale. Organisations must respond with equally sophisticated defences, combining AI-driven detection powered by human insight with tailored staff training and data driven telemetry for your enhanced behavioural analytics.

As AI-enhanced phishing becomes more prevalent, organisations must evolve their strategies to stay ahead. Defending against these advanced attacks requires more than traditional security measures, it needs a deep understanding of both the technological and psychological elements at play.

As we covered in our top cyber security tips, every employee, at every level, should be empowered to make smarter cybersecurity decisions. Take a fresh approach to your staff upskilling, building robust defences in your people, process and technology. Only by integrating these insights can organisations effectively mitigate the rising threat of AI-driven phishing in the years ahead.

Looking to build AI security skills in your business? Learn more about our cyber security training, or get in touch to discuss your requirements. 

Let's talk

Start your digital transformation journey today

Contact us today via the form or give us a call.

0113 220 7150 (UK)  

888-895-3441 (US)

 

By submitting this form, you agree to QA processing your data in accordance with our Privacy Policy and Terms & Conditions. You can unsubscribe at any time by clicking the link in our emails or contacting us directly.

Related Articles