Hackers are using LLMs to build next-generation phishing attacks

Cybersecurity experts warn that hackers are leveraging large language models (LLMs) to create sophisticated phishing attacks. These AI tools enable the generation of phishing pages on the spot, potentially making scams more dynamic and harder to detect. The trend highlights evolving threats in digital security.

In a recent article published by TechRadar on January 26, 2026, the use of large language models (LLMs) by hackers to develop advanced phishing techniques is spotlighted. The piece, titled 'Hackers are using LLMs to build the next generation of phishing attacks - here's what to look out for,' explores how these AI systems could automate and customize phishing efforts in real time.

The description poses a key question: 'What if a phishing page was generated on the spot?' This suggests a shift from static phishing sites to dynamically created ones, which could adapt to user inputs or contexts, increasing their effectiveness.

While specific examples or defenses are not detailed in the available excerpt, the article aims to inform readers on vigilance against such emerging threats. As LLMs become more accessible, cybersecurity measures must evolve to counter AI-assisted attacks, emphasizing the need for user awareness and robust detection tools.

Related Articles

Illustration of a hacker using AI to swiftly build VoidLink malware targeting Linux cloud servers, featuring rapid code generation and infiltrated systems.
Image generated by AI

AI-assisted VoidLink malware framework targets Linux cloud servers

Reported by AI Image generated by AI

Researchers at Check Point have revealed that VoidLink, a sophisticated Linux malware targeting cloud servers, was largely built by a single developer using AI tools. The framework, which includes over 30 modular plugins for long-term system access, reached 88,000 lines of code in under a week despite plans suggesting a 20-30 week timeline. This development highlights AI's potential to accelerate advanced malware creation.

Cybersecurity experts are increasingly alarmed by how artificial intelligence is reshaping cybercrime, with tools like deepfakes, AI phishing, and dark large language models enabling even novices to execute advanced scams. These developments pose significant risks to businesses in the coming year. Published insights from TechRadar underscore the scale and sophistication of these emerging threats.

Reported by AI

In 2025, cyber threats in the Philippines stuck to traditional methods like phishing and ransomware, without new forms emerging. However, artificial intelligence amplified the volume and scale of these attacks, leading to an 'industrialization of cybercrime'. Reports from various cybersecurity firms highlight increases in speed, scale, and frequency of incidents.

Nigerian businesses are being urged to focus on staff training in the face of escalating phishing threats.

Reported by AI

Google has introduced new defenses against prompt injection in its Chrome browser. The update features an AI system designed to monitor the activities of other AIs.

Security firm Varonis has identified a new method for prompt injection attacks targeting Microsoft Copilot, allowing compromise of users with just one click. This vulnerability highlights ongoing risks in AI systems. Details emerged in a recent TechRadar report.

Reported by AI

Security researchers, first reporting via TechRadar in December 2025, warn WhatsApp's 3 billion users of GhostPairing—a technique tricking victims into linking attackers' browsers to their accounts, enabling full access without breaching passwords or end-to-end encryption.

 

 

 

This website uses cookies

We use cookies for analytics to improve our site. Read our privacy policy for more information.
Decline