Cybercriminals are increasingly relying on artificial intelligence to automate and run attacks, and defenders are doing the same, according to Google Cloud’s Cybersecurity Forecast 2026. What started as tools to assist hackers is turning into autonomous systems that can plan, adapt, and execute campaigns with minimal human supervision.
AI systems now automate many steps of an attack lifecycle: they can write malware, craft and send phishing messages, imitate human behavior, rewrite their own code, and rapidly deploy new variants. Automation lets attackers scale operations cheaply — what used to require teams of people can now be carried out by a single AI model.
A specific danger highlighted in the forecast is prompt injection, where attackers trick AI models into ignoring safety rules and following hidden instructions. As businesses integrate large AI models into everyday systems, these models become attractive targets; the more connected the AI, the easier it is to manipulate.
Human‑focused scams remain effective. Criminal groups still exploit social engineering — now amplified by AI. Voice‑cloning and deepfake audio let attackers impersonate executives with convincing tone and rhythm, making vishing (voice phishing) far more persuasive. When the voice sounds real, people lower their guard.
Ransomware and data extortion continue to inflict heavy damage. In early 2025 thousands of victims appeared on data leak sites — the highest recorded — showing how widespread and costly these attacks are. The ripple effects hit suppliers, customers, and entire industries when systems are locked or data is exposed.
Defenders are turning to AI too. Security teams use AI tools to scan alerts, summarize incidents, and suggest containment steps, shifting human roles from execution to validation. That cuts response time but also concentrates risk: a single flawed AI decision can propagate quickly across systems.
Google says the defence strategy must focus on stronger guardrails and layered protections: content filters to flag risky inputs, reinforcement techniques to keep models on safe tasks, and strict confirmation steps before any critical action. Still, it’s an evolving arms race — every new control tends to inspire new exploits.
By 2026, attacks driven by AI and automation are expected to be the norm rather than the exception. The same AI advances that help secure networks are also empowering the next wave of cybercrime, and both sides are racing to keep up
Leave a comment