AI Under Siege: From Hacked Minds to Weaponized Code
Edward Liebig
vCISO | VP of Cybersecurity | IT/OT Security | U.S. Navy Veteran | CISSP, CISM
Next in our look at AI Governance and Cybersecurity
As AI becomes integral to our lives—from powering self-driving cars to securing airports—its vulnerabilities are increasingly exploited by attackers. Building on prior discussions of AI risk frameworks like ISO and NIST, and the perils of AI poisoning, this two-part article explores emerging threats. In Part 1, Hacked Minds - I’ll unpack Adversarial Machine Learning (AML) attacks that manipulate AI systems into making dangerous mistakes. In Part 2, I’ll examine how AI itself becomes a weapon for exploitation. Understanding these risks is crucial for building defenses that keep AI trustworthy and safe.
Part 1: Hacked Minds - Navigating AI Attacks—Understanding the Threat Landscape
Adversarial Machine Learning Explained
Adversarial Machine Learning (AML) refers to malicious attempts to fool or manipulate AI systems (beyond the "data" as in our Poisoning discussion), leading them to produce incorrect, biased, or dangerous outputs. Think of it like slipping a distorted lens over a camera: the AI "sees" something different from reality, often without anyone noticing. These attacks are particularly concerning because traditional security methods—like firewalls or antivirus software—often fail to detect them, leaving critical systems exposed.
Types of Adversarial ML Attacks
Real-World Impacts
Attacks targeting AI systems, such as those exploiting Adversarial Machine Learning (AML), threaten national security, public safety, and business integrity, eroding trust in AI’s reliability. The stakes are high, as evidenced by real-world data from 2024. For instance, a 2023 study found that adversarial attacks on healthcare AI could misdiagnose up to 30% of patients—imagine an AI system misidentifying cancer in a scan, delaying critical treatment with potentially fatal results. The "AI Threat Landscape 2025" report from HiddenLayer underscores this vulnerability, noting that 73% of IT leaders reported definitively knowing of an AI breach in 2024, up from 67% the previous year, signaling a surge in successful attacks on critical systems.
National security faces significant risks from these vulnerabilities. AML techniques like model evasion or backdooring could compromise defense systems—think of an AI-driven drone misclassifying enemy targets due to manipulated inputs. The report highlights GPU vulnerabilities, such as LeftoverLocals (disclosed January 2024), which allow adversaries to extract sensitive data from AI processes, making such scenarios plausible. With 51% of AI attacks originating in North America and 34% in Europe, the threat to critical infrastructure is global and pervasive.
Business integrity is equally at stake. Model theft and supply chain attacks undermine competitive advantages, with the report revealing that 90% of companies use pre-trained models from repositories like Hugging Face, yet fewer than half scan these for safety. The December 2024 Ultralytics attack, which deployed crypto miners via compromised Python packages, exemplifies this risk. The financial toll is steep—IT leaders now spend 46% of their time addressing AI risk or security, up from 15% the previous year, diverting resources from innovation to defense. These breaches foster a climate of skepticism, as 61% of IT leaders advocate for mandatory breach disclosure, yet 29% avoid reporting incidents due to backlash fears, deepening distrust in AI’s role in critical applications.
Further Reading