Persistent Issues for AI
John C. Checco, D.Sc.
Information Security Executive ∴ Innovator ∴ Firefighter ∴ Speaker
Over the course of the past decades, I have been working with various incantations of AI - from the AI Lab in IBM T. J. Watson Research Center (1982-86), Advanced Technology Labs (1990-93), NYNEX Science & Technology (1993-98), Pitney Bowes Advanced Concepts & Technology (2005-07), Bank of America Security Research & Innovation (2014-2020) and a stint at Applied Research Associates Neya Robotics division (2024).
With each generation of advancement, I have seen the same issues come up over and over again; and decided to list my Top 5 attributes when considering an AI approach to business.
(1) Embracing AI Means Understanding the Dirty State of Data.
Look beyond the AI to the solution providers themselves. when enterprises are sold security solutions, they are marketed that the solution will detect anomalies and/or drift from a baseline. But the reality is that implementing such a solution on a dirty network creates a dirty baseline … Not only does the deployment ignore existing issues, those existing issues are now considered part of the baseline.
(2) AI Modeling Transparency is Crucial to Understand Bias.
Transparency needs to be aligned to a motivational outcome. It seems that everyone mentions transparency without understanding the challenges and implications of achieving transparency. An experiment by Lakkaraju on AI-based decision modeling proved that one cannot remove bias from the decision process, even though the AI training data had bias inputs removed (https://dl.acm.org/doi/10.1145/3097983.3098066). In other words, understand we can never get rid of bias, as it is part and parcel of every decision we make; but we can understand and embrace how AI identifies bias in our society through data modeling.
(3) Entitlement Restrictions for AI is Crucial to Data Privacy.
AI systems need to know the role of the person asking the AI for data and returning results appropriate for the participants entitlements. There is a concept known as “differential privacy” (DF) which does entitlement cleansing at certain stages. If DF is done at the prompt level, then the outcome is the safest but not very useful. If DF is done at the tail end of the process, the outcomes are very detailed but very difficult to hide parts which should be restricted. Two examples where over-sharing of data privacy is being mitigated:
(4) AI Explainability in Financial Services is Crucial to Regulatory Compliance.
Nobody has even considered how the existing SEC rules affect the GenAI processing of financial instruments. When the SEC proposed more transparency in financial lending back in 2021 (Exchange Act Rule 10c-1) as a result of the mortgage crisis, I doubt they understood the enormity of what we are seeing now with AI hallucination. Beyond bad financial decisions, imagine these same AI algorithms deciding the best medical treatment for your health issues. As patients, we should demand transparency in how our medical decisions are derived.
(5) AI Verification in Threat Intelligence is Crucial to Offensive Security Operations.
Imagine our threat intelligence analysis is derived from AI-fabricated data points, resulting in mis-interpretation, mis-direction and mis-attribution. This is exactly why AI-based decision-making needs to be vetted by human experts. Let the AI give an array of alternative outcomes, and let the human determine the best path forward.
(6) AI Consumers Needs to Adopt "Verify Then Trust" Mentality.
The scrupulous over-adoption of GenAI by enterprises for every conceivable problem – relevant to the training data or not – should be tempered by continuous skepticism. My mantra for GenAI outcomes is to move from the legacy security tenet of “Trust but Verify” to a new perspective of “Verify THEN Trust” – an ode to the Volkswagen marketing slogan “Sign then Drive”. Basically assume your network is already compromised. (We can get into the efficacy of the Zero-Trust rabbit hole here too.)