Setting the record straight on explainable AI : (2nd out of N) Are ML models really black box?

Setting the record straight on explainable AI : (2nd out of N) Are ML models really black box?

A black box is defined as a system, which can be viewed in terms of its inputs and outputs (or transfer characteristics/function), without any knowledge of its internal workings. The opposite of a black box is a system where the inner components or logic are available for inspection (commonly referred to as a glass box or white box). I remember, when studying electrical/control engineering, it was quite common to deal with such terms in courses such as control theory, system identification, and digital signal processing, just to name a few; in all these topics, a system was considered a black box if it could be viewed solely in terms of its inputs and outputs. 

In this post, as a follow up to an earlier post that I wrote on explainable AI (XAI), I would like to argue that ML models are not really black box; quite the opposite in fact, as they are glass box systems whose underlying computational logic and components are fully visible to its owners, regulators and anyone who is supposed to understand and interrogate them. Of course, in some cases, due to IP and other considerations/concerns, the model can stay black box to some; this is different from ML models such neural networks being black box by nature and/or in general. I think that the common association of ML models to black-box systems can be attributed to circumstances where one: 

  1. Sees the model as equal to the actual system it attempts to model, 
  2. Lacks the necessary technical skills to understand the model’s glass-box transparency and interrogate it, 
  3. Finds the provided natural-language explanation complicated/hard to understand (regardless of technical ability) due to the model’s complexity. 
No alt text provided for this image

I described the first issue in an earlier post. Imagine that there is a system that we attempt to model (see the Figure). Due to the system being a black box, we employ ML (a neural network, for instance) to model its transfer characteristics (or, I/O logic); an ML model that is developed and can mimic the system’s transfer characteristics is not a black box. While the system can be assessed only in terms of its transfer characteristics, for the model we know everything from the assumptions and training data, to the mathematical/statistical logic and more. We knew enough to code every detail of the model, and use its software implementation to mimic the systems transfer function (illustrated in the Figure). 

The second issue tends to be dependent on the audience, i.e., rather than ML models being objectively black box, their opacity is subjective; while it is glass box to some, it can be perceived as black box by others. Let’s consider a simple hypothetical model: As input, it takes daily systolic blood pressure measurements for the current and last week (i.e., [b11, b12, …, b1?] and [b21, b22, …, b2?]), calculates their correlation; if the correlation has a P-value<0.05, the model recommends a certain action. This is an extremely simple model — hence glass box to anyone familiar with basic computations — that is not accessible/explainable to many. Therefore, one can argue that any model beyond a small number of if-then rules, has the risk of being “subjective black box” to some, despite being glass box to others.

Simple is better than complex. 
Complex is better than complicated.

Related to the third issue, in most XAI discussions, what people really mean are partial explanations, and simple (or even more appropriately, simplified) explanations. That is, regardless of the type of ML models used, when dealing with many input variables (i.e., high-dimensional input spaces), and hence many more possible combinations of these variables that it takes to make accurate predictions, any complete explanation of the model will be too lengthy and too complicated. In other words, in most cases, such simplified explanations will not qualify as complete explanations, which itself can take us to shaky ethical grounds of misleading out audience by simplifying the model’s explanation. This is why scientists in the domain have a different approach in understanding their models: They first understand and trust the maths behind their models, and then train and validate them on appropriate data, understand their models’ edge cases (e.g., where do they perform well and poorly?), see if their models’ assumptions hold, attempt to break their models / be their own models’ biggest critic, attempt to interrogate some of their models’ inner workings (a la partial explanations), assess them in a zero-to-minimum-harm live testing environment, and more; rather than explaining it to themselves in (and getting onboard based on) a few simple natural-language sentences about the model. 

In summary, ML models are not black box; rather, they are likely to be models of black box systems. When attempting to explain the ML models, one can have varying degrees of difficulties communicating them depending on their audiences’ technical skills (and their appetite for getting to know a very complex system and/or a complicated concept). A simple — yet risky and not complete/perfect — solution to bridge this gap might be such simplified/partial explanations. The ultimate solution, however, is trust; it can be built by many approaches, including, but not limited to simplified explanations of the model. 




This was the 2nd post in a series that I decided to write on XAI and a range of related topics. Given the vagueness of many definitions, and extreme heterogeneity of viewpoints re this topic, I hope to borrow from the wisdom of the crowd (through your comments, disagreements, and discussions on these posts) and eventually write a more comprehensive paper on this topic, which captures a broader set of perspectives. Please comment, share and move the conversation forward. Of course, these are personal opinions, and do not necessarily reflect the viewpoints of the institutions that I am affiliated with.



Minh-Tri Pham

Head of AI and Data Science

4 年

In my view, any neural network is just a differentable process. We define the process, so it is completely "glass" to the authors. It's just that depending on the audience's familiarity to the topic, they may find it "black". Edge cases happen because the problem is changing all the time.

回复
Brian Katz

Builder, Strategist, Negotiator | 2 Build & Flips | Insurance Startup | Claims | Product | LiDar / ADAS | xUSAA xLuminar

4 年

Reza Khorshidi, D.Phil. (Oxon) I enjoyed the article. Thank you.

When describing the issue of AI transparency, we should not focus too much on the academic definition of a black box. If a potential user is forced to blindly believe in the proper application of AI, it doesn't matter whether the problem is a classical black box situation or too much complexity. If it cannot be understood, it will not be understood. To me, XAI means, amongst others: - translating AI logic in human language, e.g. using a 3-dimensional example to explain a gradient descent - explaining how the success of a transparent algorithm depends on correct training data outlining the risks and the uncertainty that comes with probability - decomposing complex approaches into simpler, explainable steps - letting people play with different data sets and different parameters, so that they see the magnitude of the variance So, maybe we should call the complexity (and complicatedness) of AI a "grey box". In any case, the target will be to get closer to a glass box.

Zachary W.

Engineer | Data Scientist | Computational Science | Operations Research | MBSE | Autonomous Systems | (MS) Comp & Applied Math | (MS) Financial Engineering | MS Statistics | BS Applied Math Comp Sci & Econ | BS Sys Engr

4 年

Many ML models do not correspond to an explainable theoretical and can often provide results that are contrary to an accurate theoretical description of the causal dynamics and structure. This provides great fits to the function being modeled but as emergent properties take shape, ML models weight only inductive factors and not abductive or deductive factors. There are many kinds (or orders) of infinite ways to fit a given set of data. ML models can have a reliance on inductive properties and tend towards solutions that are at odds with a given set of theoretical solutions which are less accurate but far more effective for long term decision making because they take account of the properties and causal logic rather than mere induction. The allure of ML was the ability to take incomprehensible amounts of high dimensional data to adjust the parameters of a model to best fit the observations. You can look at many models and see what methods and weights were used to define a model that fits that output, but there are many infinite amounts of ways to fit the data. ML models assuming the evolutionary dynamics of the system continue as previously observed, this is often not the case in complex dynamic in human systems.

I agree with the thesis here. I'm also impressed with the concentration of wisdom to words in this passage: "understand their models’ edge cases (e.g., where do they perform well and poorly?), see if their models’ assumptions hold, attempt to break their models / be their own models’ biggest critic, attempt to interrogate some of their models’ inner workings (a la partial explanations), assess them in a zero-to-minimum-harm live testing environment"

要查看或添加评论,请登录

Reza Khorshidi, D.Phil. (Oxon)的更多文章

社区洞察

其他会员也浏览了